Loading...
Note: File does not exist in v3.5.6.
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Copyright (C) 2017-2023 Oracle. All Rights Reserved.
4 * Author: Darrick J. Wong <djwong@kernel.org>
5 */
6#ifndef __XFS_SCRUB_COMMON_H__
7#define __XFS_SCRUB_COMMON_H__
8
9/*
10 * We /could/ terminate a scrub/repair operation early. If we're not
11 * in a good place to continue (fatal signal, etc.) then bail out.
12 * Note that we're careful not to make any judgements about *error.
13 */
14static inline bool
15xchk_should_terminate(
16 struct xfs_scrub *sc,
17 int *error)
18{
19 /*
20 * If preemption is disabled, we need to yield to the scheduler every
21 * few seconds so that we don't run afoul of the soft lockup watchdog
22 * or RCU stall detector.
23 */
24 cond_resched();
25
26 if (fatal_signal_pending(current)) {
27 if (*error == 0)
28 *error = -EINTR;
29 return true;
30 }
31 return false;
32}
33
34int xchk_trans_alloc(struct xfs_scrub *sc, uint resblks);
35void xchk_trans_cancel(struct xfs_scrub *sc);
36
37bool xchk_process_error(struct xfs_scrub *sc, xfs_agnumber_t agno,
38 xfs_agblock_t bno, int *error);
39bool xchk_fblock_process_error(struct xfs_scrub *sc, int whichfork,
40 xfs_fileoff_t offset, int *error);
41
42bool xchk_xref_process_error(struct xfs_scrub *sc,
43 xfs_agnumber_t agno, xfs_agblock_t bno, int *error);
44bool xchk_fblock_xref_process_error(struct xfs_scrub *sc,
45 int whichfork, xfs_fileoff_t offset, int *error);
46
47void xchk_block_set_preen(struct xfs_scrub *sc,
48 struct xfs_buf *bp);
49void xchk_ino_set_preen(struct xfs_scrub *sc, xfs_ino_t ino);
50
51void xchk_set_corrupt(struct xfs_scrub *sc);
52void xchk_block_set_corrupt(struct xfs_scrub *sc,
53 struct xfs_buf *bp);
54void xchk_ino_set_corrupt(struct xfs_scrub *sc, xfs_ino_t ino);
55void xchk_fblock_set_corrupt(struct xfs_scrub *sc, int whichfork,
56 xfs_fileoff_t offset);
57
58void xchk_block_xref_set_corrupt(struct xfs_scrub *sc,
59 struct xfs_buf *bp);
60void xchk_ino_xref_set_corrupt(struct xfs_scrub *sc,
61 xfs_ino_t ino);
62void xchk_fblock_xref_set_corrupt(struct xfs_scrub *sc,
63 int whichfork, xfs_fileoff_t offset);
64
65void xchk_ino_set_warning(struct xfs_scrub *sc, xfs_ino_t ino);
66void xchk_fblock_set_warning(struct xfs_scrub *sc, int whichfork,
67 xfs_fileoff_t offset);
68
69void xchk_set_incomplete(struct xfs_scrub *sc);
70int xchk_checkpoint_log(struct xfs_mount *mp);
71
72/* Are we set up for a cross-referencing check? */
73bool xchk_should_check_xref(struct xfs_scrub *sc, int *error,
74 struct xfs_btree_cur **curpp);
75
76/* Setup functions */
77int xchk_setup_agheader(struct xfs_scrub *sc);
78int xchk_setup_fs(struct xfs_scrub *sc);
79int xchk_setup_ag_allocbt(struct xfs_scrub *sc);
80int xchk_setup_ag_iallocbt(struct xfs_scrub *sc);
81int xchk_setup_ag_rmapbt(struct xfs_scrub *sc);
82int xchk_setup_ag_refcountbt(struct xfs_scrub *sc);
83int xchk_setup_inode(struct xfs_scrub *sc);
84int xchk_setup_inode_bmap(struct xfs_scrub *sc);
85int xchk_setup_inode_bmap_data(struct xfs_scrub *sc);
86int xchk_setup_directory(struct xfs_scrub *sc);
87int xchk_setup_xattr(struct xfs_scrub *sc);
88int xchk_setup_symlink(struct xfs_scrub *sc);
89int xchk_setup_parent(struct xfs_scrub *sc);
90#ifdef CONFIG_XFS_RT
91int xchk_setup_rtbitmap(struct xfs_scrub *sc);
92int xchk_setup_rtsummary(struct xfs_scrub *sc);
93#else
94static inline int
95xchk_setup_rtbitmap(struct xfs_scrub *sc)
96{
97 return -ENOENT;
98}
99static inline int
100xchk_setup_rtsummary(struct xfs_scrub *sc)
101{
102 return -ENOENT;
103}
104#endif
105#ifdef CONFIG_XFS_QUOTA
106int xchk_ino_dqattach(struct xfs_scrub *sc);
107int xchk_setup_quota(struct xfs_scrub *sc);
108#else
109static inline int
110xchk_ino_dqattach(struct xfs_scrub *sc)
111{
112 return 0;
113}
114static inline int
115xchk_setup_quota(struct xfs_scrub *sc)
116{
117 return -ENOENT;
118}
119#endif
120int xchk_setup_fscounters(struct xfs_scrub *sc);
121
122void xchk_ag_free(struct xfs_scrub *sc, struct xchk_ag *sa);
123int xchk_ag_init(struct xfs_scrub *sc, xfs_agnumber_t agno,
124 struct xchk_ag *sa);
125
126/*
127 * Grab all AG resources, treating the inability to grab the perag structure as
128 * a fs corruption. This is intended for callers checking an ondisk reference
129 * to a given AG, which means that the AG must still exist.
130 */
131static inline int
132xchk_ag_init_existing(
133 struct xfs_scrub *sc,
134 xfs_agnumber_t agno,
135 struct xchk_ag *sa)
136{
137 int error = xchk_ag_init(sc, agno, sa);
138
139 return error == -ENOENT ? -EFSCORRUPTED : error;
140}
141
142int xchk_ag_read_headers(struct xfs_scrub *sc, xfs_agnumber_t agno,
143 struct xchk_ag *sa);
144void xchk_ag_btcur_free(struct xchk_ag *sa);
145void xchk_ag_btcur_init(struct xfs_scrub *sc, struct xchk_ag *sa);
146int xchk_count_rmap_ownedby_ag(struct xfs_scrub *sc, struct xfs_btree_cur *cur,
147 const struct xfs_owner_info *oinfo, xfs_filblks_t *blocks);
148
149int xchk_setup_ag_btree(struct xfs_scrub *sc, bool force_log);
150int xchk_iget_for_scrubbing(struct xfs_scrub *sc);
151int xchk_setup_inode_contents(struct xfs_scrub *sc, unsigned int resblks);
152int xchk_install_live_inode(struct xfs_scrub *sc, struct xfs_inode *ip);
153
154void xchk_ilock(struct xfs_scrub *sc, unsigned int ilock_flags);
155bool xchk_ilock_nowait(struct xfs_scrub *sc, unsigned int ilock_flags);
156void xchk_iunlock(struct xfs_scrub *sc, unsigned int ilock_flags);
157
158void xchk_buffer_recheck(struct xfs_scrub *sc, struct xfs_buf *bp);
159
160/*
161 * Grab the inode at @inum. The caller must have created a scrub transaction
162 * so that we can confirm the inumber by walking the inobt and not deadlock on
163 * a loop in the inobt.
164 */
165int xchk_iget(struct xfs_scrub *sc, xfs_ino_t inum, struct xfs_inode **ipp);
166int xchk_iget_agi(struct xfs_scrub *sc, xfs_ino_t inum,
167 struct xfs_buf **agi_bpp, struct xfs_inode **ipp);
168void xchk_irele(struct xfs_scrub *sc, struct xfs_inode *ip);
169int xchk_install_handle_inode(struct xfs_scrub *sc, struct xfs_inode *ip);
170
171/*
172 * Safe version of (untrusted) xchk_iget that uses an empty transaction to
173 * avoid deadlocking on loops in the inobt. This should only be used in a
174 * scrub or repair setup routine, and only prior to grabbing a transaction.
175 */
176static inline int
177xchk_iget_safe(struct xfs_scrub *sc, xfs_ino_t inum, struct xfs_inode **ipp)
178{
179 int error;
180
181 ASSERT(sc->tp == NULL);
182
183 error = xchk_trans_alloc(sc, 0);
184 if (error)
185 return error;
186 error = xchk_iget(sc, inum, ipp);
187 xchk_trans_cancel(sc);
188 return error;
189}
190
191/*
192 * Don't bother cross-referencing if we already found corruption or cross
193 * referencing discrepancies.
194 */
195static inline bool xchk_skip_xref(struct xfs_scrub_metadata *sm)
196{
197 return sm->sm_flags & (XFS_SCRUB_OFLAG_CORRUPT |
198 XFS_SCRUB_OFLAG_XCORRUPT);
199}
200
201bool xchk_dir_looks_zapped(struct xfs_inode *dp);
202
203#ifdef CONFIG_XFS_ONLINE_REPAIR
204/* Decide if a repair is required. */
205static inline bool xchk_needs_repair(const struct xfs_scrub_metadata *sm)
206{
207 return sm->sm_flags & (XFS_SCRUB_OFLAG_CORRUPT |
208 XFS_SCRUB_OFLAG_XCORRUPT |
209 XFS_SCRUB_OFLAG_PREEN);
210}
211
212/*
213 * "Should we prepare for a repair?"
214 *
215 * Return true if the caller permits us to repair metadata and we're not
216 * setting up for a post-repair evaluation.
217 */
218static inline bool xchk_could_repair(const struct xfs_scrub *sc)
219{
220 return (sc->sm->sm_flags & XFS_SCRUB_IFLAG_REPAIR) &&
221 !(sc->flags & XREP_ALREADY_FIXED);
222}
223#else
224# define xchk_needs_repair(sc) (false)
225# define xchk_could_repair(sc) (false)
226#endif /* CONFIG_XFS_ONLINE_REPAIR */
227
228int xchk_metadata_inode_forks(struct xfs_scrub *sc);
229
230/*
231 * Helper macros to allocate and format xfile description strings.
232 * Callers must kfree the pointer returned.
233 */
234#define xchk_xfile_descr(sc, fmt, ...) \
235 kasprintf(XCHK_GFP_FLAGS, "XFS (%s): " fmt, \
236 (sc)->mp->m_super->s_id, ##__VA_ARGS__)
237#define xchk_xfile_ag_descr(sc, fmt, ...) \
238 kasprintf(XCHK_GFP_FLAGS, "XFS (%s): AG 0x%x " fmt, \
239 (sc)->mp->m_super->s_id, \
240 (sc)->sa.pag ? (sc)->sa.pag->pag_agno : (sc)->sm->sm_agno, \
241 ##__VA_ARGS__)
242#define xchk_xfile_ino_descr(sc, fmt, ...) \
243 kasprintf(XCHK_GFP_FLAGS, "XFS (%s): inode 0x%llx " fmt, \
244 (sc)->mp->m_super->s_id, \
245 (sc)->ip ? (sc)->ip->i_ino : (sc)->sm->sm_ino, \
246 ##__VA_ARGS__)
247
248/*
249 * Setting up a hook to wait for intents to drain is costly -- we have to take
250 * the CPU hotplug lock and force an i-cache flush on all CPUs once to set it
251 * up, and again to tear it down. These costs add up quickly, so we only want
252 * to enable the drain waiter if the drain actually detected a conflict with
253 * running intent chains.
254 */
255static inline bool xchk_need_intent_drain(struct xfs_scrub *sc)
256{
257 return sc->flags & XCHK_NEED_DRAIN;
258}
259
260void xchk_fsgates_enable(struct xfs_scrub *sc, unsigned int scrub_fshooks);
261
262int xchk_inode_is_allocated(struct xfs_scrub *sc, xfs_agino_t agino,
263 bool *inuse);
264
265#endif /* __XFS_SCRUB_COMMON_H__ */