xref: /linux/fs/xfs/scrub/common.h (revision c02ce1735b150cf7c3b43790b48e23dcd17c0d46)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * Copyright (C) 2017-2023 Oracle.  All Rights Reserved.
4  * Author: Darrick J. Wong <djwong@kernel.org>
5  */
6 #ifndef __XFS_SCRUB_COMMON_H__
7 #define __XFS_SCRUB_COMMON_H__
8 
9 /*
10  * We /could/ terminate a scrub/repair operation early.  If we're not
11  * in a good place to continue (fatal signal, etc.) then bail out.
12  * Note that we're careful not to make any judgements about *error.
13  */
14 static inline bool
15 xchk_should_terminate(
16 	struct xfs_scrub	*sc,
17 	int			*error)
18 {
19 	/*
20 	 * If preemption is disabled, we need to yield to the scheduler every
21 	 * few seconds so that we don't run afoul of the soft lockup watchdog
22 	 * or RCU stall detector.
23 	 */
24 	cond_resched();
25 
26 	if (fatal_signal_pending(current)) {
27 		if (*error == 0)
28 			*error = -EINTR;
29 		return true;
30 	}
31 	return false;
32 }
33 
34 int xchk_trans_alloc(struct xfs_scrub *sc, uint resblks);
35 int xchk_trans_alloc_empty(struct xfs_scrub *sc);
36 void xchk_trans_cancel(struct xfs_scrub *sc);
37 
38 bool xchk_process_error(struct xfs_scrub *sc, xfs_agnumber_t agno,
39 		xfs_agblock_t bno, int *error);
40 bool xchk_fblock_process_error(struct xfs_scrub *sc, int whichfork,
41 		xfs_fileoff_t offset, int *error);
42 
43 bool xchk_xref_process_error(struct xfs_scrub *sc,
44 		xfs_agnumber_t agno, xfs_agblock_t bno, int *error);
45 bool xchk_fblock_xref_process_error(struct xfs_scrub *sc,
46 		int whichfork, xfs_fileoff_t offset, int *error);
47 
48 void xchk_block_set_preen(struct xfs_scrub *sc,
49 		struct xfs_buf *bp);
50 void xchk_ino_set_preen(struct xfs_scrub *sc, xfs_ino_t ino);
51 
52 void xchk_set_corrupt(struct xfs_scrub *sc);
53 void xchk_block_set_corrupt(struct xfs_scrub *sc,
54 		struct xfs_buf *bp);
55 void xchk_ino_set_corrupt(struct xfs_scrub *sc, xfs_ino_t ino);
56 void xchk_fblock_set_corrupt(struct xfs_scrub *sc, int whichfork,
57 		xfs_fileoff_t offset);
58 #ifdef CONFIG_XFS_QUOTA
59 void xchk_qcheck_set_corrupt(struct xfs_scrub *sc, unsigned int dqtype,
60 		xfs_dqid_t id);
61 #endif
62 
63 void xchk_block_xref_set_corrupt(struct xfs_scrub *sc,
64 		struct xfs_buf *bp);
65 void xchk_ino_xref_set_corrupt(struct xfs_scrub *sc,
66 		xfs_ino_t ino);
67 void xchk_fblock_xref_set_corrupt(struct xfs_scrub *sc,
68 		int whichfork, xfs_fileoff_t offset);
69 
70 void xchk_ino_set_warning(struct xfs_scrub *sc, xfs_ino_t ino);
71 void xchk_fblock_set_warning(struct xfs_scrub *sc, int whichfork,
72 		xfs_fileoff_t offset);
73 
74 void xchk_set_incomplete(struct xfs_scrub *sc);
75 int xchk_checkpoint_log(struct xfs_mount *mp);
76 
77 /* Are we set up for a cross-referencing check? */
78 bool xchk_should_check_xref(struct xfs_scrub *sc, int *error,
79 			   struct xfs_btree_cur **curpp);
80 
81 /* Setup functions */
82 int xchk_setup_agheader(struct xfs_scrub *sc);
83 int xchk_setup_fs(struct xfs_scrub *sc);
84 int xchk_setup_ag_allocbt(struct xfs_scrub *sc);
85 int xchk_setup_ag_iallocbt(struct xfs_scrub *sc);
86 int xchk_setup_ag_rmapbt(struct xfs_scrub *sc);
87 int xchk_setup_ag_refcountbt(struct xfs_scrub *sc);
88 int xchk_setup_inode(struct xfs_scrub *sc);
89 int xchk_setup_inode_bmap(struct xfs_scrub *sc);
90 int xchk_setup_inode_bmap_data(struct xfs_scrub *sc);
91 int xchk_setup_directory(struct xfs_scrub *sc);
92 int xchk_setup_xattr(struct xfs_scrub *sc);
93 int xchk_setup_symlink(struct xfs_scrub *sc);
94 int xchk_setup_parent(struct xfs_scrub *sc);
95 #ifdef CONFIG_XFS_RT
96 int xchk_setup_rtbitmap(struct xfs_scrub *sc);
97 int xchk_setup_rtsummary(struct xfs_scrub *sc);
98 #else
99 static inline int
100 xchk_setup_rtbitmap(struct xfs_scrub *sc)
101 {
102 	return -ENOENT;
103 }
104 static inline int
105 xchk_setup_rtsummary(struct xfs_scrub *sc)
106 {
107 	return -ENOENT;
108 }
109 #endif
110 #ifdef CONFIG_XFS_QUOTA
111 int xchk_ino_dqattach(struct xfs_scrub *sc);
112 int xchk_setup_quota(struct xfs_scrub *sc);
113 int xchk_setup_quotacheck(struct xfs_scrub *sc);
114 #else
115 static inline int
116 xchk_ino_dqattach(struct xfs_scrub *sc)
117 {
118 	return 0;
119 }
120 static inline int
121 xchk_setup_quota(struct xfs_scrub *sc)
122 {
123 	return -ENOENT;
124 }
125 static inline int
126 xchk_setup_quotacheck(struct xfs_scrub *sc)
127 {
128 	return -ENOENT;
129 }
130 #endif
131 int xchk_setup_fscounters(struct xfs_scrub *sc);
132 int xchk_setup_nlinks(struct xfs_scrub *sc);
133 
134 void xchk_ag_free(struct xfs_scrub *sc, struct xchk_ag *sa);
135 int xchk_ag_init(struct xfs_scrub *sc, xfs_agnumber_t agno,
136 		struct xchk_ag *sa);
137 int xchk_perag_drain_and_lock(struct xfs_scrub *sc);
138 
139 /*
140  * Grab all AG resources, treating the inability to grab the perag structure as
141  * a fs corruption.  This is intended for callers checking an ondisk reference
142  * to a given AG, which means that the AG must still exist.
143  */
144 static inline int
145 xchk_ag_init_existing(
146 	struct xfs_scrub	*sc,
147 	xfs_agnumber_t		agno,
148 	struct xchk_ag		*sa)
149 {
150 	int			error = xchk_ag_init(sc, agno, sa);
151 
152 	return error == -ENOENT ? -EFSCORRUPTED : error;
153 }
154 
155 int xchk_ag_read_headers(struct xfs_scrub *sc, xfs_agnumber_t agno,
156 		struct xchk_ag *sa);
157 void xchk_ag_btcur_free(struct xchk_ag *sa);
158 void xchk_ag_btcur_init(struct xfs_scrub *sc, struct xchk_ag *sa);
159 int xchk_count_rmap_ownedby_ag(struct xfs_scrub *sc, struct xfs_btree_cur *cur,
160 		const struct xfs_owner_info *oinfo, xfs_filblks_t *blocks);
161 
162 int xchk_setup_ag_btree(struct xfs_scrub *sc, bool force_log);
163 int xchk_iget_for_scrubbing(struct xfs_scrub *sc);
164 int xchk_setup_inode_contents(struct xfs_scrub *sc, unsigned int resblks);
165 int xchk_install_live_inode(struct xfs_scrub *sc, struct xfs_inode *ip);
166 
167 void xchk_ilock(struct xfs_scrub *sc, unsigned int ilock_flags);
168 bool xchk_ilock_nowait(struct xfs_scrub *sc, unsigned int ilock_flags);
169 void xchk_iunlock(struct xfs_scrub *sc, unsigned int ilock_flags);
170 
171 void xchk_buffer_recheck(struct xfs_scrub *sc, struct xfs_buf *bp);
172 
173 /*
174  * Grab the inode at @inum.  The caller must have created a scrub transaction
175  * so that we can confirm the inumber by walking the inobt and not deadlock on
176  * a loop in the inobt.
177  */
178 int xchk_iget(struct xfs_scrub *sc, xfs_ino_t inum, struct xfs_inode **ipp);
179 int xchk_iget_agi(struct xfs_scrub *sc, xfs_ino_t inum,
180 		struct xfs_buf **agi_bpp, struct xfs_inode **ipp);
181 void xchk_irele(struct xfs_scrub *sc, struct xfs_inode *ip);
182 int xchk_install_handle_inode(struct xfs_scrub *sc, struct xfs_inode *ip);
183 
184 /*
185  * Safe version of (untrusted) xchk_iget that uses an empty transaction to
186  * avoid deadlocking on loops in the inobt.  This should only be used in a
187  * scrub or repair setup routine, and only prior to grabbing a transaction.
188  */
189 static inline int
190 xchk_iget_safe(struct xfs_scrub *sc, xfs_ino_t inum, struct xfs_inode **ipp)
191 {
192 	int	error;
193 
194 	ASSERT(sc->tp == NULL);
195 
196 	error = xchk_trans_alloc(sc, 0);
197 	if (error)
198 		return error;
199 	error = xchk_iget(sc, inum, ipp);
200 	xchk_trans_cancel(sc);
201 	return error;
202 }
203 
204 /*
205  * Don't bother cross-referencing if we already found corruption or cross
206  * referencing discrepancies.
207  */
208 static inline bool xchk_skip_xref(struct xfs_scrub_metadata *sm)
209 {
210 	return sm->sm_flags & (XFS_SCRUB_OFLAG_CORRUPT |
211 			       XFS_SCRUB_OFLAG_XCORRUPT);
212 }
213 
214 bool xchk_dir_looks_zapped(struct xfs_inode *dp);
215 
216 #ifdef CONFIG_XFS_ONLINE_REPAIR
217 /* Decide if a repair is required. */
218 static inline bool xchk_needs_repair(const struct xfs_scrub_metadata *sm)
219 {
220 	return sm->sm_flags & (XFS_SCRUB_OFLAG_CORRUPT |
221 			       XFS_SCRUB_OFLAG_XCORRUPT |
222 			       XFS_SCRUB_OFLAG_PREEN);
223 }
224 
225 /*
226  * "Should we prepare for a repair?"
227  *
228  * Return true if the caller permits us to repair metadata and we're not
229  * setting up for a post-repair evaluation.
230  */
231 static inline bool xchk_could_repair(const struct xfs_scrub *sc)
232 {
233 	return (sc->sm->sm_flags & XFS_SCRUB_IFLAG_REPAIR) &&
234 		!(sc->flags & XREP_ALREADY_FIXED);
235 }
236 #else
237 # define xchk_needs_repair(sc)		(false)
238 # define xchk_could_repair(sc)		(false)
239 #endif /* CONFIG_XFS_ONLINE_REPAIR */
240 
241 int xchk_metadata_inode_forks(struct xfs_scrub *sc);
242 
243 /*
244  * Helper macros to allocate and format xfile description strings.
245  * Callers must kfree the pointer returned.
246  */
247 #define xchk_xfile_descr(sc, fmt, ...) \
248 	kasprintf(XCHK_GFP_FLAGS, "XFS (%s): " fmt, \
249 			(sc)->mp->m_super->s_id, ##__VA_ARGS__)
250 #define xchk_xfile_ag_descr(sc, fmt, ...) \
251 	kasprintf(XCHK_GFP_FLAGS, "XFS (%s): AG 0x%x " fmt, \
252 			(sc)->mp->m_super->s_id, \
253 			(sc)->sa.pag ? (sc)->sa.pag->pag_agno : (sc)->sm->sm_agno, \
254 			##__VA_ARGS__)
255 #define xchk_xfile_ino_descr(sc, fmt, ...) \
256 	kasprintf(XCHK_GFP_FLAGS, "XFS (%s): inode 0x%llx " fmt, \
257 			(sc)->mp->m_super->s_id, \
258 			(sc)->ip ? (sc)->ip->i_ino : (sc)->sm->sm_ino, \
259 			##__VA_ARGS__)
260 
261 /*
262  * Setting up a hook to wait for intents to drain is costly -- we have to take
263  * the CPU hotplug lock and force an i-cache flush on all CPUs once to set it
264  * up, and again to tear it down.  These costs add up quickly, so we only want
265  * to enable the drain waiter if the drain actually detected a conflict with
266  * running intent chains.
267  */
268 static inline bool xchk_need_intent_drain(struct xfs_scrub *sc)
269 {
270 	return sc->flags & XCHK_NEED_DRAIN;
271 }
272 
273 void xchk_fsgates_enable(struct xfs_scrub *sc, unsigned int scrub_fshooks);
274 
275 int xchk_inode_is_allocated(struct xfs_scrub *sc, xfs_agino_t agino,
276 		bool *inuse);
277 
278 #endif	/* __XFS_SCRUB_COMMON_H__ */
279