xref: /linux/fs/xfs/scrub/scrub.h (revision 50a0844bf8c4d38be540e423672ef9408d029252)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * Copyright (C) 2017-2023 Oracle.  All Rights Reserved.
4  * Author: Darrick J. Wong <djwong@kernel.org>
5  */
6 #ifndef __XFS_SCRUB_SCRUB_H__
7 #define __XFS_SCRUB_SCRUB_H__
8 
9 struct xfs_scrub;
10 
11 /*
12  * Standard flags for allocating memory within scrub.  NOFS context is
13  * configured by the process allocation scope.  Scrub and repair must be able
14  * to back out gracefully if there isn't enough memory.  Force-cast to avoid
15  * complaints from static checkers.
16  */
17 #define XCHK_GFP_FLAGS	((__force gfp_t)(GFP_KERNEL | __GFP_NOWARN | \
18 					 __GFP_RETRY_MAYFAIL))
19 
20 /* Type info and names for the scrub types. */
21 enum xchk_type {
22 	ST_NONE = 1,	/* disabled */
23 	ST_PERAG,	/* per-AG metadata */
24 	ST_FS,		/* per-FS metadata */
25 	ST_INODE,	/* per-inode metadata */
26 };
27 
28 struct xchk_meta_ops {
29 	/* Acquire whatever resources are needed for the operation. */
30 	int		(*setup)(struct xfs_scrub *sc);
31 
32 	/* Examine metadata for errors. */
33 	int		(*scrub)(struct xfs_scrub *);
34 
35 	/* Repair or optimize the metadata. */
36 	int		(*repair)(struct xfs_scrub *);
37 
38 	/*
39 	 * Re-scrub the metadata we repaired, in case there's extra work that
40 	 * we need to do to check our repair work.  If this is NULL, we'll use
41 	 * the ->scrub function pointer, assuming that the regular scrub is
42 	 * sufficient.
43 	 */
44 	int		(*repair_eval)(struct xfs_scrub *sc);
45 
46 	/* Decide if we even have this piece of metadata. */
47 	bool		(*has)(struct xfs_mount *);
48 
49 	/* type describing required/allowed inputs */
50 	enum xchk_type	type;
51 };
52 
53 /* Buffer pointers and btree cursors for an entire AG. */
54 struct xchk_ag {
55 	struct xfs_perag	*pag;
56 
57 	/* AG btree roots */
58 	struct xfs_buf		*agf_bp;
59 	struct xfs_buf		*agi_bp;
60 
61 	/* AG btrees */
62 	struct xfs_btree_cur	*bno_cur;
63 	struct xfs_btree_cur	*cnt_cur;
64 	struct xfs_btree_cur	*ino_cur;
65 	struct xfs_btree_cur	*fino_cur;
66 	struct xfs_btree_cur	*rmap_cur;
67 	struct xfs_btree_cur	*refc_cur;
68 };
69 
70 struct xfs_scrub {
71 	/* General scrub state. */
72 	struct xfs_mount		*mp;
73 	struct xfs_scrub_metadata	*sm;
74 	const struct xchk_meta_ops	*ops;
75 	struct xfs_trans		*tp;
76 
77 	/* File that scrub was called with. */
78 	struct file			*file;
79 
80 	/*
81 	 * File that is undergoing the scrub operation.  This can differ from
82 	 * the file that scrub was called with if we're checking file-based fs
83 	 * metadata (e.g. rt bitmaps) or if we're doing a scrub-by-handle for
84 	 * something that can't be opened directly (e.g. symlinks).
85 	 */
86 	struct xfs_inode		*ip;
87 
88 	/* Kernel memory buffer used by scrubbers; freed at teardown. */
89 	void				*buf;
90 
91 	/*
92 	 * Clean up resources owned by whatever is in the buffer.  Cleanup can
93 	 * be deferred with this hook as a means for scrub functions to pass
94 	 * data to repair functions.  This function must not free the buffer
95 	 * itself.
96 	 */
97 	void				(*buf_cleanup)(void *buf);
98 
99 	/* xfile used by the scrubbers; freed at teardown. */
100 	struct xfile			*xfile;
101 
102 	/* buffer target for in-memory btrees; also freed at teardown. */
103 	struct xfs_buftarg		*xmbtp;
104 
105 	/* Lock flags for @ip. */
106 	uint				ilock_flags;
107 
108 	/* See the XCHK/XREP state flags below. */
109 	unsigned int			flags;
110 
111 	/*
112 	 * The XFS_SICK_* flags that correspond to the metadata being scrubbed
113 	 * or repaired.  We will use this mask to update the in-core fs health
114 	 * status with whatever we find.
115 	 */
116 	unsigned int			sick_mask;
117 
118 	/* State tracking for single-AG operations. */
119 	struct xchk_ag			sa;
120 };
121 
122 /* XCHK state flags grow up from zero, XREP state flags grown down from 2^31 */
123 #define XCHK_TRY_HARDER		(1U << 0)  /* can't get resources, try again */
124 #define XCHK_HAVE_FREEZE_PROT	(1U << 1)  /* do we have freeze protection? */
125 #define XCHK_FSGATES_DRAIN	(1U << 2)  /* defer ops draining enabled */
126 #define XCHK_NEED_DRAIN		(1U << 3)  /* scrub needs to drain defer ops */
127 #define XCHK_FSGATES_QUOTA	(1U << 4)  /* quota live update enabled */
128 #define XCHK_FSGATES_DIRENTS	(1U << 5)  /* directory live update enabled */
129 #define XCHK_FSGATES_RMAP	(1U << 6)  /* rmapbt live update enabled */
130 #define XREP_RESET_PERAG_RESV	(1U << 30) /* must reset AG space reservation */
131 #define XREP_ALREADY_FIXED	(1U << 31) /* checking our repair work */
132 
133 /*
134  * The XCHK_FSGATES* flags reflect functionality in the main filesystem that
135  * are only enabled for this particular online fsck.  When not in use, the
136  * features are gated off via dynamic code patching, which is why the state
137  * must be enabled during scrub setup and can only be torn down afterwards.
138  */
139 #define XCHK_FSGATES_ALL	(XCHK_FSGATES_DRAIN | \
140 				 XCHK_FSGATES_QUOTA | \
141 				 XCHK_FSGATES_DIRENTS | \
142 				 XCHK_FSGATES_RMAP)
143 
144 /* Metadata scrubbers */
145 int xchk_tester(struct xfs_scrub *sc);
146 int xchk_superblock(struct xfs_scrub *sc);
147 int xchk_agf(struct xfs_scrub *sc);
148 int xchk_agfl(struct xfs_scrub *sc);
149 int xchk_agi(struct xfs_scrub *sc);
150 int xchk_allocbt(struct xfs_scrub *sc);
151 int xchk_iallocbt(struct xfs_scrub *sc);
152 int xchk_rmapbt(struct xfs_scrub *sc);
153 int xchk_refcountbt(struct xfs_scrub *sc);
154 int xchk_inode(struct xfs_scrub *sc);
155 int xchk_bmap_data(struct xfs_scrub *sc);
156 int xchk_bmap_attr(struct xfs_scrub *sc);
157 int xchk_bmap_cow(struct xfs_scrub *sc);
158 int xchk_directory(struct xfs_scrub *sc);
159 int xchk_xattr(struct xfs_scrub *sc);
160 int xchk_symlink(struct xfs_scrub *sc);
161 int xchk_parent(struct xfs_scrub *sc);
162 #ifdef CONFIG_XFS_RT
163 int xchk_rtbitmap(struct xfs_scrub *sc);
164 int xchk_rtsummary(struct xfs_scrub *sc);
165 #else
166 static inline int
167 xchk_rtbitmap(struct xfs_scrub *sc)
168 {
169 	return -ENOENT;
170 }
171 static inline int
172 xchk_rtsummary(struct xfs_scrub *sc)
173 {
174 	return -ENOENT;
175 }
176 #endif
177 #ifdef CONFIG_XFS_QUOTA
178 int xchk_quota(struct xfs_scrub *sc);
179 int xchk_quotacheck(struct xfs_scrub *sc);
180 #else
181 static inline int
182 xchk_quota(struct xfs_scrub *sc)
183 {
184 	return -ENOENT;
185 }
186 static inline int
187 xchk_quotacheck(struct xfs_scrub *sc)
188 {
189 	return -ENOENT;
190 }
191 #endif
192 int xchk_fscounters(struct xfs_scrub *sc);
193 int xchk_nlinks(struct xfs_scrub *sc);
194 
195 /* cross-referencing helpers */
196 void xchk_xref_is_used_space(struct xfs_scrub *sc, xfs_agblock_t agbno,
197 		xfs_extlen_t len);
198 void xchk_xref_is_not_inode_chunk(struct xfs_scrub *sc, xfs_agblock_t agbno,
199 		xfs_extlen_t len);
200 void xchk_xref_is_inode_chunk(struct xfs_scrub *sc, xfs_agblock_t agbno,
201 		xfs_extlen_t len);
202 void xchk_xref_is_only_owned_by(struct xfs_scrub *sc, xfs_agblock_t agbno,
203 		xfs_extlen_t len, const struct xfs_owner_info *oinfo);
204 void xchk_xref_is_not_owned_by(struct xfs_scrub *sc, xfs_agblock_t agbno,
205 		xfs_extlen_t len, const struct xfs_owner_info *oinfo);
206 void xchk_xref_has_no_owner(struct xfs_scrub *sc, xfs_agblock_t agbno,
207 		xfs_extlen_t len);
208 void xchk_xref_is_cow_staging(struct xfs_scrub *sc, xfs_agblock_t bno,
209 		xfs_extlen_t len);
210 void xchk_xref_is_not_shared(struct xfs_scrub *sc, xfs_agblock_t bno,
211 		xfs_extlen_t len);
212 void xchk_xref_is_not_cow_staging(struct xfs_scrub *sc, xfs_agblock_t bno,
213 		xfs_extlen_t len);
214 #ifdef CONFIG_XFS_RT
215 void xchk_xref_is_used_rt_space(struct xfs_scrub *sc, xfs_rtblock_t rtbno,
216 		xfs_extlen_t len);
217 #else
218 # define xchk_xref_is_used_rt_space(sc, rtbno, len) do { } while (0)
219 #endif
220 
221 #endif	/* __XFS_SCRUB_SCRUB_H__ */
222