xref: /linux/fs/xfs/scrub/iscan.h (revision 90d32e92011eaae8e70a9169b4e7acf4ca8f9d3a)
1 /* SPDX-License-Identifier: GPL-2.0-or-later */
2 /*
3  * Copyright (c) 2021-2024 Oracle.  All Rights Reserved.
4  * Author: Darrick J. Wong <djwong@kernel.org>
5  */
6 #ifndef __XFS_SCRUB_ISCAN_H__
7 #define __XFS_SCRUB_ISCAN_H__
8 
9 struct xchk_iscan {
10 	struct xfs_scrub	*sc;
11 
12 	/* Lock to protect the scan cursor. */
13 	struct mutex		lock;
14 
15 	/*
16 	 * This is the first inode in the inumber address space that we
17 	 * examined.  When the scan wraps around back to here, the scan is
18 	 * finished.
19 	 */
20 	xfs_ino_t		scan_start_ino;
21 
22 	/* This is the inode that will be examined next. */
23 	xfs_ino_t		cursor_ino;
24 
25 	/* If nonzero and non-NULL, skip this inode when scanning. */
26 	xfs_ino_t		skip_ino;
27 
28 	/*
29 	 * This is the last inode that we've successfully scanned, either
30 	 * because the caller scanned it, or we moved the cursor past an empty
31 	 * part of the inode address space.  Scan callers should only use the
32 	 * xchk_iscan_visit function to modify this.
33 	 */
34 	xfs_ino_t		__visited_ino;
35 
36 	/* Operational state of the livescan. */
37 	unsigned long		__opstate;
38 
39 	/* Give up on iterating @cursor_ino if we can't iget it by this time. */
40 	unsigned long		__iget_deadline;
41 
42 	/* Amount of time (in ms) that we will try to iget an inode. */
43 	unsigned int		iget_timeout;
44 
45 	/* Wait this many ms to retry an iget. */
46 	unsigned int		iget_retry_delay;
47 
48 	/*
49 	 * The scan grabs batches of inodes and stashes them here before
50 	 * handing them out with _iter.  Unallocated inodes are set in the
51 	 * mask so that all updates to that inode are selected for live
52 	 * update propagation.
53 	 */
54 	xfs_ino_t		__batch_ino;
55 	xfs_inofree_t		__skipped_inomask;
56 	struct xfs_inode	*__inodes[XFS_INODES_PER_CHUNK];
57 };
58 
59 /* Set if the scan has been aborted due to some event in the fs. */
60 #define XCHK_ISCAN_OPSTATE_ABORTED	(1)
61 
62 /* Use trylock to acquire the AGI */
63 #define XCHK_ISCAN_OPSTATE_TRYLOCK_AGI	(2)
64 
65 static inline bool
66 xchk_iscan_aborted(const struct xchk_iscan *iscan)
67 {
68 	return test_bit(XCHK_ISCAN_OPSTATE_ABORTED, &iscan->__opstate);
69 }
70 
71 static inline void
72 xchk_iscan_abort(struct xchk_iscan *iscan)
73 {
74 	set_bit(XCHK_ISCAN_OPSTATE_ABORTED, &iscan->__opstate);
75 }
76 
77 static inline bool
78 xchk_iscan_agi_needs_trylock(const struct xchk_iscan *iscan)
79 {
80 	return test_bit(XCHK_ISCAN_OPSTATE_TRYLOCK_AGI, &iscan->__opstate);
81 }
82 
83 static inline void
84 xchk_iscan_set_agi_trylock(struct xchk_iscan *iscan)
85 {
86 	set_bit(XCHK_ISCAN_OPSTATE_TRYLOCK_AGI, &iscan->__opstate);
87 }
88 
89 void xchk_iscan_start(struct xfs_scrub *sc, unsigned int iget_timeout,
90 		unsigned int iget_retry_delay, struct xchk_iscan *iscan);
91 void xchk_iscan_finish_early(struct xchk_iscan *iscan);
92 void xchk_iscan_teardown(struct xchk_iscan *iscan);
93 
94 int xchk_iscan_iter(struct xchk_iscan *iscan, struct xfs_inode **ipp);
95 void xchk_iscan_iter_finish(struct xchk_iscan *iscan);
96 
97 void xchk_iscan_mark_visited(struct xchk_iscan *iscan, struct xfs_inode *ip);
98 bool xchk_iscan_want_live_update(struct xchk_iscan *iscan, xfs_ino_t ino);
99 
100 #endif /* __XFS_SCRUB_ISCAN_H__ */
101