1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Copyright (C) 2017-2023 Oracle. All Rights Reserved. 4 * Author: Darrick J. Wong <djwong@kernel.org> 5 */ 6 #ifndef __XFS_SCRUB_COMMON_H__ 7 #define __XFS_SCRUB_COMMON_H__ 8 9 /* 10 * We /could/ terminate a scrub/repair operation early. If we're not 11 * in a good place to continue (fatal signal, etc.) then bail out. 12 * Note that we're careful not to make any judgements about *error. 13 */ 14 static inline bool 15 xchk_should_terminate( 16 struct xfs_scrub *sc, 17 int *error) 18 { 19 /* 20 * If preemption is disabled, we need to yield to the scheduler every 21 * few seconds so that we don't run afoul of the soft lockup watchdog 22 * or RCU stall detector. 23 */ 24 cond_resched(); 25 26 if (fatal_signal_pending(current)) { 27 if (*error == 0) 28 *error = -EINTR; 29 return true; 30 } 31 return false; 32 } 33 34 int xchk_trans_alloc(struct xfs_scrub *sc, uint resblks); 35 void xchk_trans_cancel(struct xfs_scrub *sc); 36 37 bool xchk_process_error(struct xfs_scrub *sc, xfs_agnumber_t agno, 38 xfs_agblock_t bno, int *error); 39 bool xchk_fblock_process_error(struct xfs_scrub *sc, int whichfork, 40 xfs_fileoff_t offset, int *error); 41 42 bool xchk_xref_process_error(struct xfs_scrub *sc, 43 xfs_agnumber_t agno, xfs_agblock_t bno, int *error); 44 bool xchk_fblock_xref_process_error(struct xfs_scrub *sc, 45 int whichfork, xfs_fileoff_t offset, int *error); 46 47 void xchk_block_set_preen(struct xfs_scrub *sc, 48 struct xfs_buf *bp); 49 void xchk_ino_set_preen(struct xfs_scrub *sc, xfs_ino_t ino); 50 51 void xchk_set_corrupt(struct xfs_scrub *sc); 52 void xchk_block_set_corrupt(struct xfs_scrub *sc, 53 struct xfs_buf *bp); 54 void xchk_ino_set_corrupt(struct xfs_scrub *sc, xfs_ino_t ino); 55 void xchk_fblock_set_corrupt(struct xfs_scrub *sc, int whichfork, 56 xfs_fileoff_t offset); 57 58 void xchk_block_xref_set_corrupt(struct xfs_scrub *sc, 59 struct xfs_buf *bp); 60 void xchk_ino_xref_set_corrupt(struct xfs_scrub *sc, 61 xfs_ino_t ino); 62 void xchk_fblock_xref_set_corrupt(struct xfs_scrub *sc, 63 int whichfork, xfs_fileoff_t offset); 64 65 void xchk_ino_set_warning(struct xfs_scrub *sc, xfs_ino_t ino); 66 void xchk_fblock_set_warning(struct xfs_scrub *sc, int whichfork, 67 xfs_fileoff_t offset); 68 69 void xchk_set_incomplete(struct xfs_scrub *sc); 70 int xchk_checkpoint_log(struct xfs_mount *mp); 71 72 /* Are we set up for a cross-referencing check? */ 73 bool xchk_should_check_xref(struct xfs_scrub *sc, int *error, 74 struct xfs_btree_cur **curpp); 75 76 /* Setup functions */ 77 int xchk_setup_agheader(struct xfs_scrub *sc); 78 int xchk_setup_fs(struct xfs_scrub *sc); 79 int xchk_setup_ag_allocbt(struct xfs_scrub *sc); 80 int xchk_setup_ag_iallocbt(struct xfs_scrub *sc); 81 int xchk_setup_ag_rmapbt(struct xfs_scrub *sc); 82 int xchk_setup_ag_refcountbt(struct xfs_scrub *sc); 83 int xchk_setup_inode(struct xfs_scrub *sc); 84 int xchk_setup_inode_bmap(struct xfs_scrub *sc); 85 int xchk_setup_inode_bmap_data(struct xfs_scrub *sc); 86 int xchk_setup_directory(struct xfs_scrub *sc); 87 int xchk_setup_xattr(struct xfs_scrub *sc); 88 int xchk_setup_symlink(struct xfs_scrub *sc); 89 int xchk_setup_parent(struct xfs_scrub *sc); 90 #ifdef CONFIG_XFS_RT 91 int xchk_setup_rtbitmap(struct xfs_scrub *sc); 92 int xchk_setup_rtsummary(struct xfs_scrub *sc); 93 #else 94 static inline int 95 xchk_setup_rtbitmap(struct xfs_scrub *sc) 96 { 97 return -ENOENT; 98 } 99 static inline int 100 xchk_setup_rtsummary(struct xfs_scrub *sc) 101 { 102 return -ENOENT; 103 } 104 #endif 105 #ifdef CONFIG_XFS_QUOTA 106 int xchk_ino_dqattach(struct xfs_scrub *sc); 107 int xchk_setup_quota(struct xfs_scrub *sc); 108 #else 109 static inline int 110 xchk_ino_dqattach(struct xfs_scrub *sc) 111 { 112 return 0; 113 } 114 static inline int 115 xchk_setup_quota(struct xfs_scrub *sc) 116 { 117 return -ENOENT; 118 } 119 #endif 120 int xchk_setup_fscounters(struct xfs_scrub *sc); 121 122 void xchk_ag_free(struct xfs_scrub *sc, struct xchk_ag *sa); 123 int xchk_ag_init(struct xfs_scrub *sc, xfs_agnumber_t agno, 124 struct xchk_ag *sa); 125 126 /* 127 * Grab all AG resources, treating the inability to grab the perag structure as 128 * a fs corruption. This is intended for callers checking an ondisk reference 129 * to a given AG, which means that the AG must still exist. 130 */ 131 static inline int 132 xchk_ag_init_existing( 133 struct xfs_scrub *sc, 134 xfs_agnumber_t agno, 135 struct xchk_ag *sa) 136 { 137 int error = xchk_ag_init(sc, agno, sa); 138 139 return error == -ENOENT ? -EFSCORRUPTED : error; 140 } 141 142 int xchk_ag_read_headers(struct xfs_scrub *sc, xfs_agnumber_t agno, 143 struct xchk_ag *sa); 144 void xchk_ag_btcur_free(struct xchk_ag *sa); 145 void xchk_ag_btcur_init(struct xfs_scrub *sc, struct xchk_ag *sa); 146 int xchk_count_rmap_ownedby_ag(struct xfs_scrub *sc, struct xfs_btree_cur *cur, 147 const struct xfs_owner_info *oinfo, xfs_filblks_t *blocks); 148 149 int xchk_setup_ag_btree(struct xfs_scrub *sc, bool force_log); 150 int xchk_iget_for_scrubbing(struct xfs_scrub *sc); 151 int xchk_setup_inode_contents(struct xfs_scrub *sc, unsigned int resblks); 152 int xchk_install_live_inode(struct xfs_scrub *sc, struct xfs_inode *ip); 153 154 void xchk_ilock(struct xfs_scrub *sc, unsigned int ilock_flags); 155 bool xchk_ilock_nowait(struct xfs_scrub *sc, unsigned int ilock_flags); 156 void xchk_iunlock(struct xfs_scrub *sc, unsigned int ilock_flags); 157 158 void xchk_buffer_recheck(struct xfs_scrub *sc, struct xfs_buf *bp); 159 160 /* 161 * Grab the inode at @inum. The caller must have created a scrub transaction 162 * so that we can confirm the inumber by walking the inobt and not deadlock on 163 * a loop in the inobt. 164 */ 165 int xchk_iget(struct xfs_scrub *sc, xfs_ino_t inum, struct xfs_inode **ipp); 166 int xchk_iget_agi(struct xfs_scrub *sc, xfs_ino_t inum, 167 struct xfs_buf **agi_bpp, struct xfs_inode **ipp); 168 void xchk_irele(struct xfs_scrub *sc, struct xfs_inode *ip); 169 int xchk_install_handle_inode(struct xfs_scrub *sc, struct xfs_inode *ip); 170 171 /* 172 * Safe version of (untrusted) xchk_iget that uses an empty transaction to 173 * avoid deadlocking on loops in the inobt. This should only be used in a 174 * scrub or repair setup routine, and only prior to grabbing a transaction. 175 */ 176 static inline int 177 xchk_iget_safe(struct xfs_scrub *sc, xfs_ino_t inum, struct xfs_inode **ipp) 178 { 179 int error; 180 181 ASSERT(sc->tp == NULL); 182 183 error = xchk_trans_alloc(sc, 0); 184 if (error) 185 return error; 186 error = xchk_iget(sc, inum, ipp); 187 xchk_trans_cancel(sc); 188 return error; 189 } 190 191 /* 192 * Don't bother cross-referencing if we already found corruption or cross 193 * referencing discrepancies. 194 */ 195 static inline bool xchk_skip_xref(struct xfs_scrub_metadata *sm) 196 { 197 return sm->sm_flags & (XFS_SCRUB_OFLAG_CORRUPT | 198 XFS_SCRUB_OFLAG_XCORRUPT); 199 } 200 201 bool xchk_dir_looks_zapped(struct xfs_inode *dp); 202 203 #ifdef CONFIG_XFS_ONLINE_REPAIR 204 /* Decide if a repair is required. */ 205 static inline bool xchk_needs_repair(const struct xfs_scrub_metadata *sm) 206 { 207 return sm->sm_flags & (XFS_SCRUB_OFLAG_CORRUPT | 208 XFS_SCRUB_OFLAG_XCORRUPT | 209 XFS_SCRUB_OFLAG_PREEN); 210 } 211 212 /* 213 * "Should we prepare for a repair?" 214 * 215 * Return true if the caller permits us to repair metadata and we're not 216 * setting up for a post-repair evaluation. 217 */ 218 static inline bool xchk_could_repair(const struct xfs_scrub *sc) 219 { 220 return (sc->sm->sm_flags & XFS_SCRUB_IFLAG_REPAIR) && 221 !(sc->flags & XREP_ALREADY_FIXED); 222 } 223 #else 224 # define xchk_needs_repair(sc) (false) 225 # define xchk_could_repair(sc) (false) 226 #endif /* CONFIG_XFS_ONLINE_REPAIR */ 227 228 int xchk_metadata_inode_forks(struct xfs_scrub *sc); 229 230 /* 231 * Helper macros to allocate and format xfile description strings. 232 * Callers must kfree the pointer returned. 233 */ 234 #define xchk_xfile_descr(sc, fmt, ...) \ 235 kasprintf(XCHK_GFP_FLAGS, "XFS (%s): " fmt, \ 236 (sc)->mp->m_super->s_id, ##__VA_ARGS__) 237 #define xchk_xfile_ag_descr(sc, fmt, ...) \ 238 kasprintf(XCHK_GFP_FLAGS, "XFS (%s): AG 0x%x " fmt, \ 239 (sc)->mp->m_super->s_id, \ 240 (sc)->sa.pag ? (sc)->sa.pag->pag_agno : (sc)->sm->sm_agno, \ 241 ##__VA_ARGS__) 242 #define xchk_xfile_ino_descr(sc, fmt, ...) \ 243 kasprintf(XCHK_GFP_FLAGS, "XFS (%s): inode 0x%llx " fmt, \ 244 (sc)->mp->m_super->s_id, \ 245 (sc)->ip ? (sc)->ip->i_ino : (sc)->sm->sm_ino, \ 246 ##__VA_ARGS__) 247 248 /* 249 * Setting up a hook to wait for intents to drain is costly -- we have to take 250 * the CPU hotplug lock and force an i-cache flush on all CPUs once to set it 251 * up, and again to tear it down. These costs add up quickly, so we only want 252 * to enable the drain waiter if the drain actually detected a conflict with 253 * running intent chains. 254 */ 255 static inline bool xchk_need_intent_drain(struct xfs_scrub *sc) 256 { 257 return sc->flags & XCHK_NEED_DRAIN; 258 } 259 260 void xchk_fsgates_enable(struct xfs_scrub *sc, unsigned int scrub_fshooks); 261 262 int xchk_inode_is_allocated(struct xfs_scrub *sc, xfs_agino_t agino, 263 bool *inuse); 264 265 #endif /* __XFS_SCRUB_COMMON_H__ */ 266