xref: /linux/fs/xfs/scrub/tempfile.c (revision 55d0969c451159cff86949b38c39171cab962069)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * Copyright (c) 2021-2024 Oracle.  All Rights Reserved.
4  * Author: Darrick J. Wong <djwong@kernel.org>
5  */
6 #include "xfs.h"
7 #include "xfs_fs.h"
8 #include "xfs_shared.h"
9 #include "xfs_format.h"
10 #include "xfs_trans_resv.h"
11 #include "xfs_mount.h"
12 #include "xfs_log_format.h"
13 #include "xfs_trans.h"
14 #include "xfs_inode.h"
15 #include "xfs_ialloc.h"
16 #include "xfs_quota.h"
17 #include "xfs_bmap.h"
18 #include "xfs_bmap_btree.h"
19 #include "xfs_trans_space.h"
20 #include "xfs_dir2.h"
21 #include "xfs_exchrange.h"
22 #include "xfs_exchmaps.h"
23 #include "xfs_defer.h"
24 #include "xfs_symlink_remote.h"
25 #include "scrub/scrub.h"
26 #include "scrub/common.h"
27 #include "scrub/repair.h"
28 #include "scrub/trace.h"
29 #include "scrub/tempfile.h"
30 #include "scrub/tempexch.h"
31 #include "scrub/xfile.h"
32 
33 /*
34  * Create a temporary file for reconstructing metadata, with the intention of
35  * atomically exchanging the temporary file's contents with the file that's
36  * being repaired.
37  */
38 int
39 xrep_tempfile_create(
40 	struct xfs_scrub	*sc,
41 	uint16_t		mode)
42 {
43 	struct xfs_icreate_args	args = {
44 		.pip		= sc->mp->m_rootip,
45 		.mode		= mode,
46 		.flags		= XFS_ICREATE_TMPFILE | XFS_ICREATE_UNLINKABLE,
47 	};
48 	struct xfs_mount	*mp = sc->mp;
49 	struct xfs_trans	*tp = NULL;
50 	struct xfs_dquot	*udqp;
51 	struct xfs_dquot	*gdqp;
52 	struct xfs_dquot	*pdqp;
53 	struct xfs_trans_res	*tres;
54 	struct xfs_inode	*dp = mp->m_rootip;
55 	xfs_ino_t		ino;
56 	unsigned int		resblks;
57 	bool			is_dir = S_ISDIR(mode);
58 	int			error;
59 
60 	if (xfs_is_shutdown(mp))
61 		return -EIO;
62 	if (xfs_is_readonly(mp))
63 		return -EROFS;
64 
65 	ASSERT(sc->tp == NULL);
66 	ASSERT(sc->tempip == NULL);
67 
68 	/*
69 	 * Make sure that we have allocated dquot(s) on disk.  The temporary
70 	 * inode should be completely root owned so that we don't fail due to
71 	 * quota limits.
72 	 */
73 	error = xfs_icreate_dqalloc(&args, &udqp, &gdqp, &pdqp);
74 	if (error)
75 		return error;
76 
77 	if (is_dir) {
78 		resblks = xfs_mkdir_space_res(mp, 0);
79 		tres = &M_RES(mp)->tr_mkdir;
80 	} else {
81 		resblks = XFS_IALLOC_SPACE_RES(mp);
82 		tres = &M_RES(mp)->tr_create_tmpfile;
83 	}
84 
85 	error = xfs_trans_alloc_icreate(mp, tres, udqp, gdqp, pdqp, resblks,
86 			&tp);
87 	if (error)
88 		goto out_release_dquots;
89 
90 	/* Allocate inode, set up directory. */
91 	error = xfs_dialloc(&tp, &args, &ino);
92 	if (error)
93 		goto out_trans_cancel;
94 	error = xfs_icreate(tp, ino, &args, &sc->tempip);
95 	if (error)
96 		goto out_trans_cancel;
97 
98 	/* We don't touch file data, so drop the realtime flags. */
99 	sc->tempip->i_diflags &= ~(XFS_DIFLAG_REALTIME | XFS_DIFLAG_RTINHERIT);
100 	xfs_trans_log_inode(tp, sc->tempip, XFS_ILOG_CORE);
101 
102 	/*
103 	 * Mark our temporary file as private so that LSMs and the ACL code
104 	 * don't try to add their own metadata or reason about these files.
105 	 * The file should never be exposed to userspace.
106 	 */
107 	VFS_I(sc->tempip)->i_flags |= S_PRIVATE;
108 	VFS_I(sc->tempip)->i_opflags &= ~IOP_XATTR;
109 
110 	if (is_dir) {
111 		error = xfs_dir_init(tp, sc->tempip, dp);
112 		if (error)
113 			goto out_trans_cancel;
114 	} else if (S_ISLNK(VFS_I(sc->tempip)->i_mode)) {
115 		/*
116 		 * Initialize the temporary symlink with a meaningless target
117 		 * that won't trip the verifiers.  Repair must rewrite the
118 		 * target with meaningful content before swapping with the file
119 		 * being repaired.  A single-byte target will not write a
120 		 * remote target block, so the owner is irrelevant.
121 		 */
122 		error = xfs_symlink_write_target(tp, sc->tempip,
123 				sc->tempip->i_ino, ".", 1, 0, 0);
124 		if (error)
125 			goto out_trans_cancel;
126 	}
127 
128 	/*
129 	 * Attach the dquot(s) to the inodes and modify them incore.
130 	 * These ids of the inode couldn't have changed since the new
131 	 * inode has been locked ever since it was created.
132 	 */
133 	xfs_qm_vop_create_dqattach(tp, sc->tempip, udqp, gdqp, pdqp);
134 
135 	/*
136 	 * Put our temp file on the unlinked list so it's purged automatically.
137 	 * All file-based metadata being reconstructed using this file must be
138 	 * atomically exchanged with the original file because the contents
139 	 * here will be purged when the inode is dropped or log recovery cleans
140 	 * out the unlinked list.
141 	 */
142 	error = xfs_iunlink(tp, sc->tempip);
143 	if (error)
144 		goto out_trans_cancel;
145 
146 	error = xfs_trans_commit(tp);
147 	if (error)
148 		goto out_release_inode;
149 
150 	trace_xrep_tempfile_create(sc);
151 
152 	xfs_qm_dqrele(udqp);
153 	xfs_qm_dqrele(gdqp);
154 	xfs_qm_dqrele(pdqp);
155 
156 	/* Finish setting up the incore / vfs context. */
157 	xfs_iunlock(sc->tempip, XFS_ILOCK_EXCL);
158 	xfs_setup_iops(sc->tempip);
159 	xfs_finish_inode_setup(sc->tempip);
160 
161 	sc->temp_ilock_flags = 0;
162 	return error;
163 
164 out_trans_cancel:
165 	xfs_trans_cancel(tp);
166 out_release_inode:
167 	/*
168 	 * Wait until after the current transaction is aborted to finish the
169 	 * setup of the inode and release the inode.  This prevents recursive
170 	 * transactions and deadlocks from xfs_inactive.
171 	 */
172 	if (sc->tempip) {
173 		xfs_iunlock(sc->tempip, XFS_ILOCK_EXCL);
174 		xfs_finish_inode_setup(sc->tempip);
175 		xchk_irele(sc, sc->tempip);
176 	}
177 out_release_dquots:
178 	xfs_qm_dqrele(udqp);
179 	xfs_qm_dqrele(gdqp);
180 	xfs_qm_dqrele(pdqp);
181 
182 	return error;
183 }
184 
185 /* Take IOLOCK_EXCL on the temporary file, maybe. */
186 bool
187 xrep_tempfile_iolock_nowait(
188 	struct xfs_scrub	*sc)
189 {
190 	if (xfs_ilock_nowait(sc->tempip, XFS_IOLOCK_EXCL)) {
191 		sc->temp_ilock_flags |= XFS_IOLOCK_EXCL;
192 		return true;
193 	}
194 
195 	return false;
196 }
197 
198 /*
199  * Take the temporary file's IOLOCK while holding a different inode's IOLOCK.
200  * In theory nobody else should hold the tempfile's IOLOCK, but we use trylock
201  * to avoid deadlocks and lockdep complaints.
202  */
203 int
204 xrep_tempfile_iolock_polled(
205 	struct xfs_scrub	*sc)
206 {
207 	int			error = 0;
208 
209 	while (!xrep_tempfile_iolock_nowait(sc)) {
210 		if (xchk_should_terminate(sc, &error))
211 			return error;
212 		delay(1);
213 	}
214 
215 	return 0;
216 }
217 
218 /* Release IOLOCK_EXCL on the temporary file. */
219 void
220 xrep_tempfile_iounlock(
221 	struct xfs_scrub	*sc)
222 {
223 	xfs_iunlock(sc->tempip, XFS_IOLOCK_EXCL);
224 	sc->temp_ilock_flags &= ~XFS_IOLOCK_EXCL;
225 }
226 
227 /* Prepare the temporary file for metadata updates by grabbing ILOCK_EXCL. */
228 void
229 xrep_tempfile_ilock(
230 	struct xfs_scrub	*sc)
231 {
232 	sc->temp_ilock_flags |= XFS_ILOCK_EXCL;
233 	xfs_ilock(sc->tempip, XFS_ILOCK_EXCL);
234 }
235 
236 /* Try to grab ILOCK_EXCL on the temporary file. */
237 bool
238 xrep_tempfile_ilock_nowait(
239 	struct xfs_scrub	*sc)
240 {
241 	if (xfs_ilock_nowait(sc->tempip, XFS_ILOCK_EXCL)) {
242 		sc->temp_ilock_flags |= XFS_ILOCK_EXCL;
243 		return true;
244 	}
245 
246 	return false;
247 }
248 
249 /* Unlock ILOCK_EXCL on the temporary file after an update. */
250 void
251 xrep_tempfile_iunlock(
252 	struct xfs_scrub	*sc)
253 {
254 	xfs_iunlock(sc->tempip, XFS_ILOCK_EXCL);
255 	sc->temp_ilock_flags &= ~XFS_ILOCK_EXCL;
256 }
257 
258 /*
259  * Begin the process of making changes to both the file being scrubbed and
260  * the temporary file by taking ILOCK_EXCL on both.
261  */
262 void
263 xrep_tempfile_ilock_both(
264 	struct xfs_scrub	*sc)
265 {
266 	xfs_lock_two_inodes(sc->ip, XFS_ILOCK_EXCL, sc->tempip, XFS_ILOCK_EXCL);
267 	sc->ilock_flags |= XFS_ILOCK_EXCL;
268 	sc->temp_ilock_flags |= XFS_ILOCK_EXCL;
269 }
270 
271 /* Unlock ILOCK_EXCL on both files. */
272 void
273 xrep_tempfile_iunlock_both(
274 	struct xfs_scrub	*sc)
275 {
276 	xrep_tempfile_iunlock(sc);
277 	xchk_iunlock(sc, XFS_ILOCK_EXCL);
278 }
279 
280 /* Release the temporary file. */
281 void
282 xrep_tempfile_rele(
283 	struct xfs_scrub	*sc)
284 {
285 	if (!sc->tempip)
286 		return;
287 
288 	if (sc->temp_ilock_flags) {
289 		xfs_iunlock(sc->tempip, sc->temp_ilock_flags);
290 		sc->temp_ilock_flags = 0;
291 	}
292 
293 	xchk_irele(sc, sc->tempip);
294 	sc->tempip = NULL;
295 }
296 
297 /*
298  * Make sure that the given range of the data fork of the temporary file is
299  * mapped to written blocks.  The caller must ensure that both inodes are
300  * joined to the transaction.
301  */
302 int
303 xrep_tempfile_prealloc(
304 	struct xfs_scrub	*sc,
305 	xfs_fileoff_t		off,
306 	xfs_filblks_t		len)
307 {
308 	struct xfs_bmbt_irec	map;
309 	xfs_fileoff_t		end = off + len;
310 	int			error;
311 
312 	ASSERT(sc->tempip != NULL);
313 	ASSERT(!XFS_NOT_DQATTACHED(sc->mp, sc->tempip));
314 
315 	for (; off < end; off = map.br_startoff + map.br_blockcount) {
316 		int		nmaps = 1;
317 
318 		/*
319 		 * If we have a real extent mapping this block then we're
320 		 * in ok shape.
321 		 */
322 		error = xfs_bmapi_read(sc->tempip, off, end - off, &map, &nmaps,
323 				XFS_DATA_FORK);
324 		if (error)
325 			return error;
326 		if (nmaps == 0) {
327 			ASSERT(nmaps != 0);
328 			return -EFSCORRUPTED;
329 		}
330 
331 		if (xfs_bmap_is_written_extent(&map))
332 			continue;
333 
334 		/*
335 		 * If we find a delalloc reservation then something is very
336 		 * very wrong.  Bail out.
337 		 */
338 		if (map.br_startblock == DELAYSTARTBLOCK)
339 			return -EFSCORRUPTED;
340 
341 		/*
342 		 * Make sure this block has a real zeroed extent allocated to
343 		 * it.
344 		 */
345 		nmaps = 1;
346 		error = xfs_bmapi_write(sc->tp, sc->tempip, off, end - off,
347 				XFS_BMAPI_CONVERT | XFS_BMAPI_ZERO, 0, &map,
348 				&nmaps);
349 		if (error)
350 			return error;
351 		if (nmaps != 1)
352 			return -EFSCORRUPTED;
353 
354 		trace_xrep_tempfile_prealloc(sc, XFS_DATA_FORK, &map);
355 
356 		/* Commit new extent and all deferred work. */
357 		error = xfs_defer_finish(&sc->tp);
358 		if (error)
359 			return error;
360 	}
361 
362 	return 0;
363 }
364 
365 /*
366  * Write data to each block of a file.  The given range of the tempfile's data
367  * fork must already be populated with written extents.
368  */
369 int
370 xrep_tempfile_copyin(
371 	struct xfs_scrub	*sc,
372 	xfs_fileoff_t		off,
373 	xfs_filblks_t		len,
374 	xrep_tempfile_copyin_fn	prep_fn,
375 	void			*data)
376 {
377 	LIST_HEAD(buffers_list);
378 	struct xfs_mount	*mp = sc->mp;
379 	struct xfs_buf		*bp;
380 	xfs_fileoff_t		flush_mask;
381 	xfs_fileoff_t		end = off + len;
382 	loff_t			pos = XFS_FSB_TO_B(mp, off);
383 	int			error = 0;
384 
385 	ASSERT(S_ISREG(VFS_I(sc->tempip)->i_mode));
386 
387 	/* Flush buffers to disk every 512K */
388 	flush_mask = XFS_B_TO_FSBT(mp, (1U << 19)) - 1;
389 
390 	for (; off < end; off++, pos += mp->m_sb.sb_blocksize) {
391 		struct xfs_bmbt_irec	map;
392 		int			nmaps = 1;
393 
394 		/* Read block mapping for this file block. */
395 		error = xfs_bmapi_read(sc->tempip, off, 1, &map, &nmaps, 0);
396 		if (error)
397 			goto out_err;
398 		if (nmaps == 0 || !xfs_bmap_is_written_extent(&map)) {
399 			error = -EFSCORRUPTED;
400 			goto out_err;
401 		}
402 
403 		/* Get the metadata buffer for this offset in the file. */
404 		error = xfs_trans_get_buf(sc->tp, mp->m_ddev_targp,
405 				XFS_FSB_TO_DADDR(mp, map.br_startblock),
406 				mp->m_bsize, 0, &bp);
407 		if (error)
408 			goto out_err;
409 
410 		trace_xrep_tempfile_copyin(sc, XFS_DATA_FORK, &map);
411 
412 		/* Read in a block's worth of data from the xfile. */
413 		error = prep_fn(sc, bp, data);
414 		if (error) {
415 			xfs_trans_brelse(sc->tp, bp);
416 			goto out_err;
417 		}
418 
419 		/* Queue buffer, and flush if we have too much dirty data. */
420 		xfs_buf_delwri_queue_here(bp, &buffers_list);
421 		xfs_trans_brelse(sc->tp, bp);
422 
423 		if (!(off & flush_mask)) {
424 			error = xfs_buf_delwri_submit(&buffers_list);
425 			if (error)
426 				goto out_err;
427 		}
428 	}
429 
430 	/*
431 	 * Write the new blocks to disk.  If the ordered list isn't empty after
432 	 * that, then something went wrong and we have to fail.  This should
433 	 * never happen, but we'll check anyway.
434 	 */
435 	error = xfs_buf_delwri_submit(&buffers_list);
436 	if (error)
437 		goto out_err;
438 
439 	if (!list_empty(&buffers_list)) {
440 		ASSERT(list_empty(&buffers_list));
441 		error = -EIO;
442 		goto out_err;
443 	}
444 
445 	return 0;
446 
447 out_err:
448 	xfs_buf_delwri_cancel(&buffers_list);
449 	return error;
450 }
451 
452 /*
453  * Set the temporary file's size.  Caller must join the tempfile to the scrub
454  * transaction and is responsible for adjusting block mappings as needed.
455  */
456 int
457 xrep_tempfile_set_isize(
458 	struct xfs_scrub	*sc,
459 	unsigned long long	isize)
460 {
461 	if (sc->tempip->i_disk_size == isize)
462 		return 0;
463 
464 	sc->tempip->i_disk_size = isize;
465 	i_size_write(VFS_I(sc->tempip), isize);
466 	return xrep_tempfile_roll_trans(sc);
467 }
468 
469 /*
470  * Roll a repair transaction involving the temporary file.  Caller must join
471  * both the temporary file and the file being scrubbed to the transaction.
472  * This function return with both inodes joined to a new scrub transaction,
473  * or the usual negative errno.
474  */
475 int
476 xrep_tempfile_roll_trans(
477 	struct xfs_scrub	*sc)
478 {
479 	int			error;
480 
481 	xfs_trans_log_inode(sc->tp, sc->tempip, XFS_ILOG_CORE);
482 	error = xrep_roll_trans(sc);
483 	if (error)
484 		return error;
485 
486 	xfs_trans_ijoin(sc->tp, sc->tempip, 0);
487 	return 0;
488 }
489 
490 /*
491  * Fill out the mapping exchange request in preparation for atomically
492  * committing the contents of a metadata file that we've rebuilt in the temp
493  * file.
494  */
495 STATIC int
496 xrep_tempexch_prep_request(
497 	struct xfs_scrub	*sc,
498 	int			whichfork,
499 	struct xrep_tempexch	*tx)
500 {
501 	struct xfs_exchmaps_req	*req = &tx->req;
502 
503 	memset(tx, 0, sizeof(struct xrep_tempexch));
504 
505 	/* COW forks don't exist on disk. */
506 	if (whichfork == XFS_COW_FORK) {
507 		ASSERT(0);
508 		return -EINVAL;
509 	}
510 
511 	/* Both files should have the relevant forks. */
512 	if (!xfs_ifork_ptr(sc->ip, whichfork) ||
513 	    !xfs_ifork_ptr(sc->tempip, whichfork)) {
514 		ASSERT(xfs_ifork_ptr(sc->ip, whichfork) != NULL);
515 		ASSERT(xfs_ifork_ptr(sc->tempip, whichfork) != NULL);
516 		return -EINVAL;
517 	}
518 
519 	/* Exchange all mappings in both forks. */
520 	req->ip1 = sc->tempip;
521 	req->ip2 = sc->ip;
522 	req->startoff1 = 0;
523 	req->startoff2 = 0;
524 	switch (whichfork) {
525 	case XFS_ATTR_FORK:
526 		req->flags |= XFS_EXCHMAPS_ATTR_FORK;
527 		break;
528 	case XFS_DATA_FORK:
529 		/* Always exchange sizes when exchanging data fork mappings. */
530 		req->flags |= XFS_EXCHMAPS_SET_SIZES;
531 		break;
532 	}
533 	req->blockcount = XFS_MAX_FILEOFF;
534 
535 	return 0;
536 }
537 
538 /*
539  * Fill out the mapping exchange resource estimation structures in preparation
540  * for exchanging the contents of a metadata file that we've rebuilt in the
541  * temp file.  Caller must hold IOLOCK_EXCL but not ILOCK_EXCL on both files.
542  */
543 STATIC int
544 xrep_tempexch_estimate(
545 	struct xfs_scrub	*sc,
546 	struct xrep_tempexch	*tx)
547 {
548 	struct xfs_exchmaps_req	*req = &tx->req;
549 	struct xfs_ifork	*ifp;
550 	struct xfs_ifork	*tifp;
551 	int			whichfork = xfs_exchmaps_reqfork(req);
552 	int			state = 0;
553 
554 	/*
555 	 * The exchmaps code only knows how to exchange file fork space
556 	 * mappings.  Any fork data in local format must be promoted to a
557 	 * single block before the exchange can take place.
558 	 */
559 	ifp = xfs_ifork_ptr(sc->ip, whichfork);
560 	if (ifp->if_format == XFS_DINODE_FMT_LOCAL)
561 		state |= 1;
562 
563 	tifp = xfs_ifork_ptr(sc->tempip, whichfork);
564 	if (tifp->if_format == XFS_DINODE_FMT_LOCAL)
565 		state |= 2;
566 
567 	switch (state) {
568 	case 0:
569 		/* Both files have mapped extents; use the regular estimate. */
570 		return xfs_exchrange_estimate(req);
571 	case 1:
572 		/*
573 		 * The file being repaired is in local format, but the temp
574 		 * file has mapped extents.  To perform the exchange, the file
575 		 * being repaired must have its shorform data converted to an
576 		 * ondisk block so that the forks will be in extents format.
577 		 * We need one resblk for the conversion; the number of
578 		 * exchanges is (worst case) the temporary file's extent count
579 		 * plus the block we converted.
580 		 */
581 		req->ip1_bcount = sc->tempip->i_nblocks;
582 		req->ip2_bcount = 1;
583 		req->nr_exchanges = 1 + tifp->if_nextents;
584 		req->resblks = 1;
585 		break;
586 	case 2:
587 		/*
588 		 * The temporary file is in local format, but the file being
589 		 * repaired has mapped extents.  To perform the exchange, the
590 		 * temp file must have its shortform data converted to an
591 		 * ondisk block, and the fork changed to extents format.  We
592 		 * need one resblk for the conversion; the number of exchanges
593 		 * is (worst case) the extent count of the file being repaired
594 		 * plus the block we converted.
595 		 */
596 		req->ip1_bcount = 1;
597 		req->ip2_bcount = sc->ip->i_nblocks;
598 		req->nr_exchanges = 1 + ifp->if_nextents;
599 		req->resblks = 1;
600 		break;
601 	case 3:
602 		/*
603 		 * Both forks are in local format.  To perform the exchange,
604 		 * both files must have their shortform data converted to
605 		 * fsblocks, and both forks must be converted to extents
606 		 * format.  We need two resblks for the two conversions, and
607 		 * the number of exchanges is 1 since there's only one block at
608 		 * fileoff 0.  Presumably, the caller could not exchange the
609 		 * two inode fork areas directly.
610 		 */
611 		req->ip1_bcount = 1;
612 		req->ip2_bcount = 1;
613 		req->nr_exchanges = 1;
614 		req->resblks = 2;
615 		break;
616 	}
617 
618 	return xfs_exchmaps_estimate_overhead(req);
619 }
620 
621 /*
622  * Obtain a quota reservation to make sure we don't hit EDQUOT.  We can skip
623  * this if quota enforcement is disabled or if both inodes' dquots are the
624  * same.  The qretry structure must be initialized to zeroes before the first
625  * call to this function.
626  */
627 STATIC int
628 xrep_tempexch_reserve_quota(
629 	struct xfs_scrub		*sc,
630 	const struct xrep_tempexch	*tx)
631 {
632 	struct xfs_trans		*tp = sc->tp;
633 	const struct xfs_exchmaps_req	*req = &tx->req;
634 	int64_t				ddelta, rdelta;
635 	int				error;
636 
637 	/*
638 	 * Don't bother with a quota reservation if we're not enforcing them
639 	 * or the two inodes have the same dquots.
640 	 */
641 	if (!XFS_IS_QUOTA_ON(tp->t_mountp) || req->ip1 == req->ip2 ||
642 	    (req->ip1->i_udquot == req->ip2->i_udquot &&
643 	     req->ip1->i_gdquot == req->ip2->i_gdquot &&
644 	     req->ip1->i_pdquot == req->ip2->i_pdquot))
645 		return 0;
646 
647 	/*
648 	 * Quota reservation for each file comes from two sources.  First, we
649 	 * need to account for any net gain in mapped blocks during the
650 	 * exchange.  Second, we need reservation for the gross gain in mapped
651 	 * blocks so that we don't trip over any quota block reservation
652 	 * assertions.  We must reserve the gross gain because the quota code
653 	 * subtracts from bcount the number of blocks that we unmap; it does
654 	 * not add that quantity back to the quota block reservation.
655 	 */
656 	ddelta = max_t(int64_t, 0, req->ip2_bcount - req->ip1_bcount);
657 	rdelta = max_t(int64_t, 0, req->ip2_rtbcount - req->ip1_rtbcount);
658 	error = xfs_trans_reserve_quota_nblks(tp, req->ip1,
659 			ddelta + req->ip1_bcount, rdelta + req->ip1_rtbcount,
660 			true);
661 	if (error)
662 		return error;
663 
664 	ddelta = max_t(int64_t, 0, req->ip1_bcount - req->ip2_bcount);
665 	rdelta = max_t(int64_t, 0, req->ip1_rtbcount - req->ip2_rtbcount);
666 	return xfs_trans_reserve_quota_nblks(tp, req->ip2,
667 			ddelta + req->ip2_bcount, rdelta + req->ip2_rtbcount,
668 			true);
669 }
670 
671 /*
672  * Prepare an existing transaction for an atomic file contents exchange.
673  *
674  * This function fills out the mapping exchange request and resource estimation
675  * structures in preparation for exchanging the contents of a metadata file
676  * that has been rebuilt in the temp file.  Next, it reserves space and quota
677  * for the transaction.
678  *
679  * The caller must hold ILOCK_EXCL of the scrub target file and the temporary
680  * file.  The caller must join both inodes to the transaction with no unlock
681  * flags, and is responsible for dropping both ILOCKs when appropriate.  Only
682  * use this when those ILOCKs cannot be dropped.
683  */
684 int
685 xrep_tempexch_trans_reserve(
686 	struct xfs_scrub	*sc,
687 	int			whichfork,
688 	struct xrep_tempexch	*tx)
689 {
690 	int			error;
691 
692 	ASSERT(sc->tp != NULL);
693 	xfs_assert_ilocked(sc->ip, XFS_ILOCK_EXCL);
694 	xfs_assert_ilocked(sc->tempip, XFS_ILOCK_EXCL);
695 
696 	error = xrep_tempexch_prep_request(sc, whichfork, tx);
697 	if (error)
698 		return error;
699 
700 	error = xfs_exchmaps_estimate(&tx->req);
701 	if (error)
702 		return error;
703 
704 	error = xfs_trans_reserve_more(sc->tp, tx->req.resblks, 0);
705 	if (error)
706 		return error;
707 
708 	return xrep_tempexch_reserve_quota(sc, tx);
709 }
710 
711 /*
712  * Create a new transaction for a file contents exchange.
713  *
714  * This function fills out the mapping excahange request and resource
715  * estimation structures in preparation for exchanging the contents of a
716  * metadata file that has been rebuilt in the temp file.  Next, it reserves
717  * space, takes ILOCK_EXCL of both inodes, joins them to the transaction and
718  * reserves quota for the transaction.
719  *
720  * The caller is responsible for dropping both ILOCKs when appropriate.
721  */
722 int
723 xrep_tempexch_trans_alloc(
724 	struct xfs_scrub	*sc,
725 	int			whichfork,
726 	struct xrep_tempexch	*tx)
727 {
728 	unsigned int		flags = 0;
729 	int			error;
730 
731 	ASSERT(sc->tp == NULL);
732 	ASSERT(xfs_has_exchange_range(sc->mp));
733 
734 	error = xrep_tempexch_prep_request(sc, whichfork, tx);
735 	if (error)
736 		return error;
737 
738 	error = xrep_tempexch_estimate(sc, tx);
739 	if (error)
740 		return error;
741 
742 	if (xfs_has_lazysbcount(sc->mp))
743 		flags |= XFS_TRANS_RES_FDBLKS;
744 
745 	error = xfs_trans_alloc(sc->mp, &M_RES(sc->mp)->tr_itruncate,
746 			tx->req.resblks, 0, flags, &sc->tp);
747 	if (error)
748 		return error;
749 
750 	sc->temp_ilock_flags |= XFS_ILOCK_EXCL;
751 	sc->ilock_flags |= XFS_ILOCK_EXCL;
752 	xfs_exchrange_ilock(sc->tp, sc->ip, sc->tempip);
753 
754 	return xrep_tempexch_reserve_quota(sc, tx);
755 }
756 
757 /*
758  * Exchange file mappings (and hence file contents) between the file being
759  * repaired and the temporary file.  Returns with both inodes locked and joined
760  * to a clean scrub transaction.
761  */
762 int
763 xrep_tempexch_contents(
764 	struct xfs_scrub	*sc,
765 	struct xrep_tempexch	*tx)
766 {
767 	int			error;
768 
769 	ASSERT(xfs_has_exchange_range(sc->mp));
770 
771 	xfs_exchange_mappings(sc->tp, &tx->req);
772 	error = xfs_defer_finish(&sc->tp);
773 	if (error)
774 		return error;
775 
776 	/*
777 	 * If we exchanged the ondisk sizes of two metadata files, we must
778 	 * exchanged the incore sizes as well.
779 	 */
780 	if (tx->req.flags & XFS_EXCHMAPS_SET_SIZES) {
781 		loff_t	temp;
782 
783 		temp = i_size_read(VFS_I(sc->ip));
784 		i_size_write(VFS_I(sc->ip), i_size_read(VFS_I(sc->tempip)));
785 		i_size_write(VFS_I(sc->tempip), temp);
786 	}
787 
788 	return 0;
789 }
790 
791 /*
792  * Write local format data from one of the temporary file's forks into the same
793  * fork of file being repaired, and exchange the file sizes, if appropriate.
794  * Caller must ensure that the file being repaired has enough fork space to
795  * hold all the bytes.
796  */
797 void
798 xrep_tempfile_copyout_local(
799 	struct xfs_scrub	*sc,
800 	int			whichfork)
801 {
802 	struct xfs_ifork	*temp_ifp;
803 	struct xfs_ifork	*ifp;
804 	unsigned int		ilog_flags = XFS_ILOG_CORE;
805 
806 	temp_ifp = xfs_ifork_ptr(sc->tempip, whichfork);
807 	ifp = xfs_ifork_ptr(sc->ip, whichfork);
808 
809 	ASSERT(temp_ifp != NULL);
810 	ASSERT(ifp != NULL);
811 	ASSERT(temp_ifp->if_format == XFS_DINODE_FMT_LOCAL);
812 	ASSERT(ifp->if_format == XFS_DINODE_FMT_LOCAL);
813 
814 	switch (whichfork) {
815 	case XFS_DATA_FORK:
816 		ASSERT(sc->tempip->i_disk_size <=
817 					xfs_inode_data_fork_size(sc->ip));
818 		break;
819 	case XFS_ATTR_FORK:
820 		ASSERT(sc->tempip->i_forkoff >= sc->ip->i_forkoff);
821 		break;
822 	default:
823 		ASSERT(0);
824 		return;
825 	}
826 
827 	/* Recreate @sc->ip's incore fork (ifp) with data from temp_ifp. */
828 	xfs_idestroy_fork(ifp);
829 	xfs_init_local_fork(sc->ip, whichfork, temp_ifp->if_data,
830 			temp_ifp->if_bytes);
831 
832 	if (whichfork == XFS_DATA_FORK) {
833 		i_size_write(VFS_I(sc->ip), i_size_read(VFS_I(sc->tempip)));
834 		sc->ip->i_disk_size = sc->tempip->i_disk_size;
835 	}
836 
837 	ilog_flags |= xfs_ilog_fdata(whichfork);
838 	xfs_trans_log_inode(sc->tp, sc->ip, ilog_flags);
839 }
840 
841 /* Decide if a given XFS inode is a temporary file for a repair. */
842 bool
843 xrep_is_tempfile(
844 	const struct xfs_inode	*ip)
845 {
846 	const struct inode	*inode = &ip->i_vnode;
847 
848 	if (IS_PRIVATE(inode) && !(inode->i_opflags & IOP_XATTR))
849 		return true;
850 
851 	return false;
852 }
853