1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Copyright (c) 2000-2003,2005 Silicon Graphics, Inc.
4 * Copyright (C) 2010 Red Hat, Inc.
5 * All Rights Reserved.
6 */
7 #include "xfs.h"
8 #include "xfs_fs.h"
9 #include "xfs_shared.h"
10 #include "xfs_format.h"
11 #include "xfs_log_format.h"
12 #include "xfs_trans_resv.h"
13 #include "xfs_mount.h"
14 #include "xfs_da_format.h"
15 #include "xfs_da_btree.h"
16 #include "xfs_inode.h"
17 #include "xfs_bmap_btree.h"
18 #include "xfs_quota.h"
19 #include "xfs_trans.h"
20 #include "xfs_qm.h"
21 #include "xfs_trans_space.h"
22 #include "xfs_rtbitmap.h"
23 #include "xfs_attr_item.h"
24 #include "xfs_log.h"
25
26 #define _ALLOC true
27 #define _FREE false
28
29 /*
30 * A buffer has a format structure overhead in the log in addition
31 * to the data, so we need to take this into account when reserving
32 * space in a transaction for a buffer. Round the space required up
33 * to a multiple of 128 bytes so that we don't change the historical
34 * reservation that has been used for this overhead.
35 */
36 STATIC uint
xfs_buf_log_overhead(void)37 xfs_buf_log_overhead(void)
38 {
39 return round_up(sizeof(struct xlog_op_header) +
40 sizeof(struct xfs_buf_log_format), 128);
41 }
42
43 /*
44 * Calculate out transaction log reservation per item in bytes.
45 *
46 * The nbufs argument is used to indicate the number of items that
47 * will be changed in a transaction. size is used to tell how many
48 * bytes should be reserved per item.
49 */
50 STATIC uint
xfs_calc_buf_res(uint nbufs,uint size)51 xfs_calc_buf_res(
52 uint nbufs,
53 uint size)
54 {
55 return nbufs * (size + xfs_buf_log_overhead());
56 }
57
58 /*
59 * Per-extent log reservation for the btree changes involved in freeing or
60 * allocating an extent. In classic XFS there were two trees that will be
61 * modified (bnobt + cntbt). With rmap enabled, there are three trees
62 * (rmapbt). The number of blocks reserved is based on the formula:
63 *
64 * num trees * ((2 blocks/level * max depth) - 1)
65 *
66 * Keep in mind that max depth is calculated separately for each type of tree.
67 */
68 uint
xfs_allocfree_block_count(struct xfs_mount * mp,uint num_ops)69 xfs_allocfree_block_count(
70 struct xfs_mount *mp,
71 uint num_ops)
72 {
73 uint blocks;
74
75 blocks = num_ops * 2 * (2 * mp->m_alloc_maxlevels - 1);
76 if (xfs_has_rmapbt(mp))
77 blocks += num_ops * (2 * mp->m_rmap_maxlevels - 1);
78
79 return blocks;
80 }
81
82 /*
83 * Per-extent log reservation for refcount btree changes. These are never done
84 * in the same transaction as an allocation or a free, so we compute them
85 * separately.
86 */
87 static unsigned int
xfs_refcountbt_block_count(struct xfs_mount * mp,unsigned int num_ops)88 xfs_refcountbt_block_count(
89 struct xfs_mount *mp,
90 unsigned int num_ops)
91 {
92 return num_ops * (2 * mp->m_refc_maxlevels - 1);
93 }
94
95 /*
96 * Logging inodes is really tricksy. They are logged in memory format,
97 * which means that what we write into the log doesn't directly translate into
98 * the amount of space they use on disk.
99 *
100 * Case in point - btree format forks in memory format use more space than the
101 * on-disk format. In memory, the buffer contains a normal btree block header so
102 * the btree code can treat it as though it is just another generic buffer.
103 * However, when we write it to the inode fork, we don't write all of this
104 * header as it isn't needed. e.g. the root is only ever in the inode, so
105 * there's no need for sibling pointers which would waste 16 bytes of space.
106 *
107 * Hence when we have an inode with a maximally sized btree format fork, then
108 * amount of information we actually log is greater than the size of the inode
109 * on disk. Hence we need an inode reservation function that calculates all this
110 * correctly. So, we log:
111 *
112 * - 4 log op headers for object
113 * - for the ilf, the inode core and 2 forks
114 * - inode log format object
115 * - the inode core
116 * - two inode forks containing bmap btree root blocks.
117 * - the btree data contained by both forks will fit into the inode size,
118 * hence when combined with the inode core above, we have a total of the
119 * actual inode size.
120 * - the BMBT headers need to be accounted separately, as they are
121 * additional to the records and pointers that fit inside the inode
122 * forks.
123 */
124 STATIC uint
xfs_calc_inode_res(struct xfs_mount * mp,uint ninodes)125 xfs_calc_inode_res(
126 struct xfs_mount *mp,
127 uint ninodes)
128 {
129 return ninodes *
130 (4 * sizeof(struct xlog_op_header) +
131 sizeof(struct xfs_inode_log_format) +
132 mp->m_sb.sb_inodesize +
133 2 * xfs_bmbt_block_len(mp));
134 }
135
136 /*
137 * Inode btree record insertion/removal modifies the inode btree and free space
138 * btrees (since the inobt does not use the agfl). This requires the following
139 * reservation:
140 *
141 * the inode btree: max depth * blocksize
142 * the allocation btrees: 2 trees * (max depth - 1) * block size
143 *
144 * The caller must account for SB and AG header modifications, etc.
145 */
146 STATIC uint
xfs_calc_inobt_res(struct xfs_mount * mp)147 xfs_calc_inobt_res(
148 struct xfs_mount *mp)
149 {
150 return xfs_calc_buf_res(M_IGEO(mp)->inobt_maxlevels,
151 XFS_FSB_TO_B(mp, 1)) +
152 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
153 XFS_FSB_TO_B(mp, 1));
154 }
155
156 /*
157 * The free inode btree is a conditional feature. The behavior differs slightly
158 * from that of the traditional inode btree in that the finobt tracks records
159 * for inode chunks with at least one free inode. A record can be removed from
160 * the tree during individual inode allocation. Therefore the finobt
161 * reservation is unconditional for both the inode chunk allocation and
162 * individual inode allocation (modify) cases.
163 *
164 * Behavior aside, the reservation for finobt modification is equivalent to the
165 * traditional inobt: cover a full finobt shape change plus block allocation.
166 */
167 STATIC uint
xfs_calc_finobt_res(struct xfs_mount * mp)168 xfs_calc_finobt_res(
169 struct xfs_mount *mp)
170 {
171 if (!xfs_has_finobt(mp))
172 return 0;
173
174 return xfs_calc_inobt_res(mp);
175 }
176
177 /*
178 * Calculate the reservation required to allocate or free an inode chunk. This
179 * includes:
180 *
181 * the allocation btrees: 2 trees * (max depth - 1) * block size
182 * the inode chunk: m_ino_geo.ialloc_blks * N
183 *
184 * The size N of the inode chunk reservation depends on whether it is for
185 * allocation or free and which type of create transaction is in use. An inode
186 * chunk free always invalidates the buffers and only requires reservation for
187 * headers (N == 0). An inode chunk allocation requires a chunk sized
188 * reservation on v4 and older superblocks to initialize the chunk. No chunk
189 * reservation is required for allocation on v5 supers, which use ordered
190 * buffers to initialize.
191 */
192 STATIC uint
xfs_calc_inode_chunk_res(struct xfs_mount * mp,bool alloc)193 xfs_calc_inode_chunk_res(
194 struct xfs_mount *mp,
195 bool alloc)
196 {
197 uint res, size = 0;
198
199 res = xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
200 XFS_FSB_TO_B(mp, 1));
201 if (alloc) {
202 /* icreate tx uses ordered buffers */
203 if (xfs_has_v3inodes(mp))
204 return res;
205 size = XFS_FSB_TO_B(mp, 1);
206 }
207
208 res += xfs_calc_buf_res(M_IGEO(mp)->ialloc_blks, size);
209 return res;
210 }
211
212 /*
213 * Per-extent log reservation for the btree changes involved in freeing or
214 * allocating a realtime extent. We have to be able to log as many rtbitmap
215 * blocks as needed to mark inuse XFS_BMBT_MAX_EXTLEN blocks' worth of realtime
216 * extents, as well as the realtime summary block.
217 */
218 static unsigned int
xfs_rtalloc_block_count(struct xfs_mount * mp,unsigned int num_ops)219 xfs_rtalloc_block_count(
220 struct xfs_mount *mp,
221 unsigned int num_ops)
222 {
223 unsigned int rtbmp_blocks;
224 xfs_rtxlen_t rtxlen;
225
226 rtxlen = xfs_extlen_to_rtxlen(mp, XFS_MAX_BMBT_EXTLEN);
227 rtbmp_blocks = xfs_rtbitmap_blockcount(mp, rtxlen);
228 return (rtbmp_blocks + 1) * num_ops;
229 }
230
231 /*
232 * Various log reservation values.
233 *
234 * These are based on the size of the file system block because that is what
235 * most transactions manipulate. Each adds in an additional 128 bytes per
236 * item logged to try to account for the overhead of the transaction mechanism.
237 *
238 * Note: Most of the reservations underestimate the number of allocation
239 * groups into which they could free extents in the xfs_defer_finish() call.
240 * This is because the number in the worst case is quite high and quite
241 * unusual. In order to fix this we need to change xfs_defer_finish() to free
242 * extents in only a single AG at a time. This will require changes to the
243 * EFI code as well, however, so that the EFI for the extents not freed is
244 * logged again in each transaction. See SGI PV #261917.
245 *
246 * Reservation functions here avoid a huge stack in xfs_trans_init due to
247 * register overflow from temporaries in the calculations.
248 */
249
250 /*
251 * Compute the log reservation required to handle the refcount update
252 * transaction. Refcount updates are always done via deferred log items.
253 *
254 * This is calculated as:
255 * Data device refcount updates (t1):
256 * the agfs of the ags containing the blocks: nr_ops * sector size
257 * the refcount btrees: nr_ops * 1 trees * (2 * max depth - 1) * block size
258 */
259 static unsigned int
xfs_calc_refcountbt_reservation(struct xfs_mount * mp,unsigned int nr_ops)260 xfs_calc_refcountbt_reservation(
261 struct xfs_mount *mp,
262 unsigned int nr_ops)
263 {
264 unsigned int blksz = XFS_FSB_TO_B(mp, 1);
265
266 if (!xfs_has_reflink(mp))
267 return 0;
268
269 return xfs_calc_buf_res(nr_ops, mp->m_sb.sb_sectsize) +
270 xfs_calc_buf_res(xfs_refcountbt_block_count(mp, nr_ops), blksz);
271 }
272
273 /*
274 * In a write transaction we can allocate a maximum of 2
275 * extents. This gives (t1):
276 * the inode getting the new extents: inode size
277 * the inode's bmap btree: max depth * block size
278 * the agfs of the ags from which the extents are allocated: 2 * sector
279 * the superblock free block counter: sector size
280 * the allocation btrees: 2 exts * 2 trees * (2 * max depth - 1) * block size
281 * Or, if we're writing to a realtime file (t2):
282 * the inode getting the new extents: inode size
283 * the inode's bmap btree: max depth * block size
284 * the agfs of the ags from which the extents are allocated: 2 * sector
285 * the superblock free block counter: sector size
286 * the realtime bitmap: ((XFS_BMBT_MAX_EXTLEN / rtextsize) / NBBY) bytes
287 * the realtime summary: 1 block
288 * the allocation btrees: 2 trees * (2 * max depth - 1) * block size
289 * And the bmap_finish transaction can free bmap blocks in a join (t3):
290 * the agfs of the ags containing the blocks: 2 * sector size
291 * the agfls of the ags containing the blocks: 2 * sector size
292 * the super block free block counter: sector size
293 * the allocation btrees: 2 exts * 2 trees * (2 * max depth - 1) * block size
294 * And any refcount updates that happen in a separate transaction (t4).
295 */
296 STATIC uint
xfs_calc_write_reservation(struct xfs_mount * mp,bool for_minlogsize)297 xfs_calc_write_reservation(
298 struct xfs_mount *mp,
299 bool for_minlogsize)
300 {
301 unsigned int t1, t2, t3, t4;
302 unsigned int blksz = XFS_FSB_TO_B(mp, 1);
303
304 t1 = xfs_calc_inode_res(mp, 1) +
305 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK), blksz) +
306 xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) +
307 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2), blksz);
308
309 if (xfs_has_realtime(mp)) {
310 t2 = xfs_calc_inode_res(mp, 1) +
311 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK),
312 blksz) +
313 xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) +
314 xfs_calc_buf_res(xfs_rtalloc_block_count(mp, 1), blksz) +
315 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1), blksz);
316 } else {
317 t2 = 0;
318 }
319
320 t3 = xfs_calc_buf_res(5, mp->m_sb.sb_sectsize) +
321 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2), blksz);
322
323 /*
324 * In the early days of reflink, we included enough reservation to log
325 * two refcountbt splits for each transaction. The codebase runs
326 * refcountbt updates in separate transactions now, so to compute the
327 * minimum log size, add the refcountbtree splits back to t1 and t3 and
328 * do not account them separately as t4. Reflink did not support
329 * realtime when the reservations were established, so no adjustment to
330 * t2 is needed.
331 */
332 if (for_minlogsize) {
333 unsigned int adj = 0;
334
335 if (xfs_has_reflink(mp))
336 adj = xfs_calc_buf_res(
337 xfs_refcountbt_block_count(mp, 2),
338 blksz);
339 t1 += adj;
340 t3 += adj;
341 return XFS_DQUOT_LOGRES + max3(t1, t2, t3);
342 }
343
344 t4 = xfs_calc_refcountbt_reservation(mp, 1);
345 return XFS_DQUOT_LOGRES + max(t4, max3(t1, t2, t3));
346 }
347
348 unsigned int
xfs_calc_write_reservation_minlogsize(struct xfs_mount * mp)349 xfs_calc_write_reservation_minlogsize(
350 struct xfs_mount *mp)
351 {
352 return xfs_calc_write_reservation(mp, true);
353 }
354
355 /*
356 * In truncating a file we free up to two extents at once. We can modify (t1):
357 * the inode being truncated: inode size
358 * the inode's bmap btree: (max depth + 1) * block size
359 * And the bmap_finish transaction can free the blocks and bmap blocks (t2):
360 * the agf for each of the ags: 4 * sector size
361 * the agfl for each of the ags: 4 * sector size
362 * the super block to reflect the freed blocks: sector size
363 * worst case split in allocation btrees per extent assuming 4 extents:
364 * 4 exts * 2 trees * (2 * max depth - 1) * block size
365 * Or, if it's a realtime file (t3):
366 * the agf for each of the ags: 2 * sector size
367 * the agfl for each of the ags: 2 * sector size
368 * the super block to reflect the freed blocks: sector size
369 * the realtime bitmap:
370 * 2 exts * ((XFS_BMBT_MAX_EXTLEN / rtextsize) / NBBY) bytes
371 * the realtime summary: 2 exts * 1 block
372 * worst case split in allocation btrees per extent assuming 2 extents:
373 * 2 exts * 2 trees * (2 * max depth - 1) * block size
374 * And any refcount updates that happen in a separate transaction (t4).
375 */
376 STATIC uint
xfs_calc_itruncate_reservation(struct xfs_mount * mp,bool for_minlogsize)377 xfs_calc_itruncate_reservation(
378 struct xfs_mount *mp,
379 bool for_minlogsize)
380 {
381 unsigned int t1, t2, t3, t4;
382 unsigned int blksz = XFS_FSB_TO_B(mp, 1);
383
384 t1 = xfs_calc_inode_res(mp, 1) +
385 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK) + 1, blksz);
386
387 t2 = xfs_calc_buf_res(9, mp->m_sb.sb_sectsize) +
388 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 4), blksz);
389
390 if (xfs_has_realtime(mp)) {
391 t3 = xfs_calc_buf_res(5, mp->m_sb.sb_sectsize) +
392 xfs_calc_buf_res(xfs_rtalloc_block_count(mp, 2), blksz) +
393 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2), blksz);
394 } else {
395 t3 = 0;
396 }
397
398 /*
399 * In the early days of reflink, we included enough reservation to log
400 * four refcountbt splits in the same transaction as bnobt/cntbt
401 * updates. The codebase runs refcountbt updates in separate
402 * transactions now, so to compute the minimum log size, add the
403 * refcount btree splits back here and do not compute them separately
404 * as t4. Reflink did not support realtime when the reservations were
405 * established, so do not adjust t3.
406 */
407 if (for_minlogsize) {
408 if (xfs_has_reflink(mp))
409 t2 += xfs_calc_buf_res(
410 xfs_refcountbt_block_count(mp, 4),
411 blksz);
412
413 return XFS_DQUOT_LOGRES + max3(t1, t2, t3);
414 }
415
416 t4 = xfs_calc_refcountbt_reservation(mp, 2);
417 return XFS_DQUOT_LOGRES + max(t4, max3(t1, t2, t3));
418 }
419
420 unsigned int
xfs_calc_itruncate_reservation_minlogsize(struct xfs_mount * mp)421 xfs_calc_itruncate_reservation_minlogsize(
422 struct xfs_mount *mp)
423 {
424 return xfs_calc_itruncate_reservation(mp, true);
425 }
426
xfs_calc_pptr_link_overhead(void)427 static inline unsigned int xfs_calc_pptr_link_overhead(void)
428 {
429 return sizeof(struct xfs_attri_log_format) +
430 xlog_calc_iovec_len(sizeof(struct xfs_parent_rec)) +
431 xlog_calc_iovec_len(MAXNAMELEN - 1);
432 }
xfs_calc_pptr_unlink_overhead(void)433 static inline unsigned int xfs_calc_pptr_unlink_overhead(void)
434 {
435 return sizeof(struct xfs_attri_log_format) +
436 xlog_calc_iovec_len(sizeof(struct xfs_parent_rec)) +
437 xlog_calc_iovec_len(MAXNAMELEN - 1);
438 }
xfs_calc_pptr_replace_overhead(void)439 static inline unsigned int xfs_calc_pptr_replace_overhead(void)
440 {
441 return sizeof(struct xfs_attri_log_format) +
442 xlog_calc_iovec_len(sizeof(struct xfs_parent_rec)) +
443 xlog_calc_iovec_len(MAXNAMELEN - 1) +
444 xlog_calc_iovec_len(sizeof(struct xfs_parent_rec)) +
445 xlog_calc_iovec_len(MAXNAMELEN - 1);
446 }
447
448 /*
449 * In renaming a files we can modify:
450 * the five inodes involved: 5 * inode size
451 * the two directory btrees: 2 * (max depth + v2) * dir block size
452 * the two directory bmap btrees: 2 * max depth * block size
453 * And the bmap_finish transaction can free dir and bmap blocks (two sets
454 * of bmap blocks) giving (t2):
455 * the agf for the ags in which the blocks live: 3 * sector size
456 * the agfl for the ags in which the blocks live: 3 * sector size
457 * the superblock for the free block count: sector size
458 * the allocation btrees: 3 exts * 2 trees * (2 * max depth - 1) * block size
459 * If parent pointers are enabled (t3), then each transaction in the chain
460 * must be capable of setting or removing the extended attribute
461 * containing the parent information. It must also be able to handle
462 * the three xattr intent items that track the progress of the parent
463 * pointer update.
464 */
465 STATIC uint
xfs_calc_rename_reservation(struct xfs_mount * mp)466 xfs_calc_rename_reservation(
467 struct xfs_mount *mp)
468 {
469 unsigned int overhead = XFS_DQUOT_LOGRES;
470 struct xfs_trans_resv *resp = M_RES(mp);
471 unsigned int t1, t2, t3 = 0;
472
473 t1 = xfs_calc_inode_res(mp, 5) +
474 xfs_calc_buf_res(2 * XFS_DIROP_LOG_COUNT(mp),
475 XFS_FSB_TO_B(mp, 1));
476
477 t2 = xfs_calc_buf_res(7, mp->m_sb.sb_sectsize) +
478 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 3),
479 XFS_FSB_TO_B(mp, 1));
480
481 if (xfs_has_parent(mp)) {
482 unsigned int rename_overhead, exchange_overhead;
483
484 t3 = max(resp->tr_attrsetm.tr_logres,
485 resp->tr_attrrm.tr_logres);
486
487 /*
488 * For a standard rename, the three xattr intent log items
489 * are (1) replacing the pptr for the source file; (2)
490 * removing the pptr on the dest file; and (3) adding a
491 * pptr for the whiteout file in the src dir.
492 *
493 * For an RENAME_EXCHANGE, there are two xattr intent
494 * items to replace the pptr for both src and dest
495 * files. Link counts don't change and there is no
496 * whiteout.
497 *
498 * In the worst case we can end up relogging all log
499 * intent items to allow the log tail to move ahead, so
500 * they become overhead added to each transaction in a
501 * processing chain.
502 */
503 rename_overhead = xfs_calc_pptr_replace_overhead() +
504 xfs_calc_pptr_unlink_overhead() +
505 xfs_calc_pptr_link_overhead();
506 exchange_overhead = 2 * xfs_calc_pptr_replace_overhead();
507
508 overhead += max(rename_overhead, exchange_overhead);
509 }
510
511 return overhead + max3(t1, t2, t3);
512 }
513
514 static inline unsigned int
xfs_rename_log_count(struct xfs_mount * mp,struct xfs_trans_resv * resp)515 xfs_rename_log_count(
516 struct xfs_mount *mp,
517 struct xfs_trans_resv *resp)
518 {
519 /* One for the rename, one more for freeing blocks */
520 unsigned int ret = XFS_RENAME_LOG_COUNT;
521
522 /*
523 * Pre-reserve enough log reservation to handle the transaction
524 * rolling needed to remove or add one parent pointer.
525 */
526 if (xfs_has_parent(mp))
527 ret += max(resp->tr_attrsetm.tr_logcount,
528 resp->tr_attrrm.tr_logcount);
529
530 return ret;
531 }
532
533 /*
534 * For removing an inode from unlinked list at first, we can modify:
535 * the agi hash list and counters: sector size
536 * the on disk inode before ours in the agi hash list: inode cluster size
537 * the on disk inode in the agi hash list: inode cluster size
538 */
539 STATIC uint
xfs_calc_iunlink_remove_reservation(struct xfs_mount * mp)540 xfs_calc_iunlink_remove_reservation(
541 struct xfs_mount *mp)
542 {
543 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
544 2 * M_IGEO(mp)->inode_cluster_size;
545 }
546
547 static inline unsigned int
xfs_link_log_count(struct xfs_mount * mp,struct xfs_trans_resv * resp)548 xfs_link_log_count(
549 struct xfs_mount *mp,
550 struct xfs_trans_resv *resp)
551 {
552 unsigned int ret = XFS_LINK_LOG_COUNT;
553
554 /*
555 * Pre-reserve enough log reservation to handle the transaction
556 * rolling needed to add one parent pointer.
557 */
558 if (xfs_has_parent(mp))
559 ret += resp->tr_attrsetm.tr_logcount;
560
561 return ret;
562 }
563
564 /*
565 * For creating a link to an inode:
566 * the parent directory inode: inode size
567 * the linked inode: inode size
568 * the directory btree could split: (max depth + v2) * dir block size
569 * the directory bmap btree could join or split: (max depth + v2) * blocksize
570 * And the bmap_finish transaction can free some bmap blocks giving:
571 * the agf for the ag in which the blocks live: sector size
572 * the agfl for the ag in which the blocks live: sector size
573 * the superblock for the free block count: sector size
574 * the allocation btrees: 2 trees * (2 * max depth - 1) * block size
575 */
576 STATIC uint
xfs_calc_link_reservation(struct xfs_mount * mp)577 xfs_calc_link_reservation(
578 struct xfs_mount *mp)
579 {
580 unsigned int overhead = XFS_DQUOT_LOGRES;
581 struct xfs_trans_resv *resp = M_RES(mp);
582 unsigned int t1, t2, t3 = 0;
583
584 overhead += xfs_calc_iunlink_remove_reservation(mp);
585 t1 = xfs_calc_inode_res(mp, 2) +
586 xfs_calc_buf_res(XFS_DIROP_LOG_COUNT(mp), XFS_FSB_TO_B(mp, 1));
587 t2 = xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) +
588 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
589 XFS_FSB_TO_B(mp, 1));
590
591 if (xfs_has_parent(mp)) {
592 t3 = resp->tr_attrsetm.tr_logres;
593 overhead += xfs_calc_pptr_link_overhead();
594 }
595
596 return overhead + max3(t1, t2, t3);
597 }
598
599 /*
600 * For adding an inode to unlinked list we can modify:
601 * the agi hash list: sector size
602 * the on disk inode: inode cluster size
603 */
604 STATIC uint
xfs_calc_iunlink_add_reservation(xfs_mount_t * mp)605 xfs_calc_iunlink_add_reservation(xfs_mount_t *mp)
606 {
607 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
608 M_IGEO(mp)->inode_cluster_size;
609 }
610
611 static inline unsigned int
xfs_remove_log_count(struct xfs_mount * mp,struct xfs_trans_resv * resp)612 xfs_remove_log_count(
613 struct xfs_mount *mp,
614 struct xfs_trans_resv *resp)
615 {
616 unsigned int ret = XFS_REMOVE_LOG_COUNT;
617
618 /*
619 * Pre-reserve enough log reservation to handle the transaction
620 * rolling needed to add one parent pointer.
621 */
622 if (xfs_has_parent(mp))
623 ret += resp->tr_attrrm.tr_logcount;
624
625 return ret;
626 }
627
628 /*
629 * For removing a directory entry we can modify:
630 * the parent directory inode: inode size
631 * the removed inode: inode size
632 * the directory btree could join: (max depth + v2) * dir block size
633 * the directory bmap btree could join or split: (max depth + v2) * blocksize
634 * And the bmap_finish transaction can free the dir and bmap blocks giving:
635 * the agf for the ag in which the blocks live: 2 * sector size
636 * the agfl for the ag in which the blocks live: 2 * sector size
637 * the superblock for the free block count: sector size
638 * the allocation btrees: 2 exts * 2 trees * (2 * max depth - 1) * block size
639 */
640 STATIC uint
xfs_calc_remove_reservation(struct xfs_mount * mp)641 xfs_calc_remove_reservation(
642 struct xfs_mount *mp)
643 {
644 unsigned int overhead = XFS_DQUOT_LOGRES;
645 struct xfs_trans_resv *resp = M_RES(mp);
646 unsigned int t1, t2, t3 = 0;
647
648 overhead += xfs_calc_iunlink_add_reservation(mp);
649
650 t1 = xfs_calc_inode_res(mp, 2) +
651 xfs_calc_buf_res(XFS_DIROP_LOG_COUNT(mp), XFS_FSB_TO_B(mp, 1));
652 t2 = xfs_calc_buf_res(4, mp->m_sb.sb_sectsize) +
653 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2),
654 XFS_FSB_TO_B(mp, 1));
655
656 if (xfs_has_parent(mp)) {
657 t3 = resp->tr_attrrm.tr_logres;
658 overhead += xfs_calc_pptr_unlink_overhead();
659 }
660
661 return overhead + max3(t1, t2, t3);
662 }
663
664 /*
665 * For create, break it in to the two cases that the transaction
666 * covers. We start with the modify case - allocation done by modification
667 * of the state of existing inodes - and the allocation case.
668 */
669
670 /*
671 * For create we can modify:
672 * the parent directory inode: inode size
673 * the new inode: inode size
674 * the inode btree entry: block size
675 * the superblock for the nlink flag: sector size
676 * the directory btree: (max depth + v2) * dir block size
677 * the directory inode's bmap btree: (max depth + v2) * block size
678 * the finobt (record modification and allocation btrees)
679 */
680 STATIC uint
xfs_calc_create_resv_modify(struct xfs_mount * mp)681 xfs_calc_create_resv_modify(
682 struct xfs_mount *mp)
683 {
684 return xfs_calc_inode_res(mp, 2) +
685 xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
686 (uint)XFS_FSB_TO_B(mp, 1) +
687 xfs_calc_buf_res(XFS_DIROP_LOG_COUNT(mp), XFS_FSB_TO_B(mp, 1)) +
688 xfs_calc_finobt_res(mp);
689 }
690
691 /*
692 * For icreate we can allocate some inodes giving:
693 * the agi and agf of the ag getting the new inodes: 2 * sectorsize
694 * the superblock for the nlink flag: sector size
695 * the inode chunk (allocation, optional init)
696 * the inobt (record insertion)
697 * the finobt (optional, record insertion)
698 */
699 STATIC uint
xfs_calc_icreate_resv_alloc(struct xfs_mount * mp)700 xfs_calc_icreate_resv_alloc(
701 struct xfs_mount *mp)
702 {
703 return xfs_calc_buf_res(2, mp->m_sb.sb_sectsize) +
704 mp->m_sb.sb_sectsize +
705 xfs_calc_inode_chunk_res(mp, _ALLOC) +
706 xfs_calc_inobt_res(mp) +
707 xfs_calc_finobt_res(mp);
708 }
709
710 static inline unsigned int
xfs_icreate_log_count(struct xfs_mount * mp,struct xfs_trans_resv * resp)711 xfs_icreate_log_count(
712 struct xfs_mount *mp,
713 struct xfs_trans_resv *resp)
714 {
715 unsigned int ret = XFS_CREATE_LOG_COUNT;
716
717 /*
718 * Pre-reserve enough log reservation to handle the transaction
719 * rolling needed to add one parent pointer.
720 */
721 if (xfs_has_parent(mp))
722 ret += resp->tr_attrsetm.tr_logcount;
723
724 return ret;
725 }
726
727 STATIC uint
xfs_calc_icreate_reservation(struct xfs_mount * mp)728 xfs_calc_icreate_reservation(
729 struct xfs_mount *mp)
730 {
731 struct xfs_trans_resv *resp = M_RES(mp);
732 unsigned int overhead = XFS_DQUOT_LOGRES;
733 unsigned int t1, t2, t3 = 0;
734
735 t1 = xfs_calc_icreate_resv_alloc(mp);
736 t2 = xfs_calc_create_resv_modify(mp);
737
738 if (xfs_has_parent(mp)) {
739 t3 = resp->tr_attrsetm.tr_logres;
740 overhead += xfs_calc_pptr_link_overhead();
741 }
742
743 return overhead + max3(t1, t2, t3);
744 }
745
746 STATIC uint
xfs_calc_create_tmpfile_reservation(struct xfs_mount * mp)747 xfs_calc_create_tmpfile_reservation(
748 struct xfs_mount *mp)
749 {
750 uint res = XFS_DQUOT_LOGRES;
751
752 res += xfs_calc_icreate_resv_alloc(mp);
753 return res + xfs_calc_iunlink_add_reservation(mp);
754 }
755
756 static inline unsigned int
xfs_mkdir_log_count(struct xfs_mount * mp,struct xfs_trans_resv * resp)757 xfs_mkdir_log_count(
758 struct xfs_mount *mp,
759 struct xfs_trans_resv *resp)
760 {
761 unsigned int ret = XFS_MKDIR_LOG_COUNT;
762
763 /*
764 * Pre-reserve enough log reservation to handle the transaction
765 * rolling needed to add one parent pointer.
766 */
767 if (xfs_has_parent(mp))
768 ret += resp->tr_attrsetm.tr_logcount;
769
770 return ret;
771 }
772
773 /*
774 * Making a new directory is the same as creating a new file.
775 */
776 STATIC uint
xfs_calc_mkdir_reservation(struct xfs_mount * mp)777 xfs_calc_mkdir_reservation(
778 struct xfs_mount *mp)
779 {
780 return xfs_calc_icreate_reservation(mp);
781 }
782
783 static inline unsigned int
xfs_symlink_log_count(struct xfs_mount * mp,struct xfs_trans_resv * resp)784 xfs_symlink_log_count(
785 struct xfs_mount *mp,
786 struct xfs_trans_resv *resp)
787 {
788 unsigned int ret = XFS_SYMLINK_LOG_COUNT;
789
790 /*
791 * Pre-reserve enough log reservation to handle the transaction
792 * rolling needed to add one parent pointer.
793 */
794 if (xfs_has_parent(mp))
795 ret += resp->tr_attrsetm.tr_logcount;
796
797 return ret;
798 }
799
800 /*
801 * Making a new symplink is the same as creating a new file, but
802 * with the added blocks for remote symlink data which can be up to 1kB in
803 * length (XFS_SYMLINK_MAXLEN).
804 */
805 STATIC uint
xfs_calc_symlink_reservation(struct xfs_mount * mp)806 xfs_calc_symlink_reservation(
807 struct xfs_mount *mp)
808 {
809 return xfs_calc_icreate_reservation(mp) +
810 xfs_calc_buf_res(1, XFS_SYMLINK_MAXLEN);
811 }
812
813 /*
814 * In freeing an inode we can modify:
815 * the inode being freed: inode size
816 * the super block free inode counter, AGF and AGFL: sector size
817 * the on disk inode (agi unlinked list removal)
818 * the inode chunk (invalidated, headers only)
819 * the inode btree
820 * the finobt (record insertion, removal or modification)
821 *
822 * Note that the inode chunk res. includes an allocfree res. for freeing of the
823 * inode chunk. This is technically extraneous because the inode chunk free is
824 * deferred (it occurs after a transaction roll). Include the extra reservation
825 * anyways since we've had reports of ifree transaction overruns due to too many
826 * agfl fixups during inode chunk frees.
827 */
828 STATIC uint
xfs_calc_ifree_reservation(struct xfs_mount * mp)829 xfs_calc_ifree_reservation(
830 struct xfs_mount *mp)
831 {
832 return XFS_DQUOT_LOGRES +
833 xfs_calc_inode_res(mp, 1) +
834 xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) +
835 xfs_calc_iunlink_remove_reservation(mp) +
836 xfs_calc_inode_chunk_res(mp, _FREE) +
837 xfs_calc_inobt_res(mp) +
838 xfs_calc_finobt_res(mp);
839 }
840
841 /*
842 * When only changing the inode we log the inode and possibly the superblock
843 * We also add a bit of slop for the transaction stuff.
844 */
845 STATIC uint
xfs_calc_ichange_reservation(struct xfs_mount * mp)846 xfs_calc_ichange_reservation(
847 struct xfs_mount *mp)
848 {
849 return XFS_DQUOT_LOGRES +
850 xfs_calc_inode_res(mp, 1) +
851 xfs_calc_buf_res(1, mp->m_sb.sb_sectsize);
852
853 }
854
855 /*
856 * Growing the data section of the filesystem.
857 * superblock
858 * agi and agf
859 * allocation btrees
860 */
861 STATIC uint
xfs_calc_growdata_reservation(struct xfs_mount * mp)862 xfs_calc_growdata_reservation(
863 struct xfs_mount *mp)
864 {
865 return xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) +
866 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
867 XFS_FSB_TO_B(mp, 1));
868 }
869
870 /*
871 * Growing the rt section of the filesystem.
872 * In the first set of transactions (ALLOC) we allocate space to the
873 * bitmap or summary files.
874 * superblock: sector size
875 * agf of the ag from which the extent is allocated: sector size
876 * bmap btree for bitmap/summary inode: max depth * blocksize
877 * bitmap/summary inode: inode size
878 * allocation btrees for 1 block alloc: 2 * (2 * maxdepth - 1) * blocksize
879 */
880 STATIC uint
xfs_calc_growrtalloc_reservation(struct xfs_mount * mp)881 xfs_calc_growrtalloc_reservation(
882 struct xfs_mount *mp)
883 {
884 return xfs_calc_buf_res(2, mp->m_sb.sb_sectsize) +
885 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK),
886 XFS_FSB_TO_B(mp, 1)) +
887 xfs_calc_inode_res(mp, 1) +
888 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
889 XFS_FSB_TO_B(mp, 1));
890 }
891
892 /*
893 * Growing the rt section of the filesystem.
894 * In the second set of transactions (ZERO) we zero the new metadata blocks.
895 * one bitmap/summary block: blocksize
896 */
897 STATIC uint
xfs_calc_growrtzero_reservation(struct xfs_mount * mp)898 xfs_calc_growrtzero_reservation(
899 struct xfs_mount *mp)
900 {
901 return xfs_calc_buf_res(1, mp->m_sb.sb_blocksize);
902 }
903
904 /*
905 * Growing the rt section of the filesystem.
906 * In the third set of transactions (FREE) we update metadata without
907 * allocating any new blocks.
908 * superblock: sector size
909 * bitmap inode: inode size
910 * summary inode: inode size
911 * one bitmap block: blocksize
912 * summary blocks: new summary size
913 */
914 STATIC uint
xfs_calc_growrtfree_reservation(struct xfs_mount * mp)915 xfs_calc_growrtfree_reservation(
916 struct xfs_mount *mp)
917 {
918 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
919 xfs_calc_inode_res(mp, 2) +
920 xfs_calc_buf_res(1, mp->m_sb.sb_blocksize) +
921 xfs_calc_buf_res(1, XFS_FSB_TO_B(mp, mp->m_rsumblocks));
922 }
923
924 /*
925 * Logging the inode modification timestamp on a synchronous write.
926 * inode
927 */
928 STATIC uint
xfs_calc_swrite_reservation(struct xfs_mount * mp)929 xfs_calc_swrite_reservation(
930 struct xfs_mount *mp)
931 {
932 return xfs_calc_inode_res(mp, 1);
933 }
934
935 /*
936 * Logging the inode mode bits when writing a setuid/setgid file
937 * inode
938 */
939 STATIC uint
xfs_calc_writeid_reservation(struct xfs_mount * mp)940 xfs_calc_writeid_reservation(
941 struct xfs_mount *mp)
942 {
943 return xfs_calc_inode_res(mp, 1);
944 }
945
946 /*
947 * Converting the inode from non-attributed to attributed.
948 * the inode being converted: inode size
949 * agf block and superblock (for block allocation)
950 * the new block (directory sized)
951 * bmap blocks for the new directory block
952 * allocation btrees
953 */
954 STATIC uint
xfs_calc_addafork_reservation(struct xfs_mount * mp)955 xfs_calc_addafork_reservation(
956 struct xfs_mount *mp)
957 {
958 return XFS_DQUOT_LOGRES +
959 xfs_calc_inode_res(mp, 1) +
960 xfs_calc_buf_res(2, mp->m_sb.sb_sectsize) +
961 xfs_calc_buf_res(1, mp->m_dir_geo->blksize) +
962 xfs_calc_buf_res(XFS_DAENTER_BMAP1B(mp, XFS_DATA_FORK) + 1,
963 XFS_FSB_TO_B(mp, 1)) +
964 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
965 XFS_FSB_TO_B(mp, 1));
966 }
967
968 /*
969 * Removing the attribute fork of a file
970 * the inode being truncated: inode size
971 * the inode's bmap btree: max depth * block size
972 * And the bmap_finish transaction can free the blocks and bmap blocks:
973 * the agf for each of the ags: 4 * sector size
974 * the agfl for each of the ags: 4 * sector size
975 * the super block to reflect the freed blocks: sector size
976 * worst case split in allocation btrees per extent assuming 4 extents:
977 * 4 exts * 2 trees * (2 * max depth - 1) * block size
978 */
979 STATIC uint
xfs_calc_attrinval_reservation(struct xfs_mount * mp)980 xfs_calc_attrinval_reservation(
981 struct xfs_mount *mp)
982 {
983 return max((xfs_calc_inode_res(mp, 1) +
984 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_ATTR_FORK),
985 XFS_FSB_TO_B(mp, 1))),
986 (xfs_calc_buf_res(9, mp->m_sb.sb_sectsize) +
987 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 4),
988 XFS_FSB_TO_B(mp, 1))));
989 }
990
991 /*
992 * Setting an attribute at mount time.
993 * the inode getting the attribute
994 * the superblock for allocations
995 * the agfs extents are allocated from
996 * the attribute btree * max depth
997 * the inode allocation btree
998 * Since attribute transaction space is dependent on the size of the attribute,
999 * the calculation is done partially at mount time and partially at runtime(see
1000 * below).
1001 */
1002 STATIC uint
xfs_calc_attrsetm_reservation(struct xfs_mount * mp)1003 xfs_calc_attrsetm_reservation(
1004 struct xfs_mount *mp)
1005 {
1006 return XFS_DQUOT_LOGRES +
1007 xfs_calc_inode_res(mp, 1) +
1008 xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
1009 xfs_calc_buf_res(XFS_DA_NODE_MAXDEPTH, XFS_FSB_TO_B(mp, 1));
1010 }
1011
1012 /*
1013 * Setting an attribute at runtime, transaction space unit per block.
1014 * the superblock for allocations: sector size
1015 * the inode bmap btree could join or split: max depth * block size
1016 * Since the runtime attribute transaction space is dependent on the total
1017 * blocks needed for the 1st bmap, here we calculate out the space unit for
1018 * one block so that the caller could figure out the total space according
1019 * to the attibute extent length in blocks by:
1020 * ext * M_RES(mp)->tr_attrsetrt.tr_logres
1021 */
1022 STATIC uint
xfs_calc_attrsetrt_reservation(struct xfs_mount * mp)1023 xfs_calc_attrsetrt_reservation(
1024 struct xfs_mount *mp)
1025 {
1026 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
1027 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_ATTR_FORK),
1028 XFS_FSB_TO_B(mp, 1));
1029 }
1030
1031 /*
1032 * Removing an attribute.
1033 * the inode: inode size
1034 * the attribute btree could join: max depth * block size
1035 * the inode bmap btree could join or split: max depth * block size
1036 * And the bmap_finish transaction can free the attr blocks freed giving:
1037 * the agf for the ag in which the blocks live: 2 * sector size
1038 * the agfl for the ag in which the blocks live: 2 * sector size
1039 * the superblock for the free block count: sector size
1040 * the allocation btrees: 2 exts * 2 trees * (2 * max depth - 1) * block size
1041 */
1042 STATIC uint
xfs_calc_attrrm_reservation(struct xfs_mount * mp)1043 xfs_calc_attrrm_reservation(
1044 struct xfs_mount *mp)
1045 {
1046 return XFS_DQUOT_LOGRES +
1047 max((xfs_calc_inode_res(mp, 1) +
1048 xfs_calc_buf_res(XFS_DA_NODE_MAXDEPTH,
1049 XFS_FSB_TO_B(mp, 1)) +
1050 (uint)XFS_FSB_TO_B(mp,
1051 XFS_BM_MAXLEVELS(mp, XFS_ATTR_FORK)) +
1052 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK), 0)),
1053 (xfs_calc_buf_res(5, mp->m_sb.sb_sectsize) +
1054 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2),
1055 XFS_FSB_TO_B(mp, 1))));
1056 }
1057
1058 /*
1059 * Clearing a bad agino number in an agi hash bucket.
1060 */
1061 STATIC uint
xfs_calc_clear_agi_bucket_reservation(struct xfs_mount * mp)1062 xfs_calc_clear_agi_bucket_reservation(
1063 struct xfs_mount *mp)
1064 {
1065 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize);
1066 }
1067
1068 /*
1069 * Adjusting quota limits.
1070 * the disk quota buffer: sizeof(struct xfs_disk_dquot)
1071 */
1072 STATIC uint
xfs_calc_qm_setqlim_reservation(void)1073 xfs_calc_qm_setqlim_reservation(void)
1074 {
1075 return xfs_calc_buf_res(1, sizeof(struct xfs_disk_dquot));
1076 }
1077
1078 /*
1079 * Allocating quota on disk if needed.
1080 * the write transaction log space for quota file extent allocation
1081 * the unit of quota allocation: one system block size
1082 */
1083 STATIC uint
xfs_calc_qm_dqalloc_reservation(struct xfs_mount * mp,bool for_minlogsize)1084 xfs_calc_qm_dqalloc_reservation(
1085 struct xfs_mount *mp,
1086 bool for_minlogsize)
1087 {
1088 return xfs_calc_write_reservation(mp, for_minlogsize) +
1089 xfs_calc_buf_res(1,
1090 XFS_FSB_TO_B(mp, XFS_DQUOT_CLUSTER_SIZE_FSB) - 1);
1091 }
1092
1093 unsigned int
xfs_calc_qm_dqalloc_reservation_minlogsize(struct xfs_mount * mp)1094 xfs_calc_qm_dqalloc_reservation_minlogsize(
1095 struct xfs_mount *mp)
1096 {
1097 return xfs_calc_qm_dqalloc_reservation(mp, true);
1098 }
1099
1100 /*
1101 * Syncing the incore super block changes to disk.
1102 * the super block to reflect the changes: sector size
1103 */
1104 STATIC uint
xfs_calc_sb_reservation(struct xfs_mount * mp)1105 xfs_calc_sb_reservation(
1106 struct xfs_mount *mp)
1107 {
1108 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize);
1109 }
1110
1111 /*
1112 * Namespace reservations.
1113 *
1114 * These get tricky when parent pointers are enabled as we have attribute
1115 * modifications occurring from within these transactions. Rather than confuse
1116 * each of these reservation calculations with the conditional attribute
1117 * reservations, add them here in a clear and concise manner. This requires that
1118 * the attribute reservations have already been calculated.
1119 *
1120 * Note that we only include the static attribute reservation here; the runtime
1121 * reservation will have to be modified by the size of the attributes being
1122 * added/removed/modified. See the comments on the attribute reservation
1123 * calculations for more details.
1124 */
1125 STATIC void
xfs_calc_namespace_reservations(struct xfs_mount * mp,struct xfs_trans_resv * resp)1126 xfs_calc_namespace_reservations(
1127 struct xfs_mount *mp,
1128 struct xfs_trans_resv *resp)
1129 {
1130 ASSERT(resp->tr_attrsetm.tr_logres > 0);
1131
1132 resp->tr_rename.tr_logres = xfs_calc_rename_reservation(mp);
1133 resp->tr_rename.tr_logcount = xfs_rename_log_count(mp, resp);
1134 resp->tr_rename.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1135
1136 resp->tr_link.tr_logres = xfs_calc_link_reservation(mp);
1137 resp->tr_link.tr_logcount = xfs_link_log_count(mp, resp);
1138 resp->tr_link.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1139
1140 resp->tr_remove.tr_logres = xfs_calc_remove_reservation(mp);
1141 resp->tr_remove.tr_logcount = xfs_remove_log_count(mp, resp);
1142 resp->tr_remove.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1143
1144 resp->tr_symlink.tr_logres = xfs_calc_symlink_reservation(mp);
1145 resp->tr_symlink.tr_logcount = xfs_symlink_log_count(mp, resp);
1146 resp->tr_symlink.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1147
1148 resp->tr_create.tr_logres = xfs_calc_icreate_reservation(mp);
1149 resp->tr_create.tr_logcount = xfs_icreate_log_count(mp, resp);
1150 resp->tr_create.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1151
1152 resp->tr_mkdir.tr_logres = xfs_calc_mkdir_reservation(mp);
1153 resp->tr_mkdir.tr_logcount = xfs_mkdir_log_count(mp, resp);
1154 resp->tr_mkdir.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1155 }
1156
1157 void
xfs_trans_resv_calc(struct xfs_mount * mp,struct xfs_trans_resv * resp)1158 xfs_trans_resv_calc(
1159 struct xfs_mount *mp,
1160 struct xfs_trans_resv *resp)
1161 {
1162 int logcount_adj = 0;
1163
1164 /*
1165 * The following transactions are logged in physical format and
1166 * require a permanent reservation on space.
1167 */
1168 resp->tr_write.tr_logres = xfs_calc_write_reservation(mp, false);
1169 resp->tr_write.tr_logcount = XFS_WRITE_LOG_COUNT;
1170 resp->tr_write.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1171
1172 resp->tr_itruncate.tr_logres = xfs_calc_itruncate_reservation(mp, false);
1173 resp->tr_itruncate.tr_logcount = XFS_ITRUNCATE_LOG_COUNT;
1174 resp->tr_itruncate.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1175
1176 resp->tr_create_tmpfile.tr_logres =
1177 xfs_calc_create_tmpfile_reservation(mp);
1178 resp->tr_create_tmpfile.tr_logcount = XFS_CREATE_TMPFILE_LOG_COUNT;
1179 resp->tr_create_tmpfile.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1180
1181 resp->tr_ifree.tr_logres = xfs_calc_ifree_reservation(mp);
1182 resp->tr_ifree.tr_logcount = XFS_INACTIVE_LOG_COUNT;
1183 resp->tr_ifree.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1184
1185 resp->tr_addafork.tr_logres = xfs_calc_addafork_reservation(mp);
1186 resp->tr_addafork.tr_logcount = XFS_ADDAFORK_LOG_COUNT;
1187 resp->tr_addafork.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1188
1189 resp->tr_attrinval.tr_logres = xfs_calc_attrinval_reservation(mp);
1190 resp->tr_attrinval.tr_logcount = XFS_ATTRINVAL_LOG_COUNT;
1191 resp->tr_attrinval.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1192
1193 resp->tr_attrsetm.tr_logres = xfs_calc_attrsetm_reservation(mp);
1194 resp->tr_attrsetm.tr_logcount = XFS_ATTRSET_LOG_COUNT;
1195 resp->tr_attrsetm.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1196
1197 resp->tr_attrrm.tr_logres = xfs_calc_attrrm_reservation(mp);
1198 resp->tr_attrrm.tr_logcount = XFS_ATTRRM_LOG_COUNT;
1199 resp->tr_attrrm.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1200
1201 resp->tr_growrtalloc.tr_logres = xfs_calc_growrtalloc_reservation(mp);
1202 resp->tr_growrtalloc.tr_logcount = XFS_DEFAULT_PERM_LOG_COUNT;
1203 resp->tr_growrtalloc.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1204
1205 resp->tr_qm_dqalloc.tr_logres = xfs_calc_qm_dqalloc_reservation(mp,
1206 false);
1207 resp->tr_qm_dqalloc.tr_logcount = XFS_WRITE_LOG_COUNT;
1208 resp->tr_qm_dqalloc.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1209
1210 xfs_calc_namespace_reservations(mp, resp);
1211
1212 /*
1213 * The following transactions are logged in logical format with
1214 * a default log count.
1215 */
1216 resp->tr_qm_setqlim.tr_logres = xfs_calc_qm_setqlim_reservation();
1217 resp->tr_qm_setqlim.tr_logcount = XFS_DEFAULT_LOG_COUNT;
1218
1219 resp->tr_sb.tr_logres = xfs_calc_sb_reservation(mp);
1220 resp->tr_sb.tr_logcount = XFS_DEFAULT_LOG_COUNT;
1221
1222 /* growdata requires permanent res; it can free space to the last AG */
1223 resp->tr_growdata.tr_logres = xfs_calc_growdata_reservation(mp);
1224 resp->tr_growdata.tr_logcount = XFS_DEFAULT_PERM_LOG_COUNT;
1225 resp->tr_growdata.tr_logflags |= XFS_TRANS_PERM_LOG_RES;
1226
1227 /* The following transaction are logged in logical format */
1228 resp->tr_ichange.tr_logres = xfs_calc_ichange_reservation(mp);
1229 resp->tr_fsyncts.tr_logres = xfs_calc_swrite_reservation(mp);
1230 resp->tr_writeid.tr_logres = xfs_calc_writeid_reservation(mp);
1231 resp->tr_attrsetrt.tr_logres = xfs_calc_attrsetrt_reservation(mp);
1232 resp->tr_clearagi.tr_logres = xfs_calc_clear_agi_bucket_reservation(mp);
1233 resp->tr_growrtzero.tr_logres = xfs_calc_growrtzero_reservation(mp);
1234 resp->tr_growrtfree.tr_logres = xfs_calc_growrtfree_reservation(mp);
1235
1236 /*
1237 * Add one logcount for BUI items that appear with rmap or reflink,
1238 * one logcount for refcount intent items, and one logcount for rmap
1239 * intent items.
1240 */
1241 if (xfs_has_reflink(mp) || xfs_has_rmapbt(mp))
1242 logcount_adj++;
1243 if (xfs_has_reflink(mp))
1244 logcount_adj++;
1245 if (xfs_has_rmapbt(mp))
1246 logcount_adj++;
1247
1248 resp->tr_itruncate.tr_logcount += logcount_adj;
1249 resp->tr_write.tr_logcount += logcount_adj;
1250 resp->tr_qm_dqalloc.tr_logcount += logcount_adj;
1251 }
1252