1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Copyright (c) 2000-2003,2005 Silicon Graphics, Inc. 4 * Copyright (C) 2010 Red Hat, Inc. 5 * All Rights Reserved. 6 */ 7 #include "xfs.h" 8 #include "xfs_fs.h" 9 #include "xfs_shared.h" 10 #include "xfs_format.h" 11 #include "xfs_log_format.h" 12 #include "xfs_trans_resv.h" 13 #include "xfs_mount.h" 14 #include "xfs_da_format.h" 15 #include "xfs_da_btree.h" 16 #include "xfs_inode.h" 17 #include "xfs_bmap_btree.h" 18 #include "xfs_quota.h" 19 #include "xfs_trans.h" 20 #include "xfs_qm.h" 21 #include "xfs_trans_space.h" 22 #include "xfs_rtbitmap.h" 23 #include "xfs_attr_item.h" 24 #include "xfs_log.h" 25 26 #define _ALLOC true 27 #define _FREE false 28 29 /* 30 * A buffer has a format structure overhead in the log in addition 31 * to the data, so we need to take this into account when reserving 32 * space in a transaction for a buffer. Round the space required up 33 * to a multiple of 128 bytes so that we don't change the historical 34 * reservation that has been used for this overhead. 35 */ 36 STATIC uint 37 xfs_buf_log_overhead(void) 38 { 39 return round_up(sizeof(struct xlog_op_header) + 40 sizeof(struct xfs_buf_log_format), 128); 41 } 42 43 /* 44 * Calculate out transaction log reservation per item in bytes. 45 * 46 * The nbufs argument is used to indicate the number of items that 47 * will be changed in a transaction. size is used to tell how many 48 * bytes should be reserved per item. 49 */ 50 STATIC uint 51 xfs_calc_buf_res( 52 uint nbufs, 53 uint size) 54 { 55 return nbufs * (size + xfs_buf_log_overhead()); 56 } 57 58 /* 59 * Per-extent log reservation for the btree changes involved in freeing or 60 * allocating an extent. In classic XFS there were two trees that will be 61 * modified (bnobt + cntbt). With rmap enabled, there are three trees 62 * (rmapbt). The number of blocks reserved is based on the formula: 63 * 64 * num trees * ((2 blocks/level * max depth) - 1) 65 * 66 * Keep in mind that max depth is calculated separately for each type of tree. 67 */ 68 uint 69 xfs_allocfree_block_count( 70 struct xfs_mount *mp, 71 uint num_ops) 72 { 73 uint blocks; 74 75 blocks = num_ops * 2 * (2 * mp->m_alloc_maxlevels - 1); 76 if (xfs_has_rmapbt(mp)) 77 blocks += num_ops * (2 * mp->m_rmap_maxlevels - 1); 78 79 return blocks; 80 } 81 82 /* 83 * Per-extent log reservation for refcount btree changes. These are never done 84 * in the same transaction as an allocation or a free, so we compute them 85 * separately. 86 */ 87 static unsigned int 88 xfs_refcountbt_block_count( 89 struct xfs_mount *mp, 90 unsigned int num_ops) 91 { 92 return num_ops * (2 * mp->m_refc_maxlevels - 1); 93 } 94 95 /* 96 * Logging inodes is really tricksy. They are logged in memory format, 97 * which means that what we write into the log doesn't directly translate into 98 * the amount of space they use on disk. 99 * 100 * Case in point - btree format forks in memory format use more space than the 101 * on-disk format. In memory, the buffer contains a normal btree block header so 102 * the btree code can treat it as though it is just another generic buffer. 103 * However, when we write it to the inode fork, we don't write all of this 104 * header as it isn't needed. e.g. the root is only ever in the inode, so 105 * there's no need for sibling pointers which would waste 16 bytes of space. 106 * 107 * Hence when we have an inode with a maximally sized btree format fork, then 108 * amount of information we actually log is greater than the size of the inode 109 * on disk. Hence we need an inode reservation function that calculates all this 110 * correctly. So, we log: 111 * 112 * - 4 log op headers for object 113 * - for the ilf, the inode core and 2 forks 114 * - inode log format object 115 * - the inode core 116 * - two inode forks containing bmap btree root blocks. 117 * - the btree data contained by both forks will fit into the inode size, 118 * hence when combined with the inode core above, we have a total of the 119 * actual inode size. 120 * - the BMBT headers need to be accounted separately, as they are 121 * additional to the records and pointers that fit inside the inode 122 * forks. 123 */ 124 STATIC uint 125 xfs_calc_inode_res( 126 struct xfs_mount *mp, 127 uint ninodes) 128 { 129 return ninodes * 130 (4 * sizeof(struct xlog_op_header) + 131 sizeof(struct xfs_inode_log_format) + 132 mp->m_sb.sb_inodesize + 133 2 * xfs_bmbt_block_len(mp)); 134 } 135 136 /* 137 * Inode btree record insertion/removal modifies the inode btree and free space 138 * btrees (since the inobt does not use the agfl). This requires the following 139 * reservation: 140 * 141 * the inode btree: max depth * blocksize 142 * the allocation btrees: 2 trees * (max depth - 1) * block size 143 * 144 * The caller must account for SB and AG header modifications, etc. 145 */ 146 STATIC uint 147 xfs_calc_inobt_res( 148 struct xfs_mount *mp) 149 { 150 return xfs_calc_buf_res(M_IGEO(mp)->inobt_maxlevels, 151 XFS_FSB_TO_B(mp, 1)) + 152 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1), 153 XFS_FSB_TO_B(mp, 1)); 154 } 155 156 /* 157 * The free inode btree is a conditional feature. The behavior differs slightly 158 * from that of the traditional inode btree in that the finobt tracks records 159 * for inode chunks with at least one free inode. A record can be removed from 160 * the tree during individual inode allocation. Therefore the finobt 161 * reservation is unconditional for both the inode chunk allocation and 162 * individual inode allocation (modify) cases. 163 * 164 * Behavior aside, the reservation for finobt modification is equivalent to the 165 * traditional inobt: cover a full finobt shape change plus block allocation. 166 */ 167 STATIC uint 168 xfs_calc_finobt_res( 169 struct xfs_mount *mp) 170 { 171 if (!xfs_has_finobt(mp)) 172 return 0; 173 174 return xfs_calc_inobt_res(mp); 175 } 176 177 /* 178 * Calculate the reservation required to allocate or free an inode chunk. This 179 * includes: 180 * 181 * the allocation btrees: 2 trees * (max depth - 1) * block size 182 * the inode chunk: m_ino_geo.ialloc_blks * N 183 * 184 * The size N of the inode chunk reservation depends on whether it is for 185 * allocation or free and which type of create transaction is in use. An inode 186 * chunk free always invalidates the buffers and only requires reservation for 187 * headers (N == 0). An inode chunk allocation requires a chunk sized 188 * reservation on v4 and older superblocks to initialize the chunk. No chunk 189 * reservation is required for allocation on v5 supers, which use ordered 190 * buffers to initialize. 191 */ 192 STATIC uint 193 xfs_calc_inode_chunk_res( 194 struct xfs_mount *mp, 195 bool alloc) 196 { 197 uint res, size = 0; 198 199 res = xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1), 200 XFS_FSB_TO_B(mp, 1)); 201 if (alloc) { 202 /* icreate tx uses ordered buffers */ 203 if (xfs_has_v3inodes(mp)) 204 return res; 205 size = XFS_FSB_TO_B(mp, 1); 206 } 207 208 res += xfs_calc_buf_res(M_IGEO(mp)->ialloc_blks, size); 209 return res; 210 } 211 212 /* 213 * Per-extent log reservation for the btree changes involved in freeing or 214 * allocating a realtime extent. We have to be able to log as many rtbitmap 215 * blocks as needed to mark inuse XFS_BMBT_MAX_EXTLEN blocks' worth of realtime 216 * extents, as well as the realtime summary block. 217 */ 218 static unsigned int 219 xfs_rtalloc_block_count( 220 struct xfs_mount *mp, 221 unsigned int num_ops) 222 { 223 unsigned int rtbmp_blocks; 224 xfs_rtxlen_t rtxlen; 225 226 rtxlen = xfs_extlen_to_rtxlen(mp, XFS_MAX_BMBT_EXTLEN); 227 rtbmp_blocks = xfs_rtbitmap_blockcount_len(mp, rtxlen); 228 return (rtbmp_blocks + 1) * num_ops; 229 } 230 231 /* 232 * Various log reservation values. 233 * 234 * These are based on the size of the file system block because that is what 235 * most transactions manipulate. Each adds in an additional 128 bytes per 236 * item logged to try to account for the overhead of the transaction mechanism. 237 * 238 * Note: Most of the reservations underestimate the number of allocation 239 * groups into which they could free extents in the xfs_defer_finish() call. 240 * This is because the number in the worst case is quite high and quite 241 * unusual. In order to fix this we need to change xfs_defer_finish() to free 242 * extents in only a single AG at a time. This will require changes to the 243 * EFI code as well, however, so that the EFI for the extents not freed is 244 * logged again in each transaction. See SGI PV #261917. 245 * 246 * Reservation functions here avoid a huge stack in xfs_trans_init due to 247 * register overflow from temporaries in the calculations. 248 */ 249 250 /* 251 * Compute the log reservation required to handle the refcount update 252 * transaction. Refcount updates are always done via deferred log items. 253 * 254 * This is calculated as: 255 * Data device refcount updates (t1): 256 * the agfs of the ags containing the blocks: nr_ops * sector size 257 * the refcount btrees: nr_ops * 1 trees * (2 * max depth - 1) * block size 258 */ 259 static unsigned int 260 xfs_calc_refcountbt_reservation( 261 struct xfs_mount *mp, 262 unsigned int nr_ops) 263 { 264 unsigned int blksz = XFS_FSB_TO_B(mp, 1); 265 266 if (!xfs_has_reflink(mp)) 267 return 0; 268 269 return xfs_calc_buf_res(nr_ops, mp->m_sb.sb_sectsize) + 270 xfs_calc_buf_res(xfs_refcountbt_block_count(mp, nr_ops), blksz); 271 } 272 273 /* 274 * In a write transaction we can allocate a maximum of 2 275 * extents. This gives (t1): 276 * the inode getting the new extents: inode size 277 * the inode's bmap btree: max depth * block size 278 * the agfs of the ags from which the extents are allocated: 2 * sector 279 * the superblock free block counter: sector size 280 * the allocation btrees: 2 exts * 2 trees * (2 * max depth - 1) * block size 281 * Or, if we're writing to a realtime file (t2): 282 * the inode getting the new extents: inode size 283 * the inode's bmap btree: max depth * block size 284 * the agfs of the ags from which the extents are allocated: 2 * sector 285 * the superblock free block counter: sector size 286 * the realtime bitmap: ((XFS_BMBT_MAX_EXTLEN / rtextsize) / NBBY) bytes 287 * the realtime summary: 1 block 288 * the allocation btrees: 2 trees * (2 * max depth - 1) * block size 289 * And the bmap_finish transaction can free bmap blocks in a join (t3): 290 * the agfs of the ags containing the blocks: 2 * sector size 291 * the agfls of the ags containing the blocks: 2 * sector size 292 * the super block free block counter: sector size 293 * the allocation btrees: 2 exts * 2 trees * (2 * max depth - 1) * block size 294 * And any refcount updates that happen in a separate transaction (t4). 295 */ 296 STATIC uint 297 xfs_calc_write_reservation( 298 struct xfs_mount *mp, 299 bool for_minlogsize) 300 { 301 unsigned int t1, t2, t3, t4; 302 unsigned int blksz = XFS_FSB_TO_B(mp, 1); 303 304 t1 = xfs_calc_inode_res(mp, 1) + 305 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK), blksz) + 306 xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) + 307 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2), blksz); 308 309 if (xfs_has_realtime(mp)) { 310 t2 = xfs_calc_inode_res(mp, 1) + 311 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK), 312 blksz) + 313 xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) + 314 xfs_calc_buf_res(xfs_rtalloc_block_count(mp, 1), blksz) + 315 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1), blksz); 316 } else { 317 t2 = 0; 318 } 319 320 t3 = xfs_calc_buf_res(5, mp->m_sb.sb_sectsize) + 321 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2), blksz); 322 323 /* 324 * In the early days of reflink, we included enough reservation to log 325 * two refcountbt splits for each transaction. The codebase runs 326 * refcountbt updates in separate transactions now, so to compute the 327 * minimum log size, add the refcountbtree splits back to t1 and t3 and 328 * do not account them separately as t4. Reflink did not support 329 * realtime when the reservations were established, so no adjustment to 330 * t2 is needed. 331 */ 332 if (for_minlogsize) { 333 unsigned int adj = 0; 334 335 if (xfs_has_reflink(mp)) 336 adj = xfs_calc_buf_res( 337 xfs_refcountbt_block_count(mp, 2), 338 blksz); 339 t1 += adj; 340 t3 += adj; 341 return XFS_DQUOT_LOGRES + max3(t1, t2, t3); 342 } 343 344 t4 = xfs_calc_refcountbt_reservation(mp, 1); 345 return XFS_DQUOT_LOGRES + max(t4, max3(t1, t2, t3)); 346 } 347 348 unsigned int 349 xfs_calc_write_reservation_minlogsize( 350 struct xfs_mount *mp) 351 { 352 return xfs_calc_write_reservation(mp, true); 353 } 354 355 /* 356 * In truncating a file we free up to two extents at once. We can modify (t1): 357 * the inode being truncated: inode size 358 * the inode's bmap btree: (max depth + 1) * block size 359 * And the bmap_finish transaction can free the blocks and bmap blocks (t2): 360 * the agf for each of the ags: 4 * sector size 361 * the agfl for each of the ags: 4 * sector size 362 * the super block to reflect the freed blocks: sector size 363 * worst case split in allocation btrees per extent assuming 4 extents: 364 * 4 exts * 2 trees * (2 * max depth - 1) * block size 365 * Or, if it's a realtime file (t3): 366 * the agf for each of the ags: 2 * sector size 367 * the agfl for each of the ags: 2 * sector size 368 * the super block to reflect the freed blocks: sector size 369 * the realtime bitmap: 370 * 2 exts * ((XFS_BMBT_MAX_EXTLEN / rtextsize) / NBBY) bytes 371 * the realtime summary: 2 exts * 1 block 372 * worst case split in allocation btrees per extent assuming 2 extents: 373 * 2 exts * 2 trees * (2 * max depth - 1) * block size 374 * And any refcount updates that happen in a separate transaction (t4). 375 */ 376 STATIC uint 377 xfs_calc_itruncate_reservation( 378 struct xfs_mount *mp, 379 bool for_minlogsize) 380 { 381 unsigned int t1, t2, t3, t4; 382 unsigned int blksz = XFS_FSB_TO_B(mp, 1); 383 384 t1 = xfs_calc_inode_res(mp, 1) + 385 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK) + 1, blksz); 386 387 t2 = xfs_calc_buf_res(9, mp->m_sb.sb_sectsize) + 388 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 4), blksz); 389 390 if (xfs_has_realtime(mp)) { 391 t3 = xfs_calc_buf_res(5, mp->m_sb.sb_sectsize) + 392 xfs_calc_buf_res(xfs_rtalloc_block_count(mp, 2), blksz) + 393 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2), blksz); 394 } else { 395 t3 = 0; 396 } 397 398 /* 399 * In the early days of reflink, we included enough reservation to log 400 * four refcountbt splits in the same transaction as bnobt/cntbt 401 * updates. The codebase runs refcountbt updates in separate 402 * transactions now, so to compute the minimum log size, add the 403 * refcount btree splits back here and do not compute them separately 404 * as t4. Reflink did not support realtime when the reservations were 405 * established, so do not adjust t3. 406 */ 407 if (for_minlogsize) { 408 if (xfs_has_reflink(mp)) 409 t2 += xfs_calc_buf_res( 410 xfs_refcountbt_block_count(mp, 4), 411 blksz); 412 413 return XFS_DQUOT_LOGRES + max3(t1, t2, t3); 414 } 415 416 t4 = xfs_calc_refcountbt_reservation(mp, 2); 417 return XFS_DQUOT_LOGRES + max(t4, max3(t1, t2, t3)); 418 } 419 420 unsigned int 421 xfs_calc_itruncate_reservation_minlogsize( 422 struct xfs_mount *mp) 423 { 424 return xfs_calc_itruncate_reservation(mp, true); 425 } 426 427 static inline unsigned int xfs_calc_pptr_link_overhead(void) 428 { 429 return sizeof(struct xfs_attri_log_format) + 430 xlog_calc_iovec_len(sizeof(struct xfs_parent_rec)) + 431 xlog_calc_iovec_len(MAXNAMELEN - 1); 432 } 433 static inline unsigned int xfs_calc_pptr_unlink_overhead(void) 434 { 435 return sizeof(struct xfs_attri_log_format) + 436 xlog_calc_iovec_len(sizeof(struct xfs_parent_rec)) + 437 xlog_calc_iovec_len(MAXNAMELEN - 1); 438 } 439 static inline unsigned int xfs_calc_pptr_replace_overhead(void) 440 { 441 return sizeof(struct xfs_attri_log_format) + 442 xlog_calc_iovec_len(sizeof(struct xfs_parent_rec)) + 443 xlog_calc_iovec_len(MAXNAMELEN - 1) + 444 xlog_calc_iovec_len(sizeof(struct xfs_parent_rec)) + 445 xlog_calc_iovec_len(MAXNAMELEN - 1); 446 } 447 448 /* 449 * In renaming a files we can modify: 450 * the five inodes involved: 5 * inode size 451 * the two directory btrees: 2 * (max depth + v2) * dir block size 452 * the two directory bmap btrees: 2 * max depth * block size 453 * And the bmap_finish transaction can free dir and bmap blocks (two sets 454 * of bmap blocks) giving (t2): 455 * the agf for the ags in which the blocks live: 3 * sector size 456 * the agfl for the ags in which the blocks live: 3 * sector size 457 * the superblock for the free block count: sector size 458 * the allocation btrees: 3 exts * 2 trees * (2 * max depth - 1) * block size 459 * If parent pointers are enabled (t3), then each transaction in the chain 460 * must be capable of setting or removing the extended attribute 461 * containing the parent information. It must also be able to handle 462 * the three xattr intent items that track the progress of the parent 463 * pointer update. 464 */ 465 STATIC uint 466 xfs_calc_rename_reservation( 467 struct xfs_mount *mp) 468 { 469 unsigned int overhead = XFS_DQUOT_LOGRES; 470 struct xfs_trans_resv *resp = M_RES(mp); 471 unsigned int t1, t2, t3 = 0; 472 473 t1 = xfs_calc_inode_res(mp, 5) + 474 xfs_calc_buf_res(2 * XFS_DIROP_LOG_COUNT(mp), 475 XFS_FSB_TO_B(mp, 1)); 476 477 t2 = xfs_calc_buf_res(7, mp->m_sb.sb_sectsize) + 478 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 3), 479 XFS_FSB_TO_B(mp, 1)); 480 481 if (xfs_has_parent(mp)) { 482 unsigned int rename_overhead, exchange_overhead; 483 484 t3 = max(resp->tr_attrsetm.tr_logres, 485 resp->tr_attrrm.tr_logres); 486 487 /* 488 * For a standard rename, the three xattr intent log items 489 * are (1) replacing the pptr for the source file; (2) 490 * removing the pptr on the dest file; and (3) adding a 491 * pptr for the whiteout file in the src dir. 492 * 493 * For an RENAME_EXCHANGE, there are two xattr intent 494 * items to replace the pptr for both src and dest 495 * files. Link counts don't change and there is no 496 * whiteout. 497 * 498 * In the worst case we can end up relogging all log 499 * intent items to allow the log tail to move ahead, so 500 * they become overhead added to each transaction in a 501 * processing chain. 502 */ 503 rename_overhead = xfs_calc_pptr_replace_overhead() + 504 xfs_calc_pptr_unlink_overhead() + 505 xfs_calc_pptr_link_overhead(); 506 exchange_overhead = 2 * xfs_calc_pptr_replace_overhead(); 507 508 overhead += max(rename_overhead, exchange_overhead); 509 } 510 511 return overhead + max3(t1, t2, t3); 512 } 513 514 static inline unsigned int 515 xfs_rename_log_count( 516 struct xfs_mount *mp, 517 struct xfs_trans_resv *resp) 518 { 519 /* One for the rename, one more for freeing blocks */ 520 unsigned int ret = XFS_RENAME_LOG_COUNT; 521 522 /* 523 * Pre-reserve enough log reservation to handle the transaction 524 * rolling needed to remove or add one parent pointer. 525 */ 526 if (xfs_has_parent(mp)) 527 ret += max(resp->tr_attrsetm.tr_logcount, 528 resp->tr_attrrm.tr_logcount); 529 530 return ret; 531 } 532 533 /* 534 * For removing an inode from unlinked list at first, we can modify: 535 * the agi hash list and counters: sector size 536 * the on disk inode before ours in the agi hash list: inode cluster size 537 * the on disk inode in the agi hash list: inode cluster size 538 */ 539 STATIC uint 540 xfs_calc_iunlink_remove_reservation( 541 struct xfs_mount *mp) 542 { 543 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) + 544 2 * M_IGEO(mp)->inode_cluster_size; 545 } 546 547 static inline unsigned int 548 xfs_link_log_count( 549 struct xfs_mount *mp, 550 struct xfs_trans_resv *resp) 551 { 552 unsigned int ret = XFS_LINK_LOG_COUNT; 553 554 /* 555 * Pre-reserve enough log reservation to handle the transaction 556 * rolling needed to add one parent pointer. 557 */ 558 if (xfs_has_parent(mp)) 559 ret += resp->tr_attrsetm.tr_logcount; 560 561 return ret; 562 } 563 564 /* 565 * For creating a link to an inode: 566 * the parent directory inode: inode size 567 * the linked inode: inode size 568 * the directory btree could split: (max depth + v2) * dir block size 569 * the directory bmap btree could join or split: (max depth + v2) * blocksize 570 * And the bmap_finish transaction can free some bmap blocks giving: 571 * the agf for the ag in which the blocks live: sector size 572 * the agfl for the ag in which the blocks live: sector size 573 * the superblock for the free block count: sector size 574 * the allocation btrees: 2 trees * (2 * max depth - 1) * block size 575 */ 576 STATIC uint 577 xfs_calc_link_reservation( 578 struct xfs_mount *mp) 579 { 580 unsigned int overhead = XFS_DQUOT_LOGRES; 581 struct xfs_trans_resv *resp = M_RES(mp); 582 unsigned int t1, t2, t3 = 0; 583 584 overhead += xfs_calc_iunlink_remove_reservation(mp); 585 t1 = xfs_calc_inode_res(mp, 2) + 586 xfs_calc_buf_res(XFS_DIROP_LOG_COUNT(mp), XFS_FSB_TO_B(mp, 1)); 587 t2 = xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) + 588 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1), 589 XFS_FSB_TO_B(mp, 1)); 590 591 if (xfs_has_parent(mp)) { 592 t3 = resp->tr_attrsetm.tr_logres; 593 overhead += xfs_calc_pptr_link_overhead(); 594 } 595 596 return overhead + max3(t1, t2, t3); 597 } 598 599 /* 600 * For adding an inode to unlinked list we can modify: 601 * the agi hash list: sector size 602 * the on disk inode: inode cluster size 603 */ 604 STATIC uint 605 xfs_calc_iunlink_add_reservation(xfs_mount_t *mp) 606 { 607 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) + 608 M_IGEO(mp)->inode_cluster_size; 609 } 610 611 static inline unsigned int 612 xfs_remove_log_count( 613 struct xfs_mount *mp, 614 struct xfs_trans_resv *resp) 615 { 616 unsigned int ret = XFS_REMOVE_LOG_COUNT; 617 618 /* 619 * Pre-reserve enough log reservation to handle the transaction 620 * rolling needed to add one parent pointer. 621 */ 622 if (xfs_has_parent(mp)) 623 ret += resp->tr_attrrm.tr_logcount; 624 625 return ret; 626 } 627 628 /* 629 * For removing a directory entry we can modify: 630 * the parent directory inode: inode size 631 * the removed inode: inode size 632 * the directory btree could join: (max depth + v2) * dir block size 633 * the directory bmap btree could join or split: (max depth + v2) * blocksize 634 * And the bmap_finish transaction can free the dir and bmap blocks giving: 635 * the agf for the ag in which the blocks live: 2 * sector size 636 * the agfl for the ag in which the blocks live: 2 * sector size 637 * the superblock for the free block count: sector size 638 * the allocation btrees: 2 exts * 2 trees * (2 * max depth - 1) * block size 639 */ 640 STATIC uint 641 xfs_calc_remove_reservation( 642 struct xfs_mount *mp) 643 { 644 unsigned int overhead = XFS_DQUOT_LOGRES; 645 struct xfs_trans_resv *resp = M_RES(mp); 646 unsigned int t1, t2, t3 = 0; 647 648 overhead += xfs_calc_iunlink_add_reservation(mp); 649 650 t1 = xfs_calc_inode_res(mp, 2) + 651 xfs_calc_buf_res(XFS_DIROP_LOG_COUNT(mp), XFS_FSB_TO_B(mp, 1)); 652 t2 = xfs_calc_buf_res(4, mp->m_sb.sb_sectsize) + 653 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2), 654 XFS_FSB_TO_B(mp, 1)); 655 656 if (xfs_has_parent(mp)) { 657 t3 = resp->tr_attrrm.tr_logres; 658 overhead += xfs_calc_pptr_unlink_overhead(); 659 } 660 661 return overhead + max3(t1, t2, t3); 662 } 663 664 /* 665 * For create, break it in to the two cases that the transaction 666 * covers. We start with the modify case - allocation done by modification 667 * of the state of existing inodes - and the allocation case. 668 */ 669 670 /* 671 * For create we can modify: 672 * the parent directory inode: inode size 673 * the new inode: inode size 674 * the inode btree entry: block size 675 * the superblock for the nlink flag: sector size 676 * the directory btree: (max depth + v2) * dir block size 677 * the directory inode's bmap btree: (max depth + v2) * block size 678 * the finobt (record modification and allocation btrees) 679 */ 680 STATIC uint 681 xfs_calc_create_resv_modify( 682 struct xfs_mount *mp) 683 { 684 return xfs_calc_inode_res(mp, 2) + 685 xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) + 686 (uint)XFS_FSB_TO_B(mp, 1) + 687 xfs_calc_buf_res(XFS_DIROP_LOG_COUNT(mp), XFS_FSB_TO_B(mp, 1)) + 688 xfs_calc_finobt_res(mp); 689 } 690 691 /* 692 * For icreate we can allocate some inodes giving: 693 * the agi and agf of the ag getting the new inodes: 2 * sectorsize 694 * the superblock for the nlink flag: sector size 695 * the inode chunk (allocation, optional init) 696 * the inobt (record insertion) 697 * the finobt (optional, record insertion) 698 */ 699 STATIC uint 700 xfs_calc_icreate_resv_alloc( 701 struct xfs_mount *mp) 702 { 703 return xfs_calc_buf_res(2, mp->m_sb.sb_sectsize) + 704 mp->m_sb.sb_sectsize + 705 xfs_calc_inode_chunk_res(mp, _ALLOC) + 706 xfs_calc_inobt_res(mp) + 707 xfs_calc_finobt_res(mp); 708 } 709 710 static inline unsigned int 711 xfs_icreate_log_count( 712 struct xfs_mount *mp, 713 struct xfs_trans_resv *resp) 714 { 715 unsigned int ret = XFS_CREATE_LOG_COUNT; 716 717 /* 718 * Pre-reserve enough log reservation to handle the transaction 719 * rolling needed to add one parent pointer. 720 */ 721 if (xfs_has_parent(mp)) 722 ret += resp->tr_attrsetm.tr_logcount; 723 724 return ret; 725 } 726 727 STATIC uint 728 xfs_calc_icreate_reservation( 729 struct xfs_mount *mp) 730 { 731 struct xfs_trans_resv *resp = M_RES(mp); 732 unsigned int overhead = XFS_DQUOT_LOGRES; 733 unsigned int t1, t2, t3 = 0; 734 735 t1 = xfs_calc_icreate_resv_alloc(mp); 736 t2 = xfs_calc_create_resv_modify(mp); 737 738 if (xfs_has_parent(mp)) { 739 t3 = resp->tr_attrsetm.tr_logres; 740 overhead += xfs_calc_pptr_link_overhead(); 741 } 742 743 return overhead + max3(t1, t2, t3); 744 } 745 746 STATIC uint 747 xfs_calc_create_tmpfile_reservation( 748 struct xfs_mount *mp) 749 { 750 uint res = XFS_DQUOT_LOGRES; 751 752 res += xfs_calc_icreate_resv_alloc(mp); 753 return res + xfs_calc_iunlink_add_reservation(mp); 754 } 755 756 static inline unsigned int 757 xfs_mkdir_log_count( 758 struct xfs_mount *mp, 759 struct xfs_trans_resv *resp) 760 { 761 unsigned int ret = XFS_MKDIR_LOG_COUNT; 762 763 /* 764 * Pre-reserve enough log reservation to handle the transaction 765 * rolling needed to add one parent pointer. 766 */ 767 if (xfs_has_parent(mp)) 768 ret += resp->tr_attrsetm.tr_logcount; 769 770 return ret; 771 } 772 773 /* 774 * Making a new directory is the same as creating a new file. 775 */ 776 STATIC uint 777 xfs_calc_mkdir_reservation( 778 struct xfs_mount *mp) 779 { 780 return xfs_calc_icreate_reservation(mp); 781 } 782 783 static inline unsigned int 784 xfs_symlink_log_count( 785 struct xfs_mount *mp, 786 struct xfs_trans_resv *resp) 787 { 788 unsigned int ret = XFS_SYMLINK_LOG_COUNT; 789 790 /* 791 * Pre-reserve enough log reservation to handle the transaction 792 * rolling needed to add one parent pointer. 793 */ 794 if (xfs_has_parent(mp)) 795 ret += resp->tr_attrsetm.tr_logcount; 796 797 return ret; 798 } 799 800 /* 801 * Making a new symplink is the same as creating a new file, but 802 * with the added blocks for remote symlink data which can be up to 1kB in 803 * length (XFS_SYMLINK_MAXLEN). 804 */ 805 STATIC uint 806 xfs_calc_symlink_reservation( 807 struct xfs_mount *mp) 808 { 809 return xfs_calc_icreate_reservation(mp) + 810 xfs_calc_buf_res(1, XFS_SYMLINK_MAXLEN); 811 } 812 813 /* 814 * In freeing an inode we can modify: 815 * the inode being freed: inode size 816 * the super block free inode counter, AGF and AGFL: sector size 817 * the on disk inode (agi unlinked list removal) 818 * the inode chunk (invalidated, headers only) 819 * the inode btree 820 * the finobt (record insertion, removal or modification) 821 * 822 * Note that the inode chunk res. includes an allocfree res. for freeing of the 823 * inode chunk. This is technically extraneous because the inode chunk free is 824 * deferred (it occurs after a transaction roll). Include the extra reservation 825 * anyways since we've had reports of ifree transaction overruns due to too many 826 * agfl fixups during inode chunk frees. 827 */ 828 STATIC uint 829 xfs_calc_ifree_reservation( 830 struct xfs_mount *mp) 831 { 832 return XFS_DQUOT_LOGRES + 833 xfs_calc_inode_res(mp, 1) + 834 xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) + 835 xfs_calc_iunlink_remove_reservation(mp) + 836 xfs_calc_inode_chunk_res(mp, _FREE) + 837 xfs_calc_inobt_res(mp) + 838 xfs_calc_finobt_res(mp); 839 } 840 841 /* 842 * When only changing the inode we log the inode and possibly the superblock 843 * We also add a bit of slop for the transaction stuff. 844 */ 845 STATIC uint 846 xfs_calc_ichange_reservation( 847 struct xfs_mount *mp) 848 { 849 return XFS_DQUOT_LOGRES + 850 xfs_calc_inode_res(mp, 1) + 851 xfs_calc_buf_res(1, mp->m_sb.sb_sectsize); 852 853 } 854 855 /* 856 * Growing the data section of the filesystem. 857 * superblock 858 * agi and agf 859 * allocation btrees 860 */ 861 STATIC uint 862 xfs_calc_growdata_reservation( 863 struct xfs_mount *mp) 864 { 865 return xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) + 866 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1), 867 XFS_FSB_TO_B(mp, 1)); 868 } 869 870 /* 871 * Growing the rt section of the filesystem. 872 * In the first set of transactions (ALLOC) we allocate space to the 873 * bitmap or summary files. 874 * superblock: sector size 875 * agf of the ag from which the extent is allocated: sector size 876 * bmap btree for bitmap/summary inode: max depth * blocksize 877 * bitmap/summary inode: inode size 878 * allocation btrees for 1 block alloc: 2 * (2 * maxdepth - 1) * blocksize 879 */ 880 STATIC uint 881 xfs_calc_growrtalloc_reservation( 882 struct xfs_mount *mp) 883 { 884 return xfs_calc_buf_res(2, mp->m_sb.sb_sectsize) + 885 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK), 886 XFS_FSB_TO_B(mp, 1)) + 887 xfs_calc_inode_res(mp, 1) + 888 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1), 889 XFS_FSB_TO_B(mp, 1)); 890 } 891 892 /* 893 * Growing the rt section of the filesystem. 894 * In the second set of transactions (ZERO) we zero the new metadata blocks. 895 * one bitmap/summary block: blocksize 896 */ 897 STATIC uint 898 xfs_calc_growrtzero_reservation( 899 struct xfs_mount *mp) 900 { 901 return xfs_calc_buf_res(1, mp->m_sb.sb_blocksize); 902 } 903 904 /* 905 * Growing the rt section of the filesystem. 906 * In the third set of transactions (FREE) we update metadata without 907 * allocating any new blocks. 908 * superblock: sector size 909 * bitmap inode: inode size 910 * summary inode: inode size 911 * one bitmap block: blocksize 912 * summary blocks: new summary size 913 */ 914 STATIC uint 915 xfs_calc_growrtfree_reservation( 916 struct xfs_mount *mp) 917 { 918 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) + 919 xfs_calc_inode_res(mp, 2) + 920 xfs_calc_buf_res(1, mp->m_sb.sb_blocksize) + 921 xfs_calc_buf_res(1, XFS_FSB_TO_B(mp, mp->m_rsumblocks)); 922 } 923 924 /* 925 * Logging the inode modification timestamp on a synchronous write. 926 * inode 927 */ 928 STATIC uint 929 xfs_calc_swrite_reservation( 930 struct xfs_mount *mp) 931 { 932 return xfs_calc_inode_res(mp, 1); 933 } 934 935 /* 936 * Logging the inode mode bits when writing a setuid/setgid file 937 * inode 938 */ 939 STATIC uint 940 xfs_calc_writeid_reservation( 941 struct xfs_mount *mp) 942 { 943 return xfs_calc_inode_res(mp, 1); 944 } 945 946 /* 947 * Converting the inode from non-attributed to attributed. 948 * the inode being converted: inode size 949 * agf block and superblock (for block allocation) 950 * the new block (directory sized) 951 * bmap blocks for the new directory block 952 * allocation btrees 953 */ 954 STATIC uint 955 xfs_calc_addafork_reservation( 956 struct xfs_mount *mp) 957 { 958 return XFS_DQUOT_LOGRES + 959 xfs_calc_inode_res(mp, 1) + 960 xfs_calc_buf_res(2, mp->m_sb.sb_sectsize) + 961 xfs_calc_buf_res(1, mp->m_dir_geo->blksize) + 962 xfs_calc_buf_res(XFS_DAENTER_BMAP1B(mp, XFS_DATA_FORK) + 1, 963 XFS_FSB_TO_B(mp, 1)) + 964 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1), 965 XFS_FSB_TO_B(mp, 1)); 966 } 967 968 /* 969 * Removing the attribute fork of a file 970 * the inode being truncated: inode size 971 * the inode's bmap btree: max depth * block size 972 * And the bmap_finish transaction can free the blocks and bmap blocks: 973 * the agf for each of the ags: 4 * sector size 974 * the agfl for each of the ags: 4 * sector size 975 * the super block to reflect the freed blocks: sector size 976 * worst case split in allocation btrees per extent assuming 4 extents: 977 * 4 exts * 2 trees * (2 * max depth - 1) * block size 978 */ 979 STATIC uint 980 xfs_calc_attrinval_reservation( 981 struct xfs_mount *mp) 982 { 983 return max((xfs_calc_inode_res(mp, 1) + 984 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_ATTR_FORK), 985 XFS_FSB_TO_B(mp, 1))), 986 (xfs_calc_buf_res(9, mp->m_sb.sb_sectsize) + 987 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 4), 988 XFS_FSB_TO_B(mp, 1)))); 989 } 990 991 /* 992 * Setting an attribute at mount time. 993 * the inode getting the attribute 994 * the superblock for allocations 995 * the agfs extents are allocated from 996 * the attribute btree * max depth 997 * the inode allocation btree 998 * Since attribute transaction space is dependent on the size of the attribute, 999 * the calculation is done partially at mount time and partially at runtime(see 1000 * below). 1001 */ 1002 STATIC uint 1003 xfs_calc_attrsetm_reservation( 1004 struct xfs_mount *mp) 1005 { 1006 return XFS_DQUOT_LOGRES + 1007 xfs_calc_inode_res(mp, 1) + 1008 xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) + 1009 xfs_calc_buf_res(XFS_DA_NODE_MAXDEPTH, XFS_FSB_TO_B(mp, 1)); 1010 } 1011 1012 /* 1013 * Setting an attribute at runtime, transaction space unit per block. 1014 * the superblock for allocations: sector size 1015 * the inode bmap btree could join or split: max depth * block size 1016 * Since the runtime attribute transaction space is dependent on the total 1017 * blocks needed for the 1st bmap, here we calculate out the space unit for 1018 * one block so that the caller could figure out the total space according 1019 * to the attibute extent length in blocks by: 1020 * ext * M_RES(mp)->tr_attrsetrt.tr_logres 1021 */ 1022 STATIC uint 1023 xfs_calc_attrsetrt_reservation( 1024 struct xfs_mount *mp) 1025 { 1026 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) + 1027 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_ATTR_FORK), 1028 XFS_FSB_TO_B(mp, 1)); 1029 } 1030 1031 /* 1032 * Removing an attribute. 1033 * the inode: inode size 1034 * the attribute btree could join: max depth * block size 1035 * the inode bmap btree could join or split: max depth * block size 1036 * And the bmap_finish transaction can free the attr blocks freed giving: 1037 * the agf for the ag in which the blocks live: 2 * sector size 1038 * the agfl for the ag in which the blocks live: 2 * sector size 1039 * the superblock for the free block count: sector size 1040 * the allocation btrees: 2 exts * 2 trees * (2 * max depth - 1) * block size 1041 */ 1042 STATIC uint 1043 xfs_calc_attrrm_reservation( 1044 struct xfs_mount *mp) 1045 { 1046 return XFS_DQUOT_LOGRES + 1047 max((xfs_calc_inode_res(mp, 1) + 1048 xfs_calc_buf_res(XFS_DA_NODE_MAXDEPTH, 1049 XFS_FSB_TO_B(mp, 1)) + 1050 (uint)XFS_FSB_TO_B(mp, 1051 XFS_BM_MAXLEVELS(mp, XFS_ATTR_FORK)) + 1052 xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK), 0)), 1053 (xfs_calc_buf_res(5, mp->m_sb.sb_sectsize) + 1054 xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2), 1055 XFS_FSB_TO_B(mp, 1)))); 1056 } 1057 1058 /* 1059 * Clearing a bad agino number in an agi hash bucket. 1060 */ 1061 STATIC uint 1062 xfs_calc_clear_agi_bucket_reservation( 1063 struct xfs_mount *mp) 1064 { 1065 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize); 1066 } 1067 1068 /* 1069 * Adjusting quota limits. 1070 * the disk quota buffer: sizeof(struct xfs_disk_dquot) 1071 */ 1072 STATIC uint 1073 xfs_calc_qm_setqlim_reservation(void) 1074 { 1075 return xfs_calc_buf_res(1, sizeof(struct xfs_disk_dquot)); 1076 } 1077 1078 /* 1079 * Allocating quota on disk if needed. 1080 * the write transaction log space for quota file extent allocation 1081 * the unit of quota allocation: one system block size 1082 */ 1083 STATIC uint 1084 xfs_calc_qm_dqalloc_reservation( 1085 struct xfs_mount *mp, 1086 bool for_minlogsize) 1087 { 1088 return xfs_calc_write_reservation(mp, for_minlogsize) + 1089 xfs_calc_buf_res(1, 1090 XFS_FSB_TO_B(mp, XFS_DQUOT_CLUSTER_SIZE_FSB) - 1); 1091 } 1092 1093 unsigned int 1094 xfs_calc_qm_dqalloc_reservation_minlogsize( 1095 struct xfs_mount *mp) 1096 { 1097 return xfs_calc_qm_dqalloc_reservation(mp, true); 1098 } 1099 1100 /* 1101 * Syncing the incore super block changes to disk. 1102 * the super block to reflect the changes: sector size 1103 */ 1104 STATIC uint 1105 xfs_calc_sb_reservation( 1106 struct xfs_mount *mp) 1107 { 1108 return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize); 1109 } 1110 1111 /* 1112 * Namespace reservations. 1113 * 1114 * These get tricky when parent pointers are enabled as we have attribute 1115 * modifications occurring from within these transactions. Rather than confuse 1116 * each of these reservation calculations with the conditional attribute 1117 * reservations, add them here in a clear and concise manner. This requires that 1118 * the attribute reservations have already been calculated. 1119 * 1120 * Note that we only include the static attribute reservation here; the runtime 1121 * reservation will have to be modified by the size of the attributes being 1122 * added/removed/modified. See the comments on the attribute reservation 1123 * calculations for more details. 1124 */ 1125 STATIC void 1126 xfs_calc_namespace_reservations( 1127 struct xfs_mount *mp, 1128 struct xfs_trans_resv *resp) 1129 { 1130 ASSERT(resp->tr_attrsetm.tr_logres > 0); 1131 1132 resp->tr_rename.tr_logres = xfs_calc_rename_reservation(mp); 1133 resp->tr_rename.tr_logcount = xfs_rename_log_count(mp, resp); 1134 resp->tr_rename.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1135 1136 resp->tr_link.tr_logres = xfs_calc_link_reservation(mp); 1137 resp->tr_link.tr_logcount = xfs_link_log_count(mp, resp); 1138 resp->tr_link.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1139 1140 resp->tr_remove.tr_logres = xfs_calc_remove_reservation(mp); 1141 resp->tr_remove.tr_logcount = xfs_remove_log_count(mp, resp); 1142 resp->tr_remove.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1143 1144 resp->tr_symlink.tr_logres = xfs_calc_symlink_reservation(mp); 1145 resp->tr_symlink.tr_logcount = xfs_symlink_log_count(mp, resp); 1146 resp->tr_symlink.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1147 1148 resp->tr_create.tr_logres = xfs_calc_icreate_reservation(mp); 1149 resp->tr_create.tr_logcount = xfs_icreate_log_count(mp, resp); 1150 resp->tr_create.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1151 1152 resp->tr_mkdir.tr_logres = xfs_calc_mkdir_reservation(mp); 1153 resp->tr_mkdir.tr_logcount = xfs_mkdir_log_count(mp, resp); 1154 resp->tr_mkdir.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1155 } 1156 1157 void 1158 xfs_trans_resv_calc( 1159 struct xfs_mount *mp, 1160 struct xfs_trans_resv *resp) 1161 { 1162 int logcount_adj = 0; 1163 1164 /* 1165 * The following transactions are logged in physical format and 1166 * require a permanent reservation on space. 1167 */ 1168 resp->tr_write.tr_logres = xfs_calc_write_reservation(mp, false); 1169 resp->tr_write.tr_logcount = XFS_WRITE_LOG_COUNT; 1170 resp->tr_write.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1171 1172 resp->tr_itruncate.tr_logres = xfs_calc_itruncate_reservation(mp, false); 1173 resp->tr_itruncate.tr_logcount = XFS_ITRUNCATE_LOG_COUNT; 1174 resp->tr_itruncate.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1175 1176 resp->tr_create_tmpfile.tr_logres = 1177 xfs_calc_create_tmpfile_reservation(mp); 1178 resp->tr_create_tmpfile.tr_logcount = XFS_CREATE_TMPFILE_LOG_COUNT; 1179 resp->tr_create_tmpfile.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1180 1181 resp->tr_ifree.tr_logres = xfs_calc_ifree_reservation(mp); 1182 resp->tr_ifree.tr_logcount = XFS_INACTIVE_LOG_COUNT; 1183 resp->tr_ifree.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1184 1185 resp->tr_addafork.tr_logres = xfs_calc_addafork_reservation(mp); 1186 resp->tr_addafork.tr_logcount = XFS_ADDAFORK_LOG_COUNT; 1187 resp->tr_addafork.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1188 1189 resp->tr_attrinval.tr_logres = xfs_calc_attrinval_reservation(mp); 1190 resp->tr_attrinval.tr_logcount = XFS_ATTRINVAL_LOG_COUNT; 1191 resp->tr_attrinval.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1192 1193 resp->tr_attrsetm.tr_logres = xfs_calc_attrsetm_reservation(mp); 1194 resp->tr_attrsetm.tr_logcount = XFS_ATTRSET_LOG_COUNT; 1195 resp->tr_attrsetm.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1196 1197 resp->tr_attrrm.tr_logres = xfs_calc_attrrm_reservation(mp); 1198 resp->tr_attrrm.tr_logcount = XFS_ATTRRM_LOG_COUNT; 1199 resp->tr_attrrm.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1200 1201 resp->tr_growrtalloc.tr_logres = xfs_calc_growrtalloc_reservation(mp); 1202 resp->tr_growrtalloc.tr_logcount = XFS_DEFAULT_PERM_LOG_COUNT; 1203 resp->tr_growrtalloc.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1204 1205 resp->tr_qm_dqalloc.tr_logres = xfs_calc_qm_dqalloc_reservation(mp, 1206 false); 1207 resp->tr_qm_dqalloc.tr_logcount = XFS_WRITE_LOG_COUNT; 1208 resp->tr_qm_dqalloc.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1209 1210 xfs_calc_namespace_reservations(mp, resp); 1211 1212 /* 1213 * The following transactions are logged in logical format with 1214 * a default log count. 1215 */ 1216 resp->tr_qm_setqlim.tr_logres = xfs_calc_qm_setqlim_reservation(); 1217 resp->tr_qm_setqlim.tr_logcount = XFS_DEFAULT_LOG_COUNT; 1218 1219 resp->tr_sb.tr_logres = xfs_calc_sb_reservation(mp); 1220 resp->tr_sb.tr_logcount = XFS_DEFAULT_LOG_COUNT; 1221 1222 /* growdata requires permanent res; it can free space to the last AG */ 1223 resp->tr_growdata.tr_logres = xfs_calc_growdata_reservation(mp); 1224 resp->tr_growdata.tr_logcount = XFS_DEFAULT_PERM_LOG_COUNT; 1225 resp->tr_growdata.tr_logflags |= XFS_TRANS_PERM_LOG_RES; 1226 1227 /* The following transaction are logged in logical format */ 1228 resp->tr_ichange.tr_logres = xfs_calc_ichange_reservation(mp); 1229 resp->tr_fsyncts.tr_logres = xfs_calc_swrite_reservation(mp); 1230 resp->tr_writeid.tr_logres = xfs_calc_writeid_reservation(mp); 1231 resp->tr_attrsetrt.tr_logres = xfs_calc_attrsetrt_reservation(mp); 1232 resp->tr_clearagi.tr_logres = xfs_calc_clear_agi_bucket_reservation(mp); 1233 resp->tr_growrtzero.tr_logres = xfs_calc_growrtzero_reservation(mp); 1234 resp->tr_growrtfree.tr_logres = xfs_calc_growrtfree_reservation(mp); 1235 1236 /* 1237 * Add one logcount for BUI items that appear with rmap or reflink, 1238 * one logcount for refcount intent items, and one logcount for rmap 1239 * intent items. 1240 */ 1241 if (xfs_has_reflink(mp) || xfs_has_rmapbt(mp)) 1242 logcount_adj++; 1243 if (xfs_has_reflink(mp)) 1244 logcount_adj++; 1245 if (xfs_has_rmapbt(mp)) 1246 logcount_adj++; 1247 1248 resp->tr_itruncate.tr_logcount += logcount_adj; 1249 resp->tr_write.tr_logcount += logcount_adj; 1250 resp->tr_qm_dqalloc.tr_logcount += logcount_adj; 1251 } 1252