1 /* 2 * Copyright (c) 2000-2002 Silicon Graphics, Inc. 3 * All Rights Reserved. 4 * 5 * This program is free software; you can redistribute it and/or 6 * modify it under the terms of the GNU General Public License as 7 * published by the Free Software Foundation. 8 * 9 * This program is distributed in the hope that it would be useful, 10 * but WITHOUT ANY WARRANTY; without even the implied warranty of 11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 12 * GNU General Public License for more details. 13 * 14 * You should have received a copy of the GNU General Public License 15 * along with this program; if not, write the Free Software Foundation, 16 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA 17 */ 18 #include "xfs.h" 19 #include "xfs_fs.h" 20 #include "xfs_log.h" 21 #include "xfs_trans.h" 22 #include "xfs_sb.h" 23 #include "xfs_ag.h" 24 #include "xfs_alloc.h" 25 #include "xfs_quota.h" 26 #include "xfs_mount.h" 27 #include "xfs_bmap_btree.h" 28 #include "xfs_inode.h" 29 #include "xfs_itable.h" 30 #include "xfs_bmap.h" 31 #include "xfs_rtalloc.h" 32 #include "xfs_error.h" 33 #include "xfs_attr.h" 34 #include "xfs_buf_item.h" 35 #include "xfs_trans_priv.h" 36 #include "xfs_qm.h" 37 38 STATIC void xfs_trans_alloc_dqinfo(xfs_trans_t *); 39 40 /* 41 * Add the locked dquot to the transaction. 42 * The dquot must be locked, and it cannot be associated with any 43 * transaction. 44 */ 45 void 46 xfs_trans_dqjoin( 47 xfs_trans_t *tp, 48 xfs_dquot_t *dqp) 49 { 50 ASSERT(dqp->q_transp != tp); 51 ASSERT(XFS_DQ_IS_LOCKED(dqp)); 52 ASSERT(dqp->q_logitem.qli_dquot == dqp); 53 54 /* 55 * Get a log_item_desc to point at the new item. 56 */ 57 xfs_trans_add_item(tp, &dqp->q_logitem.qli_item); 58 59 /* 60 * Initialize d_transp so we can later determine if this dquot is 61 * associated with this transaction. 62 */ 63 dqp->q_transp = tp; 64 } 65 66 67 /* 68 * This is called to mark the dquot as needing 69 * to be logged when the transaction is committed. The dquot must 70 * already be associated with the given transaction. 71 * Note that it marks the entire transaction as dirty. In the ordinary 72 * case, this gets called via xfs_trans_commit, after the transaction 73 * is already dirty. However, there's nothing stop this from getting 74 * called directly, as done by xfs_qm_scall_setqlim. Hence, the TRANS_DIRTY 75 * flag. 76 */ 77 void 78 xfs_trans_log_dquot( 79 xfs_trans_t *tp, 80 xfs_dquot_t *dqp) 81 { 82 ASSERT(dqp->q_transp == tp); 83 ASSERT(XFS_DQ_IS_LOCKED(dqp)); 84 85 tp->t_flags |= XFS_TRANS_DIRTY; 86 dqp->q_logitem.qli_item.li_desc->lid_flags |= XFS_LID_DIRTY; 87 } 88 89 /* 90 * Carry forward whatever is left of the quota blk reservation to 91 * the spanky new transaction 92 */ 93 void 94 xfs_trans_dup_dqinfo( 95 xfs_trans_t *otp, 96 xfs_trans_t *ntp) 97 { 98 xfs_dqtrx_t *oq, *nq; 99 int i,j; 100 xfs_dqtrx_t *oqa, *nqa; 101 102 if (!otp->t_dqinfo) 103 return; 104 105 xfs_trans_alloc_dqinfo(ntp); 106 107 /* 108 * Because the quota blk reservation is carried forward, 109 * it is also necessary to carry forward the DQ_DIRTY flag. 110 */ 111 if(otp->t_flags & XFS_TRANS_DQ_DIRTY) 112 ntp->t_flags |= XFS_TRANS_DQ_DIRTY; 113 114 for (j = 0; j < XFS_QM_TRANS_DQTYPES; j++) { 115 oqa = otp->t_dqinfo->dqs[j]; 116 nqa = ntp->t_dqinfo->dqs[j]; 117 for (i = 0; i < XFS_QM_TRANS_MAXDQS; i++) { 118 if (oqa[i].qt_dquot == NULL) 119 break; 120 oq = &oqa[i]; 121 nq = &nqa[i]; 122 123 nq->qt_dquot = oq->qt_dquot; 124 nq->qt_bcount_delta = nq->qt_icount_delta = 0; 125 nq->qt_rtbcount_delta = 0; 126 127 /* 128 * Transfer whatever is left of the reservations. 129 */ 130 nq->qt_blk_res = oq->qt_blk_res - oq->qt_blk_res_used; 131 oq->qt_blk_res = oq->qt_blk_res_used; 132 133 nq->qt_rtblk_res = oq->qt_rtblk_res - 134 oq->qt_rtblk_res_used; 135 oq->qt_rtblk_res = oq->qt_rtblk_res_used; 136 137 nq->qt_ino_res = oq->qt_ino_res - oq->qt_ino_res_used; 138 oq->qt_ino_res = oq->qt_ino_res_used; 139 140 } 141 } 142 } 143 144 /* 145 * Wrap around mod_dquot to account for both user and group quotas. 146 */ 147 void 148 xfs_trans_mod_dquot_byino( 149 xfs_trans_t *tp, 150 xfs_inode_t *ip, 151 uint field, 152 long delta) 153 { 154 xfs_mount_t *mp = tp->t_mountp; 155 156 if (!XFS_IS_QUOTA_RUNNING(mp) || 157 !XFS_IS_QUOTA_ON(mp) || 158 xfs_is_quota_inode(&mp->m_sb, ip->i_ino)) 159 return; 160 161 if (tp->t_dqinfo == NULL) 162 xfs_trans_alloc_dqinfo(tp); 163 164 if (XFS_IS_UQUOTA_ON(mp) && ip->i_udquot) 165 (void) xfs_trans_mod_dquot(tp, ip->i_udquot, field, delta); 166 if (XFS_IS_OQUOTA_ON(mp) && ip->i_gdquot) 167 (void) xfs_trans_mod_dquot(tp, ip->i_gdquot, field, delta); 168 } 169 170 STATIC struct xfs_dqtrx * 171 xfs_trans_get_dqtrx( 172 struct xfs_trans *tp, 173 struct xfs_dquot *dqp) 174 { 175 int i; 176 struct xfs_dqtrx *qa; 177 178 if (XFS_QM_ISUDQ(dqp)) 179 qa = tp->t_dqinfo->dqs[XFS_QM_TRANS_USR]; 180 else 181 qa = tp->t_dqinfo->dqs[XFS_QM_TRANS_GRP]; 182 183 for (i = 0; i < XFS_QM_TRANS_MAXDQS; i++) { 184 if (qa[i].qt_dquot == NULL || 185 qa[i].qt_dquot == dqp) 186 return &qa[i]; 187 } 188 189 return NULL; 190 } 191 192 /* 193 * Make the changes in the transaction structure. 194 * The moral equivalent to xfs_trans_mod_sb(). 195 * We don't touch any fields in the dquot, so we don't care 196 * if it's locked or not (most of the time it won't be). 197 */ 198 void 199 xfs_trans_mod_dquot( 200 xfs_trans_t *tp, 201 xfs_dquot_t *dqp, 202 uint field, 203 long delta) 204 { 205 xfs_dqtrx_t *qtrx; 206 207 ASSERT(tp); 208 ASSERT(XFS_IS_QUOTA_RUNNING(tp->t_mountp)); 209 qtrx = NULL; 210 211 if (tp->t_dqinfo == NULL) 212 xfs_trans_alloc_dqinfo(tp); 213 /* 214 * Find either the first free slot or the slot that belongs 215 * to this dquot. 216 */ 217 qtrx = xfs_trans_get_dqtrx(tp, dqp); 218 ASSERT(qtrx); 219 if (qtrx->qt_dquot == NULL) 220 qtrx->qt_dquot = dqp; 221 222 switch (field) { 223 224 /* 225 * regular disk blk reservation 226 */ 227 case XFS_TRANS_DQ_RES_BLKS: 228 qtrx->qt_blk_res += (ulong)delta; 229 break; 230 231 /* 232 * inode reservation 233 */ 234 case XFS_TRANS_DQ_RES_INOS: 235 qtrx->qt_ino_res += (ulong)delta; 236 break; 237 238 /* 239 * disk blocks used. 240 */ 241 case XFS_TRANS_DQ_BCOUNT: 242 if (qtrx->qt_blk_res && delta > 0) { 243 qtrx->qt_blk_res_used += (ulong)delta; 244 ASSERT(qtrx->qt_blk_res >= qtrx->qt_blk_res_used); 245 } 246 qtrx->qt_bcount_delta += delta; 247 break; 248 249 case XFS_TRANS_DQ_DELBCOUNT: 250 qtrx->qt_delbcnt_delta += delta; 251 break; 252 253 /* 254 * Inode Count 255 */ 256 case XFS_TRANS_DQ_ICOUNT: 257 if (qtrx->qt_ino_res && delta > 0) { 258 qtrx->qt_ino_res_used += (ulong)delta; 259 ASSERT(qtrx->qt_ino_res >= qtrx->qt_ino_res_used); 260 } 261 qtrx->qt_icount_delta += delta; 262 break; 263 264 /* 265 * rtblk reservation 266 */ 267 case XFS_TRANS_DQ_RES_RTBLKS: 268 qtrx->qt_rtblk_res += (ulong)delta; 269 break; 270 271 /* 272 * rtblk count 273 */ 274 case XFS_TRANS_DQ_RTBCOUNT: 275 if (qtrx->qt_rtblk_res && delta > 0) { 276 qtrx->qt_rtblk_res_used += (ulong)delta; 277 ASSERT(qtrx->qt_rtblk_res >= qtrx->qt_rtblk_res_used); 278 } 279 qtrx->qt_rtbcount_delta += delta; 280 break; 281 282 case XFS_TRANS_DQ_DELRTBCOUNT: 283 qtrx->qt_delrtb_delta += delta; 284 break; 285 286 default: 287 ASSERT(0); 288 } 289 tp->t_flags |= XFS_TRANS_DQ_DIRTY; 290 } 291 292 293 /* 294 * Given an array of dqtrx structures, lock all the dquots associated 295 * and join them to the transaction, provided they have been modified. 296 * We know that the highest number of dquots (of one type - usr OR grp), 297 * involved in a transaction is 2 and that both usr and grp combined - 3. 298 * So, we don't attempt to make this very generic. 299 */ 300 STATIC void 301 xfs_trans_dqlockedjoin( 302 xfs_trans_t *tp, 303 xfs_dqtrx_t *q) 304 { 305 ASSERT(q[0].qt_dquot != NULL); 306 if (q[1].qt_dquot == NULL) { 307 xfs_dqlock(q[0].qt_dquot); 308 xfs_trans_dqjoin(tp, q[0].qt_dquot); 309 } else { 310 ASSERT(XFS_QM_TRANS_MAXDQS == 2); 311 xfs_dqlock2(q[0].qt_dquot, q[1].qt_dquot); 312 xfs_trans_dqjoin(tp, q[0].qt_dquot); 313 xfs_trans_dqjoin(tp, q[1].qt_dquot); 314 } 315 } 316 317 318 /* 319 * Called by xfs_trans_commit() and similar in spirit to 320 * xfs_trans_apply_sb_deltas(). 321 * Go thru all the dquots belonging to this transaction and modify the 322 * INCORE dquot to reflect the actual usages. 323 * Unreserve just the reservations done by this transaction. 324 * dquot is still left locked at exit. 325 */ 326 void 327 xfs_trans_apply_dquot_deltas( 328 struct xfs_trans *tp) 329 { 330 int i, j; 331 struct xfs_dquot *dqp; 332 struct xfs_dqtrx *qtrx, *qa; 333 struct xfs_disk_dquot *d; 334 long totalbdelta; 335 long totalrtbdelta; 336 337 if (!(tp->t_flags & XFS_TRANS_DQ_DIRTY)) 338 return; 339 340 ASSERT(tp->t_dqinfo); 341 for (j = 0; j < XFS_QM_TRANS_DQTYPES; j++) { 342 qa = tp->t_dqinfo->dqs[j]; 343 if (qa[0].qt_dquot == NULL) 344 continue; 345 346 /* 347 * Lock all of the dquots and join them to the transaction. 348 */ 349 xfs_trans_dqlockedjoin(tp, qa); 350 351 for (i = 0; i < XFS_QM_TRANS_MAXDQS; i++) { 352 qtrx = &qa[i]; 353 /* 354 * The array of dquots is filled 355 * sequentially, not sparsely. 356 */ 357 if ((dqp = qtrx->qt_dquot) == NULL) 358 break; 359 360 ASSERT(XFS_DQ_IS_LOCKED(dqp)); 361 ASSERT(dqp->q_transp == tp); 362 363 /* 364 * adjust the actual number of blocks used 365 */ 366 d = &dqp->q_core; 367 368 /* 369 * The issue here is - sometimes we don't make a blkquota 370 * reservation intentionally to be fair to users 371 * (when the amount is small). On the other hand, 372 * delayed allocs do make reservations, but that's 373 * outside of a transaction, so we have no 374 * idea how much was really reserved. 375 * So, here we've accumulated delayed allocation blks and 376 * non-delay blks. The assumption is that the 377 * delayed ones are always reserved (outside of a 378 * transaction), and the others may or may not have 379 * quota reservations. 380 */ 381 totalbdelta = qtrx->qt_bcount_delta + 382 qtrx->qt_delbcnt_delta; 383 totalrtbdelta = qtrx->qt_rtbcount_delta + 384 qtrx->qt_delrtb_delta; 385 #ifdef DEBUG 386 if (totalbdelta < 0) 387 ASSERT(be64_to_cpu(d->d_bcount) >= 388 -totalbdelta); 389 390 if (totalrtbdelta < 0) 391 ASSERT(be64_to_cpu(d->d_rtbcount) >= 392 -totalrtbdelta); 393 394 if (qtrx->qt_icount_delta < 0) 395 ASSERT(be64_to_cpu(d->d_icount) >= 396 -qtrx->qt_icount_delta); 397 #endif 398 if (totalbdelta) 399 be64_add_cpu(&d->d_bcount, (xfs_qcnt_t)totalbdelta); 400 401 if (qtrx->qt_icount_delta) 402 be64_add_cpu(&d->d_icount, (xfs_qcnt_t)qtrx->qt_icount_delta); 403 404 if (totalrtbdelta) 405 be64_add_cpu(&d->d_rtbcount, (xfs_qcnt_t)totalrtbdelta); 406 407 /* 408 * Get any default limits in use. 409 * Start/reset the timer(s) if needed. 410 */ 411 if (d->d_id) { 412 xfs_qm_adjust_dqlimits(tp->t_mountp, dqp); 413 xfs_qm_adjust_dqtimers(tp->t_mountp, d); 414 } 415 416 dqp->dq_flags |= XFS_DQ_DIRTY; 417 /* 418 * add this to the list of items to get logged 419 */ 420 xfs_trans_log_dquot(tp, dqp); 421 /* 422 * Take off what's left of the original reservation. 423 * In case of delayed allocations, there's no 424 * reservation that a transaction structure knows of. 425 */ 426 if (qtrx->qt_blk_res != 0) { 427 if (qtrx->qt_blk_res != qtrx->qt_blk_res_used) { 428 if (qtrx->qt_blk_res > 429 qtrx->qt_blk_res_used) 430 dqp->q_res_bcount -= (xfs_qcnt_t) 431 (qtrx->qt_blk_res - 432 qtrx->qt_blk_res_used); 433 else 434 dqp->q_res_bcount -= (xfs_qcnt_t) 435 (qtrx->qt_blk_res_used - 436 qtrx->qt_blk_res); 437 } 438 } else { 439 /* 440 * These blks were never reserved, either inside 441 * a transaction or outside one (in a delayed 442 * allocation). Also, this isn't always a 443 * negative number since we sometimes 444 * deliberately skip quota reservations. 445 */ 446 if (qtrx->qt_bcount_delta) { 447 dqp->q_res_bcount += 448 (xfs_qcnt_t)qtrx->qt_bcount_delta; 449 } 450 } 451 /* 452 * Adjust the RT reservation. 453 */ 454 if (qtrx->qt_rtblk_res != 0) { 455 if (qtrx->qt_rtblk_res != qtrx->qt_rtblk_res_used) { 456 if (qtrx->qt_rtblk_res > 457 qtrx->qt_rtblk_res_used) 458 dqp->q_res_rtbcount -= (xfs_qcnt_t) 459 (qtrx->qt_rtblk_res - 460 qtrx->qt_rtblk_res_used); 461 else 462 dqp->q_res_rtbcount -= (xfs_qcnt_t) 463 (qtrx->qt_rtblk_res_used - 464 qtrx->qt_rtblk_res); 465 } 466 } else { 467 if (qtrx->qt_rtbcount_delta) 468 dqp->q_res_rtbcount += 469 (xfs_qcnt_t)qtrx->qt_rtbcount_delta; 470 } 471 472 /* 473 * Adjust the inode reservation. 474 */ 475 if (qtrx->qt_ino_res != 0) { 476 ASSERT(qtrx->qt_ino_res >= 477 qtrx->qt_ino_res_used); 478 if (qtrx->qt_ino_res > qtrx->qt_ino_res_used) 479 dqp->q_res_icount -= (xfs_qcnt_t) 480 (qtrx->qt_ino_res - 481 qtrx->qt_ino_res_used); 482 } else { 483 if (qtrx->qt_icount_delta) 484 dqp->q_res_icount += 485 (xfs_qcnt_t)qtrx->qt_icount_delta; 486 } 487 488 ASSERT(dqp->q_res_bcount >= 489 be64_to_cpu(dqp->q_core.d_bcount)); 490 ASSERT(dqp->q_res_icount >= 491 be64_to_cpu(dqp->q_core.d_icount)); 492 ASSERT(dqp->q_res_rtbcount >= 493 be64_to_cpu(dqp->q_core.d_rtbcount)); 494 } 495 } 496 } 497 498 /* 499 * Release the reservations, and adjust the dquots accordingly. 500 * This is called only when the transaction is being aborted. If by 501 * any chance we have done dquot modifications incore (ie. deltas) already, 502 * we simply throw those away, since that's the expected behavior 503 * when a transaction is curtailed without a commit. 504 */ 505 void 506 xfs_trans_unreserve_and_mod_dquots( 507 xfs_trans_t *tp) 508 { 509 int i, j; 510 xfs_dquot_t *dqp; 511 xfs_dqtrx_t *qtrx, *qa; 512 bool locked; 513 514 if (!tp->t_dqinfo || !(tp->t_flags & XFS_TRANS_DQ_DIRTY)) 515 return; 516 517 for (j = 0; j < XFS_QM_TRANS_DQTYPES; j++) { 518 qa = tp->t_dqinfo->dqs[j]; 519 520 for (i = 0; i < XFS_QM_TRANS_MAXDQS; i++) { 521 qtrx = &qa[i]; 522 /* 523 * We assume that the array of dquots is filled 524 * sequentially, not sparsely. 525 */ 526 if ((dqp = qtrx->qt_dquot) == NULL) 527 break; 528 /* 529 * Unreserve the original reservation. We don't care 530 * about the number of blocks used field, or deltas. 531 * Also we don't bother to zero the fields. 532 */ 533 locked = false; 534 if (qtrx->qt_blk_res) { 535 xfs_dqlock(dqp); 536 locked = true; 537 dqp->q_res_bcount -= 538 (xfs_qcnt_t)qtrx->qt_blk_res; 539 } 540 if (qtrx->qt_ino_res) { 541 if (!locked) { 542 xfs_dqlock(dqp); 543 locked = true; 544 } 545 dqp->q_res_icount -= 546 (xfs_qcnt_t)qtrx->qt_ino_res; 547 } 548 549 if (qtrx->qt_rtblk_res) { 550 if (!locked) { 551 xfs_dqlock(dqp); 552 locked = true; 553 } 554 dqp->q_res_rtbcount -= 555 (xfs_qcnt_t)qtrx->qt_rtblk_res; 556 } 557 if (locked) 558 xfs_dqunlock(dqp); 559 560 } 561 } 562 } 563 564 STATIC void 565 xfs_quota_warn( 566 struct xfs_mount *mp, 567 struct xfs_dquot *dqp, 568 int type) 569 { 570 /* no warnings for project quotas - we just return ENOSPC later */ 571 if (dqp->dq_flags & XFS_DQ_PROJ) 572 return; 573 quota_send_warning(make_kqid(&init_user_ns, 574 (dqp->dq_flags & XFS_DQ_USER) ? 575 USRQUOTA : GRPQUOTA, 576 be32_to_cpu(dqp->q_core.d_id)), 577 mp->m_super->s_dev, type); 578 } 579 580 /* 581 * This reserves disk blocks and inodes against a dquot. 582 * Flags indicate if the dquot is to be locked here and also 583 * if the blk reservation is for RT or regular blocks. 584 * Sending in XFS_QMOPT_FORCE_RES flag skips the quota check. 585 */ 586 STATIC int 587 xfs_trans_dqresv( 588 xfs_trans_t *tp, 589 xfs_mount_t *mp, 590 xfs_dquot_t *dqp, 591 long nblks, 592 long ninos, 593 uint flags) 594 { 595 xfs_qcnt_t hardlimit; 596 xfs_qcnt_t softlimit; 597 time_t timer; 598 xfs_qwarncnt_t warns; 599 xfs_qwarncnt_t warnlimit; 600 xfs_qcnt_t total_count; 601 xfs_qcnt_t *resbcountp; 602 xfs_quotainfo_t *q = mp->m_quotainfo; 603 604 605 xfs_dqlock(dqp); 606 607 if (flags & XFS_TRANS_DQ_RES_BLKS) { 608 hardlimit = be64_to_cpu(dqp->q_core.d_blk_hardlimit); 609 if (!hardlimit) 610 hardlimit = q->qi_bhardlimit; 611 softlimit = be64_to_cpu(dqp->q_core.d_blk_softlimit); 612 if (!softlimit) 613 softlimit = q->qi_bsoftlimit; 614 timer = be32_to_cpu(dqp->q_core.d_btimer); 615 warns = be16_to_cpu(dqp->q_core.d_bwarns); 616 warnlimit = dqp->q_mount->m_quotainfo->qi_bwarnlimit; 617 resbcountp = &dqp->q_res_bcount; 618 } else { 619 ASSERT(flags & XFS_TRANS_DQ_RES_RTBLKS); 620 hardlimit = be64_to_cpu(dqp->q_core.d_rtb_hardlimit); 621 if (!hardlimit) 622 hardlimit = q->qi_rtbhardlimit; 623 softlimit = be64_to_cpu(dqp->q_core.d_rtb_softlimit); 624 if (!softlimit) 625 softlimit = q->qi_rtbsoftlimit; 626 timer = be32_to_cpu(dqp->q_core.d_rtbtimer); 627 warns = be16_to_cpu(dqp->q_core.d_rtbwarns); 628 warnlimit = dqp->q_mount->m_quotainfo->qi_rtbwarnlimit; 629 resbcountp = &dqp->q_res_rtbcount; 630 } 631 632 if ((flags & XFS_QMOPT_FORCE_RES) == 0 && 633 dqp->q_core.d_id && 634 ((XFS_IS_UQUOTA_ENFORCED(dqp->q_mount) && XFS_QM_ISUDQ(dqp)) || 635 (XFS_IS_GQUOTA_ENFORCED(dqp->q_mount) && XFS_QM_ISGDQ(dqp)) || 636 (XFS_IS_PQUOTA_ENFORCED(dqp->q_mount) && XFS_QM_ISPDQ(dqp)))) { 637 if (nblks > 0) { 638 /* 639 * dquot is locked already. See if we'd go over the 640 * hardlimit or exceed the timelimit if we allocate 641 * nblks. 642 */ 643 total_count = *resbcountp + nblks; 644 if (hardlimit && total_count > hardlimit) { 645 xfs_quota_warn(mp, dqp, QUOTA_NL_BHARDWARN); 646 goto error_return; 647 } 648 if (softlimit && total_count > softlimit) { 649 if ((timer != 0 && get_seconds() > timer) || 650 (warns != 0 && warns >= warnlimit)) { 651 xfs_quota_warn(mp, dqp, 652 QUOTA_NL_BSOFTLONGWARN); 653 goto error_return; 654 } 655 656 xfs_quota_warn(mp, dqp, QUOTA_NL_BSOFTWARN); 657 } 658 } 659 if (ninos > 0) { 660 total_count = be64_to_cpu(dqp->q_core.d_icount) + ninos; 661 timer = be32_to_cpu(dqp->q_core.d_itimer); 662 warns = be16_to_cpu(dqp->q_core.d_iwarns); 663 warnlimit = dqp->q_mount->m_quotainfo->qi_iwarnlimit; 664 hardlimit = be64_to_cpu(dqp->q_core.d_ino_hardlimit); 665 if (!hardlimit) 666 hardlimit = q->qi_ihardlimit; 667 softlimit = be64_to_cpu(dqp->q_core.d_ino_softlimit); 668 if (!softlimit) 669 softlimit = q->qi_isoftlimit; 670 671 if (hardlimit && total_count > hardlimit) { 672 xfs_quota_warn(mp, dqp, QUOTA_NL_IHARDWARN); 673 goto error_return; 674 } 675 if (softlimit && total_count > softlimit) { 676 if ((timer != 0 && get_seconds() > timer) || 677 (warns != 0 && warns >= warnlimit)) { 678 xfs_quota_warn(mp, dqp, 679 QUOTA_NL_ISOFTLONGWARN); 680 goto error_return; 681 } 682 xfs_quota_warn(mp, dqp, QUOTA_NL_ISOFTWARN); 683 } 684 } 685 } 686 687 /* 688 * Change the reservation, but not the actual usage. 689 * Note that q_res_bcount = q_core.d_bcount + resv 690 */ 691 (*resbcountp) += (xfs_qcnt_t)nblks; 692 if (ninos != 0) 693 dqp->q_res_icount += (xfs_qcnt_t)ninos; 694 695 /* 696 * note the reservation amt in the trans struct too, 697 * so that the transaction knows how much was reserved by 698 * it against this particular dquot. 699 * We don't do this when we are reserving for a delayed allocation, 700 * because we don't have the luxury of a transaction envelope then. 701 */ 702 if (tp) { 703 ASSERT(tp->t_dqinfo); 704 ASSERT(flags & XFS_QMOPT_RESBLK_MASK); 705 if (nblks != 0) 706 xfs_trans_mod_dquot(tp, dqp, 707 flags & XFS_QMOPT_RESBLK_MASK, 708 nblks); 709 if (ninos != 0) 710 xfs_trans_mod_dquot(tp, dqp, 711 XFS_TRANS_DQ_RES_INOS, 712 ninos); 713 } 714 ASSERT(dqp->q_res_bcount >= be64_to_cpu(dqp->q_core.d_bcount)); 715 ASSERT(dqp->q_res_rtbcount >= be64_to_cpu(dqp->q_core.d_rtbcount)); 716 ASSERT(dqp->q_res_icount >= be64_to_cpu(dqp->q_core.d_icount)); 717 718 xfs_dqunlock(dqp); 719 return 0; 720 721 error_return: 722 xfs_dqunlock(dqp); 723 if (flags & XFS_QMOPT_ENOSPC) 724 return ENOSPC; 725 return EDQUOT; 726 } 727 728 729 /* 730 * Given dquot(s), make disk block and/or inode reservations against them. 731 * The fact that this does the reservation against both the usr and 732 * grp/prj quotas is important, because this follows a both-or-nothing 733 * approach. 734 * 735 * flags = XFS_QMOPT_FORCE_RES evades limit enforcement. Used by chown. 736 * XFS_QMOPT_ENOSPC returns ENOSPC not EDQUOT. Used by pquota. 737 * XFS_TRANS_DQ_RES_BLKS reserves regular disk blocks 738 * XFS_TRANS_DQ_RES_RTBLKS reserves realtime disk blocks 739 * dquots are unlocked on return, if they were not locked by caller. 740 */ 741 int 742 xfs_trans_reserve_quota_bydquots( 743 struct xfs_trans *tp, 744 struct xfs_mount *mp, 745 struct xfs_dquot *udqp, 746 struct xfs_dquot *gdqp, 747 long nblks, 748 long ninos, 749 uint flags) 750 { 751 int error; 752 753 if (!XFS_IS_QUOTA_RUNNING(mp) || !XFS_IS_QUOTA_ON(mp)) 754 return 0; 755 756 if (tp && tp->t_dqinfo == NULL) 757 xfs_trans_alloc_dqinfo(tp); 758 759 ASSERT(flags & XFS_QMOPT_RESBLK_MASK); 760 761 if (udqp) { 762 error = xfs_trans_dqresv(tp, mp, udqp, nblks, ninos, 763 (flags & ~XFS_QMOPT_ENOSPC)); 764 if (error) 765 return error; 766 } 767 768 if (gdqp) { 769 error = xfs_trans_dqresv(tp, mp, gdqp, nblks, ninos, flags); 770 if (error) 771 goto unwind_usr; 772 } 773 774 /* 775 * Didn't change anything critical, so, no need to log 776 */ 777 return 0; 778 779 unwind_usr: 780 flags |= XFS_QMOPT_FORCE_RES; 781 if (udqp) 782 xfs_trans_dqresv(tp, mp, udqp, -nblks, -ninos, flags); 783 return error; 784 } 785 786 787 /* 788 * Lock the dquot and change the reservation if we can. 789 * This doesn't change the actual usage, just the reservation. 790 * The inode sent in is locked. 791 */ 792 int 793 xfs_trans_reserve_quota_nblks( 794 struct xfs_trans *tp, 795 struct xfs_inode *ip, 796 long nblks, 797 long ninos, 798 uint flags) 799 { 800 struct xfs_mount *mp = ip->i_mount; 801 802 if (!XFS_IS_QUOTA_RUNNING(mp) || !XFS_IS_QUOTA_ON(mp)) 803 return 0; 804 if (XFS_IS_PQUOTA_ON(mp)) 805 flags |= XFS_QMOPT_ENOSPC; 806 807 ASSERT(!xfs_is_quota_inode(&mp->m_sb, ip->i_ino)); 808 809 ASSERT(xfs_isilocked(ip, XFS_ILOCK_EXCL)); 810 ASSERT((flags & ~(XFS_QMOPT_FORCE_RES | XFS_QMOPT_ENOSPC)) == 811 XFS_TRANS_DQ_RES_RTBLKS || 812 (flags & ~(XFS_QMOPT_FORCE_RES | XFS_QMOPT_ENOSPC)) == 813 XFS_TRANS_DQ_RES_BLKS); 814 815 /* 816 * Reserve nblks against these dquots, with trans as the mediator. 817 */ 818 return xfs_trans_reserve_quota_bydquots(tp, mp, 819 ip->i_udquot, ip->i_gdquot, 820 nblks, ninos, flags); 821 } 822 823 /* 824 * This routine is called to allocate a quotaoff log item. 825 */ 826 xfs_qoff_logitem_t * 827 xfs_trans_get_qoff_item( 828 xfs_trans_t *tp, 829 xfs_qoff_logitem_t *startqoff, 830 uint flags) 831 { 832 xfs_qoff_logitem_t *q; 833 834 ASSERT(tp != NULL); 835 836 q = xfs_qm_qoff_logitem_init(tp->t_mountp, startqoff, flags); 837 ASSERT(q != NULL); 838 839 /* 840 * Get a log_item_desc to point at the new item. 841 */ 842 xfs_trans_add_item(tp, &q->qql_item); 843 return q; 844 } 845 846 847 /* 848 * This is called to mark the quotaoff logitem as needing 849 * to be logged when the transaction is committed. The logitem must 850 * already be associated with the given transaction. 851 */ 852 void 853 xfs_trans_log_quotaoff_item( 854 xfs_trans_t *tp, 855 xfs_qoff_logitem_t *qlp) 856 { 857 tp->t_flags |= XFS_TRANS_DIRTY; 858 qlp->qql_item.li_desc->lid_flags |= XFS_LID_DIRTY; 859 } 860 861 STATIC void 862 xfs_trans_alloc_dqinfo( 863 xfs_trans_t *tp) 864 { 865 tp->t_dqinfo = kmem_zone_zalloc(xfs_qm_dqtrxzone, KM_SLEEP); 866 } 867 868 void 869 xfs_trans_free_dqinfo( 870 xfs_trans_t *tp) 871 { 872 if (!tp->t_dqinfo) 873 return; 874 kmem_zone_free(xfs_qm_dqtrxzone, tp->t_dqinfo); 875 tp->t_dqinfo = NULL; 876 } 877