1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * vfsv0 quota IO operations on file 4 */ 5 6 #include <linux/errno.h> 7 #include <linux/fs.h> 8 #include <linux/mount.h> 9 #include <linux/dqblk_v2.h> 10 #include <linux/kernel.h> 11 #include <linux/init.h> 12 #include <linux/module.h> 13 #include <linux/slab.h> 14 #include <linux/quotaops.h> 15 16 #include <asm/byteorder.h> 17 18 #include "quota_tree.h" 19 20 MODULE_AUTHOR("Jan Kara"); 21 MODULE_DESCRIPTION("Quota trie support"); 22 MODULE_LICENSE("GPL"); 23 24 /* 25 * Maximum quota tree depth we support. Only to limit recursion when working 26 * with the tree. 27 */ 28 #define MAX_QTREE_DEPTH 6 29 30 #define __QUOTA_QT_PARANOIA 31 32 static int __get_index(struct qtree_mem_dqinfo *info, qid_t id, int depth) 33 { 34 unsigned int epb = info->dqi_usable_bs >> 2; 35 36 depth = info->dqi_qtree_depth - depth - 1; 37 while (depth--) 38 id /= epb; 39 return id % epb; 40 } 41 42 static int get_index(struct qtree_mem_dqinfo *info, struct kqid qid, int depth) 43 { 44 qid_t id = from_kqid(&init_user_ns, qid); 45 46 return __get_index(info, id, depth); 47 } 48 49 /* Number of entries in one blocks */ 50 static int qtree_dqstr_in_blk(struct qtree_mem_dqinfo *info) 51 { 52 return (info->dqi_usable_bs - sizeof(struct qt_disk_dqdbheader)) 53 / info->dqi_entry_size; 54 } 55 56 static ssize_t read_blk(struct qtree_mem_dqinfo *info, uint blk, char *buf) 57 { 58 struct super_block *sb = info->dqi_sb; 59 60 memset(buf, 0, info->dqi_usable_bs); 61 return sb->s_op->quota_read(sb, info->dqi_type, buf, 62 info->dqi_usable_bs, (loff_t)blk << info->dqi_blocksize_bits); 63 } 64 65 static ssize_t write_blk(struct qtree_mem_dqinfo *info, uint blk, char *buf) 66 { 67 struct super_block *sb = info->dqi_sb; 68 ssize_t ret; 69 70 ret = sb->s_op->quota_write(sb, info->dqi_type, buf, 71 info->dqi_usable_bs, (loff_t)blk << info->dqi_blocksize_bits); 72 if (ret != info->dqi_usable_bs) { 73 quota_error(sb, "dquota write failed"); 74 if (ret >= 0) 75 ret = -EIO; 76 } 77 return ret; 78 } 79 80 static inline int do_check_range(struct super_block *sb, const char *val_name, 81 uint val, uint min_val, uint max_val) 82 { 83 if (val < min_val || val > max_val) { 84 quota_error(sb, "Getting %s %u out of range %u-%u", 85 val_name, val, min_val, max_val); 86 return -EUCLEAN; 87 } 88 89 return 0; 90 } 91 92 static int check_dquot_block_header(struct qtree_mem_dqinfo *info, 93 struct qt_disk_dqdbheader *dh) 94 { 95 int err = 0; 96 97 err = do_check_range(info->dqi_sb, "dqdh_next_free", 98 le32_to_cpu(dh->dqdh_next_free), 0, 99 info->dqi_blocks - 1); 100 if (err) 101 return err; 102 err = do_check_range(info->dqi_sb, "dqdh_prev_free", 103 le32_to_cpu(dh->dqdh_prev_free), 0, 104 info->dqi_blocks - 1); 105 if (err) 106 return err; 107 err = do_check_range(info->dqi_sb, "dqdh_entries", 108 le16_to_cpu(dh->dqdh_entries), 0, 109 qtree_dqstr_in_blk(info)); 110 111 return err; 112 } 113 114 /* Remove empty block from list and return it */ 115 static int get_free_dqblk(struct qtree_mem_dqinfo *info) 116 { 117 char *buf = kmalloc(info->dqi_usable_bs, GFP_KERNEL); 118 struct qt_disk_dqdbheader *dh = (struct qt_disk_dqdbheader *)buf; 119 int ret, blk; 120 121 if (!buf) 122 return -ENOMEM; 123 if (info->dqi_free_blk) { 124 blk = info->dqi_free_blk; 125 ret = read_blk(info, blk, buf); 126 if (ret < 0) 127 goto out_buf; 128 ret = check_dquot_block_header(info, dh); 129 if (ret) 130 goto out_buf; 131 info->dqi_free_blk = le32_to_cpu(dh->dqdh_next_free); 132 } 133 else { 134 memset(buf, 0, info->dqi_usable_bs); 135 /* Assure block allocation... */ 136 ret = write_blk(info, info->dqi_blocks, buf); 137 if (ret < 0) 138 goto out_buf; 139 blk = info->dqi_blocks++; 140 } 141 mark_info_dirty(info->dqi_sb, info->dqi_type); 142 ret = blk; 143 out_buf: 144 kfree(buf); 145 return ret; 146 } 147 148 /* Insert empty block to the list */ 149 static int put_free_dqblk(struct qtree_mem_dqinfo *info, char *buf, uint blk) 150 { 151 struct qt_disk_dqdbheader *dh = (struct qt_disk_dqdbheader *)buf; 152 int err; 153 154 dh->dqdh_next_free = cpu_to_le32(info->dqi_free_blk); 155 dh->dqdh_prev_free = cpu_to_le32(0); 156 dh->dqdh_entries = cpu_to_le16(0); 157 err = write_blk(info, blk, buf); 158 if (err < 0) 159 return err; 160 info->dqi_free_blk = blk; 161 mark_info_dirty(info->dqi_sb, info->dqi_type); 162 return 0; 163 } 164 165 /* Remove given block from the list of blocks with free entries */ 166 static int remove_free_dqentry(struct qtree_mem_dqinfo *info, char *buf, 167 uint blk) 168 { 169 char *tmpbuf = kmalloc(info->dqi_usable_bs, GFP_KERNEL); 170 struct qt_disk_dqdbheader *dh = (struct qt_disk_dqdbheader *)buf; 171 uint nextblk = le32_to_cpu(dh->dqdh_next_free); 172 uint prevblk = le32_to_cpu(dh->dqdh_prev_free); 173 int err; 174 175 if (!tmpbuf) 176 return -ENOMEM; 177 if (nextblk) { 178 err = read_blk(info, nextblk, tmpbuf); 179 if (err < 0) 180 goto out_buf; 181 ((struct qt_disk_dqdbheader *)tmpbuf)->dqdh_prev_free = 182 dh->dqdh_prev_free; 183 err = write_blk(info, nextblk, tmpbuf); 184 if (err < 0) 185 goto out_buf; 186 } 187 if (prevblk) { 188 err = read_blk(info, prevblk, tmpbuf); 189 if (err < 0) 190 goto out_buf; 191 ((struct qt_disk_dqdbheader *)tmpbuf)->dqdh_next_free = 192 dh->dqdh_next_free; 193 err = write_blk(info, prevblk, tmpbuf); 194 if (err < 0) 195 goto out_buf; 196 } else { 197 info->dqi_free_entry = nextblk; 198 mark_info_dirty(info->dqi_sb, info->dqi_type); 199 } 200 kfree(tmpbuf); 201 dh->dqdh_next_free = dh->dqdh_prev_free = cpu_to_le32(0); 202 /* No matter whether write succeeds block is out of list */ 203 if (write_blk(info, blk, buf) < 0) 204 quota_error(info->dqi_sb, "Can't write block (%u) " 205 "with free entries", blk); 206 return 0; 207 out_buf: 208 kfree(tmpbuf); 209 return err; 210 } 211 212 /* Insert given block to the beginning of list with free entries */ 213 static int insert_free_dqentry(struct qtree_mem_dqinfo *info, char *buf, 214 uint blk) 215 { 216 char *tmpbuf = kmalloc(info->dqi_usable_bs, GFP_KERNEL); 217 struct qt_disk_dqdbheader *dh = (struct qt_disk_dqdbheader *)buf; 218 int err; 219 220 if (!tmpbuf) 221 return -ENOMEM; 222 dh->dqdh_next_free = cpu_to_le32(info->dqi_free_entry); 223 dh->dqdh_prev_free = cpu_to_le32(0); 224 err = write_blk(info, blk, buf); 225 if (err < 0) 226 goto out_buf; 227 if (info->dqi_free_entry) { 228 err = read_blk(info, info->dqi_free_entry, tmpbuf); 229 if (err < 0) 230 goto out_buf; 231 ((struct qt_disk_dqdbheader *)tmpbuf)->dqdh_prev_free = 232 cpu_to_le32(blk); 233 err = write_blk(info, info->dqi_free_entry, tmpbuf); 234 if (err < 0) 235 goto out_buf; 236 } 237 kfree(tmpbuf); 238 info->dqi_free_entry = blk; 239 mark_info_dirty(info->dqi_sb, info->dqi_type); 240 return 0; 241 out_buf: 242 kfree(tmpbuf); 243 return err; 244 } 245 246 /* Is the entry in the block free? */ 247 int qtree_entry_unused(struct qtree_mem_dqinfo *info, char *disk) 248 { 249 int i; 250 251 for (i = 0; i < info->dqi_entry_size; i++) 252 if (disk[i]) 253 return 0; 254 return 1; 255 } 256 EXPORT_SYMBOL(qtree_entry_unused); 257 258 /* Find space for dquot */ 259 static uint find_free_dqentry(struct qtree_mem_dqinfo *info, 260 struct dquot *dquot, int *err) 261 { 262 uint blk, i; 263 struct qt_disk_dqdbheader *dh; 264 char *buf = kmalloc(info->dqi_usable_bs, GFP_KERNEL); 265 char *ddquot; 266 267 *err = 0; 268 if (!buf) { 269 *err = -ENOMEM; 270 return 0; 271 } 272 dh = (struct qt_disk_dqdbheader *)buf; 273 if (info->dqi_free_entry) { 274 blk = info->dqi_free_entry; 275 *err = read_blk(info, blk, buf); 276 if (*err < 0) 277 goto out_buf; 278 *err = check_dquot_block_header(info, dh); 279 if (*err) 280 goto out_buf; 281 } else { 282 blk = get_free_dqblk(info); 283 if ((int)blk < 0) { 284 *err = blk; 285 kfree(buf); 286 return 0; 287 } 288 memset(buf, 0, info->dqi_usable_bs); 289 /* This is enough as the block is already zeroed and the entry 290 * list is empty... */ 291 info->dqi_free_entry = blk; 292 mark_info_dirty(dquot->dq_sb, dquot->dq_id.type); 293 } 294 /* Block will be full? */ 295 if (le16_to_cpu(dh->dqdh_entries) + 1 >= qtree_dqstr_in_blk(info)) { 296 *err = remove_free_dqentry(info, buf, blk); 297 if (*err < 0) { 298 quota_error(dquot->dq_sb, "Can't remove block (%u) " 299 "from entry free list", blk); 300 goto out_buf; 301 } 302 } 303 le16_add_cpu(&dh->dqdh_entries, 1); 304 /* Find free structure in block */ 305 ddquot = buf + sizeof(struct qt_disk_dqdbheader); 306 for (i = 0; i < qtree_dqstr_in_blk(info); i++) { 307 if (qtree_entry_unused(info, ddquot)) 308 break; 309 ddquot += info->dqi_entry_size; 310 } 311 #ifdef __QUOTA_QT_PARANOIA 312 if (i == qtree_dqstr_in_blk(info)) { 313 quota_error(dquot->dq_sb, "Data block full but it shouldn't"); 314 *err = -EIO; 315 goto out_buf; 316 } 317 #endif 318 *err = write_blk(info, blk, buf); 319 if (*err < 0) { 320 quota_error(dquot->dq_sb, "Can't write quota data block %u", 321 blk); 322 goto out_buf; 323 } 324 dquot->dq_off = ((loff_t)blk << info->dqi_blocksize_bits) + 325 sizeof(struct qt_disk_dqdbheader) + 326 i * info->dqi_entry_size; 327 kfree(buf); 328 return blk; 329 out_buf: 330 kfree(buf); 331 return 0; 332 } 333 334 /* Insert reference to structure into the trie */ 335 static int do_insert_tree(struct qtree_mem_dqinfo *info, struct dquot *dquot, 336 uint *blks, int depth) 337 { 338 char *buf = kmalloc(info->dqi_usable_bs, GFP_KERNEL); 339 int ret = 0, newson = 0, newact = 0; 340 __le32 *ref; 341 uint newblk; 342 int i; 343 344 if (!buf) 345 return -ENOMEM; 346 if (!blks[depth]) { 347 ret = get_free_dqblk(info); 348 if (ret < 0) 349 goto out_buf; 350 for (i = 0; i < depth; i++) 351 if (ret == blks[i]) { 352 quota_error(dquot->dq_sb, 353 "Free block already used in tree: block %u", 354 ret); 355 ret = -EIO; 356 goto out_buf; 357 } 358 blks[depth] = ret; 359 memset(buf, 0, info->dqi_usable_bs); 360 newact = 1; 361 } else { 362 ret = read_blk(info, blks[depth], buf); 363 if (ret < 0) { 364 quota_error(dquot->dq_sb, "Can't read tree quota " 365 "block %u", blks[depth]); 366 goto out_buf; 367 } 368 } 369 ref = (__le32 *)buf; 370 newblk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]); 371 ret = do_check_range(dquot->dq_sb, "block", newblk, 0, 372 info->dqi_blocks - 1); 373 if (ret) 374 goto out_buf; 375 if (!newblk) { 376 newson = 1; 377 } else { 378 for (i = 0; i <= depth; i++) 379 if (newblk == blks[i]) { 380 quota_error(dquot->dq_sb, 381 "Cycle in quota tree detected: block %u index %u", 382 blks[depth], 383 get_index(info, dquot->dq_id, depth)); 384 ret = -EIO; 385 goto out_buf; 386 } 387 } 388 blks[depth + 1] = newblk; 389 if (depth == info->dqi_qtree_depth - 1) { 390 #ifdef __QUOTA_QT_PARANOIA 391 if (newblk) { 392 quota_error(dquot->dq_sb, "Inserting already present " 393 "quota entry (block %u)", 394 le32_to_cpu(ref[get_index(info, 395 dquot->dq_id, depth)])); 396 ret = -EIO; 397 goto out_buf; 398 } 399 #endif 400 blks[depth + 1] = find_free_dqentry(info, dquot, &ret); 401 } else { 402 ret = do_insert_tree(info, dquot, blks, depth + 1); 403 } 404 if (newson && ret >= 0) { 405 ref[get_index(info, dquot->dq_id, depth)] = 406 cpu_to_le32(blks[depth + 1]); 407 ret = write_blk(info, blks[depth], buf); 408 } else if (newact && ret < 0) { 409 put_free_dqblk(info, buf, blks[depth]); 410 } 411 out_buf: 412 kfree(buf); 413 return ret; 414 } 415 416 /* Wrapper for inserting quota structure into tree */ 417 static inline int dq_insert_tree(struct qtree_mem_dqinfo *info, 418 struct dquot *dquot) 419 { 420 uint blks[MAX_QTREE_DEPTH] = { QT_TREEOFF }; 421 422 #ifdef __QUOTA_QT_PARANOIA 423 if (info->dqi_blocks <= QT_TREEOFF) { 424 quota_error(dquot->dq_sb, "Quota tree root isn't allocated!"); 425 return -EIO; 426 } 427 #endif 428 if (info->dqi_qtree_depth >= MAX_QTREE_DEPTH) { 429 quota_error(dquot->dq_sb, "Quota tree depth too big!"); 430 return -EIO; 431 } 432 return do_insert_tree(info, dquot, blks, 0); 433 } 434 435 /* 436 * We don't have to be afraid of deadlocks as we never have quotas on quota 437 * files... 438 */ 439 int qtree_write_dquot(struct qtree_mem_dqinfo *info, struct dquot *dquot) 440 { 441 int type = dquot->dq_id.type; 442 struct super_block *sb = dquot->dq_sb; 443 ssize_t ret; 444 char *ddquot = kmalloc(info->dqi_entry_size, GFP_KERNEL); 445 446 if (!ddquot) 447 return -ENOMEM; 448 449 /* dq_off is guarded by dqio_sem */ 450 if (!dquot->dq_off) { 451 ret = dq_insert_tree(info, dquot); 452 if (ret < 0) { 453 quota_error(sb, "Error %zd occurred while creating " 454 "quota", ret); 455 kfree(ddquot); 456 return ret; 457 } 458 } 459 spin_lock(&dquot->dq_dqb_lock); 460 info->dqi_ops->mem2disk_dqblk(ddquot, dquot); 461 spin_unlock(&dquot->dq_dqb_lock); 462 ret = sb->s_op->quota_write(sb, type, ddquot, info->dqi_entry_size, 463 dquot->dq_off); 464 if (ret != info->dqi_entry_size) { 465 quota_error(sb, "dquota write failed"); 466 if (ret >= 0) 467 ret = -ENOSPC; 468 } else { 469 ret = 0; 470 } 471 dqstats_inc(DQST_WRITES); 472 kfree(ddquot); 473 474 return ret; 475 } 476 EXPORT_SYMBOL(qtree_write_dquot); 477 478 /* Free dquot entry in data block */ 479 static int free_dqentry(struct qtree_mem_dqinfo *info, struct dquot *dquot, 480 uint blk) 481 { 482 struct qt_disk_dqdbheader *dh; 483 char *buf = kmalloc(info->dqi_usable_bs, GFP_KERNEL); 484 int ret = 0; 485 486 if (!buf) 487 return -ENOMEM; 488 if (dquot->dq_off >> info->dqi_blocksize_bits != blk) { 489 quota_error(dquot->dq_sb, "Quota structure has offset to " 490 "other block (%u) than it should (%u)", blk, 491 (uint)(dquot->dq_off >> info->dqi_blocksize_bits)); 492 ret = -EIO; 493 goto out_buf; 494 } 495 ret = read_blk(info, blk, buf); 496 if (ret < 0) { 497 quota_error(dquot->dq_sb, "Can't read quota data block %u", 498 blk); 499 goto out_buf; 500 } 501 dh = (struct qt_disk_dqdbheader *)buf; 502 ret = check_dquot_block_header(info, dh); 503 if (ret) 504 goto out_buf; 505 le16_add_cpu(&dh->dqdh_entries, -1); 506 if (!le16_to_cpu(dh->dqdh_entries)) { /* Block got free? */ 507 ret = remove_free_dqentry(info, buf, blk); 508 if (ret >= 0) 509 ret = put_free_dqblk(info, buf, blk); 510 if (ret < 0) { 511 quota_error(dquot->dq_sb, "Can't move quota data block " 512 "(%u) to free list", blk); 513 goto out_buf; 514 } 515 } else { 516 memset(buf + 517 (dquot->dq_off & ((1 << info->dqi_blocksize_bits) - 1)), 518 0, info->dqi_entry_size); 519 if (le16_to_cpu(dh->dqdh_entries) == 520 qtree_dqstr_in_blk(info) - 1) { 521 /* Insert will write block itself */ 522 ret = insert_free_dqentry(info, buf, blk); 523 if (ret < 0) { 524 quota_error(dquot->dq_sb, "Can't insert quota " 525 "data block (%u) to free entry list", blk); 526 goto out_buf; 527 } 528 } else { 529 ret = write_blk(info, blk, buf); 530 if (ret < 0) { 531 quota_error(dquot->dq_sb, "Can't write quota " 532 "data block %u", blk); 533 goto out_buf; 534 } 535 } 536 } 537 dquot->dq_off = 0; /* Quota is now unattached */ 538 out_buf: 539 kfree(buf); 540 return ret; 541 } 542 543 /* Remove reference to dquot from tree */ 544 static int remove_tree(struct qtree_mem_dqinfo *info, struct dquot *dquot, 545 uint *blks, int depth) 546 { 547 char *buf = kmalloc(info->dqi_usable_bs, GFP_KERNEL); 548 int ret = 0; 549 uint newblk; 550 __le32 *ref = (__le32 *)buf; 551 int i; 552 553 if (!buf) 554 return -ENOMEM; 555 ret = read_blk(info, blks[depth], buf); 556 if (ret < 0) { 557 quota_error(dquot->dq_sb, "Can't read quota data block %u", 558 blks[depth]); 559 goto out_buf; 560 } 561 newblk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]); 562 ret = do_check_range(dquot->dq_sb, "block", newblk, QT_TREEOFF, 563 info->dqi_blocks - 1); 564 if (ret) 565 goto out_buf; 566 567 for (i = 0; i <= depth; i++) 568 if (newblk == blks[i]) { 569 quota_error(dquot->dq_sb, 570 "Cycle in quota tree detected: block %u index %u", 571 blks[depth], 572 get_index(info, dquot->dq_id, depth)); 573 ret = -EIO; 574 goto out_buf; 575 } 576 if (depth == info->dqi_qtree_depth - 1) { 577 ret = free_dqentry(info, dquot, newblk); 578 blks[depth + 1] = 0; 579 } else { 580 blks[depth + 1] = newblk; 581 ret = remove_tree(info, dquot, blks, depth + 1); 582 } 583 if (ret >= 0 && !blks[depth + 1]) { 584 ref[get_index(info, dquot->dq_id, depth)] = cpu_to_le32(0); 585 /* Block got empty? */ 586 for (i = 0; i < (info->dqi_usable_bs >> 2) && !ref[i]; i++) 587 ; 588 /* Don't put the root block into the free block list */ 589 if (i == (info->dqi_usable_bs >> 2) 590 && blks[depth] != QT_TREEOFF) { 591 put_free_dqblk(info, buf, blks[depth]); 592 blks[depth] = 0; 593 } else { 594 ret = write_blk(info, blks[depth], buf); 595 if (ret < 0) 596 quota_error(dquot->dq_sb, 597 "Can't write quota tree block %u", 598 blks[depth]); 599 } 600 } 601 out_buf: 602 kfree(buf); 603 return ret; 604 } 605 606 /* Delete dquot from tree */ 607 int qtree_delete_dquot(struct qtree_mem_dqinfo *info, struct dquot *dquot) 608 { 609 uint blks[MAX_QTREE_DEPTH] = { QT_TREEOFF }; 610 611 if (!dquot->dq_off) /* Even not allocated? */ 612 return 0; 613 if (info->dqi_qtree_depth >= MAX_QTREE_DEPTH) { 614 quota_error(dquot->dq_sb, "Quota tree depth too big!"); 615 return -EIO; 616 } 617 return remove_tree(info, dquot, blks, 0); 618 } 619 EXPORT_SYMBOL(qtree_delete_dquot); 620 621 /* Find entry in block */ 622 static loff_t find_block_dqentry(struct qtree_mem_dqinfo *info, 623 struct dquot *dquot, uint blk) 624 { 625 char *buf = kmalloc(info->dqi_usable_bs, GFP_KERNEL); 626 loff_t ret = 0; 627 int i; 628 char *ddquot; 629 630 if (!buf) 631 return -ENOMEM; 632 ret = read_blk(info, blk, buf); 633 if (ret < 0) { 634 quota_error(dquot->dq_sb, "Can't read quota tree " 635 "block %u", blk); 636 goto out_buf; 637 } 638 ddquot = buf + sizeof(struct qt_disk_dqdbheader); 639 for (i = 0; i < qtree_dqstr_in_blk(info); i++) { 640 if (info->dqi_ops->is_id(ddquot, dquot)) 641 break; 642 ddquot += info->dqi_entry_size; 643 } 644 if (i == qtree_dqstr_in_blk(info)) { 645 quota_error(dquot->dq_sb, 646 "Quota for id %u referenced but not present", 647 from_kqid(&init_user_ns, dquot->dq_id)); 648 ret = -EIO; 649 goto out_buf; 650 } else { 651 ret = ((loff_t)blk << info->dqi_blocksize_bits) + sizeof(struct 652 qt_disk_dqdbheader) + i * info->dqi_entry_size; 653 } 654 out_buf: 655 kfree(buf); 656 return ret; 657 } 658 659 /* Find entry for given id in the tree */ 660 static loff_t find_tree_dqentry(struct qtree_mem_dqinfo *info, 661 struct dquot *dquot, uint *blks, int depth) 662 { 663 char *buf = kmalloc(info->dqi_usable_bs, GFP_KERNEL); 664 loff_t ret = 0; 665 __le32 *ref = (__le32 *)buf; 666 uint blk; 667 int i; 668 669 if (!buf) 670 return -ENOMEM; 671 ret = read_blk(info, blks[depth], buf); 672 if (ret < 0) { 673 quota_error(dquot->dq_sb, "Can't read quota tree block %u", 674 blks[depth]); 675 goto out_buf; 676 } 677 ret = 0; 678 blk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]); 679 if (!blk) /* No reference? */ 680 goto out_buf; 681 ret = do_check_range(dquot->dq_sb, "block", blk, QT_TREEOFF, 682 info->dqi_blocks - 1); 683 if (ret) 684 goto out_buf; 685 686 /* Check for cycles in the tree */ 687 for (i = 0; i <= depth; i++) 688 if (blk == blks[i]) { 689 quota_error(dquot->dq_sb, 690 "Cycle in quota tree detected: block %u index %u", 691 blks[depth], 692 get_index(info, dquot->dq_id, depth)); 693 ret = -EIO; 694 goto out_buf; 695 } 696 blks[depth + 1] = blk; 697 if (depth < info->dqi_qtree_depth - 1) 698 ret = find_tree_dqentry(info, dquot, blks, depth + 1); 699 else 700 ret = find_block_dqentry(info, dquot, blk); 701 out_buf: 702 kfree(buf); 703 return ret; 704 } 705 706 /* Find entry for given id in the tree - wrapper function */ 707 static inline loff_t find_dqentry(struct qtree_mem_dqinfo *info, 708 struct dquot *dquot) 709 { 710 uint blks[MAX_QTREE_DEPTH] = { QT_TREEOFF }; 711 712 if (info->dqi_qtree_depth >= MAX_QTREE_DEPTH) { 713 quota_error(dquot->dq_sb, "Quota tree depth too big!"); 714 return -EIO; 715 } 716 return find_tree_dqentry(info, dquot, blks, 0); 717 } 718 719 int qtree_read_dquot(struct qtree_mem_dqinfo *info, struct dquot *dquot) 720 { 721 int type = dquot->dq_id.type; 722 struct super_block *sb = dquot->dq_sb; 723 loff_t offset; 724 char *ddquot; 725 int ret = 0; 726 727 #ifdef __QUOTA_QT_PARANOIA 728 /* Invalidated quota? */ 729 if (!sb_dqopt(dquot->dq_sb)->files[type]) { 730 quota_error(sb, "Quota invalidated while reading!"); 731 return -EIO; 732 } 733 #endif 734 /* Do we know offset of the dquot entry in the quota file? */ 735 if (!dquot->dq_off) { 736 offset = find_dqentry(info, dquot); 737 if (offset <= 0) { /* Entry not present? */ 738 if (offset < 0) 739 quota_error(sb,"Can't read quota structure " 740 "for id %u", 741 from_kqid(&init_user_ns, 742 dquot->dq_id)); 743 dquot->dq_off = 0; 744 set_bit(DQ_FAKE_B, &dquot->dq_flags); 745 memset(&dquot->dq_dqb, 0, sizeof(struct mem_dqblk)); 746 ret = offset; 747 goto out; 748 } 749 dquot->dq_off = offset; 750 } 751 ddquot = kmalloc(info->dqi_entry_size, GFP_KERNEL); 752 if (!ddquot) 753 return -ENOMEM; 754 ret = sb->s_op->quota_read(sb, type, ddquot, info->dqi_entry_size, 755 dquot->dq_off); 756 if (ret != info->dqi_entry_size) { 757 if (ret >= 0) 758 ret = -EIO; 759 quota_error(sb, "Error while reading quota structure for id %u", 760 from_kqid(&init_user_ns, dquot->dq_id)); 761 set_bit(DQ_FAKE_B, &dquot->dq_flags); 762 memset(&dquot->dq_dqb, 0, sizeof(struct mem_dqblk)); 763 kfree(ddquot); 764 goto out; 765 } 766 spin_lock(&dquot->dq_dqb_lock); 767 info->dqi_ops->disk2mem_dqblk(dquot, ddquot); 768 if (!dquot->dq_dqb.dqb_bhardlimit && 769 !dquot->dq_dqb.dqb_bsoftlimit && 770 !dquot->dq_dqb.dqb_ihardlimit && 771 !dquot->dq_dqb.dqb_isoftlimit) 772 set_bit(DQ_FAKE_B, &dquot->dq_flags); 773 spin_unlock(&dquot->dq_dqb_lock); 774 kfree(ddquot); 775 out: 776 dqstats_inc(DQST_READS); 777 return ret; 778 } 779 EXPORT_SYMBOL(qtree_read_dquot); 780 781 /* Check whether dquot should not be deleted. We know we are 782 * the only one operating on dquot (thanks to dq_lock) */ 783 int qtree_release_dquot(struct qtree_mem_dqinfo *info, struct dquot *dquot) 784 { 785 if (test_bit(DQ_FAKE_B, &dquot->dq_flags) && 786 !(dquot->dq_dqb.dqb_curinodes | dquot->dq_dqb.dqb_curspace)) 787 return qtree_delete_dquot(info, dquot); 788 return 0; 789 } 790 EXPORT_SYMBOL(qtree_release_dquot); 791 792 static int find_next_id(struct qtree_mem_dqinfo *info, qid_t *id, 793 unsigned int blk, int depth) 794 { 795 char *buf = kmalloc(info->dqi_usable_bs, GFP_KERNEL); 796 __le32 *ref = (__le32 *)buf; 797 ssize_t ret; 798 unsigned int epb = info->dqi_usable_bs >> 2; 799 unsigned int level_inc = 1; 800 int i; 801 802 if (!buf) 803 return -ENOMEM; 804 805 for (i = depth; i < info->dqi_qtree_depth - 1; i++) 806 level_inc *= epb; 807 808 ret = read_blk(info, blk, buf); 809 if (ret < 0) { 810 quota_error(info->dqi_sb, 811 "Can't read quota tree block %u", blk); 812 goto out_buf; 813 } 814 for (i = __get_index(info, *id, depth); i < epb; i++) { 815 uint blk_no = le32_to_cpu(ref[i]); 816 817 if (blk_no == 0) { 818 *id += level_inc; 819 continue; 820 } 821 ret = do_check_range(info->dqi_sb, "block", blk_no, 0, 822 info->dqi_blocks - 1); 823 if (ret) 824 goto out_buf; 825 if (depth == info->dqi_qtree_depth - 1) { 826 ret = 0; 827 goto out_buf; 828 } 829 ret = find_next_id(info, id, blk_no, depth + 1); 830 if (ret != -ENOENT) 831 break; 832 } 833 if (i == epb) { 834 ret = -ENOENT; 835 goto out_buf; 836 } 837 out_buf: 838 kfree(buf); 839 return ret; 840 } 841 842 int qtree_get_next_id(struct qtree_mem_dqinfo *info, struct kqid *qid) 843 { 844 qid_t id = from_kqid(&init_user_ns, *qid); 845 int ret; 846 847 ret = find_next_id(info, &id, QT_TREEOFF, 0); 848 if (ret < 0) 849 return ret; 850 *qid = make_kqid(&init_user_ns, qid->type, id); 851 return 0; 852 } 853 EXPORT_SYMBOL(qtree_get_next_id); 854