1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (C) 2012 Red Hat, Inc. 4 * 5 * This file is released under the GPL. 6 */ 7 8 #include "dm-array.h" 9 #include "dm-space-map.h" 10 #include "dm-transaction-manager.h" 11 12 #include <linux/export.h> 13 #include <linux/device-mapper.h> 14 15 #define DM_MSG_PREFIX "array" 16 17 /*----------------------------------------------------------------*/ 18 19 /* 20 * The array is implemented as a fully populated btree, which points to 21 * blocks that contain the packed values. This is more space efficient 22 * than just using a btree since we don't store 1 key per value. 23 */ 24 struct array_block { 25 __le32 csum; 26 __le32 max_entries; 27 __le32 nr_entries; 28 __le32 value_size; 29 __le64 blocknr; /* Block this node is supposed to live in. */ 30 } __packed; 31 32 /*----------------------------------------------------------------*/ 33 34 /* 35 * Validator methods. As usual we calculate a checksum, and also write the 36 * block location into the header (paranoia about ssds remapping areas by 37 * mistake). 38 */ 39 #define CSUM_XOR 595846735 40 41 static void array_block_prepare_for_write(struct dm_block_validator *v, 42 struct dm_block *b, 43 size_t size_of_block) 44 { 45 struct array_block *bh_le = dm_block_data(b); 46 47 bh_le->blocknr = cpu_to_le64(dm_block_location(b)); 48 bh_le->csum = cpu_to_le32(dm_bm_checksum(&bh_le->max_entries, 49 size_of_block - sizeof(__le32), 50 CSUM_XOR)); 51 } 52 53 static int array_block_check(struct dm_block_validator *v, 54 struct dm_block *b, 55 size_t size_of_block) 56 { 57 struct array_block *bh_le = dm_block_data(b); 58 __le32 csum_disk; 59 60 if (dm_block_location(b) != le64_to_cpu(bh_le->blocknr)) { 61 DMERR_LIMIT("%s failed: blocknr %llu != wanted %llu", __func__, 62 (unsigned long long) le64_to_cpu(bh_le->blocknr), 63 (unsigned long long) dm_block_location(b)); 64 return -ENOTBLK; 65 } 66 67 csum_disk = cpu_to_le32(dm_bm_checksum(&bh_le->max_entries, 68 size_of_block - sizeof(__le32), 69 CSUM_XOR)); 70 if (csum_disk != bh_le->csum) { 71 DMERR_LIMIT("%s failed: csum %u != wanted %u", __func__, 72 (unsigned int) le32_to_cpu(csum_disk), 73 (unsigned int) le32_to_cpu(bh_le->csum)); 74 return -EILSEQ; 75 } 76 77 return 0; 78 } 79 80 static struct dm_block_validator array_validator = { 81 .name = "array", 82 .prepare_for_write = array_block_prepare_for_write, 83 .check = array_block_check 84 }; 85 86 /*----------------------------------------------------------------*/ 87 88 /* 89 * Functions for manipulating the array blocks. 90 */ 91 92 /* 93 * Returns a pointer to a value within an array block. 94 * 95 * index - The index into _this_ specific block. 96 */ 97 static void *element_at(struct dm_array_info *info, struct array_block *ab, 98 unsigned int index) 99 { 100 unsigned char *entry = (unsigned char *) (ab + 1); 101 102 entry += index * info->value_type.size; 103 104 return entry; 105 } 106 107 /* 108 * Utility function that calls one of the value_type methods on every value 109 * in an array block. 110 */ 111 static void on_entries(struct dm_array_info *info, struct array_block *ab, 112 void (*fn)(void *, const void *, unsigned int)) 113 { 114 unsigned int nr_entries = le32_to_cpu(ab->nr_entries); 115 116 fn(info->value_type.context, element_at(info, ab, 0), nr_entries); 117 } 118 119 /* 120 * Increment every value in an array block. 121 */ 122 static void inc_ablock_entries(struct dm_array_info *info, struct array_block *ab) 123 { 124 struct dm_btree_value_type *vt = &info->value_type; 125 126 if (vt->inc) 127 on_entries(info, ab, vt->inc); 128 } 129 130 /* 131 * Decrement every value in an array block. 132 */ 133 static void dec_ablock_entries(struct dm_array_info *info, struct array_block *ab) 134 { 135 struct dm_btree_value_type *vt = &info->value_type; 136 137 if (vt->dec) 138 on_entries(info, ab, vt->dec); 139 } 140 141 /* 142 * Each array block can hold this many values. 143 */ 144 static uint32_t calc_max_entries(size_t value_size, size_t size_of_block) 145 { 146 return (size_of_block - sizeof(struct array_block)) / value_size; 147 } 148 149 /* 150 * Allocate a new array block. The caller will need to unlock block. 151 */ 152 static int alloc_ablock(struct dm_array_info *info, size_t size_of_block, 153 uint32_t max_entries, 154 struct dm_block **block, struct array_block **ab) 155 { 156 int r; 157 158 r = dm_tm_new_block(info->btree_info.tm, &array_validator, block); 159 if (r) 160 return r; 161 162 (*ab) = dm_block_data(*block); 163 (*ab)->max_entries = cpu_to_le32(max_entries); 164 (*ab)->nr_entries = cpu_to_le32(0); 165 (*ab)->value_size = cpu_to_le32(info->value_type.size); 166 167 return 0; 168 } 169 170 /* 171 * Pad an array block out with a particular value. Every instance will 172 * cause an increment of the value_type. new_nr must always be more than 173 * the current number of entries. 174 */ 175 static void fill_ablock(struct dm_array_info *info, struct array_block *ab, 176 const void *value, unsigned int new_nr) 177 { 178 uint32_t nr_entries, delta, i; 179 struct dm_btree_value_type *vt = &info->value_type; 180 181 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); 182 BUG_ON(new_nr < le32_to_cpu(ab->nr_entries)); 183 184 nr_entries = le32_to_cpu(ab->nr_entries); 185 delta = new_nr - nr_entries; 186 if (vt->inc) 187 vt->inc(vt->context, value, delta); 188 for (i = nr_entries; i < new_nr; i++) 189 memcpy(element_at(info, ab, i), value, vt->size); 190 ab->nr_entries = cpu_to_le32(new_nr); 191 } 192 193 /* 194 * Remove some entries from the back of an array block. Every value 195 * removed will be decremented. new_nr must be <= the current number of 196 * entries. 197 */ 198 static void trim_ablock(struct dm_array_info *info, struct array_block *ab, 199 unsigned int new_nr) 200 { 201 uint32_t nr_entries, delta; 202 struct dm_btree_value_type *vt = &info->value_type; 203 204 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); 205 BUG_ON(new_nr > le32_to_cpu(ab->nr_entries)); 206 207 nr_entries = le32_to_cpu(ab->nr_entries); 208 delta = nr_entries - new_nr; 209 if (vt->dec) 210 vt->dec(vt->context, element_at(info, ab, new_nr - 1), delta); 211 ab->nr_entries = cpu_to_le32(new_nr); 212 } 213 214 /* 215 * Read locks a block, and coerces it to an array block. The caller must 216 * unlock 'block' when finished. 217 */ 218 static int get_ablock(struct dm_array_info *info, dm_block_t b, 219 struct dm_block **block, struct array_block **ab) 220 { 221 int r; 222 223 r = dm_tm_read_lock(info->btree_info.tm, b, &array_validator, block); 224 if (r) 225 return r; 226 227 *ab = dm_block_data(*block); 228 return 0; 229 } 230 231 /* 232 * Unlocks an array block. 233 */ 234 static void unlock_ablock(struct dm_array_info *info, struct dm_block *block) 235 { 236 dm_tm_unlock(info->btree_info.tm, block); 237 } 238 239 /*----------------------------------------------------------------*/ 240 241 /* 242 * Btree manipulation. 243 */ 244 245 /* 246 * Looks up an array block in the btree, and then read locks it. 247 * 248 * index is the index of the index of the array_block, (ie. the array index 249 * / max_entries). 250 */ 251 static int lookup_ablock(struct dm_array_info *info, dm_block_t root, 252 unsigned int index, struct dm_block **block, 253 struct array_block **ab) 254 { 255 int r; 256 uint64_t key = index; 257 __le64 block_le; 258 259 r = dm_btree_lookup(&info->btree_info, root, &key, &block_le); 260 if (r) 261 return r; 262 263 return get_ablock(info, le64_to_cpu(block_le), block, ab); 264 } 265 266 /* 267 * Insert an array block into the btree. The block is _not_ unlocked. 268 */ 269 static int insert_ablock(struct dm_array_info *info, uint64_t index, 270 struct dm_block *block, dm_block_t *root) 271 { 272 __le64 block_le = cpu_to_le64(dm_block_location(block)); 273 274 __dm_bless_for_disk(block_le); 275 return dm_btree_insert(&info->btree_info, *root, &index, &block_le, root); 276 } 277 278 /*----------------------------------------------------------------*/ 279 280 static int __shadow_ablock(struct dm_array_info *info, dm_block_t b, 281 struct dm_block **block, struct array_block **ab) 282 { 283 int inc; 284 int r = dm_tm_shadow_block(info->btree_info.tm, b, 285 &array_validator, block, &inc); 286 if (r) 287 return r; 288 289 *ab = dm_block_data(*block); 290 if (inc) 291 inc_ablock_entries(info, *ab); 292 293 return 0; 294 } 295 296 /* 297 * The shadow op will often be a noop. Only insert if it really 298 * copied data. 299 */ 300 static int __reinsert_ablock(struct dm_array_info *info, unsigned int index, 301 struct dm_block *block, dm_block_t b, 302 dm_block_t *root) 303 { 304 int r = 0; 305 306 if (dm_block_location(block) != b) { 307 /* 308 * dm_tm_shadow_block will have already decremented the old 309 * block, but it is still referenced by the btree. We 310 * increment to stop the insert decrementing it below zero 311 * when overwriting the old value. 312 */ 313 dm_tm_inc(info->btree_info.tm, b); 314 r = insert_ablock(info, index, block, root); 315 } 316 317 return r; 318 } 319 320 /* 321 * Looks up an array block in the btree. Then shadows it, and updates the 322 * btree to point to this new shadow. 'root' is an input/output parameter 323 * for both the current root block, and the new one. 324 */ 325 static int shadow_ablock(struct dm_array_info *info, dm_block_t *root, 326 unsigned int index, struct dm_block **block, 327 struct array_block **ab) 328 { 329 int r; 330 uint64_t key = index; 331 dm_block_t b; 332 __le64 block_le; 333 334 r = dm_btree_lookup(&info->btree_info, *root, &key, &block_le); 335 if (r) 336 return r; 337 b = le64_to_cpu(block_le); 338 339 r = __shadow_ablock(info, b, block, ab); 340 if (r) 341 return r; 342 343 return __reinsert_ablock(info, index, *block, b, root); 344 } 345 346 /* 347 * Allocate an new array block, and fill it with some values. 348 */ 349 static int insert_new_ablock(struct dm_array_info *info, size_t size_of_block, 350 uint32_t max_entries, 351 unsigned int block_index, uint32_t nr, 352 const void *value, dm_block_t *root) 353 { 354 int r; 355 struct dm_block *block; 356 struct array_block *ab; 357 358 r = alloc_ablock(info, size_of_block, max_entries, &block, &ab); 359 if (r) 360 return r; 361 362 fill_ablock(info, ab, value, nr); 363 r = insert_ablock(info, block_index, block, root); 364 unlock_ablock(info, block); 365 366 return r; 367 } 368 369 static int insert_full_ablocks(struct dm_array_info *info, size_t size_of_block, 370 unsigned int begin_block, unsigned int end_block, 371 unsigned int max_entries, const void *value, 372 dm_block_t *root) 373 { 374 int r = 0; 375 376 for (; !r && begin_block != end_block; begin_block++) 377 r = insert_new_ablock(info, size_of_block, max_entries, begin_block, max_entries, value, root); 378 379 return r; 380 } 381 382 /* 383 * There are a bunch of functions involved with resizing an array. This 384 * structure holds information that commonly needed by them. Purely here 385 * to reduce parameter count. 386 */ 387 struct resize { 388 /* 389 * Describes the array. 390 */ 391 struct dm_array_info *info; 392 393 /* 394 * The current root of the array. This gets updated. 395 */ 396 dm_block_t root; 397 398 /* 399 * Metadata block size. Used to calculate the nr entries in an 400 * array block. 401 */ 402 size_t size_of_block; 403 404 /* 405 * Maximum nr entries in an array block. 406 */ 407 unsigned int max_entries; 408 409 /* 410 * nr of completely full blocks in the array. 411 * 412 * 'old' refers to before the resize, 'new' after. 413 */ 414 unsigned int old_nr_full_blocks, new_nr_full_blocks; 415 416 /* 417 * Number of entries in the final block. 0 iff only full blocks in 418 * the array. 419 */ 420 unsigned int old_nr_entries_in_last_block, new_nr_entries_in_last_block; 421 422 /* 423 * The default value used when growing the array. 424 */ 425 const void *value; 426 }; 427 428 /* 429 * Removes a consecutive set of array blocks from the btree. The values 430 * in block are decremented as a side effect of the btree remove. 431 * 432 * begin_index - the index of the first array block to remove. 433 * end_index - the one-past-the-end value. ie. this block is not removed. 434 */ 435 static int drop_blocks(struct resize *resize, unsigned int begin_index, 436 unsigned int end_index) 437 { 438 int r; 439 440 while (begin_index != end_index) { 441 uint64_t key = begin_index++; 442 443 r = dm_btree_remove(&resize->info->btree_info, resize->root, 444 &key, &resize->root); 445 if (r) 446 return r; 447 } 448 449 return 0; 450 } 451 452 /* 453 * Calculates how many blocks are needed for the array. 454 */ 455 static unsigned int total_nr_blocks_needed(unsigned int nr_full_blocks, 456 unsigned int nr_entries_in_last_block) 457 { 458 return nr_full_blocks + (nr_entries_in_last_block ? 1 : 0); 459 } 460 461 /* 462 * Shrink an array. 463 */ 464 static int shrink(struct resize *resize) 465 { 466 int r; 467 unsigned int begin, end; 468 struct dm_block *block; 469 struct array_block *ab; 470 471 /* 472 * Lose some blocks from the back? 473 */ 474 if (resize->new_nr_full_blocks < resize->old_nr_full_blocks) { 475 begin = total_nr_blocks_needed(resize->new_nr_full_blocks, 476 resize->new_nr_entries_in_last_block); 477 end = total_nr_blocks_needed(resize->old_nr_full_blocks, 478 resize->old_nr_entries_in_last_block); 479 480 r = drop_blocks(resize, begin, end); 481 if (r) 482 return r; 483 } 484 485 /* 486 * Trim the new tail block 487 */ 488 if (resize->new_nr_entries_in_last_block) { 489 r = shadow_ablock(resize->info, &resize->root, 490 resize->new_nr_full_blocks, &block, &ab); 491 if (r) 492 return r; 493 494 trim_ablock(resize->info, ab, resize->new_nr_entries_in_last_block); 495 unlock_ablock(resize->info, block); 496 } 497 498 return 0; 499 } 500 501 /* 502 * Grow an array. 503 */ 504 static int grow_extend_tail_block(struct resize *resize, uint32_t new_nr_entries) 505 { 506 int r; 507 struct dm_block *block; 508 struct array_block *ab; 509 510 r = shadow_ablock(resize->info, &resize->root, 511 resize->old_nr_full_blocks, &block, &ab); 512 if (r) 513 return r; 514 515 fill_ablock(resize->info, ab, resize->value, new_nr_entries); 516 unlock_ablock(resize->info, block); 517 518 return r; 519 } 520 521 static int grow_add_tail_block(struct resize *resize) 522 { 523 return insert_new_ablock(resize->info, resize->size_of_block, 524 resize->max_entries, 525 resize->new_nr_full_blocks, 526 resize->new_nr_entries_in_last_block, 527 resize->value, &resize->root); 528 } 529 530 static int grow_needs_more_blocks(struct resize *resize) 531 { 532 int r; 533 unsigned int old_nr_blocks = resize->old_nr_full_blocks; 534 535 if (resize->old_nr_entries_in_last_block > 0) { 536 old_nr_blocks++; 537 538 r = grow_extend_tail_block(resize, resize->max_entries); 539 if (r) 540 return r; 541 } 542 543 r = insert_full_ablocks(resize->info, resize->size_of_block, 544 old_nr_blocks, 545 resize->new_nr_full_blocks, 546 resize->max_entries, resize->value, 547 &resize->root); 548 if (r) 549 return r; 550 551 if (resize->new_nr_entries_in_last_block) 552 r = grow_add_tail_block(resize); 553 554 return r; 555 } 556 557 static int grow(struct resize *resize) 558 { 559 if (resize->new_nr_full_blocks > resize->old_nr_full_blocks) 560 return grow_needs_more_blocks(resize); 561 562 else if (resize->old_nr_entries_in_last_block) 563 return grow_extend_tail_block(resize, resize->new_nr_entries_in_last_block); 564 565 else 566 return grow_add_tail_block(resize); 567 } 568 569 /*----------------------------------------------------------------*/ 570 571 /* 572 * These are the value_type functions for the btree elements, which point 573 * to array blocks. 574 */ 575 static void block_inc(void *context, const void *value, unsigned int count) 576 { 577 const __le64 *block_le = value; 578 struct dm_array_info *info = context; 579 unsigned int i; 580 581 for (i = 0; i < count; i++, block_le++) 582 dm_tm_inc(info->btree_info.tm, le64_to_cpu(*block_le)); 583 } 584 585 static void __block_dec(void *context, const void *value) 586 { 587 int r; 588 uint64_t b; 589 __le64 block_le; 590 uint32_t ref_count; 591 struct dm_block *block; 592 struct array_block *ab; 593 struct dm_array_info *info = context; 594 595 memcpy(&block_le, value, sizeof(block_le)); 596 b = le64_to_cpu(block_le); 597 598 r = dm_tm_ref(info->btree_info.tm, b, &ref_count); 599 if (r) { 600 DMERR_LIMIT("couldn't get reference count for block %llu", 601 (unsigned long long) b); 602 return; 603 } 604 605 if (ref_count == 1) { 606 /* 607 * We're about to drop the last reference to this ablock. 608 * So we need to decrement the ref count of the contents. 609 */ 610 r = get_ablock(info, b, &block, &ab); 611 if (r) { 612 DMERR_LIMIT("couldn't get array block %llu", 613 (unsigned long long) b); 614 return; 615 } 616 617 dec_ablock_entries(info, ab); 618 unlock_ablock(info, block); 619 } 620 621 dm_tm_dec(info->btree_info.tm, b); 622 } 623 624 static void block_dec(void *context, const void *value, unsigned int count) 625 { 626 unsigned int i; 627 628 for (i = 0; i < count; i++, value += sizeof(__le64)) 629 __block_dec(context, value); 630 } 631 632 static int block_equal(void *context, const void *value1, const void *value2) 633 { 634 return !memcmp(value1, value2, sizeof(__le64)); 635 } 636 637 /*----------------------------------------------------------------*/ 638 639 void dm_array_info_init(struct dm_array_info *info, 640 struct dm_transaction_manager *tm, 641 struct dm_btree_value_type *vt) 642 { 643 struct dm_btree_value_type *bvt = &info->btree_info.value_type; 644 645 memcpy(&info->value_type, vt, sizeof(info->value_type)); 646 info->btree_info.tm = tm; 647 info->btree_info.levels = 1; 648 649 bvt->context = info; 650 bvt->size = sizeof(__le64); 651 bvt->inc = block_inc; 652 bvt->dec = block_dec; 653 bvt->equal = block_equal; 654 } 655 EXPORT_SYMBOL_GPL(dm_array_info_init); 656 657 int dm_array_empty(struct dm_array_info *info, dm_block_t *root) 658 { 659 return dm_btree_empty(&info->btree_info, root); 660 } 661 EXPORT_SYMBOL_GPL(dm_array_empty); 662 663 static int array_resize(struct dm_array_info *info, dm_block_t root, 664 uint32_t old_size, uint32_t new_size, 665 const void *value, dm_block_t *new_root) 666 { 667 int r; 668 struct resize resize; 669 670 if (old_size == new_size) { 671 *new_root = root; 672 return 0; 673 } 674 675 resize.info = info; 676 resize.root = root; 677 resize.size_of_block = dm_bm_block_size(dm_tm_get_bm(info->btree_info.tm)); 678 resize.max_entries = calc_max_entries(info->value_type.size, 679 resize.size_of_block); 680 681 resize.old_nr_full_blocks = old_size / resize.max_entries; 682 resize.old_nr_entries_in_last_block = old_size % resize.max_entries; 683 resize.new_nr_full_blocks = new_size / resize.max_entries; 684 resize.new_nr_entries_in_last_block = new_size % resize.max_entries; 685 resize.value = value; 686 687 r = ((new_size > old_size) ? grow : shrink)(&resize); 688 if (r) 689 return r; 690 691 *new_root = resize.root; 692 return 0; 693 } 694 695 int dm_array_resize(struct dm_array_info *info, dm_block_t root, 696 uint32_t old_size, uint32_t new_size, 697 const void *value, dm_block_t *new_root) 698 __dm_written_to_disk(value) 699 { 700 int r = array_resize(info, root, old_size, new_size, value, new_root); 701 702 __dm_unbless_for_disk(value); 703 return r; 704 } 705 EXPORT_SYMBOL_GPL(dm_array_resize); 706 707 static int populate_ablock_with_values(struct dm_array_info *info, struct array_block *ab, 708 value_fn fn, void *context, 709 unsigned int base, unsigned int new_nr) 710 { 711 int r; 712 unsigned int i; 713 struct dm_btree_value_type *vt = &info->value_type; 714 715 BUG_ON(le32_to_cpu(ab->nr_entries)); 716 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); 717 718 for (i = 0; i < new_nr; i++) { 719 r = fn(base + i, element_at(info, ab, i), context); 720 if (r) 721 return r; 722 723 if (vt->inc) 724 vt->inc(vt->context, element_at(info, ab, i), 1); 725 } 726 727 ab->nr_entries = cpu_to_le32(new_nr); 728 return 0; 729 } 730 731 int dm_array_new(struct dm_array_info *info, dm_block_t *root, 732 uint32_t size, value_fn fn, void *context) 733 { 734 int r; 735 struct dm_block *block; 736 struct array_block *ab; 737 unsigned int block_index, end_block, size_of_block, max_entries; 738 739 r = dm_array_empty(info, root); 740 if (r) 741 return r; 742 743 size_of_block = dm_bm_block_size(dm_tm_get_bm(info->btree_info.tm)); 744 max_entries = calc_max_entries(info->value_type.size, size_of_block); 745 end_block = dm_div_up(size, max_entries); 746 747 for (block_index = 0; block_index != end_block; block_index++) { 748 r = alloc_ablock(info, size_of_block, max_entries, &block, &ab); 749 if (r) 750 break; 751 752 r = populate_ablock_with_values(info, ab, fn, context, 753 block_index * max_entries, 754 min(max_entries, size)); 755 if (r) { 756 unlock_ablock(info, block); 757 break; 758 } 759 760 r = insert_ablock(info, block_index, block, root); 761 unlock_ablock(info, block); 762 if (r) 763 break; 764 765 size -= max_entries; 766 } 767 768 return r; 769 } 770 EXPORT_SYMBOL_GPL(dm_array_new); 771 772 int dm_array_del(struct dm_array_info *info, dm_block_t root) 773 { 774 return dm_btree_del(&info->btree_info, root); 775 } 776 EXPORT_SYMBOL_GPL(dm_array_del); 777 778 int dm_array_get_value(struct dm_array_info *info, dm_block_t root, 779 uint32_t index, void *value_le) 780 { 781 int r; 782 struct dm_block *block; 783 struct array_block *ab; 784 size_t size_of_block; 785 unsigned int entry, max_entries; 786 787 size_of_block = dm_bm_block_size(dm_tm_get_bm(info->btree_info.tm)); 788 max_entries = calc_max_entries(info->value_type.size, size_of_block); 789 790 r = lookup_ablock(info, root, index / max_entries, &block, &ab); 791 if (r) 792 return r; 793 794 entry = index % max_entries; 795 if (entry >= le32_to_cpu(ab->nr_entries)) 796 r = -ENODATA; 797 else 798 memcpy(value_le, element_at(info, ab, entry), 799 info->value_type.size); 800 801 unlock_ablock(info, block); 802 return r; 803 } 804 EXPORT_SYMBOL_GPL(dm_array_get_value); 805 806 static int array_set_value(struct dm_array_info *info, dm_block_t root, 807 uint32_t index, const void *value, dm_block_t *new_root) 808 { 809 int r; 810 struct dm_block *block; 811 struct array_block *ab; 812 size_t size_of_block; 813 unsigned int max_entries; 814 unsigned int entry; 815 void *old_value; 816 struct dm_btree_value_type *vt = &info->value_type; 817 818 size_of_block = dm_bm_block_size(dm_tm_get_bm(info->btree_info.tm)); 819 max_entries = calc_max_entries(info->value_type.size, size_of_block); 820 821 r = shadow_ablock(info, &root, index / max_entries, &block, &ab); 822 if (r) 823 return r; 824 *new_root = root; 825 826 entry = index % max_entries; 827 if (entry >= le32_to_cpu(ab->nr_entries)) { 828 r = -ENODATA; 829 goto out; 830 } 831 832 old_value = element_at(info, ab, entry); 833 if (vt->dec && 834 (!vt->equal || !vt->equal(vt->context, old_value, value))) { 835 vt->dec(vt->context, old_value, 1); 836 if (vt->inc) 837 vt->inc(vt->context, value, 1); 838 } 839 840 memcpy(old_value, value, info->value_type.size); 841 842 out: 843 unlock_ablock(info, block); 844 return r; 845 } 846 847 int dm_array_set_value(struct dm_array_info *info, dm_block_t root, 848 uint32_t index, const void *value, dm_block_t *new_root) 849 __dm_written_to_disk(value) 850 { 851 int r; 852 853 r = array_set_value(info, root, index, value, new_root); 854 __dm_unbless_for_disk(value); 855 return r; 856 } 857 EXPORT_SYMBOL_GPL(dm_array_set_value); 858 859 struct walk_info { 860 struct dm_array_info *info; 861 int (*fn)(void *context, uint64_t key, void *leaf); 862 void *context; 863 }; 864 865 static int walk_ablock(void *context, uint64_t *keys, void *leaf) 866 { 867 struct walk_info *wi = context; 868 869 int r; 870 unsigned int i; 871 __le64 block_le; 872 unsigned int nr_entries, max_entries; 873 struct dm_block *block; 874 struct array_block *ab; 875 876 memcpy(&block_le, leaf, sizeof(block_le)); 877 r = get_ablock(wi->info, le64_to_cpu(block_le), &block, &ab); 878 if (r) 879 return r; 880 881 max_entries = le32_to_cpu(ab->max_entries); 882 nr_entries = le32_to_cpu(ab->nr_entries); 883 for (i = 0; i < nr_entries; i++) { 884 r = wi->fn(wi->context, keys[0] * max_entries + i, 885 element_at(wi->info, ab, i)); 886 887 if (r) 888 break; 889 } 890 891 unlock_ablock(wi->info, block); 892 return r; 893 } 894 895 int dm_array_walk(struct dm_array_info *info, dm_block_t root, 896 int (*fn)(void *, uint64_t key, void *leaf), 897 void *context) 898 { 899 struct walk_info wi; 900 901 wi.info = info; 902 wi.fn = fn; 903 wi.context = context; 904 905 return dm_btree_walk(&info->btree_info, root, walk_ablock, &wi); 906 } 907 EXPORT_SYMBOL_GPL(dm_array_walk); 908 909 /*----------------------------------------------------------------*/ 910 911 static int load_ablock(struct dm_array_cursor *c) 912 { 913 int r; 914 __le64 value_le; 915 uint64_t key; 916 917 if (c->block) 918 unlock_ablock(c->info, c->block); 919 920 c->block = NULL; 921 c->ab = NULL; 922 c->index = 0; 923 924 r = dm_btree_cursor_get_value(&c->cursor, &key, &value_le); 925 if (r) { 926 DMERR("dm_btree_cursor_get_value failed"); 927 dm_btree_cursor_end(&c->cursor); 928 929 } else { 930 r = get_ablock(c->info, le64_to_cpu(value_le), &c->block, &c->ab); 931 if (r) { 932 DMERR("get_ablock failed"); 933 dm_btree_cursor_end(&c->cursor); 934 } 935 } 936 937 return r; 938 } 939 940 int dm_array_cursor_begin(struct dm_array_info *info, dm_block_t root, 941 struct dm_array_cursor *c) 942 { 943 int r; 944 945 memset(c, 0, sizeof(*c)); 946 c->info = info; 947 r = dm_btree_cursor_begin(&info->btree_info, root, true, &c->cursor); 948 if (r) { 949 DMERR("couldn't create btree cursor"); 950 return r; 951 } 952 953 return load_ablock(c); 954 } 955 EXPORT_SYMBOL_GPL(dm_array_cursor_begin); 956 957 void dm_array_cursor_end(struct dm_array_cursor *c) 958 { 959 if (c->block) { 960 unlock_ablock(c->info, c->block); 961 dm_btree_cursor_end(&c->cursor); 962 } 963 } 964 EXPORT_SYMBOL_GPL(dm_array_cursor_end); 965 966 int dm_array_cursor_next(struct dm_array_cursor *c) 967 { 968 int r; 969 970 if (!c->block) 971 return -ENODATA; 972 973 c->index++; 974 975 if (c->index >= le32_to_cpu(c->ab->nr_entries)) { 976 r = dm_btree_cursor_next(&c->cursor); 977 if (r) 978 return r; 979 980 r = load_ablock(c); 981 if (r) 982 return r; 983 } 984 985 return 0; 986 } 987 EXPORT_SYMBOL_GPL(dm_array_cursor_next); 988 989 int dm_array_cursor_skip(struct dm_array_cursor *c, uint32_t count) 990 { 991 int r; 992 993 do { 994 uint32_t remaining = le32_to_cpu(c->ab->nr_entries) - c->index; 995 996 if (count < remaining) { 997 c->index += count; 998 return 0; 999 } 1000 1001 count -= remaining; 1002 r = dm_array_cursor_next(c); 1003 1004 } while (!r); 1005 1006 return r; 1007 } 1008 EXPORT_SYMBOL_GPL(dm_array_cursor_skip); 1009 1010 void dm_array_cursor_get_value(struct dm_array_cursor *c, void **value_le) 1011 { 1012 *value_le = element_at(c->info, c->ab, c->index); 1013 } 1014 EXPORT_SYMBOL_GPL(dm_array_cursor_get_value); 1015 1016 /*----------------------------------------------------------------*/ 1017