1 // SPDX-License-Identifier: GPL-2.0 2 3 #include "bcachefs.h" 4 #include "bkey.h" 5 #include "bkey_cmp.h" 6 #include "bkey_methods.h" 7 #include "bset.h" 8 #include "util.h" 9 10 const struct bkey_format bch2_bkey_format_current = BKEY_FORMAT_CURRENT; 11 12 void bch2_bkey_packed_to_binary_text(struct printbuf *out, 13 const struct bkey_format *f, 14 const struct bkey_packed *k) 15 { 16 const u64 *p = high_word(f, k); 17 unsigned word_bits = 64 - high_bit_offset; 18 unsigned nr_key_bits = bkey_format_key_bits(f) + high_bit_offset; 19 u64 v = *p & (~0ULL >> high_bit_offset); 20 21 if (!nr_key_bits) { 22 prt_str(out, "(empty)"); 23 return; 24 } 25 26 while (1) { 27 unsigned next_key_bits = nr_key_bits; 28 29 if (nr_key_bits < 64) { 30 v >>= 64 - nr_key_bits; 31 next_key_bits = 0; 32 } else { 33 next_key_bits -= 64; 34 } 35 36 bch2_prt_u64_base2_nbits(out, v, min(word_bits, nr_key_bits)); 37 38 if (!next_key_bits) 39 break; 40 41 prt_char(out, ' '); 42 43 p = next_word(p); 44 v = *p; 45 word_bits = 64; 46 nr_key_bits = next_key_bits; 47 } 48 } 49 50 static void __bch2_bkey_pack_verify(const struct bkey_packed *packed, 51 const struct bkey *unpacked, 52 const struct bkey_format *format) 53 { 54 struct bkey tmp; 55 56 BUG_ON(bkeyp_val_u64s(format, packed) != 57 bkey_val_u64s(unpacked)); 58 59 BUG_ON(packed->u64s < bkeyp_key_u64s(format, packed)); 60 61 tmp = __bch2_bkey_unpack_key(format, packed); 62 63 if (memcmp(&tmp, unpacked, sizeof(struct bkey))) { 64 struct printbuf buf = PRINTBUF; 65 66 prt_printf(&buf, "keys differ: format u64s %u fields %u %u %u %u %u\n", 67 format->key_u64s, 68 format->bits_per_field[0], 69 format->bits_per_field[1], 70 format->bits_per_field[2], 71 format->bits_per_field[3], 72 format->bits_per_field[4]); 73 74 prt_printf(&buf, "compiled unpack: "); 75 bch2_bkey_to_text(&buf, unpacked); 76 prt_newline(&buf); 77 78 prt_printf(&buf, "c unpack: "); 79 bch2_bkey_to_text(&buf, &tmp); 80 prt_newline(&buf); 81 82 prt_printf(&buf, "compiled unpack: "); 83 bch2_bkey_packed_to_binary_text(&buf, &bch2_bkey_format_current, 84 (struct bkey_packed *) unpacked); 85 prt_newline(&buf); 86 87 prt_printf(&buf, "c unpack: "); 88 bch2_bkey_packed_to_binary_text(&buf, &bch2_bkey_format_current, 89 (struct bkey_packed *) &tmp); 90 prt_newline(&buf); 91 92 panic("%s", buf.buf); 93 } 94 } 95 96 static inline void bch2_bkey_pack_verify(const struct bkey_packed *packed, 97 const struct bkey *unpacked, 98 const struct bkey_format *format) 99 { 100 if (static_branch_unlikely(&bch2_debug_check_bkey_unpack)) 101 __bch2_bkey_pack_verify(packed, unpacked, format); 102 } 103 104 struct pack_state { 105 const struct bkey_format *format; 106 unsigned bits; /* bits remaining in current word */ 107 u64 w; /* current word */ 108 u64 *p; /* pointer to next word */ 109 }; 110 111 __always_inline 112 static struct pack_state pack_state_init(const struct bkey_format *format, 113 struct bkey_packed *k) 114 { 115 u64 *p = high_word(format, k); 116 117 return (struct pack_state) { 118 .format = format, 119 .bits = 64 - high_bit_offset, 120 .w = 0, 121 .p = p, 122 }; 123 } 124 125 __always_inline 126 static void pack_state_finish(struct pack_state *state, 127 struct bkey_packed *k) 128 { 129 EBUG_ON(state->p < k->_data); 130 EBUG_ON(state->p >= (u64 *) k->_data + state->format->key_u64s); 131 132 *state->p = state->w; 133 } 134 135 struct unpack_state { 136 const struct bkey_format *format; 137 unsigned bits; /* bits remaining in current word */ 138 u64 w; /* current word */ 139 const u64 *p; /* pointer to next word */ 140 }; 141 142 __always_inline 143 static struct unpack_state unpack_state_init(const struct bkey_format *format, 144 const struct bkey_packed *k) 145 { 146 const u64 *p = high_word(format, k); 147 148 return (struct unpack_state) { 149 .format = format, 150 .bits = 64 - high_bit_offset, 151 .w = *p << high_bit_offset, 152 .p = p, 153 }; 154 } 155 156 __always_inline 157 static u64 get_inc_field(struct unpack_state *state, unsigned field) 158 { 159 unsigned bits = state->format->bits_per_field[field]; 160 u64 v = 0, offset = le64_to_cpu(state->format->field_offset[field]); 161 162 if (bits >= state->bits) { 163 v = state->w >> (64 - bits); 164 bits -= state->bits; 165 166 state->p = next_word(state->p); 167 state->w = *state->p; 168 state->bits = 64; 169 } 170 171 /* avoid shift by 64 if bits is 0 - bits is never 64 here: */ 172 v |= (state->w >> 1) >> (63 - bits); 173 state->w <<= bits; 174 state->bits -= bits; 175 176 return v + offset; 177 } 178 179 __always_inline 180 static void __set_inc_field(struct pack_state *state, unsigned field, u64 v) 181 { 182 unsigned bits = state->format->bits_per_field[field]; 183 184 if (bits) { 185 if (bits > state->bits) { 186 bits -= state->bits; 187 /* avoid shift by 64 if bits is 64 - bits is never 0 here: */ 188 state->w |= (v >> 1) >> (bits - 1); 189 190 *state->p = state->w; 191 state->p = next_word(state->p); 192 state->w = 0; 193 state->bits = 64; 194 } 195 196 state->bits -= bits; 197 state->w |= v << state->bits; 198 } 199 } 200 201 __always_inline 202 static bool set_inc_field(struct pack_state *state, unsigned field, u64 v) 203 { 204 unsigned bits = state->format->bits_per_field[field]; 205 u64 offset = le64_to_cpu(state->format->field_offset[field]); 206 207 if (v < offset) 208 return false; 209 210 v -= offset; 211 212 if (fls64(v) > bits) 213 return false; 214 215 __set_inc_field(state, field, v); 216 return true; 217 } 218 219 /* 220 * Note: does NOT set out->format (we don't know what it should be here!) 221 * 222 * Also: doesn't work on extents - it doesn't preserve the invariant that 223 * if k is packed bkey_start_pos(k) will successfully pack 224 */ 225 static bool bch2_bkey_transform_key(const struct bkey_format *out_f, 226 struct bkey_packed *out, 227 const struct bkey_format *in_f, 228 const struct bkey_packed *in) 229 { 230 struct pack_state out_s = pack_state_init(out_f, out); 231 struct unpack_state in_s = unpack_state_init(in_f, in); 232 u64 *w = out->_data; 233 unsigned i; 234 235 *w = 0; 236 237 for (i = 0; i < BKEY_NR_FIELDS; i++) 238 if (!set_inc_field(&out_s, i, get_inc_field(&in_s, i))) 239 return false; 240 241 /* Can't happen because the val would be too big to unpack: */ 242 EBUG_ON(in->u64s - in_f->key_u64s + out_f->key_u64s > U8_MAX); 243 244 pack_state_finish(&out_s, out); 245 out->u64s = out_f->key_u64s + in->u64s - in_f->key_u64s; 246 out->needs_whiteout = in->needs_whiteout; 247 out->type = in->type; 248 249 return true; 250 } 251 252 bool bch2_bkey_transform(const struct bkey_format *out_f, 253 struct bkey_packed *out, 254 const struct bkey_format *in_f, 255 const struct bkey_packed *in) 256 { 257 if (!bch2_bkey_transform_key(out_f, out, in_f, in)) 258 return false; 259 260 memcpy_u64s((u64 *) out + out_f->key_u64s, 261 (u64 *) in + in_f->key_u64s, 262 (in->u64s - in_f->key_u64s)); 263 return true; 264 } 265 266 struct bkey __bch2_bkey_unpack_key(const struct bkey_format *format, 267 const struct bkey_packed *in) 268 { 269 struct unpack_state state = unpack_state_init(format, in); 270 struct bkey out; 271 272 EBUG_ON(format->nr_fields != BKEY_NR_FIELDS); 273 EBUG_ON(in->u64s < format->key_u64s); 274 EBUG_ON(in->format != KEY_FORMAT_LOCAL_BTREE); 275 EBUG_ON(in->u64s - format->key_u64s + BKEY_U64s > U8_MAX); 276 277 out.u64s = BKEY_U64s + in->u64s - format->key_u64s; 278 out.format = KEY_FORMAT_CURRENT; 279 out.needs_whiteout = in->needs_whiteout; 280 out.type = in->type; 281 out.pad[0] = 0; 282 283 #define x(id, field) out.field = get_inc_field(&state, id); 284 bkey_fields() 285 #undef x 286 287 return out; 288 } 289 290 #ifndef HAVE_BCACHEFS_COMPILED_UNPACK 291 struct bpos __bkey_unpack_pos(const struct bkey_format *format, 292 const struct bkey_packed *in) 293 { 294 struct unpack_state state = unpack_state_init(format, in); 295 struct bpos out; 296 297 EBUG_ON(format->nr_fields != BKEY_NR_FIELDS); 298 EBUG_ON(in->u64s < format->key_u64s); 299 EBUG_ON(in->format != KEY_FORMAT_LOCAL_BTREE); 300 301 out.inode = get_inc_field(&state, BKEY_FIELD_INODE); 302 out.offset = get_inc_field(&state, BKEY_FIELD_OFFSET); 303 out.snapshot = get_inc_field(&state, BKEY_FIELD_SNAPSHOT); 304 305 return out; 306 } 307 #endif 308 309 /** 310 * bch2_bkey_pack_key -- pack just the key, not the value 311 * @out: packed result 312 * @in: key to pack 313 * @format: format of packed result 314 * 315 * Returns: true on success, false on failure 316 */ 317 bool bch2_bkey_pack_key(struct bkey_packed *out, const struct bkey *in, 318 const struct bkey_format *format) 319 { 320 struct pack_state state = pack_state_init(format, out); 321 u64 *w = out->_data; 322 323 EBUG_ON((void *) in == (void *) out); 324 EBUG_ON(format->nr_fields != BKEY_NR_FIELDS); 325 EBUG_ON(in->format != KEY_FORMAT_CURRENT); 326 327 *w = 0; 328 329 #define x(id, field) if (!set_inc_field(&state, id, in->field)) return false; 330 bkey_fields() 331 #undef x 332 pack_state_finish(&state, out); 333 out->u64s = format->key_u64s + in->u64s - BKEY_U64s; 334 out->format = KEY_FORMAT_LOCAL_BTREE; 335 out->needs_whiteout = in->needs_whiteout; 336 out->type = in->type; 337 338 bch2_bkey_pack_verify(out, in, format); 339 return true; 340 } 341 342 /** 343 * bch2_bkey_unpack -- unpack the key and the value 344 * @b: btree node of @src key (for packed format) 345 * @dst: unpacked result 346 * @src: packed input 347 */ 348 void bch2_bkey_unpack(const struct btree *b, struct bkey_i *dst, 349 const struct bkey_packed *src) 350 { 351 __bkey_unpack_key(b, &dst->k, src); 352 353 memcpy_u64s(&dst->v, 354 bkeyp_val(&b->format, src), 355 bkeyp_val_u64s(&b->format, src)); 356 } 357 358 /** 359 * bch2_bkey_pack -- pack the key and the value 360 * @dst: packed result 361 * @src: unpacked input 362 * @format: format of packed result 363 * 364 * Returns: true on success, false on failure 365 */ 366 bool bch2_bkey_pack(struct bkey_packed *dst, const struct bkey_i *src, 367 const struct bkey_format *format) 368 { 369 struct bkey_packed tmp; 370 371 if (!bch2_bkey_pack_key(&tmp, &src->k, format)) 372 return false; 373 374 memmove_u64s((u64 *) dst + format->key_u64s, 375 &src->v, 376 bkey_val_u64s(&src->k)); 377 memcpy_u64s_small(dst, &tmp, format->key_u64s); 378 379 return true; 380 } 381 382 __always_inline 383 static bool set_inc_field_lossy(struct pack_state *state, unsigned field, u64 v) 384 { 385 unsigned bits = state->format->bits_per_field[field]; 386 u64 offset = le64_to_cpu(state->format->field_offset[field]); 387 bool ret = true; 388 389 EBUG_ON(v < offset); 390 v -= offset; 391 392 if (fls64(v) > bits) { 393 v = ~(~0ULL << bits); 394 ret = false; 395 } 396 397 __set_inc_field(state, field, v); 398 return ret; 399 } 400 401 static bool bkey_packed_successor(struct bkey_packed *out, 402 const struct btree *b, 403 struct bkey_packed k) 404 { 405 const struct bkey_format *f = &b->format; 406 unsigned nr_key_bits = b->nr_key_bits; 407 unsigned first_bit, offset; 408 u64 *p; 409 410 EBUG_ON(b->nr_key_bits != bkey_format_key_bits(f)); 411 412 if (!nr_key_bits) 413 return false; 414 415 *out = k; 416 417 first_bit = high_bit_offset + nr_key_bits - 1; 418 p = nth_word(high_word(f, out), first_bit >> 6); 419 offset = 63 - (first_bit & 63); 420 421 while (nr_key_bits) { 422 unsigned bits = min(64 - offset, nr_key_bits); 423 u64 mask = (~0ULL >> (64 - bits)) << offset; 424 425 if ((*p & mask) != mask) { 426 *p += 1ULL << offset; 427 EBUG_ON(bch2_bkey_cmp_packed(b, out, &k) <= 0); 428 return true; 429 } 430 431 *p &= ~mask; 432 p = prev_word(p); 433 nr_key_bits -= bits; 434 offset = 0; 435 } 436 437 return false; 438 } 439 440 static bool bkey_format_has_too_big_fields(const struct bkey_format *f) 441 { 442 for (unsigned i = 0; i < f->nr_fields; i++) { 443 unsigned unpacked_bits = bch2_bkey_format_current.bits_per_field[i]; 444 u64 unpacked_max = ~((~0ULL << 1) << (unpacked_bits - 1)); 445 u64 packed_max = f->bits_per_field[i] 446 ? ~((~0ULL << 1) << (f->bits_per_field[i] - 1)) 447 : 0; 448 u64 field_offset = le64_to_cpu(f->field_offset[i]); 449 450 if (packed_max + field_offset < packed_max || 451 packed_max + field_offset > unpacked_max) 452 return true; 453 } 454 455 return false; 456 } 457 458 /* 459 * Returns a packed key that compares <= in 460 * 461 * This is used in bset_search_tree(), where we need a packed pos in order to be 462 * able to compare against the keys in the auxiliary search tree - and it's 463 * legal to use a packed pos that isn't equivalent to the original pos, 464 * _provided_ it compares <= to the original pos. 465 */ 466 enum bkey_pack_pos_ret bch2_bkey_pack_pos_lossy(struct bkey_packed *out, 467 struct bpos in, 468 const struct btree *b) 469 { 470 const struct bkey_format *f = &b->format; 471 struct pack_state state = pack_state_init(f, out); 472 u64 *w = out->_data; 473 struct bpos orig = in; 474 bool exact = true; 475 unsigned i; 476 477 /* 478 * bch2_bkey_pack_key() will write to all of f->key_u64s, minus the 3 479 * byte header, but pack_pos() won't if the len/version fields are big 480 * enough - we need to make sure to zero them out: 481 */ 482 for (i = 0; i < f->key_u64s; i++) 483 w[i] = 0; 484 485 if (unlikely(in.snapshot < 486 le64_to_cpu(f->field_offset[BKEY_FIELD_SNAPSHOT]))) { 487 if (!in.offset-- && 488 !in.inode--) 489 return BKEY_PACK_POS_FAIL; 490 in.snapshot = KEY_SNAPSHOT_MAX; 491 exact = false; 492 } 493 494 if (unlikely(in.offset < 495 le64_to_cpu(f->field_offset[BKEY_FIELD_OFFSET]))) { 496 if (!in.inode--) 497 return BKEY_PACK_POS_FAIL; 498 in.offset = KEY_OFFSET_MAX; 499 in.snapshot = KEY_SNAPSHOT_MAX; 500 exact = false; 501 } 502 503 if (unlikely(in.inode < 504 le64_to_cpu(f->field_offset[BKEY_FIELD_INODE]))) 505 return BKEY_PACK_POS_FAIL; 506 507 if (unlikely(!set_inc_field_lossy(&state, BKEY_FIELD_INODE, in.inode))) { 508 in.offset = KEY_OFFSET_MAX; 509 in.snapshot = KEY_SNAPSHOT_MAX; 510 exact = false; 511 } 512 513 if (unlikely(!set_inc_field_lossy(&state, BKEY_FIELD_OFFSET, in.offset))) { 514 in.snapshot = KEY_SNAPSHOT_MAX; 515 exact = false; 516 } 517 518 if (unlikely(!set_inc_field_lossy(&state, BKEY_FIELD_SNAPSHOT, in.snapshot))) 519 exact = false; 520 521 pack_state_finish(&state, out); 522 out->u64s = f->key_u64s; 523 out->format = KEY_FORMAT_LOCAL_BTREE; 524 out->type = KEY_TYPE_deleted; 525 526 if (static_branch_unlikely(&bch2_debug_check_bkey_unpack)) { 527 if (exact) { 528 BUG_ON(bkey_cmp_left_packed(b, out, &orig)); 529 } else { 530 struct bkey_packed successor; 531 532 BUG_ON(bkey_cmp_left_packed(b, out, &orig) >= 0); 533 BUG_ON(bkey_packed_successor(&successor, b, *out) && 534 bkey_cmp_left_packed(b, &successor, &orig) < 0 && 535 !bkey_format_has_too_big_fields(f)); 536 } 537 } 538 539 return exact ? BKEY_PACK_POS_EXACT : BKEY_PACK_POS_SMALLER; 540 } 541 542 void bch2_bkey_format_init(struct bkey_format_state *s) 543 { 544 unsigned i; 545 546 for (i = 0; i < ARRAY_SIZE(s->field_min); i++) 547 s->field_min[i] = U64_MAX; 548 549 for (i = 0; i < ARRAY_SIZE(s->field_max); i++) 550 s->field_max[i] = 0; 551 552 /* Make sure we can store a size of 0: */ 553 s->field_min[BKEY_FIELD_SIZE] = 0; 554 } 555 556 void bch2_bkey_format_add_pos(struct bkey_format_state *s, struct bpos p) 557 { 558 unsigned field = 0; 559 560 __bkey_format_add(s, field++, p.inode); 561 __bkey_format_add(s, field++, p.offset); 562 __bkey_format_add(s, field++, p.snapshot); 563 } 564 565 /* 566 * We don't want it to be possible for the packed format to represent fields 567 * bigger than a u64... that will cause confusion and issues (like with 568 * bkey_packed_successor()) 569 */ 570 static void set_format_field(struct bkey_format *f, enum bch_bkey_fields i, 571 unsigned bits, u64 offset) 572 { 573 unsigned unpacked_bits = bch2_bkey_format_current.bits_per_field[i]; 574 u64 unpacked_max = ~((~0ULL << 1) << (unpacked_bits - 1)); 575 576 bits = min(bits, unpacked_bits); 577 578 offset = bits == unpacked_bits ? 0 : min(offset, unpacked_max - ((1ULL << bits) - 1)); 579 580 f->bits_per_field[i] = bits; 581 f->field_offset[i] = cpu_to_le64(offset); 582 } 583 584 struct bkey_format bch2_bkey_format_done(struct bkey_format_state *s) 585 { 586 unsigned i, bits = KEY_PACKED_BITS_START; 587 struct bkey_format ret = { 588 .nr_fields = BKEY_NR_FIELDS, 589 }; 590 591 for (i = 0; i < ARRAY_SIZE(s->field_min); i++) { 592 s->field_min[i] = min(s->field_min[i], s->field_max[i]); 593 594 set_format_field(&ret, i, 595 fls64(s->field_max[i] - s->field_min[i]), 596 s->field_min[i]); 597 598 bits += ret.bits_per_field[i]; 599 } 600 601 /* allow for extent merging: */ 602 if (ret.bits_per_field[BKEY_FIELD_SIZE]) { 603 unsigned b = min(4U, 32U - ret.bits_per_field[BKEY_FIELD_SIZE]); 604 605 ret.bits_per_field[BKEY_FIELD_SIZE] += b; 606 bits += b; 607 } 608 609 ret.key_u64s = DIV_ROUND_UP(bits, 64); 610 611 /* if we have enough spare bits, round fields up to nearest byte */ 612 bits = ret.key_u64s * 64 - bits; 613 614 for (i = 0; i < ARRAY_SIZE(ret.bits_per_field); i++) { 615 unsigned r = round_up(ret.bits_per_field[i], 8) - 616 ret.bits_per_field[i]; 617 618 if (r <= bits) { 619 set_format_field(&ret, i, 620 ret.bits_per_field[i] + r, 621 le64_to_cpu(ret.field_offset[i])); 622 bits -= r; 623 } 624 } 625 626 if (static_branch_unlikely(&bch2_debug_check_bkey_unpack)) { 627 struct printbuf buf = PRINTBUF; 628 629 BUG_ON(bch2_bkey_format_invalid(NULL, &ret, 0, &buf)); 630 printbuf_exit(&buf); 631 } 632 633 return ret; 634 } 635 636 int bch2_bkey_format_invalid(struct bch_fs *c, 637 struct bkey_format *f, 638 enum bch_validate_flags flags, 639 struct printbuf *err) 640 { 641 unsigned bits = KEY_PACKED_BITS_START; 642 643 if (f->nr_fields != BKEY_NR_FIELDS) { 644 prt_printf(err, "incorrect number of fields: got %u, should be %u", 645 f->nr_fields, BKEY_NR_FIELDS); 646 return -BCH_ERR_invalid; 647 } 648 649 /* 650 * Verify that the packed format can't represent fields larger than the 651 * unpacked format: 652 */ 653 for (unsigned i = 0; i < f->nr_fields; i++) { 654 if (bch2_bkey_format_field_overflows(f, i)) { 655 unsigned unpacked_bits = bch2_bkey_format_current.bits_per_field[i]; 656 u64 unpacked_max = ~((~0ULL << 1) << (unpacked_bits - 1)); 657 unsigned packed_bits = min(64, f->bits_per_field[i]); 658 u64 packed_max = packed_bits 659 ? ~((~0ULL << 1) << (packed_bits - 1)) 660 : 0; 661 662 prt_printf(err, "field %u too large: %llu + %llu > %llu", 663 i, packed_max, le64_to_cpu(f->field_offset[i]), unpacked_max); 664 return -BCH_ERR_invalid; 665 } 666 667 bits += f->bits_per_field[i]; 668 } 669 670 if (f->key_u64s != DIV_ROUND_UP(bits, 64)) { 671 prt_printf(err, "incorrect key_u64s: got %u, should be %u", 672 f->key_u64s, DIV_ROUND_UP(bits, 64)); 673 return -BCH_ERR_invalid; 674 } 675 676 return 0; 677 } 678 679 void bch2_bkey_format_to_text(struct printbuf *out, const struct bkey_format *f) 680 { 681 prt_printf(out, "u64s %u fields ", f->key_u64s); 682 683 for (unsigned i = 0; i < ARRAY_SIZE(f->bits_per_field); i++) { 684 if (i) 685 prt_str(out, ", "); 686 prt_printf(out, "%u:%llu", 687 f->bits_per_field[i], 688 le64_to_cpu(f->field_offset[i])); 689 } 690 } 691 692 /* 693 * Most significant differing bit 694 * Bits are indexed from 0 - return is [0, nr_key_bits) 695 */ 696 __pure 697 unsigned bch2_bkey_greatest_differing_bit(const struct btree *b, 698 const struct bkey_packed *l_k, 699 const struct bkey_packed *r_k) 700 { 701 const u64 *l = high_word(&b->format, l_k); 702 const u64 *r = high_word(&b->format, r_k); 703 unsigned nr_key_bits = b->nr_key_bits; 704 unsigned word_bits = 64 - high_bit_offset; 705 u64 l_v, r_v; 706 707 EBUG_ON(b->nr_key_bits != bkey_format_key_bits(&b->format)); 708 709 /* for big endian, skip past header */ 710 l_v = *l & (~0ULL >> high_bit_offset); 711 r_v = *r & (~0ULL >> high_bit_offset); 712 713 while (nr_key_bits) { 714 if (nr_key_bits < word_bits) { 715 l_v >>= word_bits - nr_key_bits; 716 r_v >>= word_bits - nr_key_bits; 717 nr_key_bits = 0; 718 } else { 719 nr_key_bits -= word_bits; 720 } 721 722 if (l_v != r_v) 723 return fls64(l_v ^ r_v) - 1 + nr_key_bits; 724 725 l = next_word(l); 726 r = next_word(r); 727 728 l_v = *l; 729 r_v = *r; 730 word_bits = 64; 731 } 732 733 return 0; 734 } 735 736 /* 737 * First set bit 738 * Bits are indexed from 0 - return is [0, nr_key_bits) 739 */ 740 __pure 741 unsigned bch2_bkey_ffs(const struct btree *b, const struct bkey_packed *k) 742 { 743 const u64 *p = high_word(&b->format, k); 744 unsigned nr_key_bits = b->nr_key_bits; 745 unsigned ret = 0, offset; 746 747 EBUG_ON(b->nr_key_bits != bkey_format_key_bits(&b->format)); 748 749 offset = nr_key_bits; 750 while (offset > 64) { 751 p = next_word(p); 752 offset -= 64; 753 } 754 755 offset = 64 - offset; 756 757 while (nr_key_bits) { 758 unsigned bits = nr_key_bits + offset < 64 759 ? nr_key_bits 760 : 64 - offset; 761 762 u64 mask = (~0ULL >> (64 - bits)) << offset; 763 764 if (*p & mask) 765 return ret + __ffs64(*p & mask) - offset; 766 767 p = prev_word(p); 768 nr_key_bits -= bits; 769 ret += bits; 770 offset = 0; 771 } 772 773 return 0; 774 } 775 776 #ifdef HAVE_BCACHEFS_COMPILED_UNPACK 777 778 #define I(_x) (*(out)++ = (_x)) 779 #define I1(i0) I(i0) 780 #define I2(i0, i1) (I1(i0), I(i1)) 781 #define I3(i0, i1, i2) (I2(i0, i1), I(i2)) 782 #define I4(i0, i1, i2, i3) (I3(i0, i1, i2), I(i3)) 783 #define I5(i0, i1, i2, i3, i4) (I4(i0, i1, i2, i3), I(i4)) 784 785 static u8 *compile_bkey_field(const struct bkey_format *format, u8 *out, 786 enum bch_bkey_fields field, 787 unsigned dst_offset, unsigned dst_size, 788 bool *eax_zeroed) 789 { 790 unsigned bits = format->bits_per_field[field]; 791 u64 offset = le64_to_cpu(format->field_offset[field]); 792 unsigned i, byte, bit_offset, align, shl, shr; 793 794 if (!bits && !offset) { 795 if (!*eax_zeroed) { 796 /* xor eax, eax */ 797 I2(0x31, 0xc0); 798 } 799 800 *eax_zeroed = true; 801 goto set_field; 802 } 803 804 if (!bits) { 805 /* just return offset: */ 806 807 switch (dst_size) { 808 case 8: 809 if (offset > S32_MAX) { 810 /* mov [rdi + dst_offset], offset */ 811 I3(0xc7, 0x47, dst_offset); 812 memcpy(out, &offset, 4); 813 out += 4; 814 815 I3(0xc7, 0x47, dst_offset + 4); 816 memcpy(out, (void *) &offset + 4, 4); 817 out += 4; 818 } else { 819 /* mov [rdi + dst_offset], offset */ 820 /* sign extended */ 821 I4(0x48, 0xc7, 0x47, dst_offset); 822 memcpy(out, &offset, 4); 823 out += 4; 824 } 825 break; 826 case 4: 827 /* mov [rdi + dst_offset], offset */ 828 I3(0xc7, 0x47, dst_offset); 829 memcpy(out, &offset, 4); 830 out += 4; 831 break; 832 default: 833 BUG(); 834 } 835 836 return out; 837 } 838 839 bit_offset = format->key_u64s * 64; 840 for (i = 0; i <= field; i++) 841 bit_offset -= format->bits_per_field[i]; 842 843 byte = bit_offset / 8; 844 bit_offset -= byte * 8; 845 846 *eax_zeroed = false; 847 848 if (bit_offset == 0 && bits == 8) { 849 /* movzx eax, BYTE PTR [rsi + imm8] */ 850 I4(0x0f, 0xb6, 0x46, byte); 851 } else if (bit_offset == 0 && bits == 16) { 852 /* movzx eax, WORD PTR [rsi + imm8] */ 853 I4(0x0f, 0xb7, 0x46, byte); 854 } else if (bit_offset + bits <= 32) { 855 align = min(4 - DIV_ROUND_UP(bit_offset + bits, 8), byte & 3); 856 byte -= align; 857 bit_offset += align * 8; 858 859 BUG_ON(bit_offset + bits > 32); 860 861 /* mov eax, [rsi + imm8] */ 862 I3(0x8b, 0x46, byte); 863 864 if (bit_offset) { 865 /* shr eax, imm8 */ 866 I3(0xc1, 0xe8, bit_offset); 867 } 868 869 if (bit_offset + bits < 32) { 870 unsigned mask = ~0U >> (32 - bits); 871 872 /* and eax, imm32 */ 873 I1(0x25); 874 memcpy(out, &mask, 4); 875 out += 4; 876 } 877 } else if (bit_offset + bits <= 64) { 878 align = min(8 - DIV_ROUND_UP(bit_offset + bits, 8), byte & 7); 879 byte -= align; 880 bit_offset += align * 8; 881 882 BUG_ON(bit_offset + bits > 64); 883 884 /* mov rax, [rsi + imm8] */ 885 I4(0x48, 0x8b, 0x46, byte); 886 887 shl = 64 - bit_offset - bits; 888 shr = bit_offset + shl; 889 890 if (shl) { 891 /* shl rax, imm8 */ 892 I4(0x48, 0xc1, 0xe0, shl); 893 } 894 895 if (shr) { 896 /* shr rax, imm8 */ 897 I4(0x48, 0xc1, 0xe8, shr); 898 } 899 } else { 900 align = min(4 - DIV_ROUND_UP(bit_offset + bits, 8), byte & 3); 901 byte -= align; 902 bit_offset += align * 8; 903 904 BUG_ON(bit_offset + bits > 96); 905 906 /* mov rax, [rsi + byte] */ 907 I4(0x48, 0x8b, 0x46, byte); 908 909 /* mov edx, [rsi + byte + 8] */ 910 I3(0x8b, 0x56, byte + 8); 911 912 /* bits from next word: */ 913 shr = bit_offset + bits - 64; 914 BUG_ON(shr > bit_offset); 915 916 /* shr rax, bit_offset */ 917 I4(0x48, 0xc1, 0xe8, shr); 918 919 /* shl rdx, imm8 */ 920 I4(0x48, 0xc1, 0xe2, 64 - shr); 921 922 /* or rax, rdx */ 923 I3(0x48, 0x09, 0xd0); 924 925 shr = bit_offset - shr; 926 927 if (shr) { 928 /* shr rax, imm8 */ 929 I4(0x48, 0xc1, 0xe8, shr); 930 } 931 } 932 933 /* rax += offset: */ 934 if (offset > S32_MAX) { 935 /* mov rdx, imm64 */ 936 I2(0x48, 0xba); 937 memcpy(out, &offset, 8); 938 out += 8; 939 /* add %rdx, %rax */ 940 I3(0x48, 0x01, 0xd0); 941 } else if (offset + (~0ULL >> (64 - bits)) > U32_MAX) { 942 /* add rax, imm32 */ 943 I2(0x48, 0x05); 944 memcpy(out, &offset, 4); 945 out += 4; 946 } else if (offset) { 947 /* add eax, imm32 */ 948 I1(0x05); 949 memcpy(out, &offset, 4); 950 out += 4; 951 } 952 set_field: 953 switch (dst_size) { 954 case 8: 955 /* mov [rdi + dst_offset], rax */ 956 I4(0x48, 0x89, 0x47, dst_offset); 957 break; 958 case 4: 959 /* mov [rdi + dst_offset], eax */ 960 I3(0x89, 0x47, dst_offset); 961 break; 962 default: 963 BUG(); 964 } 965 966 return out; 967 } 968 969 int bch2_compile_bkey_format(const struct bkey_format *format, void *_out) 970 { 971 bool eax_zeroed = false; 972 u8 *out = _out; 973 974 /* 975 * rdi: dst - unpacked key 976 * rsi: src - packed key 977 */ 978 979 /* k->u64s, k->format, k->type */ 980 981 /* mov eax, [rsi] */ 982 I2(0x8b, 0x06); 983 984 /* add eax, BKEY_U64s - format->key_u64s */ 985 I5(0x05, BKEY_U64s - format->key_u64s, KEY_FORMAT_CURRENT, 0, 0); 986 987 /* and eax, imm32: mask out k->pad: */ 988 I5(0x25, 0xff, 0xff, 0xff, 0); 989 990 /* mov [rdi], eax */ 991 I2(0x89, 0x07); 992 993 #define x(id, field) \ 994 out = compile_bkey_field(format, out, id, \ 995 offsetof(struct bkey, field), \ 996 sizeof(((struct bkey *) NULL)->field), \ 997 &eax_zeroed); 998 bkey_fields() 999 #undef x 1000 1001 /* retq */ 1002 I1(0xc3); 1003 1004 return (void *) out - _out; 1005 } 1006 1007 #else 1008 #endif 1009 1010 __pure 1011 int __bch2_bkey_cmp_packed_format_checked(const struct bkey_packed *l, 1012 const struct bkey_packed *r, 1013 const struct btree *b) 1014 { 1015 return __bch2_bkey_cmp_packed_format_checked_inlined(l, r, b); 1016 } 1017 1018 __pure __flatten 1019 int __bch2_bkey_cmp_left_packed_format_checked(const struct btree *b, 1020 const struct bkey_packed *l, 1021 const struct bpos *r) 1022 { 1023 return bpos_cmp(bkey_unpack_pos_format_checked(b, l), *r); 1024 } 1025 1026 __pure __flatten 1027 int bch2_bkey_cmp_packed(const struct btree *b, 1028 const struct bkey_packed *l, 1029 const struct bkey_packed *r) 1030 { 1031 return bch2_bkey_cmp_packed_inlined(b, l, r); 1032 } 1033 1034 __pure __flatten 1035 int __bch2_bkey_cmp_left_packed(const struct btree *b, 1036 const struct bkey_packed *l, 1037 const struct bpos *r) 1038 { 1039 const struct bkey *l_unpacked; 1040 1041 return unlikely(l_unpacked = packed_to_bkey_c(l)) 1042 ? bpos_cmp(l_unpacked->p, *r) 1043 : __bch2_bkey_cmp_left_packed_format_checked(b, l, r); 1044 } 1045 1046 void bch2_bpos_swab(struct bpos *p) 1047 { 1048 u8 *l = (u8 *) p; 1049 u8 *h = ((u8 *) &p[1]) - 1; 1050 1051 while (l < h) { 1052 swap(*l, *h); 1053 l++; 1054 --h; 1055 } 1056 } 1057 1058 void bch2_bkey_swab_key(const struct bkey_format *_f, struct bkey_packed *k) 1059 { 1060 const struct bkey_format *f = bkey_packed(k) ? _f : &bch2_bkey_format_current; 1061 u8 *l = k->key_start; 1062 u8 *h = (u8 *) ((u64 *) k->_data + f->key_u64s) - 1; 1063 1064 while (l < h) { 1065 swap(*l, *h); 1066 l++; 1067 --h; 1068 } 1069 } 1070 1071 #ifdef CONFIG_BCACHEFS_DEBUG 1072 void bch2_bkey_pack_test(void) 1073 { 1074 struct bkey t = KEY(4134ULL, 1250629070527416633ULL, 0); 1075 struct bkey_packed p; 1076 1077 struct bkey_format test_format = { 1078 .key_u64s = 3, 1079 .nr_fields = BKEY_NR_FIELDS, 1080 .bits_per_field = { 1081 13, 1082 64, 1083 32, 1084 }, 1085 }; 1086 1087 struct unpack_state in_s = 1088 unpack_state_init(&bch2_bkey_format_current, (void *) &t); 1089 struct pack_state out_s = pack_state_init(&test_format, &p); 1090 unsigned i; 1091 1092 for (i = 0; i < out_s.format->nr_fields; i++) { 1093 u64 a, v = get_inc_field(&in_s, i); 1094 1095 switch (i) { 1096 #define x(id, field) case id: a = t.field; break; 1097 bkey_fields() 1098 #undef x 1099 default: 1100 BUG(); 1101 } 1102 1103 if (a != v) 1104 panic("got %llu actual %llu i %u\n", v, a, i); 1105 1106 if (!set_inc_field(&out_s, i, v)) 1107 panic("failed at %u\n", i); 1108 } 1109 1110 BUG_ON(!bch2_bkey_pack_key(&p, &t, &test_format)); 1111 } 1112 #endif 1113