xref: /linux/fs/bcachefs/bkey.h (revision 031fba65fc202abf1f193e321be7a2c274fd88ba)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _BCACHEFS_BKEY_H
3 #define _BCACHEFS_BKEY_H
4 
5 #include <linux/bug.h>
6 #include "bcachefs_format.h"
7 
8 #include "btree_types.h"
9 #include "util.h"
10 #include "vstructs.h"
11 
12 enum bkey_invalid_flags {
13 	BKEY_INVALID_WRITE		= (1U << 0),
14 	BKEY_INVALID_COMMIT		= (1U << 1),
15 	BKEY_INVALID_JOURNAL		= (1U << 2),
16 };
17 
18 #if 0
19 
20 /*
21  * compiled unpack functions are disabled, pending a new interface for
22  * dynamically allocating executable memory:
23  */
24 
25 #ifdef CONFIG_X86_64
26 #define HAVE_BCACHEFS_COMPILED_UNPACK	1
27 #endif
28 #endif
29 
30 void bch2_bkey_packed_to_binary_text(struct printbuf *,
31 				     const struct bkey_format *,
32 				     const struct bkey_packed *);
33 
34 /* bkey with split value, const */
35 struct bkey_s_c {
36 	const struct bkey	*k;
37 	const struct bch_val	*v;
38 };
39 
40 /* bkey with split value */
41 struct bkey_s {
42 	union {
43 	struct {
44 		struct bkey	*k;
45 		struct bch_val	*v;
46 	};
47 	struct bkey_s_c		s_c;
48 	};
49 };
50 
51 #define bkey_p_next(_k)		vstruct_next(_k)
52 
53 static inline struct bkey_i *bkey_next(struct bkey_i *k)
54 {
55 	return (struct bkey_i *) ((u64 *) k->_data + k->k.u64s);
56 }
57 
58 #define bkey_val_u64s(_k)	((_k)->u64s - BKEY_U64s)
59 
60 static inline size_t bkey_val_bytes(const struct bkey *k)
61 {
62 	return bkey_val_u64s(k) * sizeof(u64);
63 }
64 
65 static inline void set_bkey_val_u64s(struct bkey *k, unsigned val_u64s)
66 {
67 	unsigned u64s = BKEY_U64s + val_u64s;
68 
69 	BUG_ON(u64s > U8_MAX);
70 	k->u64s = u64s;
71 }
72 
73 static inline void set_bkey_val_bytes(struct bkey *k, unsigned bytes)
74 {
75 	set_bkey_val_u64s(k, DIV_ROUND_UP(bytes, sizeof(u64)));
76 }
77 
78 #define bkey_val_end(_k)	((void *) (((u64 *) (_k).v) + bkey_val_u64s((_k).k)))
79 
80 #define bkey_deleted(_k)	((_k)->type == KEY_TYPE_deleted)
81 
82 #define bkey_whiteout(_k)				\
83 	((_k)->type == KEY_TYPE_deleted || (_k)->type == KEY_TYPE_whiteout)
84 
85 enum bkey_lr_packed {
86 	BKEY_PACKED_BOTH,
87 	BKEY_PACKED_RIGHT,
88 	BKEY_PACKED_LEFT,
89 	BKEY_PACKED_NONE,
90 };
91 
92 #define bkey_lr_packed(_l, _r)						\
93 	((_l)->format + ((_r)->format << 1))
94 
95 #define bkey_copy(_dst, _src)					\
96 do {								\
97 	BUILD_BUG_ON(!type_is(_dst, struct bkey_i *) &&		\
98 		     !type_is(_dst, struct bkey_packed *));	\
99 	BUILD_BUG_ON(!type_is(_src, struct bkey_i *) &&		\
100 		     !type_is(_src, struct bkey_packed *));	\
101 	EBUG_ON((u64 *) (_dst) > (u64 *) (_src) &&		\
102 		(u64 *) (_dst) < (u64 *) (_src) +		\
103 		((struct bkey *) (_src))->u64s);		\
104 								\
105 	memcpy_u64s_small((_dst), (_src),			\
106 			  ((struct bkey *) (_src))->u64s);	\
107 } while (0)
108 
109 struct btree;
110 
111 __pure
112 unsigned bch2_bkey_greatest_differing_bit(const struct btree *,
113 					  const struct bkey_packed *,
114 					  const struct bkey_packed *);
115 __pure
116 unsigned bch2_bkey_ffs(const struct btree *, const struct bkey_packed *);
117 
118 __pure
119 int __bch2_bkey_cmp_packed_format_checked(const struct bkey_packed *,
120 				     const struct bkey_packed *,
121 				     const struct btree *);
122 
123 __pure
124 int __bch2_bkey_cmp_left_packed_format_checked(const struct btree *,
125 					  const struct bkey_packed *,
126 					  const struct bpos *);
127 
128 __pure
129 int bch2_bkey_cmp_packed(const struct btree *,
130 			 const struct bkey_packed *,
131 			 const struct bkey_packed *);
132 
133 __pure
134 int __bch2_bkey_cmp_left_packed(const struct btree *,
135 				const struct bkey_packed *,
136 				const struct bpos *);
137 
138 static inline __pure
139 int bkey_cmp_left_packed(const struct btree *b,
140 			 const struct bkey_packed *l, const struct bpos *r)
141 {
142 	return __bch2_bkey_cmp_left_packed(b, l, r);
143 }
144 
145 /*
146  * The compiler generates better code when we pass bpos by ref, but it's often
147  * enough terribly convenient to pass it by val... as much as I hate c++, const
148  * ref would be nice here:
149  */
150 __pure __flatten
151 static inline int bkey_cmp_left_packed_byval(const struct btree *b,
152 					     const struct bkey_packed *l,
153 					     struct bpos r)
154 {
155 	return bkey_cmp_left_packed(b, l, &r);
156 }
157 
158 static __always_inline bool bpos_eq(struct bpos l, struct bpos r)
159 {
160 	return  !((l.inode	^ r.inode) |
161 		  (l.offset	^ r.offset) |
162 		  (l.snapshot	^ r.snapshot));
163 }
164 
165 static __always_inline bool bpos_lt(struct bpos l, struct bpos r)
166 {
167 	return  l.inode	!= r.inode ? l.inode < r.inode :
168 		l.offset != r.offset ? l.offset < r.offset :
169 		l.snapshot != r.snapshot ? l.snapshot < r.snapshot : false;
170 }
171 
172 static __always_inline bool bpos_le(struct bpos l, struct bpos r)
173 {
174 	return  l.inode	!= r.inode ? l.inode < r.inode :
175 		l.offset != r.offset ? l.offset < r.offset :
176 		l.snapshot != r.snapshot ? l.snapshot < r.snapshot : true;
177 }
178 
179 static __always_inline bool bpos_gt(struct bpos l, struct bpos r)
180 {
181 	return bpos_lt(r, l);
182 }
183 
184 static __always_inline bool bpos_ge(struct bpos l, struct bpos r)
185 {
186 	return bpos_le(r, l);
187 }
188 
189 static __always_inline int bpos_cmp(struct bpos l, struct bpos r)
190 {
191 	return  cmp_int(l.inode,    r.inode) ?:
192 		cmp_int(l.offset,   r.offset) ?:
193 		cmp_int(l.snapshot, r.snapshot);
194 }
195 
196 static inline struct bpos bpos_min(struct bpos l, struct bpos r)
197 {
198 	return bpos_lt(l, r) ? l : r;
199 }
200 
201 static inline struct bpos bpos_max(struct bpos l, struct bpos r)
202 {
203 	return bpos_gt(l, r) ? l : r;
204 }
205 
206 static __always_inline bool bkey_eq(struct bpos l, struct bpos r)
207 {
208 	return  !((l.inode	^ r.inode) |
209 		  (l.offset	^ r.offset));
210 }
211 
212 static __always_inline bool bkey_lt(struct bpos l, struct bpos r)
213 {
214 	return  l.inode	!= r.inode
215 		? l.inode < r.inode
216 		: l.offset < r.offset;
217 }
218 
219 static __always_inline bool bkey_le(struct bpos l, struct bpos r)
220 {
221 	return  l.inode	!= r.inode
222 		? l.inode < r.inode
223 		: l.offset <= r.offset;
224 }
225 
226 static __always_inline bool bkey_gt(struct bpos l, struct bpos r)
227 {
228 	return bkey_lt(r, l);
229 }
230 
231 static __always_inline bool bkey_ge(struct bpos l, struct bpos r)
232 {
233 	return bkey_le(r, l);
234 }
235 
236 static __always_inline int bkey_cmp(struct bpos l, struct bpos r)
237 {
238 	return  cmp_int(l.inode,    r.inode) ?:
239 		cmp_int(l.offset,   r.offset);
240 }
241 
242 static inline struct bpos bkey_min(struct bpos l, struct bpos r)
243 {
244 	return bkey_lt(l, r) ? l : r;
245 }
246 
247 static inline struct bpos bkey_max(struct bpos l, struct bpos r)
248 {
249 	return bkey_gt(l, r) ? l : r;
250 }
251 
252 void bch2_bpos_swab(struct bpos *);
253 void bch2_bkey_swab_key(const struct bkey_format *, struct bkey_packed *);
254 
255 static __always_inline int bversion_cmp(struct bversion l, struct bversion r)
256 {
257 	return  cmp_int(l.hi, r.hi) ?:
258 		cmp_int(l.lo, r.lo);
259 }
260 
261 #define ZERO_VERSION	((struct bversion) { .hi = 0, .lo = 0 })
262 #define MAX_VERSION	((struct bversion) { .hi = ~0, .lo = ~0ULL })
263 
264 static __always_inline int bversion_zero(struct bversion v)
265 {
266 	return !bversion_cmp(v, ZERO_VERSION);
267 }
268 
269 #ifdef CONFIG_BCACHEFS_DEBUG
270 /* statement expressions confusing unlikely()? */
271 #define bkey_packed(_k)							\
272 	({ EBUG_ON((_k)->format > KEY_FORMAT_CURRENT);			\
273 	 (_k)->format != KEY_FORMAT_CURRENT; })
274 #else
275 #define bkey_packed(_k)		((_k)->format != KEY_FORMAT_CURRENT)
276 #endif
277 
278 /*
279  * It's safe to treat an unpacked bkey as a packed one, but not the reverse
280  */
281 static inline struct bkey_packed *bkey_to_packed(struct bkey_i *k)
282 {
283 	return (struct bkey_packed *) k;
284 }
285 
286 static inline const struct bkey_packed *bkey_to_packed_c(const struct bkey_i *k)
287 {
288 	return (const struct bkey_packed *) k;
289 }
290 
291 static inline struct bkey_i *packed_to_bkey(struct bkey_packed *k)
292 {
293 	return bkey_packed(k) ? NULL : (struct bkey_i *) k;
294 }
295 
296 static inline const struct bkey *packed_to_bkey_c(const struct bkey_packed *k)
297 {
298 	return bkey_packed(k) ? NULL : (const struct bkey *) k;
299 }
300 
301 static inline unsigned bkey_format_key_bits(const struct bkey_format *format)
302 {
303 	return format->bits_per_field[BKEY_FIELD_INODE] +
304 		format->bits_per_field[BKEY_FIELD_OFFSET] +
305 		format->bits_per_field[BKEY_FIELD_SNAPSHOT];
306 }
307 
308 static inline struct bpos bpos_successor(struct bpos p)
309 {
310 	if (!++p.snapshot &&
311 	    !++p.offset &&
312 	    !++p.inode)
313 		BUG();
314 
315 	return p;
316 }
317 
318 static inline struct bpos bpos_predecessor(struct bpos p)
319 {
320 	if (!p.snapshot-- &&
321 	    !p.offset-- &&
322 	    !p.inode--)
323 		BUG();
324 
325 	return p;
326 }
327 
328 static inline struct bpos bpos_nosnap_successor(struct bpos p)
329 {
330 	p.snapshot = 0;
331 
332 	if (!++p.offset &&
333 	    !++p.inode)
334 		BUG();
335 
336 	return p;
337 }
338 
339 static inline struct bpos bpos_nosnap_predecessor(struct bpos p)
340 {
341 	p.snapshot = 0;
342 
343 	if (!p.offset-- &&
344 	    !p.inode--)
345 		BUG();
346 
347 	return p;
348 }
349 
350 static inline u64 bkey_start_offset(const struct bkey *k)
351 {
352 	return k->p.offset - k->size;
353 }
354 
355 static inline struct bpos bkey_start_pos(const struct bkey *k)
356 {
357 	return (struct bpos) {
358 		.inode		= k->p.inode,
359 		.offset		= bkey_start_offset(k),
360 		.snapshot	= k->p.snapshot,
361 	};
362 }
363 
364 /* Packed helpers */
365 
366 static inline unsigned bkeyp_key_u64s(const struct bkey_format *format,
367 				      const struct bkey_packed *k)
368 {
369 	unsigned ret = bkey_packed(k) ? format->key_u64s : BKEY_U64s;
370 
371 	EBUG_ON(k->u64s < ret);
372 	return ret;
373 }
374 
375 static inline unsigned bkeyp_key_bytes(const struct bkey_format *format,
376 				       const struct bkey_packed *k)
377 {
378 	return bkeyp_key_u64s(format, k) * sizeof(u64);
379 }
380 
381 static inline unsigned bkeyp_val_u64s(const struct bkey_format *format,
382 				      const struct bkey_packed *k)
383 {
384 	return k->u64s - bkeyp_key_u64s(format, k);
385 }
386 
387 static inline size_t bkeyp_val_bytes(const struct bkey_format *format,
388 				     const struct bkey_packed *k)
389 {
390 	return bkeyp_val_u64s(format, k) * sizeof(u64);
391 }
392 
393 static inline void set_bkeyp_val_u64s(const struct bkey_format *format,
394 				      struct bkey_packed *k, unsigned val_u64s)
395 {
396 	k->u64s = bkeyp_key_u64s(format, k) + val_u64s;
397 }
398 
399 #define bkeyp_val(_format, _k)						\
400 	 ((struct bch_val *) ((u64 *) (_k)->_data + bkeyp_key_u64s(_format, _k)))
401 
402 extern const struct bkey_format bch2_bkey_format_current;
403 
404 bool bch2_bkey_transform(const struct bkey_format *,
405 			 struct bkey_packed *,
406 			 const struct bkey_format *,
407 			 const struct bkey_packed *);
408 
409 struct bkey __bch2_bkey_unpack_key(const struct bkey_format *,
410 				   const struct bkey_packed *);
411 
412 #ifndef HAVE_BCACHEFS_COMPILED_UNPACK
413 struct bpos __bkey_unpack_pos(const struct bkey_format *,
414 			      const struct bkey_packed *);
415 #endif
416 
417 bool bch2_bkey_pack_key(struct bkey_packed *, const struct bkey *,
418 		   const struct bkey_format *);
419 
420 enum bkey_pack_pos_ret {
421 	BKEY_PACK_POS_EXACT,
422 	BKEY_PACK_POS_SMALLER,
423 	BKEY_PACK_POS_FAIL,
424 };
425 
426 enum bkey_pack_pos_ret bch2_bkey_pack_pos_lossy(struct bkey_packed *, struct bpos,
427 					   const struct btree *);
428 
429 static inline bool bkey_pack_pos(struct bkey_packed *out, struct bpos in,
430 				 const struct btree *b)
431 {
432 	return bch2_bkey_pack_pos_lossy(out, in, b) == BKEY_PACK_POS_EXACT;
433 }
434 
435 void bch2_bkey_unpack(const struct btree *, struct bkey_i *,
436 		 const struct bkey_packed *);
437 bool bch2_bkey_pack(struct bkey_packed *, const struct bkey_i *,
438 	       const struct bkey_format *);
439 
440 typedef void (*compiled_unpack_fn)(struct bkey *, const struct bkey_packed *);
441 
442 static inline void
443 __bkey_unpack_key_format_checked(const struct btree *b,
444 			       struct bkey *dst,
445 			       const struct bkey_packed *src)
446 {
447 	if (IS_ENABLED(HAVE_BCACHEFS_COMPILED_UNPACK)) {
448 		compiled_unpack_fn unpack_fn = b->aux_data;
449 		unpack_fn(dst, src);
450 
451 		if (IS_ENABLED(CONFIG_BCACHEFS_DEBUG) &&
452 		    bch2_expensive_debug_checks) {
453 			struct bkey dst2 = __bch2_bkey_unpack_key(&b->format, src);
454 
455 			BUG_ON(memcmp(dst, &dst2, sizeof(*dst)));
456 		}
457 	} else {
458 		*dst = __bch2_bkey_unpack_key(&b->format, src);
459 	}
460 }
461 
462 static inline struct bkey
463 bkey_unpack_key_format_checked(const struct btree *b,
464 			       const struct bkey_packed *src)
465 {
466 	struct bkey dst;
467 
468 	__bkey_unpack_key_format_checked(b, &dst, src);
469 	return dst;
470 }
471 
472 static inline void __bkey_unpack_key(const struct btree *b,
473 				     struct bkey *dst,
474 				     const struct bkey_packed *src)
475 {
476 	if (likely(bkey_packed(src)))
477 		__bkey_unpack_key_format_checked(b, dst, src);
478 	else
479 		*dst = *packed_to_bkey_c(src);
480 }
481 
482 /**
483  * bkey_unpack_key -- unpack just the key, not the value
484  */
485 static inline struct bkey bkey_unpack_key(const struct btree *b,
486 					  const struct bkey_packed *src)
487 {
488 	return likely(bkey_packed(src))
489 		? bkey_unpack_key_format_checked(b, src)
490 		: *packed_to_bkey_c(src);
491 }
492 
493 static inline struct bpos
494 bkey_unpack_pos_format_checked(const struct btree *b,
495 			       const struct bkey_packed *src)
496 {
497 #ifdef HAVE_BCACHEFS_COMPILED_UNPACK
498 	return bkey_unpack_key_format_checked(b, src).p;
499 #else
500 	return __bkey_unpack_pos(&b->format, src);
501 #endif
502 }
503 
504 static inline struct bpos bkey_unpack_pos(const struct btree *b,
505 					  const struct bkey_packed *src)
506 {
507 	return likely(bkey_packed(src))
508 		? bkey_unpack_pos_format_checked(b, src)
509 		: packed_to_bkey_c(src)->p;
510 }
511 
512 /* Disassembled bkeys */
513 
514 static inline struct bkey_s_c bkey_disassemble(const struct btree *b,
515 					       const struct bkey_packed *k,
516 					       struct bkey *u)
517 {
518 	__bkey_unpack_key(b, u, k);
519 
520 	return (struct bkey_s_c) { u, bkeyp_val(&b->format, k), };
521 }
522 
523 /* non const version: */
524 static inline struct bkey_s __bkey_disassemble(const struct btree *b,
525 					       struct bkey_packed *k,
526 					       struct bkey *u)
527 {
528 	__bkey_unpack_key(b, u, k);
529 
530 	return (struct bkey_s) { .k = u, .v = bkeyp_val(&b->format, k), };
531 }
532 
533 static inline u64 bkey_field_max(const struct bkey_format *f,
534 				 enum bch_bkey_fields nr)
535 {
536 	return f->bits_per_field[nr] < 64
537 		? (le64_to_cpu(f->field_offset[nr]) +
538 		   ~(~0ULL << f->bits_per_field[nr]))
539 		: U64_MAX;
540 }
541 
542 #ifdef HAVE_BCACHEFS_COMPILED_UNPACK
543 
544 int bch2_compile_bkey_format(const struct bkey_format *, void *);
545 
546 #else
547 
548 static inline int bch2_compile_bkey_format(const struct bkey_format *format,
549 					  void *out) { return 0; }
550 
551 #endif
552 
553 static inline void bkey_reassemble(struct bkey_i *dst,
554 				   struct bkey_s_c src)
555 {
556 	dst->k = *src.k;
557 	memcpy_u64s_small(&dst->v, src.v, bkey_val_u64s(src.k));
558 }
559 
560 #define bkey_s_null		((struct bkey_s)   { .k = NULL })
561 #define bkey_s_c_null		((struct bkey_s_c) { .k = NULL })
562 
563 #define bkey_s_err(err)		((struct bkey_s)   { .k = ERR_PTR(err) })
564 #define bkey_s_c_err(err)	((struct bkey_s_c) { .k = ERR_PTR(err) })
565 
566 static inline struct bkey_s bkey_to_s(struct bkey *k)
567 {
568 	return (struct bkey_s) { .k = k, .v = NULL };
569 }
570 
571 static inline struct bkey_s_c bkey_to_s_c(const struct bkey *k)
572 {
573 	return (struct bkey_s_c) { .k = k, .v = NULL };
574 }
575 
576 static inline struct bkey_s bkey_i_to_s(struct bkey_i *k)
577 {
578 	return (struct bkey_s) { .k = &k->k, .v = &k->v };
579 }
580 
581 static inline struct bkey_s_c bkey_i_to_s_c(const struct bkey_i *k)
582 {
583 	return (struct bkey_s_c) { .k = &k->k, .v = &k->v };
584 }
585 
586 /*
587  * For a given type of value (e.g. struct bch_extent), generates the types for
588  * bkey + bch_extent - inline, split, split const - and also all the conversion
589  * functions, which also check that the value is of the correct type.
590  *
591  * We use anonymous unions for upcasting - e.g. converting from e.g. a
592  * bkey_i_extent to a bkey_i - since that's always safe, instead of conversion
593  * functions.
594  */
595 #define x(name, ...)					\
596 struct bkey_i_##name {							\
597 	union {								\
598 		struct bkey		k;				\
599 		struct bkey_i		k_i;				\
600 	};								\
601 	struct bch_##name		v;				\
602 };									\
603 									\
604 struct bkey_s_c_##name {						\
605 	union {								\
606 	struct {							\
607 		const struct bkey	*k;				\
608 		const struct bch_##name	*v;				\
609 	};								\
610 	struct bkey_s_c			s_c;				\
611 	};								\
612 };									\
613 									\
614 struct bkey_s_##name {							\
615 	union {								\
616 	struct {							\
617 		struct bkey		*k;				\
618 		struct bch_##name	*v;				\
619 	};								\
620 	struct bkey_s_c_##name		c;				\
621 	struct bkey_s			s;				\
622 	struct bkey_s_c			s_c;				\
623 	};								\
624 };									\
625 									\
626 static inline struct bkey_i_##name *bkey_i_to_##name(struct bkey_i *k)	\
627 {									\
628 	EBUG_ON(!IS_ERR_OR_NULL(k) && k->k.type != KEY_TYPE_##name);	\
629 	return container_of(&k->k, struct bkey_i_##name, k);		\
630 }									\
631 									\
632 static inline const struct bkey_i_##name *				\
633 bkey_i_to_##name##_c(const struct bkey_i *k)				\
634 {									\
635 	EBUG_ON(!IS_ERR_OR_NULL(k) && k->k.type != KEY_TYPE_##name);	\
636 	return container_of(&k->k, struct bkey_i_##name, k);		\
637 }									\
638 									\
639 static inline struct bkey_s_##name bkey_s_to_##name(struct bkey_s k)	\
640 {									\
641 	EBUG_ON(!IS_ERR_OR_NULL(k.k) && k.k->type != KEY_TYPE_##name);	\
642 	return (struct bkey_s_##name) {					\
643 		.k = k.k,						\
644 		.v = container_of(k.v, struct bch_##name, v),		\
645 	};								\
646 }									\
647 									\
648 static inline struct bkey_s_c_##name bkey_s_c_to_##name(struct bkey_s_c k)\
649 {									\
650 	EBUG_ON(!IS_ERR_OR_NULL(k.k) && k.k->type != KEY_TYPE_##name);	\
651 	return (struct bkey_s_c_##name) {				\
652 		.k = k.k,						\
653 		.v = container_of(k.v, struct bch_##name, v),		\
654 	};								\
655 }									\
656 									\
657 static inline struct bkey_s_##name name##_i_to_s(struct bkey_i_##name *k)\
658 {									\
659 	return (struct bkey_s_##name) {					\
660 		.k = &k->k,						\
661 		.v = &k->v,						\
662 	};								\
663 }									\
664 									\
665 static inline struct bkey_s_c_##name					\
666 name##_i_to_s_c(const struct bkey_i_##name *k)				\
667 {									\
668 	return (struct bkey_s_c_##name) {				\
669 		.k = &k->k,						\
670 		.v = &k->v,						\
671 	};								\
672 }									\
673 									\
674 static inline struct bkey_s_##name bkey_i_to_s_##name(struct bkey_i *k)	\
675 {									\
676 	EBUG_ON(!IS_ERR_OR_NULL(k) && k->k.type != KEY_TYPE_##name);	\
677 	return (struct bkey_s_##name) {					\
678 		.k = &k->k,						\
679 		.v = container_of(&k->v, struct bch_##name, v),		\
680 	};								\
681 }									\
682 									\
683 static inline struct bkey_s_c_##name					\
684 bkey_i_to_s_c_##name(const struct bkey_i *k)				\
685 {									\
686 	EBUG_ON(!IS_ERR_OR_NULL(k) && k->k.type != KEY_TYPE_##name);	\
687 	return (struct bkey_s_c_##name) {				\
688 		.k = &k->k,						\
689 		.v = container_of(&k->v, struct bch_##name, v),		\
690 	};								\
691 }									\
692 									\
693 static inline struct bkey_i_##name *bkey_##name##_init(struct bkey_i *_k)\
694 {									\
695 	struct bkey_i_##name *k =					\
696 		container_of(&_k->k, struct bkey_i_##name, k);		\
697 									\
698 	bkey_init(&k->k);						\
699 	memset(&k->v, 0, sizeof(k->v));					\
700 	k->k.type = KEY_TYPE_##name;					\
701 	set_bkey_val_bytes(&k->k, sizeof(k->v));			\
702 									\
703 	return k;							\
704 }
705 
706 BCH_BKEY_TYPES();
707 #undef x
708 
709 /* byte order helpers */
710 
711 #if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
712 
713 static inline unsigned high_word_offset(const struct bkey_format *f)
714 {
715 	return f->key_u64s - 1;
716 }
717 
718 #define high_bit_offset		0
719 #define nth_word(p, n)		((p) - (n))
720 
721 #elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
722 
723 static inline unsigned high_word_offset(const struct bkey_format *f)
724 {
725 	return 0;
726 }
727 
728 #define high_bit_offset		KEY_PACKED_BITS_START
729 #define nth_word(p, n)		((p) + (n))
730 
731 #else
732 #error edit for your odd byteorder.
733 #endif
734 
735 #define high_word(f, k)		((u64 *) (k)->_data + high_word_offset(f))
736 #define next_word(p)		nth_word(p, 1)
737 #define prev_word(p)		nth_word(p, -1)
738 
739 #ifdef CONFIG_BCACHEFS_DEBUG
740 void bch2_bkey_pack_test(void);
741 #else
742 static inline void bch2_bkey_pack_test(void) {}
743 #endif
744 
745 #define bkey_fields()							\
746 	x(BKEY_FIELD_INODE,		p.inode)			\
747 	x(BKEY_FIELD_OFFSET,		p.offset)			\
748 	x(BKEY_FIELD_SNAPSHOT,		p.snapshot)			\
749 	x(BKEY_FIELD_SIZE,		size)				\
750 	x(BKEY_FIELD_VERSION_HI,	version.hi)			\
751 	x(BKEY_FIELD_VERSION_LO,	version.lo)
752 
753 struct bkey_format_state {
754 	u64 field_min[BKEY_NR_FIELDS];
755 	u64 field_max[BKEY_NR_FIELDS];
756 };
757 
758 void bch2_bkey_format_init(struct bkey_format_state *);
759 
760 static inline void __bkey_format_add(struct bkey_format_state *s, unsigned field, u64 v)
761 {
762 	s->field_min[field] = min(s->field_min[field], v);
763 	s->field_max[field] = max(s->field_max[field], v);
764 }
765 
766 /*
767  * Changes @format so that @k can be successfully packed with @format
768  */
769 static inline void bch2_bkey_format_add_key(struct bkey_format_state *s, const struct bkey *k)
770 {
771 #define x(id, field) __bkey_format_add(s, id, k->field);
772 	bkey_fields()
773 #undef x
774 }
775 
776 void bch2_bkey_format_add_pos(struct bkey_format_state *, struct bpos);
777 struct bkey_format bch2_bkey_format_done(struct bkey_format_state *);
778 int bch2_bkey_format_invalid(struct bch_fs *, struct bkey_format *,
779 			     enum bkey_invalid_flags, struct printbuf *);
780 void bch2_bkey_format_to_text(struct printbuf *, const struct bkey_format *);
781 
782 #endif /* _BCACHEFS_BKEY_H */
783