1 /* SPDX-License-Identifier: GPL-2.0 */ 2 #ifndef _BCACHEFS_BACKPOINTERS_H 3 #define _BCACHEFS_BACKPOINTERS_H 4 5 #include "btree_cache.h" 6 #include "btree_iter.h" 7 #include "btree_update.h" 8 #include "buckets.h" 9 #include "error.h" 10 #include "super.h" 11 12 static inline u64 swab40(u64 x) 13 { 14 return (((x & 0x00000000ffULL) << 32)| 15 ((x & 0x000000ff00ULL) << 16)| 16 ((x & 0x0000ff0000ULL) >> 0)| 17 ((x & 0x00ff000000ULL) >> 16)| 18 ((x & 0xff00000000ULL) >> 32)); 19 } 20 21 int bch2_backpointer_validate(struct bch_fs *, struct bkey_s_c k, 22 struct bkey_validate_context); 23 void bch2_backpointer_to_text(struct printbuf *, struct bch_fs *, struct bkey_s_c); 24 void bch2_backpointer_swab(struct bkey_s); 25 26 #define bch2_bkey_ops_backpointer ((struct bkey_ops) { \ 27 .key_validate = bch2_backpointer_validate, \ 28 .val_to_text = bch2_backpointer_to_text, \ 29 .swab = bch2_backpointer_swab, \ 30 .min_val_size = 32, \ 31 }) 32 33 #define MAX_EXTENT_COMPRESS_RATIO_SHIFT 10 34 35 /* 36 * Convert from pos in backpointer btree to pos of corresponding bucket in alloc 37 * btree: 38 */ 39 static inline struct bpos bp_pos_to_bucket(const struct bch_dev *ca, struct bpos bp_pos) 40 { 41 u64 bucket_sector = bp_pos.offset >> MAX_EXTENT_COMPRESS_RATIO_SHIFT; 42 43 return POS(bp_pos.inode, sector_to_bucket(ca, bucket_sector)); 44 } 45 46 static inline struct bpos bp_pos_to_bucket_and_offset(const struct bch_dev *ca, struct bpos bp_pos, 47 u32 *bucket_offset) 48 { 49 u64 bucket_sector = bp_pos.offset >> MAX_EXTENT_COMPRESS_RATIO_SHIFT; 50 51 return POS(bp_pos.inode, sector_to_bucket_and_offset(ca, bucket_sector, bucket_offset)); 52 } 53 54 static inline bool bp_pos_to_bucket_nodev_noerror(struct bch_fs *c, struct bpos bp_pos, struct bpos *bucket) 55 { 56 guard(rcu)(); 57 struct bch_dev *ca = bch2_dev_rcu_noerror(c, bp_pos.inode); 58 if (ca) 59 *bucket = bp_pos_to_bucket(ca, bp_pos); 60 return ca != NULL; 61 } 62 63 static inline struct bpos bucket_pos_to_bp_noerror(const struct bch_dev *ca, 64 struct bpos bucket, 65 u64 bucket_offset) 66 { 67 return POS(bucket.inode, 68 (bucket_to_sector(ca, bucket.offset) << 69 MAX_EXTENT_COMPRESS_RATIO_SHIFT) + bucket_offset); 70 } 71 72 /* 73 * Convert from pos in alloc btree + bucket offset to pos in backpointer btree: 74 */ 75 static inline struct bpos bucket_pos_to_bp(const struct bch_dev *ca, 76 struct bpos bucket, 77 u64 bucket_offset) 78 { 79 struct bpos ret = bucket_pos_to_bp_noerror(ca, bucket, bucket_offset); 80 EBUG_ON(!bkey_eq(bucket, bp_pos_to_bucket(ca, ret))); 81 return ret; 82 } 83 84 static inline struct bpos bucket_pos_to_bp_start(const struct bch_dev *ca, struct bpos bucket) 85 { 86 return bucket_pos_to_bp(ca, bucket, 0); 87 } 88 89 static inline struct bpos bucket_pos_to_bp_end(const struct bch_dev *ca, struct bpos bucket) 90 { 91 return bpos_nosnap_predecessor(bucket_pos_to_bp(ca, bpos_nosnap_successor(bucket), 0)); 92 } 93 94 int bch2_bucket_backpointer_mod_nowritebuffer(struct btree_trans *, 95 struct bkey_s_c, 96 struct bkey_i_backpointer *, 97 bool); 98 99 static inline int bch2_bucket_backpointer_mod(struct btree_trans *trans, 100 struct bkey_s_c orig_k, 101 struct bkey_i_backpointer *bp, 102 bool insert) 103 { 104 if (static_branch_unlikely(&bch2_backpointers_no_use_write_buffer)) 105 return bch2_bucket_backpointer_mod_nowritebuffer(trans, orig_k, bp, insert); 106 107 if (!insert) { 108 bp->k.type = KEY_TYPE_deleted; 109 set_bkey_val_u64s(&bp->k, 0); 110 } 111 112 return bch2_trans_update_buffered(trans, BTREE_ID_backpointers, &bp->k_i); 113 } 114 115 static inline enum bch_data_type bch2_bkey_ptr_data_type(struct bkey_s_c k, 116 struct extent_ptr_decoded p, 117 const union bch_extent_entry *entry) 118 { 119 switch (k.k->type) { 120 case KEY_TYPE_btree_ptr: 121 case KEY_TYPE_btree_ptr_v2: 122 return BCH_DATA_btree; 123 case KEY_TYPE_extent: 124 case KEY_TYPE_reflink_v: 125 if (p.has_ec) 126 return BCH_DATA_stripe; 127 if (p.ptr.cached) 128 return BCH_DATA_cached; 129 else 130 return BCH_DATA_user; 131 case KEY_TYPE_stripe: { 132 const struct bch_extent_ptr *ptr = &entry->ptr; 133 struct bkey_s_c_stripe s = bkey_s_c_to_stripe(k); 134 135 BUG_ON(ptr < s.v->ptrs || 136 ptr >= s.v->ptrs + s.v->nr_blocks); 137 138 return ptr >= s.v->ptrs + s.v->nr_blocks - s.v->nr_redundant 139 ? BCH_DATA_parity 140 : BCH_DATA_user; 141 } 142 default: 143 BUG(); 144 } 145 } 146 147 static inline void bch2_extent_ptr_to_bp(struct bch_fs *c, 148 enum btree_id btree_id, unsigned level, 149 struct bkey_s_c k, struct extent_ptr_decoded p, 150 const union bch_extent_entry *entry, 151 struct bkey_i_backpointer *bp) 152 { 153 bkey_backpointer_init(&bp->k_i); 154 bp->k.p.inode = p.ptr.dev; 155 156 if (k.k->type != KEY_TYPE_stripe) 157 bp->k.p.offset = ((u64) p.ptr.offset << MAX_EXTENT_COMPRESS_RATIO_SHIFT) + p.crc.offset; 158 else { 159 /* 160 * Put stripe backpointers where they won't collide with the 161 * extent backpointers within the stripe: 162 */ 163 struct bkey_s_c_stripe s = bkey_s_c_to_stripe(k); 164 bp->k.p.offset = ((u64) (p.ptr.offset + le16_to_cpu(s.v->sectors)) << 165 MAX_EXTENT_COMPRESS_RATIO_SHIFT) - 1; 166 } 167 168 bp->v = (struct bch_backpointer) { 169 .btree_id = btree_id, 170 .level = level, 171 .data_type = bch2_bkey_ptr_data_type(k, p, entry), 172 .bucket_gen = p.ptr.gen, 173 .bucket_len = ptr_disk_sectors(level ? btree_sectors(c) : k.k->size, p), 174 .pos = k.k->p, 175 }; 176 } 177 178 struct bkey_buf; 179 struct bkey_s_c bch2_backpointer_get_key(struct btree_trans *, struct bkey_s_c_backpointer, 180 struct btree_iter *, unsigned, struct bkey_buf *); 181 struct btree *bch2_backpointer_get_node(struct btree_trans *, struct bkey_s_c_backpointer, 182 struct btree_iter *, struct bkey_buf *); 183 184 int bch2_check_bucket_backpointer_mismatch(struct btree_trans *, struct bch_dev *, u64, 185 bool, struct bkey_buf *); 186 187 int bch2_check_btree_backpointers(struct bch_fs *); 188 int bch2_check_extents_to_backpointers(struct bch_fs *); 189 int bch2_check_backpointers_to_extents(struct bch_fs *); 190 191 static inline bool bch2_bucket_bitmap_test(struct bucket_bitmap *b, u64 i) 192 { 193 unsigned long *bitmap = READ_ONCE(b->buckets); 194 return bitmap && test_bit(i, bitmap); 195 } 196 197 int bch2_bucket_bitmap_resize(struct bch_dev *, struct bucket_bitmap *, u64, u64); 198 void bch2_bucket_bitmap_free(struct bucket_bitmap *); 199 200 #endif /* _BCACHEFS_BACKPOINTERS_BACKGROUND_H */ 201