xref: /linux/fs/bcachefs/ec.h (revision e7c3696d4692e8046d25f6e63f983e934e12f2c5)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _BCACHEFS_EC_H
3 #define _BCACHEFS_EC_H
4 
5 #include "ec_types.h"
6 #include "buckets_types.h"
7 #include "extents_types.h"
8 
9 enum bch_validate_flags;
10 
11 int bch2_stripe_invalid(struct bch_fs *, struct bkey_s_c,
12 			enum bch_validate_flags, struct printbuf *);
13 void bch2_stripe_to_text(struct printbuf *, struct bch_fs *,
14 			 struct bkey_s_c);
15 int bch2_trigger_stripe(struct btree_trans *, enum btree_id, unsigned,
16 			struct bkey_s_c, struct bkey_s,
17 			enum btree_iter_update_trigger_flags);
18 
19 #define bch2_bkey_ops_stripe ((struct bkey_ops) {	\
20 	.key_invalid	= bch2_stripe_invalid,		\
21 	.val_to_text	= bch2_stripe_to_text,		\
22 	.swab		= bch2_ptr_swab,		\
23 	.trigger	= bch2_trigger_stripe,		\
24 	.min_val_size	= 8,				\
25 })
26 
27 static inline unsigned stripe_csums_per_device(const struct bch_stripe *s)
28 {
29 	return DIV_ROUND_UP(le16_to_cpu(s->sectors),
30 			    1 << s->csum_granularity_bits);
31 }
32 
33 static inline unsigned stripe_csum_offset(const struct bch_stripe *s,
34 					  unsigned dev, unsigned csum_idx)
35 {
36 	EBUG_ON(s->csum_type >= BCH_CSUM_NR);
37 
38 	unsigned csum_bytes = bch_crc_bytes[s->csum_type];
39 
40 	return sizeof(struct bch_stripe) +
41 		sizeof(struct bch_extent_ptr) * s->nr_blocks +
42 		(dev * stripe_csums_per_device(s) + csum_idx) * csum_bytes;
43 }
44 
45 static inline unsigned stripe_blockcount_offset(const struct bch_stripe *s,
46 						unsigned idx)
47 {
48 	return stripe_csum_offset(s, s->nr_blocks, 0) +
49 		sizeof(u16) * idx;
50 }
51 
52 static inline unsigned stripe_blockcount_get(const struct bch_stripe *s,
53 					     unsigned idx)
54 {
55 	return le16_to_cpup((void *) s + stripe_blockcount_offset(s, idx));
56 }
57 
58 static inline void stripe_blockcount_set(struct bch_stripe *s,
59 					 unsigned idx, unsigned v)
60 {
61 	__le16 *p = (void *) s + stripe_blockcount_offset(s, idx);
62 
63 	*p = cpu_to_le16(v);
64 }
65 
66 static inline unsigned stripe_val_u64s(const struct bch_stripe *s)
67 {
68 	return DIV_ROUND_UP(stripe_blockcount_offset(s, s->nr_blocks),
69 			    sizeof(u64));
70 }
71 
72 static inline void *stripe_csum(struct bch_stripe *s,
73 				unsigned block, unsigned csum_idx)
74 {
75 	EBUG_ON(block >= s->nr_blocks);
76 	EBUG_ON(csum_idx >= stripe_csums_per_device(s));
77 
78 	return (void *) s + stripe_csum_offset(s, block, csum_idx);
79 }
80 
81 static inline struct bch_csum stripe_csum_get(struct bch_stripe *s,
82 				   unsigned block, unsigned csum_idx)
83 {
84 	struct bch_csum csum = { 0 };
85 
86 	memcpy(&csum, stripe_csum(s, block, csum_idx), bch_crc_bytes[s->csum_type]);
87 	return csum;
88 }
89 
90 static inline void stripe_csum_set(struct bch_stripe *s,
91 				   unsigned block, unsigned csum_idx,
92 				   struct bch_csum csum)
93 {
94 	memcpy(stripe_csum(s, block, csum_idx), &csum, bch_crc_bytes[s->csum_type]);
95 }
96 
97 static inline bool __bch2_ptr_matches_stripe(const struct bch_extent_ptr *stripe_ptr,
98 					     const struct bch_extent_ptr *data_ptr,
99 					     unsigned sectors)
100 {
101 	return  data_ptr->dev    == stripe_ptr->dev &&
102 		data_ptr->gen    == stripe_ptr->gen &&
103 		data_ptr->offset >= stripe_ptr->offset &&
104 		data_ptr->offset  < stripe_ptr->offset + sectors;
105 }
106 
107 static inline bool bch2_ptr_matches_stripe(const struct bch_stripe *s,
108 					   struct extent_ptr_decoded p)
109 {
110 	unsigned nr_data = s->nr_blocks - s->nr_redundant;
111 
112 	BUG_ON(!p.has_ec);
113 
114 	if (p.ec.block >= nr_data)
115 		return false;
116 
117 	return __bch2_ptr_matches_stripe(&s->ptrs[p.ec.block], &p.ptr,
118 					 le16_to_cpu(s->sectors));
119 }
120 
121 static inline bool bch2_ptr_matches_stripe_m(const struct gc_stripe *m,
122 					     struct extent_ptr_decoded p)
123 {
124 	unsigned nr_data = m->nr_blocks - m->nr_redundant;
125 
126 	BUG_ON(!p.has_ec);
127 
128 	if (p.ec.block >= nr_data)
129 		return false;
130 
131 	return __bch2_ptr_matches_stripe(&m->ptrs[p.ec.block], &p.ptr,
132 					 m->sectors);
133 }
134 
135 struct bch_read_bio;
136 
137 struct ec_stripe_buf {
138 	/* might not be buffering the entire stripe: */
139 	unsigned		offset;
140 	unsigned		size;
141 	unsigned long		valid[BITS_TO_LONGS(BCH_BKEY_PTRS_MAX)];
142 
143 	void			*data[BCH_BKEY_PTRS_MAX];
144 
145 	__BKEY_PADDED(key, 255);
146 };
147 
148 struct ec_stripe_head;
149 
150 enum ec_stripe_ref {
151 	STRIPE_REF_io,
152 	STRIPE_REF_stripe,
153 	STRIPE_REF_NR
154 };
155 
156 struct ec_stripe_new {
157 	struct bch_fs		*c;
158 	struct ec_stripe_head	*h;
159 	struct mutex		lock;
160 	struct list_head	list;
161 
162 	struct hlist_node	hash;
163 	u64			idx;
164 
165 	struct closure		iodone;
166 
167 	atomic_t		ref[STRIPE_REF_NR];
168 
169 	int			err;
170 
171 	u8			nr_data;
172 	u8			nr_parity;
173 	bool			allocated;
174 	bool			pending;
175 	bool			have_existing_stripe;
176 
177 	unsigned long		blocks_gotten[BITS_TO_LONGS(BCH_BKEY_PTRS_MAX)];
178 	unsigned long		blocks_allocated[BITS_TO_LONGS(BCH_BKEY_PTRS_MAX)];
179 	open_bucket_idx_t	blocks[BCH_BKEY_PTRS_MAX];
180 	struct disk_reservation	res;
181 
182 	struct ec_stripe_buf	new_stripe;
183 	struct ec_stripe_buf	existing_stripe;
184 };
185 
186 struct ec_stripe_head {
187 	struct list_head	list;
188 	struct mutex		lock;
189 
190 	unsigned		target;
191 	unsigned		algo;
192 	unsigned		redundancy;
193 	enum bch_watermark	watermark;
194 
195 	struct bch_devs_mask	devs;
196 	unsigned		nr_active_devs;
197 
198 	unsigned		blocksize;
199 
200 	struct dev_stripe_state	block_stripe;
201 	struct dev_stripe_state	parity_stripe;
202 
203 	struct ec_stripe_new	*s;
204 };
205 
206 int bch2_ec_read_extent(struct btree_trans *, struct bch_read_bio *);
207 
208 void *bch2_writepoint_ec_buf(struct bch_fs *, struct write_point *);
209 
210 void bch2_ec_bucket_cancel(struct bch_fs *, struct open_bucket *);
211 
212 int bch2_ec_stripe_new_alloc(struct bch_fs *, struct ec_stripe_head *);
213 
214 void bch2_ec_stripe_head_put(struct bch_fs *, struct ec_stripe_head *);
215 struct ec_stripe_head *bch2_ec_stripe_head_get(struct btree_trans *,
216 			unsigned, unsigned, unsigned,
217 			enum bch_watermark, struct closure *);
218 
219 void bch2_stripes_heap_update(struct bch_fs *, struct stripe *, size_t);
220 void bch2_stripes_heap_del(struct bch_fs *, struct stripe *, size_t);
221 void bch2_stripes_heap_insert(struct bch_fs *, struct stripe *, size_t);
222 
223 void bch2_do_stripe_deletes(struct bch_fs *);
224 void bch2_ec_do_stripe_creates(struct bch_fs *);
225 void bch2_ec_stripe_new_free(struct bch_fs *, struct ec_stripe_new *);
226 
227 static inline void ec_stripe_new_get(struct ec_stripe_new *s,
228 				     enum ec_stripe_ref ref)
229 {
230 	atomic_inc(&s->ref[ref]);
231 }
232 
233 static inline void ec_stripe_new_put(struct bch_fs *c, struct ec_stripe_new *s,
234 				     enum ec_stripe_ref ref)
235 {
236 	BUG_ON(atomic_read(&s->ref[ref]) <= 0);
237 
238 	if (atomic_dec_and_test(&s->ref[ref]))
239 		switch (ref) {
240 		case STRIPE_REF_stripe:
241 			bch2_ec_stripe_new_free(c, s);
242 			break;
243 		case STRIPE_REF_io:
244 			bch2_ec_do_stripe_creates(c);
245 			break;
246 		default:
247 			BUG();
248 		}
249 }
250 
251 void bch2_ec_stop_dev(struct bch_fs *, struct bch_dev *);
252 void bch2_fs_ec_stop(struct bch_fs *);
253 void bch2_fs_ec_flush(struct bch_fs *);
254 
255 int bch2_stripes_read(struct bch_fs *);
256 
257 void bch2_stripes_heap_to_text(struct printbuf *, struct bch_fs *);
258 void bch2_new_stripes_to_text(struct printbuf *, struct bch_fs *);
259 
260 void bch2_fs_ec_exit(struct bch_fs *);
261 void bch2_fs_ec_init_early(struct bch_fs *);
262 int bch2_fs_ec_init(struct bch_fs *);
263 
264 #endif /* _BCACHEFS_EC_H */
265