xref: /linux/fs/bcachefs/sb-clean.c (revision b8e4b0529d59a3ccd0b25a31d3cfc8b0f3b34068)
1 // SPDX-License-Identifier: GPL-2.0
2 
3 #include "bcachefs.h"
4 #include "btree_update_interior.h"
5 #include "buckets.h"
6 #include "error.h"
7 #include "journal_io.h"
8 #include "replicas.h"
9 #include "sb-clean.h"
10 #include "super-io.h"
11 
12 /*
13  * BCH_SB_FIELD_clean:
14  *
15  * Btree roots, and a few other things, are recovered from the journal after an
16  * unclean shutdown - but after a clean shutdown, to avoid having to read the
17  * journal, we can store them in the superblock.
18  *
19  * bch_sb_field_clean simply contains a list of journal entries, stored exactly
20  * as they would be in the journal:
21  */
22 
23 int bch2_sb_clean_validate_late(struct bch_fs *c, struct bch_sb_field_clean *clean,
24 				int write)
25 {
26 	struct jset_entry *entry;
27 	int ret;
28 
29 	for (entry = clean->start;
30 	     entry < (struct jset_entry *) vstruct_end(&clean->field);
31 	     entry = vstruct_next(entry)) {
32 		if (vstruct_end(entry) > vstruct_end(&clean->field)) {
33 			bch_err(c, "journal entry (u64s %u) overran end of superblock clean section (u64s %u) by %zu",
34 				le16_to_cpu(entry->u64s), le32_to_cpu(clean->field.u64s),
35 				(u64 *) vstruct_end(entry) - (u64 *) vstruct_end(&clean->field));
36 			bch2_sb_error_count(c, BCH_FSCK_ERR_sb_clean_entry_overrun);
37 			return -BCH_ERR_fsck_repair_unimplemented;
38 		}
39 
40 		ret = bch2_journal_entry_validate(c, NULL, entry,
41 						  le16_to_cpu(c->disk_sb.sb->version),
42 						  BCH_SB_BIG_ENDIAN(c->disk_sb.sb),
43 						  write);
44 		if (ret)
45 			return ret;
46 	}
47 
48 	return 0;
49 }
50 
51 static struct bkey_i *btree_root_find(struct bch_fs *c,
52 				      struct bch_sb_field_clean *clean,
53 				      struct jset *j,
54 				      enum btree_id id, unsigned *level)
55 {
56 	struct bkey_i *k;
57 	struct jset_entry *entry, *start, *end;
58 
59 	if (clean) {
60 		start = clean->start;
61 		end = vstruct_end(&clean->field);
62 	} else {
63 		start = j->start;
64 		end = vstruct_last(j);
65 	}
66 
67 	for (entry = start; entry < end; entry = vstruct_next(entry))
68 		if (entry->type == BCH_JSET_ENTRY_btree_root &&
69 		    entry->btree_id == id)
70 			goto found;
71 
72 	return NULL;
73 found:
74 	if (!entry->u64s)
75 		return ERR_PTR(-EINVAL);
76 
77 	k = entry->start;
78 	*level = entry->level;
79 	return k;
80 }
81 
82 int bch2_verify_superblock_clean(struct bch_fs *c,
83 				 struct bch_sb_field_clean **cleanp,
84 				 struct jset *j)
85 {
86 	unsigned i;
87 	struct bch_sb_field_clean *clean = *cleanp;
88 	struct printbuf buf1 = PRINTBUF;
89 	struct printbuf buf2 = PRINTBUF;
90 	int ret = 0;
91 
92 	if (mustfix_fsck_err_on(j->seq != clean->journal_seq, c,
93 			sb_clean_journal_seq_mismatch,
94 			"superblock journal seq (%llu) doesn't match journal (%llu) after clean shutdown",
95 			le64_to_cpu(clean->journal_seq),
96 			le64_to_cpu(j->seq))) {
97 		kfree(clean);
98 		*cleanp = NULL;
99 		return 0;
100 	}
101 
102 	for (i = 0; i < BTREE_ID_NR; i++) {
103 		struct bkey_i *k1, *k2;
104 		unsigned l1 = 0, l2 = 0;
105 
106 		k1 = btree_root_find(c, clean, NULL, i, &l1);
107 		k2 = btree_root_find(c, NULL, j, i, &l2);
108 
109 		if (!k1 && !k2)
110 			continue;
111 
112 		printbuf_reset(&buf1);
113 		printbuf_reset(&buf2);
114 
115 		if (k1)
116 			bch2_bkey_val_to_text(&buf1, c, bkey_i_to_s_c(k1));
117 		else
118 			prt_printf(&buf1, "(none)");
119 
120 		if (k2)
121 			bch2_bkey_val_to_text(&buf2, c, bkey_i_to_s_c(k2));
122 		else
123 			prt_printf(&buf2, "(none)");
124 
125 		mustfix_fsck_err_on(!k1 || !k2 ||
126 				    IS_ERR(k1) ||
127 				    IS_ERR(k2) ||
128 				    k1->k.u64s != k2->k.u64s ||
129 				    memcmp(k1, k2, bkey_bytes(&k1->k)) ||
130 				    l1 != l2, c,
131 			sb_clean_btree_root_mismatch,
132 			"superblock btree root %u doesn't match journal after clean shutdown\n"
133 			"sb:      l=%u %s\n"
134 			"journal: l=%u %s\n", i,
135 			l1, buf1.buf,
136 			l2, buf2.buf);
137 	}
138 fsck_err:
139 	printbuf_exit(&buf2);
140 	printbuf_exit(&buf1);
141 	return ret;
142 }
143 
144 struct bch_sb_field_clean *bch2_read_superblock_clean(struct bch_fs *c)
145 {
146 	struct bch_sb_field_clean *clean, *sb_clean;
147 	int ret;
148 
149 	mutex_lock(&c->sb_lock);
150 	sb_clean = bch2_sb_field_get(c->disk_sb.sb, clean);
151 
152 	if (fsck_err_on(!sb_clean, c,
153 			sb_clean_missing,
154 			"superblock marked clean but clean section not present")) {
155 		SET_BCH_SB_CLEAN(c->disk_sb.sb, false);
156 		c->sb.clean = false;
157 		mutex_unlock(&c->sb_lock);
158 		return NULL;
159 	}
160 
161 	clean = kmemdup(sb_clean, vstruct_bytes(&sb_clean->field),
162 			GFP_KERNEL);
163 	if (!clean) {
164 		mutex_unlock(&c->sb_lock);
165 		return ERR_PTR(-BCH_ERR_ENOMEM_read_superblock_clean);
166 	}
167 
168 	ret = bch2_sb_clean_validate_late(c, clean, READ);
169 	if (ret) {
170 		mutex_unlock(&c->sb_lock);
171 		return ERR_PTR(ret);
172 	}
173 
174 	mutex_unlock(&c->sb_lock);
175 
176 	return clean;
177 fsck_err:
178 	mutex_unlock(&c->sb_lock);
179 	return ERR_PTR(ret);
180 }
181 
182 void bch2_journal_super_entries_add_common(struct bch_fs *c,
183 					   struct jset_entry **end,
184 					   u64 journal_seq)
185 {
186 	{
187 		struct jset_entry_usage *u =
188 			container_of(jset_entry_init(end, sizeof(*u)),
189 				     struct jset_entry_usage, entry);
190 
191 		u->entry.type	= BCH_JSET_ENTRY_usage;
192 		u->entry.btree_id = BCH_FS_USAGE_key_version;
193 		u->v		= cpu_to_le64(atomic64_read(&c->key_version));
194 	}
195 
196 	for (unsigned i = 0; i < 2; i++) {
197 		struct jset_entry_clock *clock =
198 			container_of(jset_entry_init(end, sizeof(*clock)),
199 				     struct jset_entry_clock, entry);
200 
201 		clock->entry.type = BCH_JSET_ENTRY_clock;
202 		clock->rw	= i;
203 		clock->time	= cpu_to_le64(atomic64_read(&c->io_clock[i].now));
204 	}
205 }
206 
207 static int bch2_sb_clean_validate(struct bch_sb *sb, struct bch_sb_field *f,
208 				  enum bch_validate_flags flags, struct printbuf *err)
209 {
210 	struct bch_sb_field_clean *clean = field_to_type(f, clean);
211 
212 	if (vstruct_bytes(&clean->field) < sizeof(*clean)) {
213 		prt_printf(err, "wrong size (got %zu should be %zu)",
214 		       vstruct_bytes(&clean->field), sizeof(*clean));
215 		return -BCH_ERR_invalid_sb_clean;
216 	}
217 
218 	for (struct jset_entry *entry = clean->start;
219 	     entry != vstruct_end(&clean->field);
220 	     entry = vstruct_next(entry)) {
221 		if ((void *) vstruct_next(entry) > vstruct_end(&clean->field)) {
222 			prt_str(err, "entry type ");
223 			bch2_prt_jset_entry_type(err, entry->type);
224 			prt_str(err, " overruns end of section");
225 			return -BCH_ERR_invalid_sb_clean;
226 		}
227 	}
228 
229 	return 0;
230 }
231 
232 static void bch2_sb_clean_to_text(struct printbuf *out, struct bch_sb *sb,
233 				  struct bch_sb_field *f)
234 {
235 	struct bch_sb_field_clean *clean = field_to_type(f, clean);
236 	struct jset_entry *entry;
237 
238 	prt_printf(out, "flags:          %x\n",		le32_to_cpu(clean->flags));
239 	prt_printf(out, "journal_seq:    %llu\n",	le64_to_cpu(clean->journal_seq));
240 
241 	for (entry = clean->start;
242 	     entry != vstruct_end(&clean->field);
243 	     entry = vstruct_next(entry)) {
244 		if ((void *) vstruct_next(entry) > vstruct_end(&clean->field))
245 			break;
246 
247 		if (entry->type == BCH_JSET_ENTRY_btree_keys &&
248 		    !entry->u64s)
249 			continue;
250 
251 		bch2_journal_entry_to_text(out, NULL, entry);
252 		prt_newline(out);
253 	}
254 }
255 
256 const struct bch_sb_field_ops bch_sb_field_ops_clean = {
257 	.validate	= bch2_sb_clean_validate,
258 	.to_text	= bch2_sb_clean_to_text,
259 };
260 
261 int bch2_fs_mark_dirty(struct bch_fs *c)
262 {
263 	int ret;
264 
265 	/*
266 	 * Unconditionally write superblock, to verify it hasn't changed before
267 	 * we go rw:
268 	 */
269 
270 	mutex_lock(&c->sb_lock);
271 	SET_BCH_SB_CLEAN(c->disk_sb.sb, false);
272 	c->disk_sb.sb->features[0] |= cpu_to_le64(BCH_SB_FEATURES_ALWAYS);
273 
274 	ret = bch2_write_super(c);
275 	mutex_unlock(&c->sb_lock);
276 
277 	return ret;
278 }
279 
280 void bch2_fs_mark_clean(struct bch_fs *c)
281 {
282 	struct bch_sb_field_clean *sb_clean;
283 	struct jset_entry *entry;
284 	unsigned u64s;
285 	int ret;
286 
287 	mutex_lock(&c->sb_lock);
288 	if (BCH_SB_CLEAN(c->disk_sb.sb))
289 		goto out;
290 
291 	SET_BCH_SB_CLEAN(c->disk_sb.sb, true);
292 
293 	c->disk_sb.sb->compat[0] |= cpu_to_le64(1ULL << BCH_COMPAT_alloc_info);
294 	c->disk_sb.sb->compat[0] |= cpu_to_le64(1ULL << BCH_COMPAT_alloc_metadata);
295 	c->disk_sb.sb->features[0] &= cpu_to_le64(~(1ULL << BCH_FEATURE_extents_above_btree_updates));
296 	c->disk_sb.sb->features[0] &= cpu_to_le64(~(1ULL << BCH_FEATURE_btree_updates_journalled));
297 
298 	u64s = sizeof(*sb_clean) / sizeof(u64) + c->journal.entry_u64s_reserved;
299 
300 	sb_clean = bch2_sb_field_resize(&c->disk_sb, clean, u64s);
301 	if (!sb_clean) {
302 		bch_err(c, "error resizing superblock while setting filesystem clean");
303 		goto out;
304 	}
305 
306 	sb_clean->flags		= 0;
307 	sb_clean->journal_seq	= cpu_to_le64(atomic64_read(&c->journal.seq));
308 
309 	/* Trying to catch outstanding bug: */
310 	BUG_ON(le64_to_cpu(sb_clean->journal_seq) > S64_MAX);
311 
312 	entry = sb_clean->start;
313 	bch2_journal_super_entries_add_common(c, &entry, 0);
314 	entry = bch2_btree_roots_to_journal_entries(c, entry, 0);
315 	BUG_ON((void *) entry > vstruct_end(&sb_clean->field));
316 
317 	memset(entry, 0,
318 	       vstruct_end(&sb_clean->field) - (void *) entry);
319 
320 	/*
321 	 * this should be in the write path, and we should be validating every
322 	 * superblock section:
323 	 */
324 	ret = bch2_sb_clean_validate_late(c, sb_clean, WRITE);
325 	if (ret) {
326 		bch_err(c, "error writing marking filesystem clean: validate error");
327 		goto out;
328 	}
329 
330 	bch2_journal_pos_from_member_info_set(c);
331 
332 	bch2_write_super(c);
333 out:
334 	mutex_unlock(&c->sb_lock);
335 }
336