xref: /linux/fs/bcachefs/buckets_waiting_for_journal.c (revision 94b481f7671f412b571e4ee5e3b24f33a889812d)
1 // SPDX-License-Identifier: GPL-2.0
2 
3 #include "bcachefs.h"
4 #include "buckets_waiting_for_journal.h"
5 #include <linux/hash.h>
6 #include <linux/random.h>
7 
8 static inline struct bucket_hashed *
bucket_hash(struct buckets_waiting_for_journal_table * t,unsigned hash_seed_idx,u64 dev_bucket)9 bucket_hash(struct buckets_waiting_for_journal_table *t,
10 	    unsigned hash_seed_idx, u64 dev_bucket)
11 {
12 	return t->d + hash_64(dev_bucket ^ t->hash_seeds[hash_seed_idx], t->bits);
13 }
14 
bucket_table_init(struct buckets_waiting_for_journal_table * t,size_t bits)15 static void bucket_table_init(struct buckets_waiting_for_journal_table *t, size_t bits)
16 {
17 	unsigned i;
18 
19 	t->bits = bits;
20 	for (i = 0; i < ARRAY_SIZE(t->hash_seeds); i++)
21 		get_random_bytes(&t->hash_seeds[i], sizeof(t->hash_seeds[i]));
22 	memset(t->d, 0, sizeof(t->d[0]) << t->bits);
23 }
24 
bch2_bucket_journal_seq_ready(struct buckets_waiting_for_journal * b,unsigned dev,u64 bucket)25 u64 bch2_bucket_journal_seq_ready(struct buckets_waiting_for_journal *b,
26 				  unsigned dev, u64 bucket)
27 {
28 	struct buckets_waiting_for_journal_table *t;
29 	u64 dev_bucket = (u64) dev << 56 | bucket;
30 	u64 ret = 0;
31 
32 	mutex_lock(&b->lock);
33 	t = b->t;
34 
35 	for (unsigned i = 0; i < ARRAY_SIZE(t->hash_seeds); i++) {
36 		struct bucket_hashed *h = bucket_hash(t, i, dev_bucket);
37 
38 		if (h->dev_bucket == dev_bucket) {
39 			ret = h->journal_seq;
40 			break;
41 		}
42 	}
43 
44 	mutex_unlock(&b->lock);
45 
46 	return ret;
47 }
48 
bucket_table_insert(struct buckets_waiting_for_journal_table * t,struct bucket_hashed * new,u64 flushed_seq)49 static bool bucket_table_insert(struct buckets_waiting_for_journal_table *t,
50 				struct bucket_hashed *new,
51 				u64 flushed_seq)
52 {
53 	struct bucket_hashed *last_evicted = NULL;
54 	unsigned tries, i;
55 
56 	for (tries = 0; tries < 10; tries++) {
57 		struct bucket_hashed *old, *victim = NULL;
58 
59 		for (i = 0; i < ARRAY_SIZE(t->hash_seeds); i++) {
60 			old = bucket_hash(t, i, new->dev_bucket);
61 
62 			if (old->dev_bucket == new->dev_bucket ||
63 			    old->journal_seq <= flushed_seq) {
64 				*old = *new;
65 				return true;
66 			}
67 
68 			if (last_evicted != old)
69 				victim = old;
70 		}
71 
72 		/* hashed to same slot 3 times: */
73 		if (!victim)
74 			break;
75 
76 		/* Failed to find an empty slot: */
77 		swap(*new, *victim);
78 		last_evicted = victim;
79 	}
80 
81 	return false;
82 }
83 
bch2_set_bucket_needs_journal_commit(struct buckets_waiting_for_journal * b,u64 flushed_seq,unsigned dev,u64 bucket,u64 journal_seq)84 int bch2_set_bucket_needs_journal_commit(struct buckets_waiting_for_journal *b,
85 					 u64 flushed_seq,
86 					 unsigned dev, u64 bucket,
87 					 u64 journal_seq)
88 {
89 	struct buckets_waiting_for_journal_table *t, *n;
90 	struct bucket_hashed tmp, new = {
91 		.dev_bucket	= (u64) dev << 56 | bucket,
92 		.journal_seq	= journal_seq,
93 	};
94 	size_t i, size, new_bits, nr_elements = 1, nr_rehashes = 0, nr_rehashes_this_size = 0;
95 	int ret = 0;
96 
97 	mutex_lock(&b->lock);
98 
99 	if (likely(bucket_table_insert(b->t, &new, flushed_seq)))
100 		goto out;
101 
102 	t = b->t;
103 	size = 1UL << t->bits;
104 	for (i = 0; i < size; i++)
105 		nr_elements += t->d[i].journal_seq > flushed_seq;
106 
107 	new_bits = ilog2(roundup_pow_of_two(nr_elements * 3));
108 realloc:
109 	n = kvmalloc(sizeof(*n) + (sizeof(n->d[0]) << new_bits), GFP_KERNEL);
110 	if (!n) {
111 		ret = -BCH_ERR_ENOMEM_buckets_waiting_for_journal_set;
112 		goto out;
113 	}
114 
115 retry_rehash:
116 	if (nr_rehashes_this_size == 3) {
117 		new_bits++;
118 		nr_rehashes_this_size = 0;
119 		kvfree(n);
120 		goto realloc;
121 	}
122 
123 	nr_rehashes++;
124 	nr_rehashes_this_size++;
125 
126 	bucket_table_init(n, new_bits);
127 
128 	tmp = new;
129 	BUG_ON(!bucket_table_insert(n, &tmp, flushed_seq));
130 
131 	for (i = 0; i < 1UL << t->bits; i++) {
132 		if (t->d[i].journal_seq <= flushed_seq)
133 			continue;
134 
135 		tmp = t->d[i];
136 		if (!bucket_table_insert(n, &tmp, flushed_seq))
137 			goto retry_rehash;
138 	}
139 
140 	b->t = n;
141 	kvfree(t);
142 
143 	pr_debug("took %zu rehashes, table at %zu/%lu elements",
144 		 nr_rehashes, nr_elements, 1UL << b->t->bits);
145 out:
146 	mutex_unlock(&b->lock);
147 
148 	return ret;
149 }
150 
bch2_fs_buckets_waiting_for_journal_exit(struct bch_fs * c)151 void bch2_fs_buckets_waiting_for_journal_exit(struct bch_fs *c)
152 {
153 	struct buckets_waiting_for_journal *b = &c->buckets_waiting_for_journal;
154 
155 	kvfree(b->t);
156 }
157 
158 #define INITIAL_TABLE_BITS		3
159 
bch2_fs_buckets_waiting_for_journal_init(struct bch_fs * c)160 int bch2_fs_buckets_waiting_for_journal_init(struct bch_fs *c)
161 {
162 	struct buckets_waiting_for_journal *b = &c->buckets_waiting_for_journal;
163 
164 	mutex_init(&b->lock);
165 
166 	b->t = kvmalloc(sizeof(*b->t) +
167 			(sizeof(b->t->d[0]) << INITIAL_TABLE_BITS), GFP_KERNEL);
168 	if (!b->t)
169 		return -BCH_ERR_ENOMEM_buckets_waiting_for_journal_init;
170 
171 	bucket_table_init(b->t, INITIAL_TABLE_BITS);
172 	return 0;
173 }
174