1 /* SPDX-License-Identifier: GPL-2.0 */ 2 #ifndef _BCACHEFS_BTREE_CACHE_H 3 #define _BCACHEFS_BTREE_CACHE_H 4 5 #include "bcachefs.h" 6 #include "btree_types.h" 7 #include "bkey_methods.h" 8 9 extern const char * const bch2_btree_node_flags[]; 10 11 struct btree_iter; 12 13 void bch2_recalc_btree_reserve(struct bch_fs *); 14 15 void bch2_btree_node_to_freelist(struct bch_fs *, struct btree *); 16 17 void __bch2_btree_node_hash_remove(struct btree_cache *, struct btree *); 18 void bch2_btree_node_hash_remove(struct btree_cache *, struct btree *); 19 20 int __bch2_btree_node_hash_insert(struct btree_cache *, struct btree *); 21 int bch2_btree_node_hash_insert(struct btree_cache *, struct btree *, 22 unsigned, enum btree_id); 23 24 void bch2_node_pin(struct bch_fs *, struct btree *); 25 void bch2_btree_cache_unpin(struct bch_fs *); 26 27 void bch2_btree_node_update_key_early(struct btree_trans *, enum btree_id, unsigned, 28 struct bkey_s_c, struct bkey_i *); 29 30 void bch2_btree_cache_cannibalize_unlock(struct btree_trans *); 31 int bch2_btree_cache_cannibalize_lock(struct btree_trans *, struct closure *); 32 33 struct btree *__bch2_btree_node_mem_alloc(struct bch_fs *); 34 struct btree *bch2_btree_node_mem_alloc(struct btree_trans *, bool); 35 36 struct btree *bch2_btree_node_get(struct btree_trans *, struct btree_path *, 37 const struct bkey_i *, unsigned, 38 enum six_lock_type, unsigned long); 39 40 struct btree *bch2_btree_node_get_noiter(struct btree_trans *, const struct bkey_i *, 41 enum btree_id, unsigned, bool); 42 43 int bch2_btree_node_prefetch(struct btree_trans *, struct btree_path *, 44 const struct bkey_i *, enum btree_id, unsigned); 45 46 void bch2_btree_node_evict(struct btree_trans *, const struct bkey_i *); 47 48 void bch2_fs_btree_cache_exit(struct bch_fs *); 49 int bch2_fs_btree_cache_init(struct bch_fs *); 50 void bch2_fs_btree_cache_init_early(struct btree_cache *); 51 52 static inline u64 btree_ptr_hash_val(const struct bkey_i *k) 53 { 54 switch (k->k.type) { 55 case KEY_TYPE_btree_ptr: 56 return *((u64 *) bkey_i_to_btree_ptr_c(k)->v.start); 57 case KEY_TYPE_btree_ptr_v2: 58 /* 59 * The cast/deref is only necessary to avoid sparse endianness 60 * warnings: 61 */ 62 return *((u64 *) &bkey_i_to_btree_ptr_v2_c(k)->v.seq); 63 default: 64 return 0; 65 } 66 } 67 68 static inline struct btree *btree_node_mem_ptr(const struct bkey_i *k) 69 { 70 return k->k.type == KEY_TYPE_btree_ptr_v2 71 ? (void *)(unsigned long)bkey_i_to_btree_ptr_v2_c(k)->v.mem_ptr 72 : NULL; 73 } 74 75 /* is btree node in hash table? */ 76 static inline bool btree_node_hashed(struct btree *b) 77 { 78 return b->hash_val != 0; 79 } 80 81 #define for_each_cached_btree(_b, _c, _tbl, _iter, _pos) \ 82 for ((_tbl) = rht_dereference_rcu((_c)->btree_cache.table.tbl, \ 83 &(_c)->btree_cache.table), \ 84 _iter = 0; _iter < (_tbl)->size; _iter++) \ 85 rht_for_each_entry_rcu((_b), (_pos), _tbl, _iter, hash) 86 87 static inline size_t btree_buf_bytes(const struct btree *b) 88 { 89 return 1UL << b->byte_order; 90 } 91 92 static inline size_t btree_buf_max_u64s(const struct btree *b) 93 { 94 return (btree_buf_bytes(b) - sizeof(struct btree_node)) / sizeof(u64); 95 } 96 97 static inline size_t btree_max_u64s(const struct bch_fs *c) 98 { 99 return (c->opts.btree_node_size - sizeof(struct btree_node)) / sizeof(u64); 100 } 101 102 static inline size_t btree_sectors(const struct bch_fs *c) 103 { 104 return c->opts.btree_node_size >> SECTOR_SHIFT; 105 } 106 107 static inline unsigned btree_blocks(const struct bch_fs *c) 108 { 109 return btree_sectors(c) >> c->block_bits; 110 } 111 112 #define BTREE_SPLIT_THRESHOLD(c) (btree_max_u64s(c) * 2 / 3) 113 114 #define BTREE_FOREGROUND_MERGE_THRESHOLD(c) (btree_max_u64s(c) * 1 / 3) 115 #define BTREE_FOREGROUND_MERGE_HYSTERESIS(c) \ 116 (BTREE_FOREGROUND_MERGE_THRESHOLD(c) + \ 117 (BTREE_FOREGROUND_MERGE_THRESHOLD(c) >> 2)) 118 119 static inline unsigned btree_id_nr_alive(struct bch_fs *c) 120 { 121 return BTREE_ID_NR + c->btree_roots_extra.nr; 122 } 123 124 static inline struct btree_root *bch2_btree_id_root(struct bch_fs *c, unsigned id) 125 { 126 if (likely(id < BTREE_ID_NR)) { 127 return &c->btree_roots_known[id]; 128 } else { 129 unsigned idx = id - BTREE_ID_NR; 130 131 EBUG_ON(idx >= c->btree_roots_extra.nr); 132 return &c->btree_roots_extra.data[idx]; 133 } 134 } 135 136 static inline struct btree *btree_node_root(struct bch_fs *c, struct btree *b) 137 { 138 return bch2_btree_id_root(c, b->c.btree_id)->b; 139 } 140 141 const char *bch2_btree_id_str(enum btree_id); 142 void bch2_btree_id_to_text(struct printbuf *, enum btree_id); 143 144 void bch2_btree_pos_to_text(struct printbuf *, struct bch_fs *, const struct btree *); 145 void bch2_btree_node_to_text(struct printbuf *, struct bch_fs *, const struct btree *); 146 void bch2_btree_cache_to_text(struct printbuf *, const struct btree_cache *); 147 148 #endif /* _BCACHEFS_BTREE_CACHE_H */ 149