| /linux/fs/hpfs/ |
| H A D | anode.c | 15 struct bplus_header *btree, unsigned sec, in hpfs_bplus_lookup() argument 24 if (bp_internal(btree)) { in hpfs_bplus_lookup() 25 for (i = 0; i < btree->n_used_nodes; i++) in hpfs_bplus_lookup() 26 if (le32_to_cpu(btree->u.internal[i].file_secno) > sec) { in hpfs_bplus_lookup() 27 a = le32_to_cpu(btree->u.internal[i].down); in hpfs_bplus_lookup() 30 btree = GET_BTREE_PTR(&anode->btree); in hpfs_bplus_lookup() 37 for (i = 0; i < btree->n_used_nodes; i++) in hpfs_bplus_lookup() 38 if (le32_to_cpu(btree->u.external[i].file_secno) <= sec && in hpfs_bplus_lookup() 39 … le32_to_cpu(btree->u.external[i].file_secno) + le32_to_cpu(btree->u.external[i].length) > sec) { in hpfs_bplus_lookup() 40 …a = le32_to_cpu(btree->u.external[i].disk_secno) + sec - le32_to_cpu(btree->u.external[i].file_sec… in hpfs_bplus_lookup() [all …]
|
| H A D | map.c | 180 if ((unsigned)fnode->btree.n_used_nodes + (unsigned)fnode->btree.n_free_nodes != in hpfs_map_fnode() 181 (bp_internal(GET_BTREE_PTR(&fnode->btree)) ? 12 : 8)) { in hpfs_map_fnode() 187 if (le16_to_cpu(fnode->btree.first_free) != in hpfs_map_fnode() 188 8 + fnode->btree.n_used_nodes * (bp_internal(GET_BTREE_PTR(&fnode->btree)) ? 8 : 12)) { in hpfs_map_fnode() 235 if ((unsigned)anode->btree.n_used_nodes + (unsigned)anode->btree.n_free_nodes != in hpfs_map_anode() 236 (bp_internal(GET_BTREE_PTR(&anode->btree)) ? 60 : 40)) { in hpfs_map_anode() 240 if (le16_to_cpu(anode->btree.first_free) != in hpfs_map_anode() 241 8 + anode->btree.n_used_nodes * (bp_internal(GET_BTREE_PTR(&anode->btree)) ? 8 : 12)) { in hpfs_map_anode()
|
| H A D | alloc.c | 466 f->btree.n_free_nodes = 8; in hpfs_alloc_fnode() 467 f->btree.first_free = cpu_to_le16(8); in hpfs_alloc_fnode() 483 a->btree.n_free_nodes = 40; in hpfs_alloc_anode() 484 a->btree.n_used_nodes = 0; in hpfs_alloc_anode() 485 a->btree.first_free = cpu_to_le16(8); in hpfs_alloc_anode()
|
| H A D | hpfs.h | 474 struct bplus_header_fixed btree; /* b+ tree, 8 extents or 12 subtrees */ member 516 struct bplus_header_fixed btree; /* b+tree, 40 extents or 60 subtrees */ member
|
| H A D | namei.c | 86 fnode->btree.n_free_nodes = 7; in hpfs_mkdir() 87 fnode->btree.n_used_nodes = 1; in hpfs_mkdir() 88 fnode->btree.first_free = cpu_to_le16(0x14); in hpfs_mkdir()
|
| /linux/fs/nilfs2/ |
| H A D | btree.c | 58 static int nilfs_btree_get_new_block(const struct nilfs_bmap *btree, in nilfs_btree_get_new_block() argument 61 struct inode *btnc_inode = NILFS_BMAP_I(btree)->i_assoc_inode; in nilfs_btree_get_new_block() 112 static int nilfs_btree_node_size(const struct nilfs_bmap *btree) in nilfs_btree_node_size() argument 114 return i_blocksize(btree->b_inode); in nilfs_btree_node_size() 117 static int nilfs_btree_nchildren_per_block(const struct nilfs_bmap *btree) in nilfs_btree_nchildren_per_block() argument 119 return btree->b_nchildren_per_block; in nilfs_btree_nchildren_per_block() 411 nilfs_btree_get_root(const struct nilfs_bmap *btree) in nilfs_btree_get_root() argument 413 return (struct nilfs_btree_node *)btree->b_u.u_data; in nilfs_btree_get_root() 428 static int nilfs_btree_height(const struct nilfs_bmap *btree) in nilfs_btree_height() argument 430 return nilfs_btree_node_get_level(nilfs_btree_get_root(btree)) + 1; in nilfs_btree_height() [all …]
|
| H A D | Makefile | 4 btnode.o bmap.o btree.o direct.o dat.o recovery.o \
|
| /linux/drivers/md/bcache/ |
| H A D | btree.h | 117 struct btree { struct 127 struct btree *parent; argument 152 static inline bool btree_node_ ## flag(struct btree *b) \ argument 155 static inline void set_btree_node_ ## flag(struct btree *b) \ 170 static inline struct btree_write *btree_current_write(struct btree *b) in btree_current_write() 175 static inline struct btree_write *btree_prev_write(struct btree *b) in btree_prev_write() 180 static inline struct bset *btree_bset_first(struct btree *b) in btree_bset_first() 185 static inline struct bset *btree_bset_last(struct btree *b) in btree_bset_last() 190 static inline unsigned int bset_block_offset(struct btree *b, struct bset *i) in bset_block_offset() 248 static inline void rw_lock(bool w, struct btree *b, int level) in rw_lock() [all …]
|
| H A D | btree.c | 107 static inline struct bset *write_block(struct btree *b) in write_block() 112 static void bch_btree_init_next(struct btree *b) in bch_btree_init_next() 139 static uint64_t btree_csum_set(struct btree *b, struct bset *i) in btree_csum_set() 148 void bch_btree_node_read_done(struct btree *b) in bch_btree_node_read_done() 243 static void bch_btree_node_read(struct btree *b) in bch_btree_node_read() 281 static void btree_complete_write(struct btree *b, struct btree_write *w) in btree_complete_write() 298 closure_type(b, struct btree, io); in CLOSURE_CALLBACK() 305 closure_type(b, struct btree, io); in CLOSURE_CALLBACK() 320 closure_type(b, struct btree, io); in CLOSURE_CALLBACK() 329 struct btree *b = container_of(cl, struct btree, io); in btree_node_write_endio() [all …]
|
| H A D | extents.c | 128 struct btree *b = container_of(keys, struct btree, keys); in bch_bkey_dump() 168 struct btree *b = container_of(bk, struct btree, keys); in bch_btree_ptr_invalid() 173 static bool btree_ptr_bad_expensive(struct btree *b, const struct bkey *k) in btree_ptr_bad_expensive() 207 struct btree *b = container_of(bk, struct btree, keys); in bch_btree_ptr_bad() 232 struct btree *b = container_of(bk, struct btree, keys); in bch_btree_ptr_insert_fixup() 328 struct cache_set *c = container_of(b, struct btree, keys)->c; in bch_extent_insert_fixup() 502 struct btree *b = container_of(bk, struct btree, keys); in bch_extent_invalid() 507 static bool bch_extent_bad_expensive(struct btree *b, const struct bkey *k, in bch_extent_bad_expensive() 539 struct btree *b = container_of(bk, struct btree, keys); in bch_extent_bad() 585 struct btree *b = container_of(bk, struct btree, keys); in bch_extent_merge()
|
| H A D | debug.h | 11 void bch_btree_verify(struct btree *b); 20 static inline void bch_btree_verify(struct btree *b) {} in bch_btree_verify()
|
| H A D | Makefile | 5 bcache-y := alloc.o bset.o btree.o debug.o extents.o io.o\
|
| H A D | bcache.h | 223 struct btree; 280 int (*cache_miss)(struct btree *b, struct search *s, 671 struct btree *root; 674 struct btree *verify_data;
|
| H A D | debug.c | 31 void bch_btree_verify(struct btree *b) in bch_btree_verify() 33 struct btree *v = b->c->verify_data; in bch_btree_verify()
|
| H A D | Kconfig | 10 a btree for indexing and the layout is optimized for SSDs.
|
| H A D | sysfs.c | 621 static int bch_btree_bset_stats(struct btree_op *b_op, struct btree *b) in bch_btree_bset_stats() 661 struct btree *b; in bch_root_usage() 684 struct btree *b; in bch_cache_size()
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | persistent-data.rst | 14 - Another btree-based caching target posted to dm-devel 72 dm-btree.[hc] 73 dm-btree-remove.c 74 dm-btree-spine.c 75 dm-btree-internal.h 77 Currently there is only one data structure, a hierarchical btree. 81 The btree is 'hierarchical' in that you can define it to be composed 83 thin-provisioning target uses a btree with two levels of nesting.
|
| /linux/drivers/md/persistent-data/ |
| H A D | Makefile | 11 dm-btree.o \ 12 dm-btree-remove.o \ 13 dm-btree-spine.o
|
| /linux/include/trace/events/ |
| H A D | bcache.h | 64 TP_PROTO(struct btree *b), 258 TP_PROTO(struct btree *b), 263 TP_PROTO(struct btree *b), 283 TP_PROTO(struct btree *b), 293 TP_PROTO(struct btree *b), 333 TP_PROTO(struct btree *b, struct bkey *k, unsigned op, unsigned status), 366 TP_PROTO(struct btree *b, unsigned keys), 383 TP_PROTO(struct btree *b, unsigned keys), 388 TP_PROTO(struct btree *b, unsigned keys), 393 TP_PROTO(struct btree *b),
|
| /linux/fs/befs/ |
| H A D | ChangeLog | 27 * Did the string comparison really right this time (btree.c) [WD] 30 a pointer value. (btree.c) [WD] 38 keys within btree nodes, rather than the linear search we were using 39 before. (btree.c) [Sergey S. Kostyliov <rathamahata@php4.ru>] 56 (btree.c) [WD] 105 * Removed notion of btree handle from btree.c. It was unnecessary, as the 128 (btree.c) [WD] 133 seekleaf() in btree.c [WD] 148 (datastream.c, btree.c super.c inode.c) [WD] 253 * Fix bug with reading an empty directory. (btree.c and dir.c) [all …]
|
| H A D | Makefile | 8 befs-objs := datastream.o btree.o super.o inode.o debug.o io.o linuxvfs.o
|
| /linux/fs/xfs/scrub/ |
| H A D | dabtree.c | 110 return hdr.btree + blk->index; in xchk_da_btree_node_entry() 337 struct xfs_da_node_entry *btree; in xchk_da_btree_block() local 441 btree = nodehdr.btree; in xchk_da_btree_block() 443 blk->hashval = be32_to_cpu(btree[*pmaxrecs - 1].hashval); in xchk_da_btree_block()
|
| /linux/Documentation/filesystems/xfs/ |
| H A D | xfs-self-describing-metadata.rst | 53 pointers in a btree end up with loops in them) are the key to understanding what 108 determine the scope of the corruption. For example, if we have a extent btree 118 freespace btree blocks are owned by an allocation group. Hence the size and 121 freespace btree block written to the wrong AG). 133 when the free space btree block that contains the block was last written 211 - short btree blocks have a 32 bit owner (ag number) and a 32 bit block
|
| /linux/fs/xfs/libxfs/ |
| H A D | xfs_ag.c | 55 uint64_t btree = 0; in xfs_initialize_perag_data() local 77 btree += pag->pagf_btreeblks; in xfs_initialize_perag_data() 80 fdblocks = bfree + bfreelst + btree; in xfs_initialize_perag_data()
|
| /linux/Documentation/filesystems/ |
| H A D | hfs.rst | 87 in btree routines derived from Brad Boyer's hfsplus driver.
|