| /linux/tools/include/linux/ |
| H A D | rbtree.h | 75 #define rb_entry_safe(ptr, type, member) \ macro 98 for (pos = rb_entry_safe(rb_first_postorder(root), typeof(*pos), field); \ 99 pos && ({ n = rb_entry_safe(rb_next_postorder(&pos->field), \
|
| /linux/include/linux/ |
| H A D | rbtree.h | 105 #define rb_entry_safe(ptr, type, member) \ macro 128 for (pos = rb_entry_safe(rb_first_postorder(root), typeof(*pos), field); \ 129 pos && ({ n = rb_entry_safe(rb_next_postorder(&pos->field), \
|
| H A D | timerqueue.h | 27 return rb_entry_safe(leftmost, struct timerqueue_node, node); in timerqueue_getnext()
|
| H A D | skbuff.h | 4126 #define rb_to_skb(rb) rb_entry_safe(rb, struct sk_buff, rbnode)
|
| /linux/fs/btrfs/ |
| H A D | ref-verify.c | 104 return rb_entry_safe(node, struct block_entry, node); in insert_block_entry() 112 return rb_entry_safe(node, struct block_entry, node); in lookup_block_entry() 141 return rb_entry_safe(node, struct root_entry, node); in insert_root_entry() 179 return rb_entry_safe(node, struct ref_entry, node); in insert_ref_entry() 187 return rb_entry_safe(node, struct root_entry, node); in lookup_root_entry()
|
| H A D | ulist.c | 150 return rb_entry_safe(node, struct ulist_node, rb_node); in ulist_rbtree_search()
|
| H A D | delayed-inode.c | 396 return rb_entry_safe(node, struct btrfs_delayed_item, rb_node); in __btrfs_lookup_delayed_item() 483 return rb_entry_safe(p, struct btrfs_delayed_item, rb_node); in __btrfs_first_delayed_insertion_item() 491 return rb_entry_safe(p, struct btrfs_delayed_item, rb_node); in __btrfs_first_delayed_deletion_item() 499 return rb_entry_safe(p, struct btrfs_delayed_item, rb_node); in __btrfs_next_delayed_item()
|
| H A D | extent_map.c | 364 merge = rb_entry_safe(rb, struct extent_map, rb_node); in try_merge_map() 382 merge = rb_entry_safe(rb, struct extent_map, rb_node); in try_merge_map()
|
| H A D | extent-io-tree.c | 208 return rb_entry_safe(next, struct extent_state, rb_node); in next_state() 215 return rb_entry_safe(next, struct extent_state, rb_node); in prev_state()
|
| H A D | qgroup.c | 183 return rb_entry_safe(node, struct btrfs_qgroup, node); in find_qgroup_rb()
|
| /linux/fs/f2fs/ |
| H A D | extent_cache.c | 263 *next_entry = rb_entry_safe(tmp_node, struct extent_node, rb_node); in __lookup_extent_node_ret() 268 *prev_entry = rb_entry_safe(tmp_node, struct extent_node, rb_node); in __lookup_extent_node_ret() 275 *prev_entry = rb_entry_safe(tmp_node, in __lookup_extent_node_ret() 281 *next_entry = rb_entry_safe(tmp_node, in __lookup_extent_node_ret() 761 next_en = rb_entry_safe(node, struct extent_node, in __update_extent_tree_range()
|
| H A D | segment.c | 1083 *next_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node); in __lookup_discard_cmd_ret() 1088 *prev_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node); in __lookup_discard_cmd_ret() 1094 *prev_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node); in __lookup_discard_cmd_ret() 1098 *next_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node); in __lookup_discard_cmd_ret() 1558 next_dc = rb_entry_safe(node, struct discard_cmd, rb_node); in __update_discard_tree_range() 1634 dc = rb_entry_safe(node, struct discard_cmd, rb_node); in __issue_discard_cmd_orderly() 3477 dc = rb_entry_safe(node, struct discard_cmd, rb_node); in __issue_discard_cmd_range()
|
| H A D | gc.c | 581 ve = rb_entry_safe(node, struct victim_entry, rb_node); in atgc_lookup_victim()
|
| /linux/fs/proc/ |
| H A D | generic.c | 58 return rb_entry_safe(rb_first(&dir->subdir), struct proc_dir_entry, in pde_subdir_first() 64 return rb_entry_safe(rb_next(&dir->subdir_node), struct proc_dir_entry, in pde_subdir_next()
|
| /linux/mm/ |
| H A D | vmalloc.c | 1054 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size() 4780 return rb_entry_safe(n, struct vmap_area, rb_node); in node_to_va()
|
| /linux/kernel/events/ |
| H A D | core.c | 1865 for (event = rb_entry_safe(rb_first(&((groups)->tree)), \ 1867 event = rb_entry_safe(rb_next(&event->group_node), \
|