Searched refs:mas_next (Results 1 – 6 of 6) sorted by relevance
| /linux/lib/ |
| H A D | test_maple_tree.c | 1371 entry = mas_next(&mas, limit); in check_next_entry() 1570 ptr = mas_next(&mas, ULONG_MAX); in check_root_expand() 1588 ptr = mas_next(&mas, ULONG_MAX); in check_root_expand() 1693 mas_next(&mas, ULONG_MAX); in check_gap_combining() 1694 entry = mas_next(&mas, ULONG_MAX); in check_gap_combining() 1724 entry = mas_next(&mas, ULONG_MAX); in check_gap_combining() 1726 mas_next(&mas, ULONG_MAX); /* go to the next entry. */ in check_gap_combining() 2222 mas_next(&mas, 1000); in next_prev_test() 2252 val = mas_next(&mas, 1000); in next_prev_test() 2262 val = mas_next(&mas, 1000); in next_prev_test() [all …]
|
| H A D | maple_tree.c | 5073 void *mas_next(struct ma_state *mas, unsigned long max) in mas_next() function 5083 EXPORT_SYMBOL_GPL(mas_next); 5126 entry = mas_next(&mas, max); in mt_next()
|
| /linux/Documentation/core-api/ |
| H A D | maple_tree.rst | 173 Using a maple state allows mas_next() and mas_prev() to function as if the 175 performance penalty is outweighed by cache optimization. mas_next() will
|
| /linux/include/linux/ |
| H A D | maple_tree.h | 576 void *mas_next(struct ma_state *mas, unsigned long max);
|
| /linux/mm/ |
| H A D | vma.c | 1513 vma_test = mas_next(&test, vms->vma_count - 1); in vms_gather_munmap_vmas()
|
| /linux/tools/testing/radix-tree/ |
| H A D | maple.c | 358 entry = mas_next(&tmp, mas_end->last); in mas_ce2_over_count() 367 entry = mas_next(&tmp, mas_end->last); in mas_ce2_over_count() 34838 MT_BUG_ON(mt, mas_next(&mas_reader, ULONG_MAX) != xa_mk_value(val)); in check_rcu_simulated() 34853 mas_next(&mas_reader, ULONG_MAX); in check_rcu_simulated() 35949 mas_next(&mas, ULONG_MAX); in check_erase_rebalance() 35955 mas_next(&mas, ULONG_MAX); in check_erase_rebalance() 35961 mas_next(&mas, ULONG_MAX); in check_erase_rebalance() 36015 mas_next(&mas, ULONG_MAX); in check_erase_rebalance()
|