| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | linked_list.c | 12 struct bpf_spin_lock lock; 21 private(C) struct bpf_spin_lock glock_c; 28 int list_push_pop(struct bpf_spin_lock *lock, struct bpf_list_head *head, bool leave_in_map) in list_push_pop() 37 bpf_spin_lock(lock); in list_push_pop() 46 bpf_spin_lock(lock); in list_push_pop() 56 bpf_spin_lock(lock); in list_push_pop() 62 bpf_spin_lock(lock); in list_push_pop() 73 bpf_spin_lock(lock); in list_push_pop() 77 bpf_spin_lock(lock); in list_push_pop() 89 bpf_spin_lock(lock); in list_push_pop() [all …]
|
| H A D | verifier_spin_lock.c | 10 struct bpf_spin_lock l; 47 __imm(bpf_spin_lock), in spin_lock_test1_success() 80 __imm(bpf_spin_lock), in lock_test2_direct_ld_st() 114 __imm(bpf_spin_lock), in __flag() 148 __imm(bpf_spin_lock), in __flag() 182 __imm(bpf_spin_lock), in call_within_a_locked_region() 216 __imm(bpf_spin_lock), in spin_lock_test6_missing_unlock() 250 __imm(bpf_spin_lock), in lock_test7_unlock_without_lock() 286 __imm(bpf_spin_lock), in spin_lock_test8_double_lock() 325 __imm(bpf_spin_lock), in spin_lock_test9_different_lock() [all …]
|
| H A D | test_spin_lock.c | 10 struct bpf_spin_lock lock; 22 struct bpf_spin_lock lock; 33 struct bpf_spin_lock lock; 70 bpf_spin_lock(&val->lock); in bpf_spin_lock_test() 84 bpf_spin_lock(&q->lock); in bpf_spin_lock_test() 97 bpf_spin_lock(&cls->lock); in bpf_spin_lock_test() 105 struct bpf_spin_lock lockA __hidden SEC(".data.A"); 123 bpf_spin_lock(&lockA); in static_subprog_lock() 142 bpf_spin_lock(&lockA); in lock_static_subprog_call() 164 bpf_spin_lock(&lockA); in lock_static_subprog_unlock()
|
| H A D | rbtree_fail.c | 16 private(A) struct bpf_spin_lock glock; 32 __failure __msg("bpf_spin_lock at off=16 must be held for bpf_rb_root") 46 __failure __msg("bpf_spin_lock at off=16 must be held for bpf_rb_root") 55 bpf_spin_lock(&glock); in rbtree_api_nolock_remove() 64 __failure __msg("bpf_spin_lock at off=16 must be held for bpf_rb_root") 88 bpf_spin_lock(&glock); in rbtree_api_remove_unadded_node() 115 bpf_spin_lock(&glock); in rbtree_api_remove_no_drop() 146 bpf_spin_lock(&glock); in rbtree_api_add_to_multiple_trees() 161 bpf_spin_lock(&glock); in rbtree_api_use_unchecked_remove_retval() 170 bpf_spin_lock( in rbtree_api_use_unchecked_remove_retval() [all...] |
| H A D | refcounted_kptr.c | 41 private(A) struct bpf_spin_lock lock; 45 private(B) struct bpf_spin_lock alock; 48 private(C) struct bpf_spin_lock block; 75 struct bpf_spin_lock *lock) in __insert_in_tree_and_list() 87 bpf_spin_lock(lock); in __insert_in_tree_and_list() 96 bpf_spin_lock(lock); in __insert_in_tree_and_list() 107 struct bpf_spin_lock *lock) in __stash_map_insert_tree() 130 bpf_spin_lock(lock); in __stash_map_insert_tree() 141 struct bpf_spin_lock *lock, in __read_from_tree() 148 bpf_spin_lock(loc in __read_from_tree() [all...] |
| H A D | linked_list_fail.c | 100 bpf_spin_lock(lexpr); \ 137 bpf_spin_lock(lexpr); \ 363 bpf_spin_lock(&glock); in use_after_unlock() 394 bpf_spin_lock(&glock); in list_double_add() 427 bpf_spin_lock(&glock); in no_node_value_type() 442 bpf_spin_lock(&glock); in incorrect_value_type() 457 bpf_spin_lock(&glock); in incorrect_node_var_off() 472 bpf_spin_lock(&glock); in incorrect_node_off1() 487 bpf_spin_lock(&glock); in incorrect_node_off2() 502 bpf_spin_lock(&glock); in no_head_type() [all …]
|
| H A D | rbtree.c | 17 struct bpf_spin_lock glock; 30 private(A) struct bpf_spin_lock glock; 48 static long __add_three(struct bpf_rb_root *root, struct bpf_spin_lock *lock) in __add_three() 64 bpf_spin_lock(lock); in __add_three() 74 bpf_spin_lock(lock); in __add_three() 108 bpf_spin_lock(&glock); in rbtree_add_and_remove() 148 bpf_spin_lock(&glock); in rbtree_add_and_remove_array() 213 bpf_spin_lock(&glock); in rbtree_first_and_remove() 237 bpf_spin_lock(&glock); in rbtree_first_and_remove() 269 bpf_spin_lock(&glock); in rbtree_api_release_aliasing() [all …]
|
| H A D | linked_list.h | 17 struct bpf_spin_lock lock; 23 struct bpf_spin_lock lock; 52 private(A) struct bpf_spin_lock glock; 54 private(B) struct bpf_spin_lock glock2;
|
| H A D | test_map_lock.c | 10 struct bpf_spin_lock lock; 22 struct bpf_spin_lock lock; 45 bpf_spin_lock(&val->lock); in bpf_map_lock_test() 54 bpf_spin_lock(&q->lock); in bpf_map_lock_test()
|
| H A D | verifier_helper_restricted.c | 10 struct bpf_spin_lock l; 205 __imm(bpf_spin_lock), in in_bpf_prog_type_kprobe_3() 228 __imm(bpf_spin_lock), in in_bpf_prog_type_tracepoint_3() 251 __imm(bpf_spin_lock), in bpf_prog_type_perf_event_3() 274 __imm(bpf_spin_lock), in bpf_prog_type_raw_tracepoint_3()
|
| H A D | refcounted_kptr_fail.c | 20 private(A) struct bpf_spin_lock glock; 44 bpf_spin_lock(&glock); in rbtree_refcounted_node_ref_escapes() 89 bpf_spin_lock(&glock); in rbtree_refcounted_node_ref_escapes_owning_input() 110 bpf_spin_lock(&glock); in BPF_PROG()
|
| H A D | tracing_failure.c | 11 int BPF_PROG(test_spin_lock, struct bpf_spin_lock *lock) in BPF_PROG() 17 int BPF_PROG(test_spin_unlock, struct bpf_spin_lock *lock) in BPF_PROG()
|
| H A D | freplace_attach_probe.c | 12 struct bpf_spin_lock lock; 33 bpf_spin_lock(&val->lock); in new_handle_kprobe()
|
| H A D | rbtree_btf_fail__wrong_node_type.c | 20 private(A) struct bpf_spin_lock glock; 32 bpf_spin_lock(&glock); in rbtree_api_add__wrong_node_type()
|
| H A D | normal_map_btf.c | 17 struct bpf_spin_lock lock; 50 bpf_spin_lock(&value->lock); in add_to_list_in_array()
|
| H A D | exceptions_fail.c | 29 private(A) struct bpf_spin_lock lock; 111 bpf_spin_lock(&lock); in subprog_lock() 121 bpf_spin_lock(&lock); in reject_with_lock() 172 bpf_spin_lock(&lock); in reject_with_rbtree_add_throw()
|
| H A D | rbtree_btf_fail__add_wrong_type.c | 34 private(A) struct bpf_spin_lock glock; 46 bpf_spin_lock(&glock); in rbtree_api_add__add_wrong_type()
|
| H A D | map_in_map_btf.c | 17 struct bpf_spin_lock lock; 67 bpf_spin_lock(&value->lock); in add_to_list_in_inner_array()
|
| H A D | test_helper_restricted.c | 11 struct bpf_spin_lock l; 54 bpf_spin_lock(&lock->l); in spin_lock_work()
|
| H A D | local_kptr_stash.c | 27 struct bpf_spin_lock l; 45 struct bpf_spin_lock l; 175 bpf_spin_lock(&res->l); in stash_local_with_root()
|
| H A D | htab_reuse.c | 9 struct bpf_spin_lock lock;
|
| H A D | timer_crash.c | 9 struct bpf_spin_lock lock;
|
| /linux/tools/sched_ext/ |
| H A D | scx_flatcg.bpf.c | 107 private(CGV_TREE) struct bpf_spin_lock cgv_tree_lock; 232 bpf_spin_lock(&cgv_tree_lock); in cgrp_refresh_hweight() 300 bpf_spin_lock(&cgv_tree_lock); in cgrp_enqueued() 474 bpf_spin_lock(&cgv_tree_lock); in update_active_weight_sums() 600 bpf_spin_lock(&cgv_tree_lock); in BPF_STRUCT_OPS() 617 bpf_spin_lock(&cgv_tree_lock); in try_pick_next_cgroup() 683 bpf_spin_lock(&cgv_tree_lock); in try_pick_next_cgroup() 708 bpf_spin_lock(&cgv_tree_lock); in try_pick_next_cgroup() 769 bpf_spin_lock(&cgv_tree_lock); in BPF_STRUCT_OPS()
|
| /linux/samples/bpf/ |
| H A D | hbm.h | 12 struct bpf_spin_lock lock;
|
| /linux/Documentation/bpf/ |
| H A D | graph_ds_impl.rst | 70 struct bpf_spin_lock glock; 74 which also contains a ``bpf_spin_lock`` - in the above example both global 92 bpf_spin_lock(&lock); 125 bpf_spin_lock(&lock); 136 * Graph data structure APIs can only be used when the ``bpf_spin_lock`` 146 Because the associated ``bpf_spin_lock`` must be held by any program adding 217 bpf_spin_lock(&lock);
|