Home
last modified time | relevance | path

Searched refs:bpf_obj_new (Results 1 – 22 of 22) sorted by relevance

/linux/tools/testing/selftests/bpf/progs/
H A Dlinked_list_fail.c31 f = bpf_obj_new(typeof(*f)); \
35 f2 = bpf_obj_new(typeof(*f2)); \
40 b = bpf_obj_new(typeof(*b)); \
229 (void)bpf_obj_new(union { int data; unsigned udata; }); in obj_new_no_struct()
238 f = bpf_obj_new(struct foo); in obj_drop_non_zero_off()
248 return bpf_obj_new(struct foo)->data; in new_null_ret()
254 (void)bpf_obj_new(struct foo); in obj_new_acq()
263 f = bpf_obj_new(typeof(*f)); in use_after_drop()
279 p = bpf_obj_new(typeof(*p)); in ptr_walk_scalar()
291 f = bpf_obj_new(typeof(*f)); in direct_read_lock()
[all …]
H A Drbtree.c52 n = bpf_obj_new(typeof(*n)); in __add_three()
57 m = bpf_obj_new(typeof(*m)); in __add_three()
69 n = bpf_obj_new(typeof(*n)); in __add_three()
98 n = bpf_obj_new(typeof(*n)); in rbtree_add_and_remove()
103 m = bpf_obj_new(typeof(*m)); in rbtree_add_and_remove()
141 nodes[i][j] = bpf_obj_new(typeof(*nodes[i][j])); in rbtree_add_and_remove_array()
195 n = bpf_obj_new(typeof(*n)); in rbtree_first_and_remove()
201 m = bpf_obj_new(typeof(*m)); in rbtree_first_and_remove()
207 o = bpf_obj_new(typeof(*o)); in rbtree_first_and_remove()
263 n = bpf_obj_new(typeof(*n)); in rbtree_api_release_aliasing()
H A Drbtree_fail.c37 n = bpf_obj_new(typeof(*n)); in rbtree_api_nolock_add()
51 n = bpf_obj_new(typeof(*n)); in rbtree_api_nolock_remove()
78 n = bpf_obj_new(typeof(*n)); in rbtree_api_remove_unadded_node()
82 m = bpf_obj_new(typeof(*m)); in rbtree_api_remove_unadded_node()
142 n = bpf_obj_new(typeof(*n)); in rbtree_api_add_to_multiple_trees()
187 n = bpf_obj_new(typeof(*n)); in rbtree_api_add_release_unlock_escape()
273 n = bpf_obj_new(typeof(*n)); in add_with_cb()
H A Dlocal_kptr_stash.c99 inner_local_kptr = bpf_obj_new(typeof(*inner_local_kptr)); in create_and_stash()
103 res = bpf_obj_new(typeof(*res)); in create_and_stash()
141 res = bpf_obj_new(typeof(*res)); in stash_plain()
164 res = bpf_obj_new(typeof(*res)); in stash_local_with_root()
169 n = bpf_obj_new(typeof(*n)); in stash_local_with_root()
271 p = bpf_obj_new(typeof(*p)); in stash_refcounted_node()
H A Drefcounted_kptr_fail.c40 n = bpf_obj_new(typeof(*n)); in rbtree_refcounted_node_ref_escapes()
62 n = bpf_obj_new(typeof(*n)); in refcount_acquire_maybe_null()
81 n = bpf_obj_new(typeof(*n)); in rbtree_refcounted_node_ref_escapes_owning_input()
104 n = bpf_obj_new(typeof(*n)); in BPF_PROG()
H A Drefcounted_kptr.c79 n = bpf_obj_new(typeof(*n)); in __insert_in_tree_and_list()
116 n = bpf_obj_new(typeof(*n)); in __stash_map_insert_tree()
376 n = bpf_obj_new(typeof(*n)); in rbtree_refcounted_node_ref_escapes()
398 n = bpf_obj_new(typeof(*n)); in rbtree_refcounted_node_ref_escapes_owning_input()
435 n = bpf_obj_new(typeof(*n)); in rbtree_wrong_owner_remove_fail_a1()
512 n = bpf_obj_new(typeof(*n)); in BPF_PROG()
546 n = bpf_obj_new(typeof(*n)); in BPF_PROG()
H A Dexceptions_fail.c169 f = bpf_obj_new(typeof(*f)); in reject_with_rbtree_add_throw()
184 f = bpf_obj_new(typeof(*f)); in reject_with_reference()
195 f = bpf_obj_new(typeof(*f)); in subprog_ref()
214 f = bpf_obj_new(typeof(*f)); in reject_with_cb_reference()
H A Dlocal_kptr_stash_fail.c54 res = bpf_obj_new(typeof(*res)); in stash_rb_nodes()
77 res = bpf_obj_new(typeof(*res)); in drop_rb_node_off()
H A Dlinked_list.c33 f = bpf_obj_new(typeof(*f)); in list_push_pop()
122 f[i] = bpf_obj_new(typeof(**f)); in list_push_pop_multiple()
127 f[i + 1] = bpf_obj_new(typeof(**f)); in list_push_pop_multiple()
198 f = bpf_obj_new(typeof(*f)); in list_in_list()
202 b = bpf_obj_new(typeof(*b)); in list_in_list()
H A Drbtree_btf_fail__wrong_node_type.c28 n = bpf_obj_new(typeof(*n)); in rbtree_api_add__wrong_node_type()
H A Dnormal_map_btf.c46 new = bpf_obj_new(typeof(*new)); in add_to_list_in_array()
H A Drbtree_btf_fail__add_wrong_type.c42 n = bpf_obj_new(typeof(*n)); in rbtree_api_add__add_wrong_type()
H A Dmap_in_map_btf.c63 new = bpf_obj_new(typeof(*new)); in add_to_list_in_inner_array()
H A Dpreempted_bpf_ma_op.c55 new = bpf_obj_new(typeof(*new)); in add_array()
H A Dpercpu_alloc_fail.c132 p = bpf_obj_new(struct val_t); in BPF_PROG()
H A Dtask_kfunc_success.c172 local = bpf_obj_new(typeof(*local)); in BPF_PROG()
H A Dirq.c412 p = bpf_obj_new(typeof(*p)); in irq_ooo_refs_array()
/linux/Documentation/bpf/
H A Dgraph_ds_impl.rst90 struct node_data *n = bpf_obj_new(typeof(*n)); /* ACQUIRED */
98 From the verifier's perspective, the pointer ``n`` returned from ``bpf_obj_new``
123 struct node_data *n = bpf_obj_new(typeof(*n)); /* ACQUIRED */
141 * Because graph nodes are allocated with ``bpf_obj_new`` and
154 returned by ``bpf_obj_new`` is accordingly considered an *owning reference*.
188 ``free``'d, and reused via bpf_obj_new would point to an entirely different thing.
215 n = bpf_obj_new(typeof(*n)); /* 1 */
H A Dbpf_design_QA.rst343 bpf_obj_new for user defined types). Will the kernel preserve backwards
/linux/tools/testing/selftests/bpf/
H A Dbpf_experimental.h26 #define bpf_obj_new(type) ((type *)bpf_obj_new_impl(bpf_core_type_id_local(type), NULL)) macro
/linux/tools/sched_ext/include/scx/
H A Dcommon.bpf.h290 #define bpf_obj_new(type) ((type *)bpf_obj_new_impl(bpf_core_type_id_local(type), NULL))
288 #define bpf_obj_new(type) ((type *)bpf_obj_new_ global() macro
/linux/tools/sched_ext/
H A Dscx_flatcg.bpf.c874 cgv_node = bpf_obj_new(struct cgv_node); in BPF_STRUCT_OPS_SLEEPABLE()