/linux/tools/testing/selftests/bpf/progs/ |
H A D | linked_list_fail.c | 31 f = bpf_obj_new(typeof(*f)); \ 35 f2 = bpf_obj_new(typeof(*f2)); \ 40 b = bpf_obj_new(typeof(*b)); \ 229 (void)bpf_obj_new(union { int data; unsigned udata; }); in obj_new_no_struct() 238 f = bpf_obj_new(struct foo); in obj_drop_non_zero_off() 248 return bpf_obj_new(struct foo)->data; in new_null_ret() 254 (void)bpf_obj_new(struct foo); in obj_new_acq() 263 f = bpf_obj_new(typeof(*f)); in use_after_drop() 279 p = bpf_obj_new(typeof(*p)); in ptr_walk_scalar() 291 f = bpf_obj_new(typeof(*f)); in direct_read_lock() [all …]
|
H A D | rbtree.c | 52 n = bpf_obj_new(typeof(*n)); in __add_three() 57 m = bpf_obj_new(typeof(*m)); in __add_three() 69 n = bpf_obj_new(typeof(*n)); in __add_three() 98 n = bpf_obj_new(typeof(*n)); in rbtree_add_and_remove() 103 m = bpf_obj_new(typeof(*m)); in rbtree_add_and_remove() 141 nodes[i][j] = bpf_obj_new(typeof(*nodes[i][j])); in rbtree_add_and_remove_array() 195 n = bpf_obj_new(typeof(*n)); in rbtree_first_and_remove() 201 m = bpf_obj_new(typeof(*m)); in rbtree_first_and_remove() 207 o = bpf_obj_new(typeof(*o)); in rbtree_first_and_remove() 263 n = bpf_obj_new(typeof(*n)); in rbtree_api_release_aliasing()
|
H A D | rbtree_fail.c | 37 n = bpf_obj_new(typeof(*n)); in rbtree_api_nolock_add() 51 n = bpf_obj_new(typeof(*n)); in rbtree_api_nolock_remove() 78 n = bpf_obj_new(typeof(*n)); in rbtree_api_remove_unadded_node() 82 m = bpf_obj_new(typeof(*m)); in rbtree_api_remove_unadded_node() 142 n = bpf_obj_new(typeof(*n)); in rbtree_api_add_to_multiple_trees() 187 n = bpf_obj_new(typeof(*n)); in rbtree_api_add_release_unlock_escape() 273 n = bpf_obj_new(typeof(*n)); in add_with_cb()
|
H A D | local_kptr_stash.c | 99 inner_local_kptr = bpf_obj_new(typeof(*inner_local_kptr)); in create_and_stash() 103 res = bpf_obj_new(typeof(*res)); in create_and_stash() 141 res = bpf_obj_new(typeof(*res)); in stash_plain() 164 res = bpf_obj_new(typeof(*res)); in stash_local_with_root() 169 n = bpf_obj_new(typeof(*n)); in stash_local_with_root() 271 p = bpf_obj_new(typeof(*p)); in stash_refcounted_node()
|
H A D | refcounted_kptr_fail.c | 40 n = bpf_obj_new(typeof(*n)); in rbtree_refcounted_node_ref_escapes() 62 n = bpf_obj_new(typeof(*n)); in refcount_acquire_maybe_null() 81 n = bpf_obj_new(typeof(*n)); in rbtree_refcounted_node_ref_escapes_owning_input() 104 n = bpf_obj_new(typeof(*n)); in BPF_PROG()
|
H A D | refcounted_kptr.c | 79 n = bpf_obj_new(typeof(*n)); in __insert_in_tree_and_list() 116 n = bpf_obj_new(typeof(*n)); in __stash_map_insert_tree() 376 n = bpf_obj_new(typeof(*n)); in rbtree_refcounted_node_ref_escapes() 398 n = bpf_obj_new(typeof(*n)); in rbtree_refcounted_node_ref_escapes_owning_input() 435 n = bpf_obj_new(typeof(*n)); in rbtree_wrong_owner_remove_fail_a1() 512 n = bpf_obj_new(typeof(*n)); in BPF_PROG() 546 n = bpf_obj_new(typeof(*n)); in BPF_PROG()
|
H A D | exceptions_fail.c | 169 f = bpf_obj_new(typeof(*f)); in reject_with_rbtree_add_throw() 184 f = bpf_obj_new(typeof(*f)); in reject_with_reference() 195 f = bpf_obj_new(typeof(*f)); in subprog_ref() 214 f = bpf_obj_new(typeof(*f)); in reject_with_cb_reference()
|
H A D | local_kptr_stash_fail.c | 54 res = bpf_obj_new(typeof(*res)); in stash_rb_nodes() 77 res = bpf_obj_new(typeof(*res)); in drop_rb_node_off()
|
H A D | linked_list.c | 33 f = bpf_obj_new(typeof(*f)); in list_push_pop() 122 f[i] = bpf_obj_new(typeof(**f)); in list_push_pop_multiple() 127 f[i + 1] = bpf_obj_new(typeof(**f)); in list_push_pop_multiple() 198 f = bpf_obj_new(typeof(*f)); in list_in_list() 202 b = bpf_obj_new(typeof(*b)); in list_in_list()
|
H A D | rbtree_btf_fail__wrong_node_type.c | 28 n = bpf_obj_new(typeof(*n)); in rbtree_api_add__wrong_node_type()
|
H A D | normal_map_btf.c | 46 new = bpf_obj_new(typeof(*new)); in add_to_list_in_array()
|
H A D | rbtree_btf_fail__add_wrong_type.c | 42 n = bpf_obj_new(typeof(*n)); in rbtree_api_add__add_wrong_type()
|
H A D | map_in_map_btf.c | 63 new = bpf_obj_new(typeof(*new)); in add_to_list_in_inner_array()
|
H A D | preempted_bpf_ma_op.c | 55 new = bpf_obj_new(typeof(*new)); in add_array()
|
H A D | percpu_alloc_fail.c | 132 p = bpf_obj_new(struct val_t); in BPF_PROG()
|
H A D | task_kfunc_success.c | 172 local = bpf_obj_new(typeof(*local)); in BPF_PROG()
|
H A D | irq.c | 412 p = bpf_obj_new(typeof(*p)); in irq_ooo_refs_array()
|
/linux/Documentation/bpf/ |
H A D | graph_ds_impl.rst | 90 struct node_data *n = bpf_obj_new(typeof(*n)); /* ACQUIRED */ 98 From the verifier's perspective, the pointer ``n`` returned from ``bpf_obj_new`` 123 struct node_data *n = bpf_obj_new(typeof(*n)); /* ACQUIRED */ 141 * Because graph nodes are allocated with ``bpf_obj_new`` and 154 returned by ``bpf_obj_new`` is accordingly considered an *owning reference*. 188 ``free``'d, and reused via bpf_obj_new would point to an entirely different thing. 215 n = bpf_obj_new(typeof(*n)); /* 1 */
|
H A D | bpf_design_QA.rst | 343 bpf_obj_new for user defined types). Will the kernel preserve backwards
|
/linux/tools/testing/selftests/bpf/ |
H A D | bpf_experimental.h | 26 #define bpf_obj_new(type) ((type *)bpf_obj_new_impl(bpf_core_type_id_local(type), NULL)) macro
|
/linux/tools/sched_ext/include/scx/ |
H A D | common.bpf.h | 290 #define bpf_obj_new(type) ((type *)bpf_obj_new_impl(bpf_core_type_id_local(type), NULL)) 288 #define bpf_obj_new(type) ((type *)bpf_obj_new_ global() macro
|
/linux/tools/sched_ext/ |
H A D | scx_flatcg.bpf.c | 874 cgv_node = bpf_obj_new(struct cgv_node); in BPF_STRUCT_OPS_SLEEPABLE()
|