| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | btf_map_in_map.c | 17 err = bpf_map_get_info_by_fd(bpf_map__fd(map), &info, &info_len); in bpf_map_id() 38 map1_fd = bpf_map__fd(skel->maps.inner_map1); in test_lookup_update() 39 map2_fd = bpf_map__fd(skel->maps.inner_map2); in test_lookup_update() 40 map3_fd = bpf_map__fd(skel->maps.inner_map3); in test_lookup_update() 41 map4_fd = bpf_map__fd(skel->maps.inner_map4); in test_lookup_update() 42 map5_fd = bpf_map__fd(skel->maps.inner_map5); in test_lookup_update() 43 outer_arr_dyn_fd = bpf_map__fd(skel->maps.outer_arr_dyn); in test_lookup_update() 44 outer_arr_fd = bpf_map__fd(skel->maps.outer_arr); in test_lookup_update() 45 outer_hash_fd = bpf_map__fd(skel->maps.outer_hash); in test_lookup_update() 118 inner_map_fd = bpf_map__fd(skel->maps.sockarr_sz2); in test_diff_size() [all …]
|
| H A D | tailcalls.c | 48 map_fd = bpf_map__fd(prog_array); in test_tailcall_1() 184 map_fd = bpf_map__fd(prog_array); in test_tailcall_2() 261 map_fd = bpf_map__fd(prog_array); in test_tailcall_count() 334 data_fd = bpf_map__fd(data_map); in test_tailcall_count() 349 data_fd = bpf_map__fd(data_map); in test_tailcall_count() 366 data_fd = bpf_map__fd(data_map); in test_tailcall_count() 445 map_fd = bpf_map__fd(prog_array); in test_tailcall_4() 453 data_fd = bpf_map__fd(data_map); in test_tailcall_4() 535 map_fd = bpf_map__fd(prog_array); in test_tailcall_5() 543 data_fd = bpf_map__fd(data_ma in test_tailcall_5() [all...] |
| H A D | recursion.c | 23 bpf_map_delete_elem(bpf_map__fd(skel->maps.hash1), &key); in test_recursion() 25 bpf_map_delete_elem(bpf_map__fd(skel->maps.hash1), &key); in test_recursion() 29 bpf_map_delete_elem(bpf_map__fd(skel->maps.hash2), &key); in test_recursion() 31 bpf_map_delete_elem(bpf_map__fd(skel->maps.hash2), &key); in test_recursion()
|
| H A D | unpriv_bpf_disabled.c | 78 perfbuf = perf_buffer__new(bpf_map__fd(skel->maps.perfbuf), 8, process_perfbuf, NULL, NULL, in test_unpriv_bpf_disabled_positive() 83 ringbuf = ring_buffer__new(bpf_map__fd(skel->maps.ringbuf), process_ringbuf, NULL, NULL); in test_unpriv_bpf_disabled_positive() 247 map_fds[0] = bpf_map__fd(skel->maps.array); in test_unpriv_bpf_disabled() 248 map_fds[1] = bpf_map__fd(skel->maps.percpu_array); in test_unpriv_bpf_disabled() 249 map_fds[2] = bpf_map__fd(skel->maps.hash); in test_unpriv_bpf_disabled() 250 map_fds[3] = bpf_map__fd(skel->maps.percpu_hash); in test_unpriv_bpf_disabled() 251 map_fds[4] = bpf_map__fd(skel->maps.perfbuf); in test_unpriv_bpf_disabled() 252 map_fds[5] = bpf_map__fd(skel->maps.ringbuf); in test_unpriv_bpf_disabled() 253 map_fds[6] = bpf_map__fd(skel->maps.prog_array); in test_unpriv_bpf_disabled()
|
| H A D | xdp_noinline.c | 44 bpf_map_update_elem(bpf_map__fd(skel->maps.vip_map), &key, &value, 0); in test_xdp_noinline() 45 bpf_map_update_elem(bpf_map__fd(skel->maps.ch_rings), &ch_key, &real_num, 0); in test_xdp_noinline() 46 bpf_map_update_elem(bpf_map__fd(skel->maps.reals), &real_num, &real_def, 0); in test_xdp_noinline() 65 bpf_map_lookup_elem(bpf_map__fd(skel->maps.stats), &stats_key, stats); in test_xdp_noinline()
|
| H A D | sockmap_basic.c | 75 src_fd = bpf_map__fd(src); in compare_cookies() 76 dst_fd = bpf_map__fd(dst); in compare_cookies() 165 map = bpf_map__fd(skel->maps.sock_map); in test_skmsg_helpers() 193 map = bpf_map__fd(skel->maps.sock_map); in test_skmsg_helpers_with_link() 260 src = bpf_map__fd(skel->maps.src); in test_sockmap_update() 326 src_fd = bpf_map__fd(src); in test_sockmap_copy() 388 map = bpf_map__fd(skel->maps.sock_map); in test_sockmap_skb_verdict_attach() 415 map = bpf_map__fd(skel->maps.sock_map); in test_sockmap_skb_verdict_attach_with_link() 464 map_fd = bpf_map__fd(skel->maps.sock_map); in test_sockmap_progs_query() 510 map = bpf_map__fd(skel->maps.sock_map_rx); in test_sockmap_skb_verdict_shutdown() [all …]
|
| H A D | stacktrace_build_id.c | 27 control_map_fd = bpf_map__fd(skel->maps.control_map); in test_stacktrace_build_id() 28 stackid_hmap_fd = bpf_map__fd(skel->maps.stackid_hmap); in test_stacktrace_build_id() 29 stackmap_fd = bpf_map__fd(skel->maps.stackmap); in test_stacktrace_build_id() 30 stack_amap_fd = bpf_map__fd(skel->maps.stack_amap); in test_stacktrace_build_id()
|
| H A D | verify_pkcs7_sig.c | 318 ret = bpf_map_update_elem(bpf_map__fd(map), &zero, &data, BPF_ANY); in test_verify_pkcs7_sig_from_map() 327 ret = bpf_map_update_elem(bpf_map__fd(map), &zero, &data, BPF_ANY); in test_verify_pkcs7_sig_from_map() 336 ret = bpf_map_update_elem(bpf_map__fd(map), &zero, &data, BPF_ANY); in test_verify_pkcs7_sig_from_map() 347 ret = bpf_map_update_elem(bpf_map__fd(map), &zero, &data, BPF_ANY); in test_verify_pkcs7_sig_from_map() 361 ret = bpf_map_update_elem(bpf_map__fd(map), &zero, &data, BPF_ANY); in test_verify_pkcs7_sig_from_map() 369 ret = bpf_map_update_elem(bpf_map__fd(map), &zero, &data, BPF_ANY); in test_verify_pkcs7_sig_from_map() 382 ret = bpf_map_update_elem(bpf_map__fd(map), &zero, &data, in test_verify_pkcs7_sig_from_map() 389 ret = bpf_map_update_elem(bpf_map__fd(map), &zero, &data, in test_verify_pkcs7_sig_from_map() 396 ret = bpf_map_update_elem(bpf_map__fd(map), &zero, &data, in test_verify_pkcs7_sig_from_map()
|
| H A D | metadata.c | 76 bpf_map__fd(obj->maps.rodata)); in test_metadata_unused() 93 bpf_map__fd(obj->maps.rodata), NULL); in test_metadata_unused() 110 bpf_map__fd(obj->maps.rodata)); in test_metadata_used() 127 bpf_map__fd(obj->maps.rodata), NULL); in test_metadata_used()
|
| H A D | tcp_hdr_options.c | 300 hdr_stg_map_fd = bpf_map__fd(skel->maps.hdr_stg_map); in fastopen_estab() 301 lport_linum_map_fd = bpf_map__fd(skel->maps.lport_linum_map); in fastopen_estab() 337 hdr_stg_map_fd = bpf_map__fd(skel->maps.hdr_stg_map); in syncookie_estab() 338 lport_linum_map_fd = bpf_map__fd(skel->maps.lport_linum_map); in syncookie_estab() 381 hdr_stg_map_fd = bpf_map__fd(skel->maps.hdr_stg_map); in fin() 382 lport_linum_map_fd = bpf_map__fd(skel->maps.lport_linum_map); in fin() 413 hdr_stg_map_fd = bpf_map__fd(skel->maps.hdr_stg_map); in __simple_estab() 414 lport_linum_map_fd = bpf_map__fd(skel->maps.lport_linum_map); in __simple_estab() 466 lport_linum_map_fd = bpf_map__fd(misc_skel->maps.lport_linum_map); in misc()
|
| H A D | stacktrace_map_skip.c | 18 stackid_hmap_fd = bpf_map__fd(skel->maps.stackid_hmap); in test_stacktrace_map_skip() 22 stackmap_fd = bpf_map__fd(skel->maps.stackmap); in test_stacktrace_map_skip() 26 stack_amap_fd = bpf_map__fd(skel->maps.stack_amap); in test_stacktrace_map_skip()
|
| H A D | map_lookup_percpu_elem.c | 37 ret = bpf_map_update_elem(bpf_map__fd(skel->maps.percpu_array_map), &key, buf, 0); in test_map_lookup_percpu_elem() 40 ret = bpf_map_update_elem(bpf_map__fd(skel->maps.percpu_hash_map), &key, buf, 0); in test_map_lookup_percpu_elem() 43 ret = bpf_map_update_elem(bpf_map__fd(skel->maps.percpu_lru_hash_map), &key, buf, 0); in test_map_lookup_percpu_elem()
|
| H A D | sockmap_listen.c | 730 int verdict_map = bpf_map__fd(skel->maps.verdict_map); in test_skb_redir_to_connected() 731 int sock_map = bpf_map__fd(inner_map); in test_skb_redir_to_connected() 754 int verdict_map = bpf_map__fd(skel->maps.verdict_map); in test_msg_redir_to_connected() 755 int sock_map = bpf_map__fd(inner_map); in test_msg_redir_to_connected() 772 int verdict_map = bpf_map__fd(skel->maps.verdict_map); in test_msg_redir_to_connected_with_link() 773 int sock_map = bpf_map__fd(inner_map); in test_msg_redir_to_connected_with_link() 850 int verdict_map = bpf_map__fd(skel->maps.verdict_map); in test_skb_redir_to_listening() 851 int sock_map = bpf_map__fd(inner_map); in test_skb_redir_to_listening() 874 int verdict_map = bpf_map__fd(skel->maps.verdict_map); in test_msg_redir_to_listening() 875 int sock_map = bpf_map__fd(inner_map); in test_msg_redir_to_listening() [all …]
|
| H A D | sk_storage_tracing.c | 30 err = bpf_map_lookup_elem(bpf_map__fd(skel->maps.sk_stg_map), &sk_fd, in check_sk_stg() 63 err = bpf_map_update_elem(bpf_map__fd(skel->maps.del_sk_stg_map), in do_test() 83 err = bpf_map_lookup_elem(bpf_map__fd(skel->maps.del_sk_stg_map), in do_test()
|
| H A D | stacktrace_build_id_nmi.c | 55 control_map_fd = bpf_map__fd(skel->maps.control_map); in test_stacktrace_build_id_nmi() 56 stackid_hmap_fd = bpf_map__fd(skel->maps.stackid_hmap); in test_stacktrace_build_id_nmi() 57 stackmap_fd = bpf_map__fd(skel->maps.stackmap); in test_stacktrace_build_id_nmi()
|
| H A D | htab_update.c | 46 err = bpf_map_update_elem(bpf_map__fd(skel->maps.htab), &key, value, BPF_ANY); in test_reenter_update() 58 err = bpf_map_update_elem(bpf_map__fd(skel->maps.htab), &key, value, BPF_ANY); in test_reenter_update() 106 ctx.fd = bpf_map__fd(skel->maps.htab); in test_concurrent_update()
|
| H A D | test_local_storage.c | 111 if (!check_syscall_operations(bpf_map__fd(skel->maps.task_storage_map), in test_test_local_storage() 128 if (!check_syscall_operations(bpf_map__fd(skel->maps.inode_storage_map), in test_test_local_storage() 160 if (!check_syscall_operations(bpf_map__fd(skel->maps.sk_storage_map), in test_test_local_storage()
|
| H A D | legacy_printk.c | 24 map_fd = bpf_map__fd(skel->maps.my_pid_map); in execute_one_variant() 40 map_fd = bpf_map__fd(skel->maps.res_map); in execute_one_variant()
|
| H A D | cgrp_local_storage.c | 50 err = bpf_map_update_elem(bpf_map__fd(skel->maps.map_b), &cgroup_fd, &val1, BPF_ANY); in test_tp_btf() 55 err = bpf_map_lookup_elem(bpf_map__fd(skel->maps.map_b), &cgroup_fd, &val2); in test_tp_btf() 62 err = bpf_map_delete_elem(bpf_map__fd(skel->maps.map_b), &cgroup_fd); in test_tp_btf() 121 err = bpf_map_lookup_elem(bpf_map__fd(skel->maps.socket_cookies), in test_attach_cgroup()
|
| H A D | map_btf.c | 33 new_fd = dup(bpf_map__fd(skel->maps.array)); in do_test_normal_map_btf() 74 new_fd = dup(bpf_map__fd(skel->maps.inner_array)); in do_test_map_in_map_btf()
|
| H A D | libbpf_get_fd_by_id_opts.c | 32 ret = bpf_map_get_info_by_fd(bpf_map__fd(skel->maps.data_input), in test_libbpf_get_fd_by_id_opts() 63 ret = bpf_map_update_elem(bpf_map__fd(skel->maps.data_input), &zero, in test_libbpf_get_fd_by_id_opts()
|
| /linux/tools/perf/util/ |
| H A D | bpf_lock_contention.c | 83 fd = bpf_map__fd(skel->maps.slab_caches); in run_slab_cache_iter() 327 fd = bpf_map__fd(skel->maps.cpu_filter); in lock_contention_prepare() 339 fd = bpf_map__fd(skel->maps.task_filter); in lock_contention_prepare() 351 fd = bpf_map__fd(skel->maps.task_filter); in lock_contention_prepare() 358 fd = bpf_map__fd(skel->maps.type_filter); in lock_contention_prepare() 367 fd = bpf_map__fd(skel->maps.addr_filter); in lock_contention_prepare() 376 fd = bpf_map__fd(skel->maps.cgroup_filter); in lock_contention_prepare() 383 fd = bpf_map__fd(skel->maps.lock_delays); in lock_contention_prepare() 404 fd = bpf_map__fd(skel->maps.slab_filter); in lock_contention_prepare() 407 cache_fd = bpf_map__fd(skel->maps.slab_caches); in lock_contention_prepare() [all …]
|
| H A D | bpf_off_cpu.c | 79 fd = bpf_map__fd(skel->maps.task_filter); in off_cpu_start() 238 fd = bpf_map__fd(skel->maps.cpu_filter); in off_cpu_prepare() 249 fd = bpf_map__fd(skel->maps.task_filter); in off_cpu_prepare() 267 fd = bpf_map__fd(skel->maps.task_filter); in off_cpu_prepare() 279 fd = bpf_map__fd(skel->maps.cgroup_filter); in off_cpu_prepare() 356 fd = bpf_map__fd(skel->maps.off_cpu); in off_cpu_write() 357 stack = bpf_map__fd(skel->maps.stacks); in off_cpu_write()
|
| H A D | bpf_trace_augment.c | 106 *enter_fd = bpf_map__fd(skel->maps.syscalls_sys_enter); in augmented_syscalls__get_map_fds() 107 *exit_fd = bpf_map__fd(skel->maps.syscalls_sys_exit); in augmented_syscalls__get_map_fds() 108 *beauty_fd = bpf_map__fd(skel->maps.beauty_map_enter); in augmented_syscalls__get_map_fds()
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | flow_dissector_load.h | 39 prog_array_fd = bpf_map__fd(prog_array); in bpf_flow_load() 48 *keys_fd = bpf_map__fd(keys); in bpf_flow_load()
|