| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | tracing_struct.c | 5 #include "tracing_struct.skel.h" 6 #include "tracing_struct_many_args.skel.h" 10 struct tracing_struct *skel; in test_struct_args() local 13 skel = tracing_struct__open_and_load(); in test_struct_args() 14 if (!ASSERT_OK_PTR(skel, "tracing_struct__open_and_load")) in test_struct_args() 17 err = tracing_struct__attach(skel); in test_struct_args() 23 ASSERT_EQ(skel->bss->t1_a_a, 2, "t1:a.a"); in test_struct_args() 24 ASSERT_EQ(skel->bss->t1_a_b, 3, "t1:a.b"); in test_struct_args() 25 ASSERT_EQ(skel->bss->t1_b, 1, "t1:b"); in test_struct_args() 26 ASSERT_EQ(skel in test_struct_args() 65 struct tracing_struct_many_args *skel; test_struct_many_args() local [all...] |
| H A D | attach_probe.c | 3 #include "test_attach_kprobe_sleepable.skel.h" 4 #include "test_attach_probe_manual.skel.h" 5 #include "test_attach_probe.skel.h" 6 #include "kprobe_write_ctx.skel.h" 44 struct test_attach_probe_manual *skel; in test_attach_probe_manual() 47 skel = test_attach_probe_manual__open_and_load(); in test_attach_probe_manual() 48 if (!ASSERT_OK_PTR(skel, "skel_kprobe_manual_open_and_load")) in test_attach_probe_manual() 58 kprobe_link = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe, in test_attach_probe_manual() 63 skel->links.handle_kprobe = kprobe_link; in test_attach_probe_manual() 66 kretprobe_link = bpf_program__attach_kprobe_opts(skel in test_attach_probe_manual() 43 struct test_attach_probe_manual *skel; test_attach_probe_manual() local 125 test_attach_probe_auto(struct test_attach_probe * skel) test_attach_probe_auto() argument 159 test_uprobe_lib(struct test_attach_probe * skel) test_uprobe_lib() argument 195 test_uprobe_ref_ctr(struct test_attach_probe * skel) test_uprobe_ref_ctr() argument 238 struct test_attach_kprobe_sleepable *skel; test_kprobe_sleepable() local 261 test_uprobe_sleepable(struct test_attach_probe * skel) test_uprobe_sleepable() argument 295 struct test_attach_probe *skel; test_attach_probe() local [all...] |
| H A D | atomics.c | 7 static void test_add(struct atomics_lskel *skel) in test_add() argument 13 prog_fd = skel->progs.add.prog_fd; in test_add() 20 ASSERT_EQ(skel->data->add64_value, 3, "add64_value"); in test_add() 21 ASSERT_EQ(skel->bss->add64_result, 1, "add64_result"); in test_add() 23 ASSERT_EQ(skel->data->add32_value, 3, "add32_value"); in test_add() 24 ASSERT_EQ(skel->bss->add32_result, 1, "add32_result"); in test_add() 26 ASSERT_EQ(skel->bss->add_stack_value_copy, 3, "add_stack_value"); in test_add() 27 ASSERT_EQ(skel->bss->add_stack_result, 1, "add_stack_result"); in test_add() 29 ASSERT_EQ(skel->data->add_noreturn_value, 3, "add_noreturn_value"); in test_add() 32 static void test_sub(struct atomics_lskel *skel) in test_sub() argument 57 test_and(struct atomics_lskel * skel) test_and() argument 79 test_or(struct atomics_lskel * skel) test_or() argument 101 test_xor(struct atomics_lskel * skel) test_xor() argument 123 test_cmpxchg(struct atomics_lskel * skel) test_cmpxchg() argument 145 test_xchg(struct atomics_lskel * skel) test_xchg() argument 167 struct atomics_lskel *skel; test_atomics() local [all...] |
| H A D | test_struct_ops_module.c | 39 static int attach_ops_and_check(struct struct_ops_module *skel, in attach_ops_and_check() argument 51 ASSERT_EQ(skel->bss->test_1_result, 0xdeadbeef, "test_1_result"); in attach_ops_and_check() 52 ASSERT_EQ(skel->bss->test_2_result, expected_test_2_result, "test_2_result"); in attach_ops_and_check() 60 struct struct_ops_module *skel; in test_struct_ops_load() local 65 skel = struct_ops_module__open(); in test_struct_ops_load() 66 if (!ASSERT_OK_PTR(skel, "struct_ops_module_open")) in test_struct_ops_load() 69 skel->struct_ops.testmod_1->data = 13; in test_struct_ops_load() 70 skel->struct_ops.testmod_1->test_2 = skel->progs.test_3; in test_struct_ops_load() 74 bpf_program__set_autoload(skel->progs.test_2, false); in test_struct_ops_load() 75 bpf_map__set_autocreate(skel->maps.testmod_zeroed, false); in test_struct_ops_load() [all …]
|
| H A D | test_strncmp.c | 6 static int trigger_strncmp(const struct strncmp_test *skel) in trigger_strncmp() argument 12 cmp = skel->bss->cmp_ret; in trigger_strncmp() 24 static void strncmp_full_str_cmp(struct strncmp_test *skel, const char *name, in strncmp_full_str_cmp() argument 27 size_t nr = sizeof(skel->bss->str); in strncmp_full_str_cmp() 28 char *str = skel->bss->str; in strncmp_full_str_cmp() 33 memcpy(str, skel->rodata->target, nr); in strncmp_full_str_cmp() 37 got = trigger_strncmp(skel); in strncmp_full_str_cmp() 46 struct strncmp_test *skel; in test_strncmp_ret() local 49 skel = strncmp_test__open(); in test_strncmp_ret() 50 if (!ASSERT_OK_PTR(skel, "strncmp_test open")) in test_strncmp_ret() [all …]
|
| H A D | arena_atomics.c | 4 #include "arena_atomics.skel.h" 6 static void test_add(struct arena_atomics *skel) in test_add() argument 12 prog_fd = bpf_program__fd(skel->progs.add); in test_add() 19 ASSERT_EQ(skel->arena->add64_value, 3, "add64_value"); in test_add() 20 ASSERT_EQ(skel->arena->add64_result, 1, "add64_result"); in test_add() 22 ASSERT_EQ(skel->arena->add32_value, 3, "add32_value"); in test_add() 23 ASSERT_EQ(skel->arena->add32_result, 1, "add32_result"); in test_add() 25 ASSERT_EQ(skel->arena->add_stack_value_copy, 3, "add_stack_value"); in test_add() 26 ASSERT_EQ(skel->arena->add_stack_result, 1, "add_stack_result"); in test_add() 28 ASSERT_EQ(skel in test_add() 31 test_sub(struct arena_atomics * skel) test_sub() argument 56 test_and(struct arena_atomics * skel) test_and() argument 73 test_or(struct arena_atomics * skel) test_or() argument 90 test_xor(struct arena_atomics * skel) test_xor() argument 107 test_cmpxchg(struct arena_atomics * skel) test_cmpxchg() argument 129 test_xchg(struct arena_atomics * skel) test_xchg() argument 149 test_uaf(struct arena_atomics * skel) test_uaf() argument 167 struct arena_atomics *skel; test_arena_atomics() local [all...] |
| H A D | bpf_iter.c | 39 struct bpf_iter_test_kern3 *skel; in test_btf_id_or_null() local 41 skel = bpf_iter_test_kern3__open_and_load(); in test_btf_id_or_null() 42 if (!ASSERT_ERR_PTR(skel, "bpf_iter_test_kern3__open_and_load")) { in test_btf_id_or_null() 43 bpf_iter_test_kern3__destroy(skel); in test_btf_id_or_null() 78 static void do_read_map_iter_fd(struct bpf_object_skeleton **skel, struct bpf_program *prog, in do_read_map_iter_fd() argument 103 bpf_object__destroy_skeleton(*skel); in do_read_map_iter_fd() 104 *skel = NULL; in do_read_map_iter_fd() 140 struct bpf_iter_ipv6_route *skel; in test_ipv6_route() local 142 skel = bpf_iter_ipv6_route__open_and_load(); in test_ipv6_route() 143 if (!ASSERT_OK_PTR(skel, "bpf_iter_ipv6_route__open_and_load")) in test_ipv6_route() [all …]
|
| H A D | bpf_loop.c | 8 static void check_nr_loops(struct bpf_loop *skel) in check_nr_loops() argument 12 link = bpf_program__attach(skel->progs.test_prog); in check_nr_loops() 17 skel->bss->nr_loops = 0; in check_nr_loops() 21 ASSERT_EQ(skel->bss->nr_loops_returned, skel->bss->nr_loops, in check_nr_loops() 25 skel->bss->nr_loops = 500; in check_nr_loops() 29 ASSERT_EQ(skel->bss->nr_loops_returned, skel->bss->nr_loops, in check_nr_loops() 31 ASSERT_EQ(skel->bss->g_output, (500 * 499) / 2, "g_output"); in check_nr_loops() 34 skel->bss->nr_loops = -1; in check_nr_loops() 38 ASSERT_EQ(skel->bss->err, -E2BIG, "over max limit"); in check_nr_loops() 43 static void check_callback_fn_stop(struct bpf_loop *skel) in check_callback_fn_stop() argument [all …]
|
| H A D | struct_ops_autocreate.c | 9 struct struct_ops_autocreate *skel; in cant_load_full_object() local 13 skel = struct_ops_autocreate__open(); in cant_load_full_object() 14 if (!ASSERT_OK_PTR(skel, "struct_ops_autocreate__open")) in cant_load_full_object() 24 err = struct_ops_autocreate__load(skel); in cant_load_full_object() 34 struct_ops_autocreate__destroy(skel); in cant_load_full_object() 37 static int check_test_1_link(struct struct_ops_autocreate *skel, struct bpf_map *map) in check_test_1_link() argument 42 link = bpf_map__attach_struct_ops(skel->maps.testmod_1); in check_test_1_link() 47 err = ASSERT_EQ(skel->bss->test_1_result, 42, "test_1_result"); in check_test_1_link() 54 struct struct_ops_autocreate *skel; in can_load_partial_object() local 57 skel = struct_ops_autocreate__open(); in can_load_partial_object() [all …]
|
| H A D | tc_links.c | 11 #include "test_tc_link.skel.h" 22 struct test_tc_link *skel; in serial_test_tc_links_basic() local 26 skel = test_tc_link__open_and_load(); in serial_test_tc_links_basic() 27 if (!ASSERT_OK_PTR(skel, "skel_load")) in serial_test_tc_links_basic() 30 pid1 = id_from_prog_fd(bpf_program__fd(skel->progs.tc1)); in serial_test_tc_links_basic() 31 pid2 = id_from_prog_fd(bpf_program__fd(skel->progs.tc2)); in serial_test_tc_links_basic() 38 ASSERT_EQ(skel->bss->seen_tc1, false, "seen_tc1"); in serial_test_tc_links_basic() 39 ASSERT_EQ(skel->bss->seen_tc2, false, "seen_tc2"); in serial_test_tc_links_basic() 41 link = bpf_program__attach_tcx(skel->progs.tc1, loopback, &optl); in serial_test_tc_links_basic() 45 skel in serial_test_tc_links_basic() 122 struct test_tc_link *skel; test_tc_links_before_target() local 276 struct test_tc_link *skel; test_tc_links_after_target() local 429 struct test_tc_link *skel; test_tc_links_revision_target() local 530 struct test_tc_link *skel; test_tc_chain_classic() local 635 struct test_tc_link *skel; test_tc_links_replace_target() local 860 struct test_tc_link *skel; test_tc_links_invalid_target() local 1174 struct test_tc_link *skel; test_tc_links_prepend_target() local 1330 struct test_tc_link *skel; test_tc_links_append_target() local 1484 struct test_tc_link *skel; test_tc_links_dev_cleanup_target() local 1582 struct test_tc_link *skel; test_tc_chain_mixed() local 1696 struct test_tc_link *skel; test_tc_links_ingress() local 1857 struct test_tc_link *skel; test_tc_links_dev_mixed() local [all...] |
| H A D | map_ops.c | 46 static int setup(struct test_map_ops **skel) in setup() argument 50 if (!skel) in setup() 53 *skel = test_map_ops__open(); in setup() 54 if (!ASSERT_OK_PTR(*skel, "test_map_ops__open")) in setup() 57 (*skel)->rodata->pid = getpid(); in setup() 59 err = test_map_ops__load(*skel); in setup() 63 err = test_map_ops__attach(*skel); in setup() 70 static void teardown(struct test_map_ops **skel) in teardown() argument 72 if (skel && *skel) in teardown() 73 test_map_ops__destroy(*skel); in teardown() [all …]
|
| H A D | iters.c | 27 struct iters_num *skel; in subtest_num_iters() local 30 skel = iters_num__open_and_load(); in subtest_num_iters() 31 if (!ASSERT_OK_PTR(skel, "skel_open_and_load")) in subtest_num_iters() 34 err = iters_num__attach(skel); in subtest_num_iters() 39 iters_num__detach(skel); in subtest_num_iters() 42 ASSERT_EQ(skel->bss->res_##case_name, \ in subtest_num_iters() 43 skel->rodata->exp_##case_name, \ in subtest_num_iters() 67 iters_num__destroy(skel); in subtest_num_iters() 72 struct iters_testmod_seq *skel; in subtest_testmod_seq_iters() local 80 skel = iters_testmod_seq__open_and_load(); in subtest_testmod_seq_iters() [all …]
|
| H A D | bpf_cookie.c | 12 #include "test_bpf_cookie.skel.h" 13 #include "kprobe_multi.skel.h" 14 #include "uprobe_multi.skel.h" 22 static void kprobe_subtest(struct test_bpf_cookie *skel) in kprobe_subtest() argument 31 link1 = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe, in kprobe_subtest() 38 link2 = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe, in kprobe_subtest() 46 retlink1 = bpf_program__attach_kprobe_opts(skel->progs.handle_kretprobe, in kprobe_subtest() 53 retlink2 = bpf_program__attach_kprobe_opts(skel->progs.handle_kretprobe, in kprobe_subtest() 61 ASSERT_EQ(skel->bss->kprobe_res, 0x1 | 0x2, "kprobe_res"); in kprobe_subtest() 62 ASSERT_EQ(skel in kprobe_subtest() 71 kprobe_multi_test_run(struct kprobe_multi * skel) kprobe_multi_test_run() argument 103 struct kprobe_multi *skel = NULL; kprobe_multi_link_api_subtest() local 182 struct kprobe_multi *skel = NULL; kprobe_multi_attach_api_subtest() local 249 uprobe_multi_test_run(struct uprobe_multi * skel) uprobe_multi_test_run() argument 274 struct uprobe_multi *skel = NULL; uprobe_multi_attach_api_subtest() local 318 uprobe_subtest(struct test_bpf_cookie * skel) uprobe_subtest() argument 372 tp_subtest(struct test_bpf_cookie * skel) tp_subtest() argument 441 pe_subtest(struct test_bpf_cookie * skel) pe_subtest() argument 510 tracing_subtest(struct test_bpf_cookie * skel) tracing_subtest() argument 567 lsm_subtest(struct test_bpf_cookie * skel) lsm_subtest() argument 598 tp_btf_subtest(struct test_bpf_cookie * skel) tp_btf_subtest() argument 679 raw_tp_subtest(struct test_bpf_cookie * skel) raw_tp_subtest() argument 733 struct test_bpf_cookie *skel; test_bpf_cookie() local [all...] |
| H A D | uprobe_multi_test.c | 194 static void uprobe_multi_test_run(struct uprobe_multi *skel, struct child *child) in uprobe_multi_test_run() argument 196 skel->bss->uprobe_multi_func_1_addr = (__u64) uprobe_multi_func_1; in uprobe_multi_test_run() 197 skel->bss->uprobe_multi_func_2_addr = (__u64) uprobe_multi_func_2; in uprobe_multi_test_run() 198 skel->bss->uprobe_multi_func_3_addr = (__u64) uprobe_multi_func_3; in uprobe_multi_test_run() 200 skel->bss->user_ptr = test_data; in uprobe_multi_test_run() 207 skel->bss->pid = child ? 0 : getpid(); in uprobe_multi_test_run() 208 skel->bss->expect_pid = child ? child->pid : 0; in uprobe_multi_test_run() 229 ASSERT_EQ(skel->bss->uprobe_multi_func_1_result, 2, "uprobe_multi_func_1_result"); in uprobe_multi_test_run() 230 ASSERT_EQ(skel->bss->uprobe_multi_func_2_result, 2, "uprobe_multi_func_2_result"); in uprobe_multi_test_run() 231 ASSERT_EQ(skel->bss->uprobe_multi_func_3_result, 2, "uprobe_multi_func_3_result"); in uprobe_multi_test_run() [all …]
|
| H A D | find_vma.c | 10 static void test_and_reset_skel(struct find_vma *skel, int expected_find_zero_ret, bool need_test) in test_and_reset_skel() argument 13 ASSERT_EQ(skel->bss->found_vm_exec, 1, "found_vm_exec"); in test_and_reset_skel() 14 ASSERT_EQ(skel->data->find_addr_ret, 0, "find_addr_ret"); in test_and_reset_skel() 15 ASSERT_EQ(skel->data->find_zero_ret, expected_find_zero_ret, "find_zero_ret"); in test_and_reset_skel() 16 ASSERT_OK_PTR(strstr(skel->bss->d_iname, "test_progs"), "find_test_progs"); in test_and_reset_skel() 19 skel->bss->found_vm_exec = 0; in test_and_reset_skel() 20 skel->data->find_addr_ret = -1; in test_and_reset_skel() 21 skel->data->find_zero_ret = -1; in test_and_reset_skel() 22 skel->bss->d_iname[0] = 0; in test_and_reset_skel() 41 static bool find_vma_pe_condition(struct find_vma *skel) in find_vma_pe_condition() argument [all …]
|
| H A D | cgroup1_hierarchy.c | 10 static void bpf_cgroup1(struct test_cgroup1_hierarchy *skel) in bpf_cgroup1() argument 16 lsm_link = bpf_program__attach_lsm(skel->progs.lsm_run); in bpf_cgroup1() 21 fentry_link = bpf_program__attach_trace(skel->progs.fentry_run); in bpf_cgroup1() 28 static void bpf_cgroup1_sleepable(struct test_cgroup1_hierarchy *skel) in bpf_cgroup1_sleepable() argument 34 lsm_link = bpf_program__attach_lsm(skel->progs.lsm_s_run); in bpf_cgroup1_sleepable() 39 fentry_link = bpf_program__attach_trace(skel->progs.fentry_run); in bpf_cgroup1_sleepable() 46 static void bpf_cgroup1_invalid_id(struct test_cgroup1_hierarchy *skel) in bpf_cgroup1_invalid_id() argument 52 lsm_link = bpf_program__attach_lsm(skel->progs.lsm_run); in bpf_cgroup1_invalid_id() 57 fentry_link = bpf_program__attach_trace(skel->progs.fentry_run); in bpf_cgroup1_invalid_id() 71 struct test_cgroup1_hierarchy *skel; in test_cgroup1_hierarchy() local [all …]
|
| H A D | btf_tag.c | 17 struct test_btf_decl_tag *skel; in test_btf_decl_tag() local 19 skel = test_btf_decl_tag__open_and_load(); in test_btf_decl_tag() 20 if (!ASSERT_OK_PTR(skel, "btf_decl_tag")) in test_btf_decl_tag() 23 if (skel->rodata->skip_tests) { in test_btf_decl_tag() 28 test_btf_decl_tag__destroy(skel); in test_btf_decl_tag() 33 struct btf_type_tag *skel; in test_btf_type_tag() local 35 skel = btf_type_tag__open_and_load(); in test_btf_type_tag() 36 if (!ASSERT_OK_PTR(skel, "btf_type_tag")) in test_btf_type_tag() 39 if (skel->rodata->skip_tests) { in test_btf_type_tag() 44 btf_type_tag__destroy(skel); in test_btf_type_tag() [all …]
|
| H A D | global_map_resize.c | 22 struct test_global_map_resize *skel; in global_map_resize_bss_subtest() local 24 const __u32 desired_sz = sizeof(skel->bss->sum) + sysconf(_SC_PAGE_SIZE) * 2; in global_map_resize_bss_subtest() 27 skel = test_global_map_resize__open(); in global_map_resize_bss_subtest() 28 if (!ASSERT_OK_PTR(skel, "test_global_map_resize__open")) in global_map_resize_bss_subtest() 35 skel->bss->array[0] = 1; in global_map_resize_bss_subtest() 38 map = skel->maps.bss; in global_map_resize_bss_subtest() 45 new_sz = sizeof(skel->data_percpu_arr->percpu_arr[0]) * libbpf_num_possible_cpus(); in global_map_resize_bss_subtest() 46 err = bpf_map__set_value_size(skel->maps.data_percpu_arr, new_sz); in global_map_resize_bss_subtest() 50 array_len = (desired_sz - sizeof(skel->bss->sum)) / sizeof(skel->bss->array[0]); in global_map_resize_bss_subtest() 54 skel->bss = bpf_map__initial_value(skel->maps.bss, &actual_sz); in global_map_resize_bss_subtest() [all …]
|
| H A D | rbtree.c | 7 #include "rbtree.skel.h" 8 #include "rbtree_fail.skel.h" 9 #include "rbtree_btf_fail__wrong_node_type.skel.h" 10 #include "rbtree_btf_fail__add_wrong_type.skel.h" 11 #include "rbtree_search.skel.h" 20 struct rbtree *skel; in test_rbtree_add_nodes() 23 skel = rbtree__open_and_load(); in test_rbtree_add_nodes() 24 if (!ASSERT_OK_PTR(skel, "rbtree__open_and_load")) in test_rbtree_add_nodes() 27 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.rbtree_add_nodes), &opts); in test_rbtree_add_nodes() 30 ASSERT_EQ(skel in test_rbtree_add_nodes() 19 struct rbtree *skel; test_rbtree_add_nodes() local 41 struct rbtree *skel; test_rbtree_add_nodes_nested() local 63 struct rbtree *skel; test_rbtree_add_and_remove() local 85 struct rbtree *skel; test_rbtree_add_and_remove_array() local 106 struct rbtree *skel; test_rbtree_first_and_remove() local 130 struct rbtree *skel; test_rbtree_api_release_aliasing() local [all...] |
| H A D | percpu_alloc.c | 9 struct percpu_alloc_array *skel; in test_array() local 13 skel = percpu_alloc_array__open(); in test_array() 14 if (!ASSERT_OK_PTR(skel, "percpu_alloc_array__open")) in test_array() 17 bpf_program__set_autoload(skel->progs.test_array_map_1, true); in test_array() 18 bpf_program__set_autoload(skel->progs.test_array_map_2, true); in test_array() 19 bpf_program__set_autoload(skel->progs.test_array_map_3, true); in test_array() 20 bpf_program__set_autoload(skel->progs.test_array_map_4, true); in test_array() 22 skel->bss->my_pid = getpid(); in test_array() 23 skel->rodata->nr_cpus = libbpf_num_possible_cpus(); in test_array() 25 err = percpu_alloc_array__load(skel); in test_array() [all …]
|
| H A D | check_mtu.c | 6 #include "test_check_mtu.skel.h" 43 struct test_check_mtu *skel; in test_check_mtu_xdp_attach() local 49 skel = test_check_mtu__open_and_load(); in test_check_mtu_xdp_attach() 50 if (CHECK(!skel, "open and load skel", "failed")) in test_check_mtu_xdp_attach() 53 prog = skel->progs.xdp_use_helper_basic; in test_check_mtu_xdp_attach() 58 skel->links.xdp_use_helper_basic = link; in test_check_mtu_xdp_attach() 75 test_check_mtu__destroy(skel); in test_check_mtu_xdp_attach() 78 static void test_check_mtu_run_xdp(struct test_check_mtu *skel, in test_check_mtu_run_xdp() argument 99 mtu_result = skel in test_check_mtu_run_xdp() 106 struct test_check_mtu *skel; test_check_mtu_xdp() local 131 test_check_mtu_run_tc(struct test_check_mtu * skel,struct bpf_program * prog,__u32 mtu_expect) test_check_mtu_run_tc() argument 159 struct test_check_mtu *skel; test_check_mtu_tc() local [all...] |
| /linux/tools/sched_ext/include/scx/ |
| H A D | enums.autogen.h | 7 #define SCX_ENUM_INIT(skel) do { \ argument 8 SCX_ENUM_SET(skel, scx_public_consts, SCX_OPS_NAME_LEN); \ 9 SCX_ENUM_SET(skel, scx_public_consts, SCX_SLICE_DFL); \ 10 SCX_ENUM_SET(skel, scx_public_consts, SCX_SLICE_INF); \ 11 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_ONLINE); \ 12 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_CAN_STOP_TICK); \ 13 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_BAL_PENDING); \ 14 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_BAL_KEEP); \ 15 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_BYPASSING); \ 16 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_CLK_VALID); \ [all …]
|
| /linux/kernel/bpf/preload/iterators/ |
| H A D | iterators.lskel-little-endian.h | 24 iterators_bpf__dump_bpf_map__attach(struct iterators_bpf *skel) in iterators_bpf__dump_bpf_map__attach() argument 26 int prog_fd = skel->progs.dump_bpf_map.prog_fd; in iterators_bpf__dump_bpf_map__attach() 30 skel->links.dump_bpf_map_fd = fd; in iterators_bpf__dump_bpf_map__attach() 35 iterators_bpf__dump_bpf_prog__attach(struct iterators_bpf *skel) in iterators_bpf__dump_bpf_prog__attach() argument 37 int prog_fd = skel->progs.dump_bpf_prog.prog_fd; in iterators_bpf__dump_bpf_prog__attach() 41 skel->links.dump_bpf_prog_fd = fd; in iterators_bpf__dump_bpf_prog__attach() 46 iterators_bpf__attach(struct iterators_bpf *skel) in iterators_bpf__attach() argument 50 ret = ret < 0 ? ret : iterators_bpf__dump_bpf_map__attach(skel); in iterators_bpf__attach() 51 ret = ret < 0 ? ret : iterators_bpf__dump_bpf_prog__attach(skel); in iterators_bpf__attach() 56 iterators_bpf__detach(struct iterators_bpf *skel) in iterators_bpf__detach() argument [all …]
|
| H A D | iterators.lskel-big-endian.h | 24 iterators_bpf__dump_bpf_map__attach(struct iterators_bpf *skel) in iterators_bpf__dump_bpf_map__attach() argument 26 int prog_fd = skel->progs.dump_bpf_map.prog_fd; in iterators_bpf__dump_bpf_map__attach() 30 skel->links.dump_bpf_map_fd = fd; in iterators_bpf__dump_bpf_map__attach() 35 iterators_bpf__dump_bpf_prog__attach(struct iterators_bpf *skel) in iterators_bpf__dump_bpf_prog__attach() argument 37 int prog_fd = skel->progs.dump_bpf_prog.prog_fd; in iterators_bpf__dump_bpf_prog__attach() 41 skel->links.dump_bpf_prog_fd = fd; in iterators_bpf__dump_bpf_prog__attach() 46 iterators_bpf__attach(struct iterators_bpf *skel) in iterators_bpf__attach() argument 50 ret = ret < 0 ? ret : iterators_bpf__dump_bpf_map__attach(skel); in iterators_bpf__attach() 51 ret = ret < 0 ? ret : iterators_bpf__dump_bpf_prog__attach(skel); in iterators_bpf__attach() 56 iterators_bpf__detach(struct iterators_bpf *skel) in iterators_bpf__detach() argument [all …]
|
| /linux/tools/sched_ext/ |
| H A D | scx_qmap.c | 58 struct scx_qmap *skel; in main() local 66 skel = SCX_OPS_OPEN(qmap_ops, scx_qmap); in main() 68 skel->rodata->slice_ns = __COMPAT_ENUM_OR_ZERO("scx_public_consts", "SCX_SLICE_DFL"); in main() 73 skel->rodata->slice_ns = strtoull(optarg, NULL, 0) * 1000; in main() 76 skel->bss->test_error_cnt = strtoul(optarg, NULL, 0); in main() 79 skel->rodata->stall_user_nth = strtoul(optarg, NULL, 0); in main() 82 skel->rodata->stall_kernel_nth = strtoul(optarg, NULL, 0); in main() 85 skel->rodata->dsp_inf_loop_after = strtoul(optarg, NULL, 0); in main() 88 skel->rodata->dsp_batch = strtoul(optarg, NULL, 0); in main() 91 skel->rodata->print_dsqs_and_events = true; in main() [all …]
|