Home
last modified time | relevance | path

Searched refs:skel (Results 1 – 25 of 327) sorted by relevance

12345678910>>...14

/linux/tools/testing/selftests/bpf/prog_tests/
H A Dtracing_struct.c5 #include "tracing_struct.skel.h"
6 #include "tracing_struct_many_args.skel.h"
10 struct tracing_struct *skel; in test_struct_args() local
13 skel = tracing_struct__open_and_load(); in test_struct_args()
14 if (!ASSERT_OK_PTR(skel, "tracing_struct__open_and_load")) in test_struct_args()
17 err = tracing_struct__attach(skel); in test_struct_args()
23 ASSERT_EQ(skel->bss->t1_a_a, 2, "t1:a.a"); in test_struct_args()
24 ASSERT_EQ(skel->bss->t1_a_b, 3, "t1:a.b"); in test_struct_args()
25 ASSERT_EQ(skel->bss->t1_b, 1, "t1:b"); in test_struct_args()
26 ASSERT_EQ(skel in test_struct_args()
65 struct tracing_struct_many_args *skel; test_struct_many_args() local
[all...]
H A Dattach_probe.c3 #include "test_attach_kprobe_sleepable.skel.h"
4 #include "test_attach_probe_manual.skel.h"
5 #include "test_attach_probe.skel.h"
6 #include "kprobe_write_ctx.skel.h"
44 struct test_attach_probe_manual *skel; in test_attach_probe_manual()
47 skel = test_attach_probe_manual__open_and_load(); in test_attach_probe_manual()
48 if (!ASSERT_OK_PTR(skel, "skel_kprobe_manual_open_and_load")) in test_attach_probe_manual()
58 kprobe_link = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe, in test_attach_probe_manual()
63 skel->links.handle_kprobe = kprobe_link; in test_attach_probe_manual()
66 kretprobe_link = bpf_program__attach_kprobe_opts(skel in test_attach_probe_manual()
43 struct test_attach_probe_manual *skel; test_attach_probe_manual() local
125 test_attach_probe_auto(struct test_attach_probe * skel) test_attach_probe_auto() argument
159 test_uprobe_lib(struct test_attach_probe * skel) test_uprobe_lib() argument
195 test_uprobe_ref_ctr(struct test_attach_probe * skel) test_uprobe_ref_ctr() argument
238 struct test_attach_kprobe_sleepable *skel; test_kprobe_sleepable() local
261 test_uprobe_sleepable(struct test_attach_probe * skel) test_uprobe_sleepable() argument
295 struct test_attach_probe *skel; test_attach_probe() local
[all...]
H A Datomics.c7 static void test_add(struct atomics_lskel *skel) in test_add() argument
13 prog_fd = skel->progs.add.prog_fd; in test_add()
20 ASSERT_EQ(skel->data->add64_value, 3, "add64_value"); in test_add()
21 ASSERT_EQ(skel->bss->add64_result, 1, "add64_result"); in test_add()
23 ASSERT_EQ(skel->data->add32_value, 3, "add32_value"); in test_add()
24 ASSERT_EQ(skel->bss->add32_result, 1, "add32_result"); in test_add()
26 ASSERT_EQ(skel->bss->add_stack_value_copy, 3, "add_stack_value"); in test_add()
27 ASSERT_EQ(skel->bss->add_stack_result, 1, "add_stack_result"); in test_add()
29 ASSERT_EQ(skel->data->add_noreturn_value, 3, "add_noreturn_value"); in test_add()
32 static void test_sub(struct atomics_lskel *skel) in test_sub() argument
57 test_and(struct atomics_lskel * skel) test_and() argument
79 test_or(struct atomics_lskel * skel) test_or() argument
101 test_xor(struct atomics_lskel * skel) test_xor() argument
123 test_cmpxchg(struct atomics_lskel * skel) test_cmpxchg() argument
145 test_xchg(struct atomics_lskel * skel) test_xchg() argument
167 struct atomics_lskel *skel; test_atomics() local
[all...]
H A Dtest_struct_ops_module.c39 static int attach_ops_and_check(struct struct_ops_module *skel, in attach_ops_and_check() argument
51 ASSERT_EQ(skel->bss->test_1_result, 0xdeadbeef, "test_1_result"); in attach_ops_and_check()
52 ASSERT_EQ(skel->bss->test_2_result, expected_test_2_result, "test_2_result"); in attach_ops_and_check()
60 struct struct_ops_module *skel; in test_struct_ops_load() local
65 skel = struct_ops_module__open(); in test_struct_ops_load()
66 if (!ASSERT_OK_PTR(skel, "struct_ops_module_open")) in test_struct_ops_load()
69 skel->struct_ops.testmod_1->data = 13; in test_struct_ops_load()
70 skel->struct_ops.testmod_1->test_2 = skel->progs.test_3; in test_struct_ops_load()
74 bpf_program__set_autoload(skel->progs.test_2, false); in test_struct_ops_load()
75 bpf_map__set_autocreate(skel->maps.testmod_zeroed, false); in test_struct_ops_load()
[all …]
H A Dtest_strncmp.c6 static int trigger_strncmp(const struct strncmp_test *skel) in trigger_strncmp() argument
12 cmp = skel->bss->cmp_ret; in trigger_strncmp()
24 static void strncmp_full_str_cmp(struct strncmp_test *skel, const char *name, in strncmp_full_str_cmp() argument
27 size_t nr = sizeof(skel->bss->str); in strncmp_full_str_cmp()
28 char *str = skel->bss->str; in strncmp_full_str_cmp()
33 memcpy(str, skel->rodata->target, nr); in strncmp_full_str_cmp()
37 got = trigger_strncmp(skel); in strncmp_full_str_cmp()
46 struct strncmp_test *skel; in test_strncmp_ret() local
49 skel = strncmp_test__open(); in test_strncmp_ret()
50 if (!ASSERT_OK_PTR(skel, "strncmp_test open")) in test_strncmp_ret()
[all …]
H A Darena_atomics.c4 #include "arena_atomics.skel.h"
6 static void test_add(struct arena_atomics *skel) in test_add() argument
12 prog_fd = bpf_program__fd(skel->progs.add); in test_add()
19 ASSERT_EQ(skel->arena->add64_value, 3, "add64_value"); in test_add()
20 ASSERT_EQ(skel->arena->add64_result, 1, "add64_result"); in test_add()
22 ASSERT_EQ(skel->arena->add32_value, 3, "add32_value"); in test_add()
23 ASSERT_EQ(skel->arena->add32_result, 1, "add32_result"); in test_add()
25 ASSERT_EQ(skel->arena->add_stack_value_copy, 3, "add_stack_value"); in test_add()
26 ASSERT_EQ(skel->arena->add_stack_result, 1, "add_stack_result"); in test_add()
28 ASSERT_EQ(skel in test_add()
31 test_sub(struct arena_atomics * skel) test_sub() argument
56 test_and(struct arena_atomics * skel) test_and() argument
73 test_or(struct arena_atomics * skel) test_or() argument
90 test_xor(struct arena_atomics * skel) test_xor() argument
107 test_cmpxchg(struct arena_atomics * skel) test_cmpxchg() argument
129 test_xchg(struct arena_atomics * skel) test_xchg() argument
149 test_uaf(struct arena_atomics * skel) test_uaf() argument
167 struct arena_atomics *skel; test_arena_atomics() local
[all...]
H A Dbpf_iter.c39 struct bpf_iter_test_kern3 *skel; in test_btf_id_or_null() local
41 skel = bpf_iter_test_kern3__open_and_load(); in test_btf_id_or_null()
42 if (!ASSERT_ERR_PTR(skel, "bpf_iter_test_kern3__open_and_load")) { in test_btf_id_or_null()
43 bpf_iter_test_kern3__destroy(skel); in test_btf_id_or_null()
78 static void do_read_map_iter_fd(struct bpf_object_skeleton **skel, struct bpf_program *prog, in do_read_map_iter_fd() argument
103 bpf_object__destroy_skeleton(*skel); in do_read_map_iter_fd()
104 *skel = NULL; in do_read_map_iter_fd()
140 struct bpf_iter_ipv6_route *skel; in test_ipv6_route() local
142 skel = bpf_iter_ipv6_route__open_and_load(); in test_ipv6_route()
143 if (!ASSERT_OK_PTR(skel, "bpf_iter_ipv6_route__open_and_load")) in test_ipv6_route()
[all …]
H A Dbpf_loop.c8 static void check_nr_loops(struct bpf_loop *skel) in check_nr_loops() argument
12 link = bpf_program__attach(skel->progs.test_prog); in check_nr_loops()
17 skel->bss->nr_loops = 0; in check_nr_loops()
21 ASSERT_EQ(skel->bss->nr_loops_returned, skel->bss->nr_loops, in check_nr_loops()
25 skel->bss->nr_loops = 500; in check_nr_loops()
29 ASSERT_EQ(skel->bss->nr_loops_returned, skel->bss->nr_loops, in check_nr_loops()
31 ASSERT_EQ(skel->bss->g_output, (500 * 499) / 2, "g_output"); in check_nr_loops()
34 skel->bss->nr_loops = -1; in check_nr_loops()
38 ASSERT_EQ(skel->bss->err, -E2BIG, "over max limit"); in check_nr_loops()
43 static void check_callback_fn_stop(struct bpf_loop *skel) in check_callback_fn_stop() argument
[all …]
H A Dstruct_ops_autocreate.c9 struct struct_ops_autocreate *skel; in cant_load_full_object() local
13 skel = struct_ops_autocreate__open(); in cant_load_full_object()
14 if (!ASSERT_OK_PTR(skel, "struct_ops_autocreate__open")) in cant_load_full_object()
24 err = struct_ops_autocreate__load(skel); in cant_load_full_object()
34 struct_ops_autocreate__destroy(skel); in cant_load_full_object()
37 static int check_test_1_link(struct struct_ops_autocreate *skel, struct bpf_map *map) in check_test_1_link() argument
42 link = bpf_map__attach_struct_ops(skel->maps.testmod_1); in check_test_1_link()
47 err = ASSERT_EQ(skel->bss->test_1_result, 42, "test_1_result"); in check_test_1_link()
54 struct struct_ops_autocreate *skel; in can_load_partial_object() local
57 skel = struct_ops_autocreate__open(); in can_load_partial_object()
[all …]
H A Dtc_links.c11 #include "test_tc_link.skel.h"
22 struct test_tc_link *skel; in serial_test_tc_links_basic() local
26 skel = test_tc_link__open_and_load(); in serial_test_tc_links_basic()
27 if (!ASSERT_OK_PTR(skel, "skel_load")) in serial_test_tc_links_basic()
30 pid1 = id_from_prog_fd(bpf_program__fd(skel->progs.tc1)); in serial_test_tc_links_basic()
31 pid2 = id_from_prog_fd(bpf_program__fd(skel->progs.tc2)); in serial_test_tc_links_basic()
38 ASSERT_EQ(skel->bss->seen_tc1, false, "seen_tc1"); in serial_test_tc_links_basic()
39 ASSERT_EQ(skel->bss->seen_tc2, false, "seen_tc2"); in serial_test_tc_links_basic()
41 link = bpf_program__attach_tcx(skel->progs.tc1, loopback, &optl); in serial_test_tc_links_basic()
45 skel in serial_test_tc_links_basic()
122 struct test_tc_link *skel; test_tc_links_before_target() local
276 struct test_tc_link *skel; test_tc_links_after_target() local
429 struct test_tc_link *skel; test_tc_links_revision_target() local
530 struct test_tc_link *skel; test_tc_chain_classic() local
635 struct test_tc_link *skel; test_tc_links_replace_target() local
860 struct test_tc_link *skel; test_tc_links_invalid_target() local
1174 struct test_tc_link *skel; test_tc_links_prepend_target() local
1330 struct test_tc_link *skel; test_tc_links_append_target() local
1484 struct test_tc_link *skel; test_tc_links_dev_cleanup_target() local
1582 struct test_tc_link *skel; test_tc_chain_mixed() local
1696 struct test_tc_link *skel; test_tc_links_ingress() local
1857 struct test_tc_link *skel; test_tc_links_dev_mixed() local
[all...]
H A Dmap_ops.c46 static int setup(struct test_map_ops **skel) in setup() argument
50 if (!skel) in setup()
53 *skel = test_map_ops__open(); in setup()
54 if (!ASSERT_OK_PTR(*skel, "test_map_ops__open")) in setup()
57 (*skel)->rodata->pid = getpid(); in setup()
59 err = test_map_ops__load(*skel); in setup()
63 err = test_map_ops__attach(*skel); in setup()
70 static void teardown(struct test_map_ops **skel) in teardown() argument
72 if (skel && *skel) in teardown()
73 test_map_ops__destroy(*skel); in teardown()
[all …]
H A Diters.c27 struct iters_num *skel; in subtest_num_iters() local
30 skel = iters_num__open_and_load(); in subtest_num_iters()
31 if (!ASSERT_OK_PTR(skel, "skel_open_and_load")) in subtest_num_iters()
34 err = iters_num__attach(skel); in subtest_num_iters()
39 iters_num__detach(skel); in subtest_num_iters()
42 ASSERT_EQ(skel->bss->res_##case_name, \ in subtest_num_iters()
43 skel->rodata->exp_##case_name, \ in subtest_num_iters()
67 iters_num__destroy(skel); in subtest_num_iters()
72 struct iters_testmod_seq *skel; in subtest_testmod_seq_iters() local
80 skel = iters_testmod_seq__open_and_load(); in subtest_testmod_seq_iters()
[all …]
H A Dbpf_cookie.c12 #include "test_bpf_cookie.skel.h"
13 #include "kprobe_multi.skel.h"
14 #include "uprobe_multi.skel.h"
22 static void kprobe_subtest(struct test_bpf_cookie *skel) in kprobe_subtest() argument
31 link1 = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe, in kprobe_subtest()
38 link2 = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe, in kprobe_subtest()
46 retlink1 = bpf_program__attach_kprobe_opts(skel->progs.handle_kretprobe, in kprobe_subtest()
53 retlink2 = bpf_program__attach_kprobe_opts(skel->progs.handle_kretprobe, in kprobe_subtest()
61 ASSERT_EQ(skel->bss->kprobe_res, 0x1 | 0x2, "kprobe_res"); in kprobe_subtest()
62 ASSERT_EQ(skel in kprobe_subtest()
71 kprobe_multi_test_run(struct kprobe_multi * skel) kprobe_multi_test_run() argument
103 struct kprobe_multi *skel = NULL; kprobe_multi_link_api_subtest() local
182 struct kprobe_multi *skel = NULL; kprobe_multi_attach_api_subtest() local
249 uprobe_multi_test_run(struct uprobe_multi * skel) uprobe_multi_test_run() argument
274 struct uprobe_multi *skel = NULL; uprobe_multi_attach_api_subtest() local
318 uprobe_subtest(struct test_bpf_cookie * skel) uprobe_subtest() argument
372 tp_subtest(struct test_bpf_cookie * skel) tp_subtest() argument
441 pe_subtest(struct test_bpf_cookie * skel) pe_subtest() argument
510 tracing_subtest(struct test_bpf_cookie * skel) tracing_subtest() argument
567 lsm_subtest(struct test_bpf_cookie * skel) lsm_subtest() argument
598 tp_btf_subtest(struct test_bpf_cookie * skel) tp_btf_subtest() argument
679 raw_tp_subtest(struct test_bpf_cookie * skel) raw_tp_subtest() argument
733 struct test_bpf_cookie *skel; test_bpf_cookie() local
[all...]
H A Duprobe_multi_test.c194 static void uprobe_multi_test_run(struct uprobe_multi *skel, struct child *child) in uprobe_multi_test_run() argument
196 skel->bss->uprobe_multi_func_1_addr = (__u64) uprobe_multi_func_1; in uprobe_multi_test_run()
197 skel->bss->uprobe_multi_func_2_addr = (__u64) uprobe_multi_func_2; in uprobe_multi_test_run()
198 skel->bss->uprobe_multi_func_3_addr = (__u64) uprobe_multi_func_3; in uprobe_multi_test_run()
200 skel->bss->user_ptr = test_data; in uprobe_multi_test_run()
207 skel->bss->pid = child ? 0 : getpid(); in uprobe_multi_test_run()
208 skel->bss->expect_pid = child ? child->pid : 0; in uprobe_multi_test_run()
229 ASSERT_EQ(skel->bss->uprobe_multi_func_1_result, 2, "uprobe_multi_func_1_result"); in uprobe_multi_test_run()
230 ASSERT_EQ(skel->bss->uprobe_multi_func_2_result, 2, "uprobe_multi_func_2_result"); in uprobe_multi_test_run()
231 ASSERT_EQ(skel->bss->uprobe_multi_func_3_result, 2, "uprobe_multi_func_3_result"); in uprobe_multi_test_run()
[all …]
H A Dfind_vma.c10 static void test_and_reset_skel(struct find_vma *skel, int expected_find_zero_ret, bool need_test) in test_and_reset_skel() argument
13 ASSERT_EQ(skel->bss->found_vm_exec, 1, "found_vm_exec"); in test_and_reset_skel()
14 ASSERT_EQ(skel->data->find_addr_ret, 0, "find_addr_ret"); in test_and_reset_skel()
15 ASSERT_EQ(skel->data->find_zero_ret, expected_find_zero_ret, "find_zero_ret"); in test_and_reset_skel()
16 ASSERT_OK_PTR(strstr(skel->bss->d_iname, "test_progs"), "find_test_progs"); in test_and_reset_skel()
19 skel->bss->found_vm_exec = 0; in test_and_reset_skel()
20 skel->data->find_addr_ret = -1; in test_and_reset_skel()
21 skel->data->find_zero_ret = -1; in test_and_reset_skel()
22 skel->bss->d_iname[0] = 0; in test_and_reset_skel()
41 static bool find_vma_pe_condition(struct find_vma *skel) in find_vma_pe_condition() argument
[all …]
H A Dcgroup1_hierarchy.c10 static void bpf_cgroup1(struct test_cgroup1_hierarchy *skel) in bpf_cgroup1() argument
16 lsm_link = bpf_program__attach_lsm(skel->progs.lsm_run); in bpf_cgroup1()
21 fentry_link = bpf_program__attach_trace(skel->progs.fentry_run); in bpf_cgroup1()
28 static void bpf_cgroup1_sleepable(struct test_cgroup1_hierarchy *skel) in bpf_cgroup1_sleepable() argument
34 lsm_link = bpf_program__attach_lsm(skel->progs.lsm_s_run); in bpf_cgroup1_sleepable()
39 fentry_link = bpf_program__attach_trace(skel->progs.fentry_run); in bpf_cgroup1_sleepable()
46 static void bpf_cgroup1_invalid_id(struct test_cgroup1_hierarchy *skel) in bpf_cgroup1_invalid_id() argument
52 lsm_link = bpf_program__attach_lsm(skel->progs.lsm_run); in bpf_cgroup1_invalid_id()
57 fentry_link = bpf_program__attach_trace(skel->progs.fentry_run); in bpf_cgroup1_invalid_id()
71 struct test_cgroup1_hierarchy *skel; in test_cgroup1_hierarchy() local
[all …]
H A Dbtf_tag.c17 struct test_btf_decl_tag *skel; in test_btf_decl_tag() local
19 skel = test_btf_decl_tag__open_and_load(); in test_btf_decl_tag()
20 if (!ASSERT_OK_PTR(skel, "btf_decl_tag")) in test_btf_decl_tag()
23 if (skel->rodata->skip_tests) { in test_btf_decl_tag()
28 test_btf_decl_tag__destroy(skel); in test_btf_decl_tag()
33 struct btf_type_tag *skel; in test_btf_type_tag() local
35 skel = btf_type_tag__open_and_load(); in test_btf_type_tag()
36 if (!ASSERT_OK_PTR(skel, "btf_type_tag")) in test_btf_type_tag()
39 if (skel->rodata->skip_tests) { in test_btf_type_tag()
44 btf_type_tag__destroy(skel); in test_btf_type_tag()
[all …]
H A Dglobal_map_resize.c22 struct test_global_map_resize *skel; in global_map_resize_bss_subtest() local
24 const __u32 desired_sz = sizeof(skel->bss->sum) + sysconf(_SC_PAGE_SIZE) * 2; in global_map_resize_bss_subtest()
27 skel = test_global_map_resize__open(); in global_map_resize_bss_subtest()
28 if (!ASSERT_OK_PTR(skel, "test_global_map_resize__open")) in global_map_resize_bss_subtest()
35 skel->bss->array[0] = 1; in global_map_resize_bss_subtest()
38 map = skel->maps.bss; in global_map_resize_bss_subtest()
45 new_sz = sizeof(skel->data_percpu_arr->percpu_arr[0]) * libbpf_num_possible_cpus(); in global_map_resize_bss_subtest()
46 err = bpf_map__set_value_size(skel->maps.data_percpu_arr, new_sz); in global_map_resize_bss_subtest()
50 array_len = (desired_sz - sizeof(skel->bss->sum)) / sizeof(skel->bss->array[0]); in global_map_resize_bss_subtest()
54 skel->bss = bpf_map__initial_value(skel->maps.bss, &actual_sz); in global_map_resize_bss_subtest()
[all …]
H A Drbtree.c7 #include "rbtree.skel.h"
8 #include "rbtree_fail.skel.h"
9 #include "rbtree_btf_fail__wrong_node_type.skel.h"
10 #include "rbtree_btf_fail__add_wrong_type.skel.h"
11 #include "rbtree_search.skel.h"
20 struct rbtree *skel; in test_rbtree_add_nodes()
23 skel = rbtree__open_and_load(); in test_rbtree_add_nodes()
24 if (!ASSERT_OK_PTR(skel, "rbtree__open_and_load")) in test_rbtree_add_nodes()
27 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.rbtree_add_nodes), &opts); in test_rbtree_add_nodes()
30 ASSERT_EQ(skel in test_rbtree_add_nodes()
19 struct rbtree *skel; test_rbtree_add_nodes() local
41 struct rbtree *skel; test_rbtree_add_nodes_nested() local
63 struct rbtree *skel; test_rbtree_add_and_remove() local
85 struct rbtree *skel; test_rbtree_add_and_remove_array() local
106 struct rbtree *skel; test_rbtree_first_and_remove() local
130 struct rbtree *skel; test_rbtree_api_release_aliasing() local
[all...]
H A Dpercpu_alloc.c9 struct percpu_alloc_array *skel; in test_array() local
13 skel = percpu_alloc_array__open(); in test_array()
14 if (!ASSERT_OK_PTR(skel, "percpu_alloc_array__open")) in test_array()
17 bpf_program__set_autoload(skel->progs.test_array_map_1, true); in test_array()
18 bpf_program__set_autoload(skel->progs.test_array_map_2, true); in test_array()
19 bpf_program__set_autoload(skel->progs.test_array_map_3, true); in test_array()
20 bpf_program__set_autoload(skel->progs.test_array_map_4, true); in test_array()
22 skel->bss->my_pid = getpid(); in test_array()
23 skel->rodata->nr_cpus = libbpf_num_possible_cpus(); in test_array()
25 err = percpu_alloc_array__load(skel); in test_array()
[all …]
H A Dcheck_mtu.c6 #include "test_check_mtu.skel.h"
43 struct test_check_mtu *skel; in test_check_mtu_xdp_attach() local
49 skel = test_check_mtu__open_and_load(); in test_check_mtu_xdp_attach()
50 if (CHECK(!skel, "open and load skel", "failed")) in test_check_mtu_xdp_attach()
53 prog = skel->progs.xdp_use_helper_basic; in test_check_mtu_xdp_attach()
58 skel->links.xdp_use_helper_basic = link; in test_check_mtu_xdp_attach()
75 test_check_mtu__destroy(skel); in test_check_mtu_xdp_attach()
78 static void test_check_mtu_run_xdp(struct test_check_mtu *skel, in test_check_mtu_run_xdp() argument
99 mtu_result = skel in test_check_mtu_run_xdp()
106 struct test_check_mtu *skel; test_check_mtu_xdp() local
131 test_check_mtu_run_tc(struct test_check_mtu * skel,struct bpf_program * prog,__u32 mtu_expect) test_check_mtu_run_tc() argument
159 struct test_check_mtu *skel; test_check_mtu_tc() local
[all...]
/linux/tools/sched_ext/include/scx/
H A Denums.autogen.h7 #define SCX_ENUM_INIT(skel) do { \ argument
8 SCX_ENUM_SET(skel, scx_public_consts, SCX_OPS_NAME_LEN); \
9 SCX_ENUM_SET(skel, scx_public_consts, SCX_SLICE_DFL); \
10 SCX_ENUM_SET(skel, scx_public_consts, SCX_SLICE_INF); \
11 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_ONLINE); \
12 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_CAN_STOP_TICK); \
13 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_BAL_PENDING); \
14 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_BAL_KEEP); \
15 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_BYPASSING); \
16 SCX_ENUM_SET(skel, scx_rq_flags, SCX_RQ_CLK_VALID); \
[all …]
/linux/kernel/bpf/preload/iterators/
H A Diterators.lskel-little-endian.h24 iterators_bpf__dump_bpf_map__attach(struct iterators_bpf *skel) in iterators_bpf__dump_bpf_map__attach() argument
26 int prog_fd = skel->progs.dump_bpf_map.prog_fd; in iterators_bpf__dump_bpf_map__attach()
30 skel->links.dump_bpf_map_fd = fd; in iterators_bpf__dump_bpf_map__attach()
35 iterators_bpf__dump_bpf_prog__attach(struct iterators_bpf *skel) in iterators_bpf__dump_bpf_prog__attach() argument
37 int prog_fd = skel->progs.dump_bpf_prog.prog_fd; in iterators_bpf__dump_bpf_prog__attach()
41 skel->links.dump_bpf_prog_fd = fd; in iterators_bpf__dump_bpf_prog__attach()
46 iterators_bpf__attach(struct iterators_bpf *skel) in iterators_bpf__attach() argument
50 ret = ret < 0 ? ret : iterators_bpf__dump_bpf_map__attach(skel); in iterators_bpf__attach()
51 ret = ret < 0 ? ret : iterators_bpf__dump_bpf_prog__attach(skel); in iterators_bpf__attach()
56 iterators_bpf__detach(struct iterators_bpf *skel) in iterators_bpf__detach() argument
[all …]
H A Diterators.lskel-big-endian.h24 iterators_bpf__dump_bpf_map__attach(struct iterators_bpf *skel) in iterators_bpf__dump_bpf_map__attach() argument
26 int prog_fd = skel->progs.dump_bpf_map.prog_fd; in iterators_bpf__dump_bpf_map__attach()
30 skel->links.dump_bpf_map_fd = fd; in iterators_bpf__dump_bpf_map__attach()
35 iterators_bpf__dump_bpf_prog__attach(struct iterators_bpf *skel) in iterators_bpf__dump_bpf_prog__attach() argument
37 int prog_fd = skel->progs.dump_bpf_prog.prog_fd; in iterators_bpf__dump_bpf_prog__attach()
41 skel->links.dump_bpf_prog_fd = fd; in iterators_bpf__dump_bpf_prog__attach()
46 iterators_bpf__attach(struct iterators_bpf *skel) in iterators_bpf__attach() argument
50 ret = ret < 0 ? ret : iterators_bpf__dump_bpf_map__attach(skel); in iterators_bpf__attach()
51 ret = ret < 0 ? ret : iterators_bpf__dump_bpf_prog__attach(skel); in iterators_bpf__attach()
56 iterators_bpf__detach(struct iterators_bpf *skel) in iterators_bpf__detach() argument
[all …]
/linux/tools/sched_ext/
H A Dscx_qmap.c58 struct scx_qmap *skel; in main() local
66 skel = SCX_OPS_OPEN(qmap_ops, scx_qmap); in main()
68 skel->rodata->slice_ns = __COMPAT_ENUM_OR_ZERO("scx_public_consts", "SCX_SLICE_DFL"); in main()
73 skel->rodata->slice_ns = strtoull(optarg, NULL, 0) * 1000; in main()
76 skel->bss->test_error_cnt = strtoul(optarg, NULL, 0); in main()
79 skel->rodata->stall_user_nth = strtoul(optarg, NULL, 0); in main()
82 skel->rodata->stall_kernel_nth = strtoul(optarg, NULL, 0); in main()
85 skel->rodata->dsp_inf_loop_after = strtoul(optarg, NULL, 0); in main()
88 skel->rodata->dsp_batch = strtoul(optarg, NULL, 0); in main()
91 skel->rodata->print_dsqs_and_events = true; in main()
[all …]

12345678910>>...14