xref: /linux/tools/testing/selftests/bpf/prog_tests/map_kptr.c (revision 4f13d0dabc87fb585b96d90cc4b29f67a2995405)
1 // SPDX-License-Identifier: GPL-2.0
2 #include <test_progs.h>
3 #include <network_helpers.h>
4 
5 #include "map_kptr.skel.h"
6 #include "map_kptr_fail.skel.h"
7 #include "rcu_tasks_trace_gp.skel.h"
8 
test_map_kptr_success(bool test_run)9 static void test_map_kptr_success(bool test_run)
10 {
11 	LIBBPF_OPTS(bpf_test_run_opts, lopts);
12 	LIBBPF_OPTS(bpf_test_run_opts, opts,
13 		.data_in = &pkt_v4,
14 		.data_size_in = sizeof(pkt_v4),
15 		.repeat = 1,
16 	);
17 	int key = 0, ret, cpu;
18 	struct map_kptr *skel;
19 	char buf[16], *pbuf;
20 
21 	skel = map_kptr__open_and_load();
22 	if (!ASSERT_OK_PTR(skel, "map_kptr__open_and_load"))
23 		return;
24 
25 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref1), &opts);
26 	ASSERT_OK(ret, "test_map_kptr_ref1 refcount");
27 	ASSERT_OK(opts.retval, "test_map_kptr_ref1 retval");
28 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref2), &opts);
29 	ASSERT_OK(ret, "test_map_kptr_ref2 refcount");
30 	ASSERT_OK(opts.retval, "test_map_kptr_ref2 retval");
31 
32 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_ls_map_kptr_ref1), &lopts);
33 	ASSERT_OK(ret, "test_ls_map_kptr_ref1 refcount");
34 	ASSERT_OK(lopts.retval, "test_ls_map_kptr_ref1 retval");
35 
36 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_ls_map_kptr_ref2), &lopts);
37 	ASSERT_OK(ret, "test_ls_map_kptr_ref2 refcount");
38 	ASSERT_OK(lopts.retval, "test_ls_map_kptr_ref2 retval");
39 
40 	if (test_run)
41 		goto exit;
42 
43 	cpu = libbpf_num_possible_cpus();
44 	if (!ASSERT_GT(cpu, 0, "libbpf_num_possible_cpus"))
45 		goto exit;
46 
47 	pbuf = calloc(cpu, sizeof(buf));
48 	if (!ASSERT_OK_PTR(pbuf, "calloc(pbuf)"))
49 		goto exit;
50 
51 	ret = bpf_map__update_elem(skel->maps.array_map,
52 				   &key, sizeof(key), buf, sizeof(buf), 0);
53 	ASSERT_OK(ret, "array_map update");
54 	skel->data->ref--;
55 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts);
56 	ASSERT_OK(ret, "test_map_kptr_ref3 refcount");
57 	ASSERT_OK(opts.retval, "test_map_kptr_ref3 retval");
58 
59 	ret = bpf_map__update_elem(skel->maps.pcpu_array_map,
60 				   &key, sizeof(key), pbuf, cpu * sizeof(buf), 0);
61 	ASSERT_OK(ret, "pcpu_array_map update");
62 	skel->data->ref--;
63 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts);
64 	ASSERT_OK(ret, "test_map_kptr_ref3 refcount");
65 	ASSERT_OK(opts.retval, "test_map_kptr_ref3 retval");
66 
67 	ret = bpf_map__delete_elem(skel->maps.hash_map, &key, sizeof(key), 0);
68 	ASSERT_OK(ret, "hash_map delete");
69 	skel->data->ref--;
70 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts);
71 	ASSERT_OK(ret, "test_map_kptr_ref3 refcount");
72 	ASSERT_OK(opts.retval, "test_map_kptr_ref3 retval");
73 
74 	ret = bpf_map__delete_elem(skel->maps.pcpu_hash_map, &key, sizeof(key), 0);
75 	ASSERT_OK(ret, "pcpu_hash_map delete");
76 	skel->data->ref--;
77 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts);
78 	ASSERT_OK(ret, "test_map_kptr_ref3 refcount");
79 	ASSERT_OK(opts.retval, "test_map_kptr_ref3 retval");
80 
81 	ret = bpf_map__delete_elem(skel->maps.hash_malloc_map, &key, sizeof(key), 0);
82 	ASSERT_OK(ret, "hash_malloc_map delete");
83 	skel->data->ref--;
84 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts);
85 	ASSERT_OK(ret, "test_map_kptr_ref3 refcount");
86 	ASSERT_OK(opts.retval, "test_map_kptr_ref3 retval");
87 
88 	ret = bpf_map__delete_elem(skel->maps.pcpu_hash_malloc_map, &key, sizeof(key), 0);
89 	ASSERT_OK(ret, "pcpu_hash_malloc_map delete");
90 	skel->data->ref--;
91 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts);
92 	ASSERT_OK(ret, "test_map_kptr_ref3 refcount");
93 	ASSERT_OK(opts.retval, "test_map_kptr_ref3 retval");
94 
95 	ret = bpf_map__delete_elem(skel->maps.lru_hash_map, &key, sizeof(key), 0);
96 	ASSERT_OK(ret, "lru_hash_map delete");
97 	skel->data->ref--;
98 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts);
99 	ASSERT_OK(ret, "test_map_kptr_ref3 refcount");
100 	ASSERT_OK(opts.retval, "test_map_kptr_ref3 retval");
101 
102 	ret = bpf_map__delete_elem(skel->maps.lru_pcpu_hash_map, &key, sizeof(key), 0);
103 	ASSERT_OK(ret, "lru_pcpu_hash_map delete");
104 	skel->data->ref--;
105 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts);
106 	ASSERT_OK(ret, "test_map_kptr_ref3 refcount");
107 	ASSERT_OK(opts.retval, "test_map_kptr_ref3 retval");
108 
109 	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_ls_map_kptr_ref_del), &lopts);
110 	ASSERT_OK(ret, "test_ls_map_kptr_ref_del delete");
111 	skel->data->ref--;
112 	ASSERT_OK(lopts.retval, "test_ls_map_kptr_ref_del retval");
113 
114 	free(pbuf);
115 exit:
116 	map_kptr__destroy(skel);
117 }
118 
kern_sync_rcu_tasks_trace(struct rcu_tasks_trace_gp * rcu)119 static int kern_sync_rcu_tasks_trace(struct rcu_tasks_trace_gp *rcu)
120 {
121 	LIBBPF_OPTS(bpf_test_run_opts, opts);
122 	int ret;
123 
124 	WRITE_ONCE(rcu->bss->done, 0);
125 	ret = bpf_prog_test_run_opts(bpf_program__fd(rcu->progs.call_rcu_tasks_trace), &opts);
126 	if (!ASSERT_OK(ret, "call_rcu_tasks_trace"))
127 		return -EFAULT;
128 	if (!ASSERT_OK(opts.retval, "call_rcu_tasks_trace retval"))
129 		return -EFAULT;
130 	while (!READ_ONCE(rcu->bss->done))
131 		sched_yield();
132 	return 0;
133 }
134 
wait_for_map_release(void)135 static void wait_for_map_release(void)
136 {
137 	LIBBPF_OPTS(bpf_test_run_opts, lopts);
138 	struct map_kptr *skel;
139 	int ret;
140 
141 	skel = map_kptr__open_and_load();
142 	if (!ASSERT_OK_PTR(skel, "map_kptr__open_and_load"))
143 		return;
144 
145 	do {
146 		ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.count_ref), &lopts);
147 		ASSERT_OK(ret, "count_ref ret");
148 		ASSERT_OK(lopts.retval, "count_ref retval");
149 	} while (skel->bss->num_of_refs != 2);
150 
151 	map_kptr__destroy(skel);
152 }
153 
serial_test_map_kptr(void)154 void serial_test_map_kptr(void)
155 {
156 	struct rcu_tasks_trace_gp *skel;
157 
158 	RUN_TESTS(map_kptr_fail);
159 
160 	skel = rcu_tasks_trace_gp__open_and_load();
161 	if (!ASSERT_OK_PTR(skel, "rcu_tasks_trace_gp__open_and_load"))
162 		return;
163 
164 	if (test__start_subtest("success-map")) {
165 		test_map_kptr_success(true);
166 
167 		ASSERT_OK(kern_sync_rcu_tasks_trace(skel), "sync rcu_tasks_trace");
168 		ASSERT_OK(kern_sync_rcu(), "sync rcu");
169 		wait_for_map_release();
170 
171 		/* Observe refcount dropping to 1 on bpf_map_free_deferred */
172 		test_map_kptr_success(false);
173 
174 		ASSERT_OK(kern_sync_rcu_tasks_trace(skel), "sync rcu_tasks_trace");
175 		ASSERT_OK(kern_sync_rcu(), "sync rcu");
176 		wait_for_map_release();
177 
178 		/* Observe refcount dropping to 1 on synchronous delete elem */
179 		test_map_kptr_success(true);
180 	}
181 
182 	rcu_tasks_trace_gp__destroy(skel);
183 }
184