xref: /linux/tools/perf/tests/sample-parsing.c (revision 8520a98dbab61e9e340cdfb72dd17ccc8a98961e)
1 // SPDX-License-Identifier: GPL-2.0
2 #include <stdbool.h>
3 #include <inttypes.h>
4 #include <stdlib.h>
5 #include <string.h>
6 #include <linux/bitops.h>
7 #include <linux/kernel.h>
8 #include <linux/types.h>
9 
10 #include "branch.h"
11 #include "util.h"
12 #include "event.h"
13 #include "evsel.h"
14 #include "debug.h"
15 
16 #include "tests.h"
17 
18 #define COMP(m) do {					\
19 	if (s1->m != s2->m) {				\
20 		pr_debug("Samples differ at '"#m"'\n");	\
21 		return false;				\
22 	}						\
23 } while (0)
24 
25 #define MCOMP(m) do {					\
26 	if (memcmp(&s1->m, &s2->m, sizeof(s1->m))) {	\
27 		pr_debug("Samples differ at '"#m"'\n");	\
28 		return false;				\
29 	}						\
30 } while (0)
31 
32 static bool samples_same(const struct perf_sample *s1,
33 			 const struct perf_sample *s2,
34 			 u64 type, u64 read_format)
35 {
36 	size_t i;
37 
38 	if (type & PERF_SAMPLE_IDENTIFIER)
39 		COMP(id);
40 
41 	if (type & PERF_SAMPLE_IP)
42 		COMP(ip);
43 
44 	if (type & PERF_SAMPLE_TID) {
45 		COMP(pid);
46 		COMP(tid);
47 	}
48 
49 	if (type & PERF_SAMPLE_TIME)
50 		COMP(time);
51 
52 	if (type & PERF_SAMPLE_ADDR)
53 		COMP(addr);
54 
55 	if (type & PERF_SAMPLE_ID)
56 		COMP(id);
57 
58 	if (type & PERF_SAMPLE_STREAM_ID)
59 		COMP(stream_id);
60 
61 	if (type & PERF_SAMPLE_CPU)
62 		COMP(cpu);
63 
64 	if (type & PERF_SAMPLE_PERIOD)
65 		COMP(period);
66 
67 	if (type & PERF_SAMPLE_READ) {
68 		if (read_format & PERF_FORMAT_GROUP)
69 			COMP(read.group.nr);
70 		else
71 			COMP(read.one.value);
72 		if (read_format & PERF_FORMAT_TOTAL_TIME_ENABLED)
73 			COMP(read.time_enabled);
74 		if (read_format & PERF_FORMAT_TOTAL_TIME_RUNNING)
75 			COMP(read.time_running);
76 		/* PERF_FORMAT_ID is forced for PERF_SAMPLE_READ */
77 		if (read_format & PERF_FORMAT_GROUP) {
78 			for (i = 0; i < s1->read.group.nr; i++)
79 				MCOMP(read.group.values[i]);
80 		} else {
81 			COMP(read.one.id);
82 		}
83 	}
84 
85 	if (type & PERF_SAMPLE_CALLCHAIN) {
86 		COMP(callchain->nr);
87 		for (i = 0; i < s1->callchain->nr; i++)
88 			COMP(callchain->ips[i]);
89 	}
90 
91 	if (type & PERF_SAMPLE_RAW) {
92 		COMP(raw_size);
93 		if (memcmp(s1->raw_data, s2->raw_data, s1->raw_size)) {
94 			pr_debug("Samples differ at 'raw_data'\n");
95 			return false;
96 		}
97 	}
98 
99 	if (type & PERF_SAMPLE_BRANCH_STACK) {
100 		COMP(branch_stack->nr);
101 		for (i = 0; i < s1->branch_stack->nr; i++)
102 			MCOMP(branch_stack->entries[i]);
103 	}
104 
105 	if (type & PERF_SAMPLE_REGS_USER) {
106 		size_t sz = hweight_long(s1->user_regs.mask) * sizeof(u64);
107 
108 		COMP(user_regs.mask);
109 		COMP(user_regs.abi);
110 		if (s1->user_regs.abi &&
111 		    (!s1->user_regs.regs || !s2->user_regs.regs ||
112 		     memcmp(s1->user_regs.regs, s2->user_regs.regs, sz))) {
113 			pr_debug("Samples differ at 'user_regs'\n");
114 			return false;
115 		}
116 	}
117 
118 	if (type & PERF_SAMPLE_STACK_USER) {
119 		COMP(user_stack.size);
120 		if (memcmp(s1->user_stack.data, s2->user_stack.data,
121 			   s1->user_stack.size)) {
122 			pr_debug("Samples differ at 'user_stack'\n");
123 			return false;
124 		}
125 	}
126 
127 	if (type & PERF_SAMPLE_WEIGHT)
128 		COMP(weight);
129 
130 	if (type & PERF_SAMPLE_DATA_SRC)
131 		COMP(data_src);
132 
133 	if (type & PERF_SAMPLE_TRANSACTION)
134 		COMP(transaction);
135 
136 	if (type & PERF_SAMPLE_REGS_INTR) {
137 		size_t sz = hweight_long(s1->intr_regs.mask) * sizeof(u64);
138 
139 		COMP(intr_regs.mask);
140 		COMP(intr_regs.abi);
141 		if (s1->intr_regs.abi &&
142 		    (!s1->intr_regs.regs || !s2->intr_regs.regs ||
143 		     memcmp(s1->intr_regs.regs, s2->intr_regs.regs, sz))) {
144 			pr_debug("Samples differ at 'intr_regs'\n");
145 			return false;
146 		}
147 	}
148 
149 	if (type & PERF_SAMPLE_PHYS_ADDR)
150 		COMP(phys_addr);
151 
152 	return true;
153 }
154 
155 static int do_test(u64 sample_type, u64 sample_regs, u64 read_format)
156 {
157 	struct evsel evsel = {
158 		.needs_swap = false,
159 		.core = {
160 			. attr = {
161 				.sample_type = sample_type,
162 				.read_format = read_format,
163 			},
164 		},
165 	};
166 	union perf_event *event;
167 	union {
168 		struct ip_callchain callchain;
169 		u64 data[64];
170 	} callchain = {
171 		/* 3 ips */
172 		.data = {3, 201, 202, 203},
173 	};
174 	union {
175 		struct branch_stack branch_stack;
176 		u64 data[64];
177 	} branch_stack = {
178 		/* 1 branch_entry */
179 		.data = {1, 211, 212, 213},
180 	};
181 	u64 regs[64];
182 	const u64 raw_data[] = {0x123456780a0b0c0dULL, 0x1102030405060708ULL};
183 	const u64 data[] = {0x2211443366558877ULL, 0, 0xaabbccddeeff4321ULL};
184 	struct perf_sample sample = {
185 		.ip		= 101,
186 		.pid		= 102,
187 		.tid		= 103,
188 		.time		= 104,
189 		.addr		= 105,
190 		.id		= 106,
191 		.stream_id	= 107,
192 		.period		= 108,
193 		.weight		= 109,
194 		.cpu		= 110,
195 		.raw_size	= sizeof(raw_data),
196 		.data_src	= 111,
197 		.transaction	= 112,
198 		.raw_data	= (void *)raw_data,
199 		.callchain	= &callchain.callchain,
200 		.branch_stack	= &branch_stack.branch_stack,
201 		.user_regs	= {
202 			.abi	= PERF_SAMPLE_REGS_ABI_64,
203 			.mask	= sample_regs,
204 			.regs	= regs,
205 		},
206 		.user_stack	= {
207 			.size	= sizeof(data),
208 			.data	= (void *)data,
209 		},
210 		.read		= {
211 			.time_enabled = 0x030a59d664fca7deULL,
212 			.time_running = 0x011b6ae553eb98edULL,
213 		},
214 		.intr_regs	= {
215 			.abi	= PERF_SAMPLE_REGS_ABI_64,
216 			.mask	= sample_regs,
217 			.regs	= regs,
218 		},
219 		.phys_addr	= 113,
220 	};
221 	struct sample_read_value values[] = {{1, 5}, {9, 3}, {2, 7}, {6, 4},};
222 	struct perf_sample sample_out;
223 	size_t i, sz, bufsz;
224 	int err, ret = -1;
225 
226 	if (sample_type & PERF_SAMPLE_REGS_USER)
227 		evsel.core.attr.sample_regs_user = sample_regs;
228 
229 	if (sample_type & PERF_SAMPLE_REGS_INTR)
230 		evsel.core.attr.sample_regs_intr = sample_regs;
231 
232 	for (i = 0; i < sizeof(regs); i++)
233 		*(i + (u8 *)regs) = i & 0xfe;
234 
235 	if (read_format & PERF_FORMAT_GROUP) {
236 		sample.read.group.nr     = 4;
237 		sample.read.group.values = values;
238 	} else {
239 		sample.read.one.value = 0x08789faeb786aa87ULL;
240 		sample.read.one.id    = 99;
241 	}
242 
243 	sz = perf_event__sample_event_size(&sample, sample_type, read_format);
244 	bufsz = sz + 4096; /* Add a bit for overrun checking */
245 	event = malloc(bufsz);
246 	if (!event) {
247 		pr_debug("malloc failed\n");
248 		return -1;
249 	}
250 
251 	memset(event, 0xff, bufsz);
252 	event->header.type = PERF_RECORD_SAMPLE;
253 	event->header.misc = 0;
254 	event->header.size = sz;
255 
256 	err = perf_event__synthesize_sample(event, sample_type, read_format,
257 					    &sample);
258 	if (err) {
259 		pr_debug("%s failed for sample_type %#"PRIx64", error %d\n",
260 			 "perf_event__synthesize_sample", sample_type, err);
261 		goto out_free;
262 	}
263 
264 	/* The data does not contain 0xff so we use that to check the size */
265 	for (i = bufsz; i > 0; i--) {
266 		if (*(i - 1 + (u8 *)event) != 0xff)
267 			break;
268 	}
269 	if (i != sz) {
270 		pr_debug("Event size mismatch: actual %zu vs expected %zu\n",
271 			 i, sz);
272 		goto out_free;
273 	}
274 
275 	evsel.sample_size = __perf_evsel__sample_size(sample_type);
276 
277 	err = perf_evsel__parse_sample(&evsel, event, &sample_out);
278 	if (err) {
279 		pr_debug("%s failed for sample_type %#"PRIx64", error %d\n",
280 			 "perf_evsel__parse_sample", sample_type, err);
281 		goto out_free;
282 	}
283 
284 	if (!samples_same(&sample, &sample_out, sample_type, read_format)) {
285 		pr_debug("parsing failed for sample_type %#"PRIx64"\n",
286 			 sample_type);
287 		goto out_free;
288 	}
289 
290 	ret = 0;
291 out_free:
292 	free(event);
293 	if (ret && read_format)
294 		pr_debug("read_format %#"PRIx64"\n", read_format);
295 	return ret;
296 }
297 
298 /**
299  * test__sample_parsing - test sample parsing.
300  *
301  * This function implements a test that synthesizes a sample event, parses it
302  * and then checks that the parsed sample matches the original sample.  The test
303  * checks sample format bits separately and together.  If the test passes %0 is
304  * returned, otherwise %-1 is returned.
305  */
306 int test__sample_parsing(struct test *test __maybe_unused, int subtest __maybe_unused)
307 {
308 	const u64 rf[] = {4, 5, 6, 7, 12, 13, 14, 15};
309 	u64 sample_type;
310 	u64 sample_regs;
311 	size_t i;
312 	int err;
313 
314 	/*
315 	 * Fail the test if it has not been updated when new sample format bits
316 	 * were added.  Please actually update the test rather than just change
317 	 * the condition below.
318 	 */
319 	if (PERF_SAMPLE_MAX > PERF_SAMPLE_PHYS_ADDR << 1) {
320 		pr_debug("sample format has changed, some new PERF_SAMPLE_ bit was introduced - test needs updating\n");
321 		return -1;
322 	}
323 
324 	/* Test each sample format bit separately */
325 	for (sample_type = 1; sample_type != PERF_SAMPLE_MAX;
326 	     sample_type <<= 1) {
327 		/* Test read_format variations */
328 		if (sample_type == PERF_SAMPLE_READ) {
329 			for (i = 0; i < ARRAY_SIZE(rf); i++) {
330 				err = do_test(sample_type, 0, rf[i]);
331 				if (err)
332 					return err;
333 			}
334 			continue;
335 		}
336 		sample_regs = 0;
337 
338 		if (sample_type == PERF_SAMPLE_REGS_USER)
339 			sample_regs = 0x3fff;
340 
341 		if (sample_type == PERF_SAMPLE_REGS_INTR)
342 			sample_regs = 0xff0fff;
343 
344 		err = do_test(sample_type, sample_regs, 0);
345 		if (err)
346 			return err;
347 	}
348 
349 	/* Test all sample format bits together */
350 	sample_type = PERF_SAMPLE_MAX - 1;
351 	sample_regs = 0x3fff; /* shared yb intr and user regs */
352 	for (i = 0; i < ARRAY_SIZE(rf); i++) {
353 		err = do_test(sample_type, sample_regs, rf[i]);
354 		if (err)
355 			return err;
356 	}
357 
358 	return 0;
359 }
360