1 // SPDX-License-Identifier: GPL-2.0 2 #include <stdbool.h> 3 #include <inttypes.h> 4 #include <stdlib.h> 5 #include <string.h> 6 #include <linux/bitops.h> 7 #include <linux/kernel.h> 8 #include <linux/types.h> 9 10 #include "map_symbol.h" 11 #include "branch.h" 12 #include "event.h" 13 #include "evsel.h" 14 #include "debug.h" 15 #include "util/synthetic-events.h" 16 #include "util/util.h" 17 18 #include "tests.h" 19 20 #define COMP(m) do { \ 21 if (s1->m != s2->m) { \ 22 pr_debug("Samples differ at '"#m"'\n"); \ 23 return false; \ 24 } \ 25 } while (0) 26 27 #define MCOMP(m) do { \ 28 if (memcmp(&s1->m, &s2->m, sizeof(s1->m))) { \ 29 pr_debug("Samples differ at '"#m"'\n"); \ 30 return false; \ 31 } \ 32 } while (0) 33 34 /* 35 * Hardcode the expected values for branch_entry flags. 36 * These are based on the input value (213) specified 37 * in branch_stack variable. 38 */ 39 #define BS_EXPECTED_BE 0xa000d00000000000 40 #define BS_EXPECTED_LE 0x1aa00000000 41 #define FLAG(s) s->branch_stack->entries[i].flags 42 43 static bool samples_same(struct perf_sample *s1, 44 struct perf_sample *s2, 45 u64 type, u64 read_format, bool needs_swap) 46 { 47 size_t i; 48 49 if (type & PERF_SAMPLE_IDENTIFIER) 50 COMP(id); 51 52 if (type & PERF_SAMPLE_IP) 53 COMP(ip); 54 55 if (type & PERF_SAMPLE_TID) { 56 COMP(pid); 57 COMP(tid); 58 } 59 60 if (type & PERF_SAMPLE_TIME) 61 COMP(time); 62 63 if (type & PERF_SAMPLE_ADDR) 64 COMP(addr); 65 66 if (type & PERF_SAMPLE_ID) 67 COMP(id); 68 69 if (type & PERF_SAMPLE_STREAM_ID) 70 COMP(stream_id); 71 72 if (type & PERF_SAMPLE_CPU) 73 COMP(cpu); 74 75 if (type & PERF_SAMPLE_PERIOD) 76 COMP(period); 77 78 if (type & PERF_SAMPLE_READ) { 79 if (read_format & PERF_FORMAT_GROUP) 80 COMP(read.group.nr); 81 else 82 COMP(read.one.value); 83 if (read_format & PERF_FORMAT_TOTAL_TIME_ENABLED) 84 COMP(read.time_enabled); 85 if (read_format & PERF_FORMAT_TOTAL_TIME_RUNNING) 86 COMP(read.time_running); 87 /* PERF_FORMAT_ID is forced for PERF_SAMPLE_READ */ 88 if (read_format & PERF_FORMAT_GROUP) { 89 for (i = 0; i < s1->read.group.nr; i++) { 90 /* FIXME: check values without LOST */ 91 if (read_format & PERF_FORMAT_LOST) 92 MCOMP(read.group.values[i]); 93 } 94 } else { 95 COMP(read.one.id); 96 if (read_format & PERF_FORMAT_LOST) 97 COMP(read.one.lost); 98 } 99 } 100 101 if (type & PERF_SAMPLE_CALLCHAIN) { 102 COMP(callchain->nr); 103 for (i = 0; i < s1->callchain->nr; i++) 104 COMP(callchain->ips[i]); 105 } 106 107 if (type & PERF_SAMPLE_RAW) { 108 COMP(raw_size); 109 if (memcmp(s1->raw_data, s2->raw_data, s1->raw_size)) { 110 pr_debug("Samples differ at 'raw_data'\n"); 111 return false; 112 } 113 } 114 115 if (type & PERF_SAMPLE_BRANCH_STACK) { 116 COMP(branch_stack->nr); 117 COMP(branch_stack->hw_idx); 118 for (i = 0; i < s1->branch_stack->nr; i++) { 119 if (needs_swap) 120 return ((host_is_bigendian()) ? 121 (FLAG(s2).value == BS_EXPECTED_BE) : 122 (FLAG(s2).value == BS_EXPECTED_LE)); 123 else 124 MCOMP(branch_stack->entries[i]); 125 } 126 } 127 128 if (type & PERF_SAMPLE_REGS_USER) { 129 struct regs_dump *s1_regs = perf_sample__user_regs(s1); 130 struct regs_dump *s2_regs = perf_sample__user_regs(s2); 131 size_t sz = hweight_long(s1_regs->mask) * sizeof(u64); 132 133 COMP(user_regs->mask); 134 COMP(user_regs->abi); 135 if (s1_regs->abi && 136 (!s1_regs->regs || !s2_regs->regs || 137 memcmp(s1_regs->regs, s2_regs->regs, sz))) { 138 pr_debug("Samples differ at 'user_regs'\n"); 139 return false; 140 } 141 } 142 143 if (type & PERF_SAMPLE_STACK_USER) { 144 COMP(user_stack.size); 145 if (memcmp(s1->user_stack.data, s2->user_stack.data, 146 s1->user_stack.size)) { 147 pr_debug("Samples differ at 'user_stack'\n"); 148 return false; 149 } 150 } 151 152 if (type & PERF_SAMPLE_WEIGHT) 153 COMP(weight); 154 155 if (type & PERF_SAMPLE_DATA_SRC) 156 COMP(data_src); 157 158 if (type & PERF_SAMPLE_TRANSACTION) 159 COMP(transaction); 160 161 if (type & PERF_SAMPLE_REGS_INTR) { 162 struct regs_dump *s1_regs = perf_sample__intr_regs(s1); 163 struct regs_dump *s2_regs = perf_sample__intr_regs(s2); 164 size_t sz = hweight_long(s1_regs->mask) * sizeof(u64); 165 166 COMP(intr_regs->mask); 167 COMP(intr_regs->abi); 168 if (s1_regs->abi && 169 (!s1_regs->regs || !s2_regs->regs || 170 memcmp(s1_regs->regs, s2_regs->regs, sz))) { 171 pr_debug("Samples differ at 'intr_regs'\n"); 172 return false; 173 } 174 } 175 176 if (type & PERF_SAMPLE_PHYS_ADDR) 177 COMP(phys_addr); 178 179 if (type & PERF_SAMPLE_CGROUP) 180 COMP(cgroup); 181 182 if (type & PERF_SAMPLE_DATA_PAGE_SIZE) 183 COMP(data_page_size); 184 185 if (type & PERF_SAMPLE_CODE_PAGE_SIZE) 186 COMP(code_page_size); 187 188 if (type & PERF_SAMPLE_AUX) { 189 COMP(aux_sample.size); 190 if (memcmp(s1->aux_sample.data, s2->aux_sample.data, 191 s1->aux_sample.size)) { 192 pr_debug("Samples differ at 'aux_sample'\n"); 193 return false; 194 } 195 } 196 197 return true; 198 } 199 200 static int do_test(u64 sample_type, u64 sample_regs, u64 read_format) 201 { 202 struct evsel evsel = { 203 .needs_swap = false, 204 .core = { 205 . attr = { 206 .sample_type = sample_type, 207 .read_format = read_format, 208 }, 209 }, 210 }; 211 union perf_event *event; 212 union { 213 struct ip_callchain callchain; 214 u64 data[64]; 215 } callchain = { 216 /* 3 ips */ 217 .data = {3, 201, 202, 203}, 218 }; 219 union { 220 struct branch_stack branch_stack; 221 u64 data[64]; 222 } branch_stack = { 223 /* 1 branch_entry */ 224 .data = {1, -1ULL, 211, 212, 213}, 225 }; 226 u64 regs[64]; 227 const u32 raw_data[] = {0x12345678, 0x0a0b0c0d, 0x11020304, 0x05060708, 0 }; 228 const u64 data[] = {0x2211443366558877ULL, 0, 0xaabbccddeeff4321ULL}; 229 const u64 aux_data[] = {0xa55a, 0, 0xeeddee, 0x0282028202820282}; 230 struct regs_dump user_regs = { 231 .abi = PERF_SAMPLE_REGS_ABI_64, 232 .mask = sample_regs, 233 .regs = regs, 234 }; 235 struct regs_dump intr_regs = { 236 .abi = PERF_SAMPLE_REGS_ABI_64, 237 .mask = sample_regs, 238 .regs = regs, 239 }; 240 struct perf_sample sample = { 241 .ip = 101, 242 .pid = 102, 243 .tid = 103, 244 .time = 104, 245 .addr = 105, 246 .id = 106, 247 .stream_id = 107, 248 .period = 108, 249 .weight = 109, 250 .cpu = 110, 251 .raw_size = sizeof(raw_data), 252 .data_src = 111, 253 .transaction = 112, 254 .raw_data = (void *)raw_data, 255 .callchain = &callchain.callchain, 256 .no_hw_idx = false, 257 .branch_stack = &branch_stack.branch_stack, 258 .user_regs = &user_regs, 259 .user_stack = { 260 .size = sizeof(data), 261 .data = (void *)data, 262 }, 263 .read = { 264 .time_enabled = 0x030a59d664fca7deULL, 265 .time_running = 0x011b6ae553eb98edULL, 266 }, 267 .intr_regs = &intr_regs, 268 .phys_addr = 113, 269 .cgroup = 114, 270 .data_page_size = 115, 271 .code_page_size = 116, 272 .aux_sample = { 273 .size = sizeof(aux_data), 274 .data = (void *)aux_data, 275 }, 276 }; 277 struct sample_read_value values[] = {{1, 5, 0}, {9, 3, 0}, {2, 7, 0}, {6, 4, 1},}; 278 struct perf_sample sample_out, sample_out_endian; 279 size_t i, sz, bufsz; 280 int err, ret = -1; 281 282 perf_sample__init(&sample_out, /*all=*/false); 283 perf_sample__init(&sample_out_endian, /*all=*/false); 284 if (sample_type & PERF_SAMPLE_REGS_USER) 285 evsel.core.attr.sample_regs_user = sample_regs; 286 287 if (sample_type & PERF_SAMPLE_REGS_INTR) 288 evsel.core.attr.sample_regs_intr = sample_regs; 289 290 if (sample_type & PERF_SAMPLE_BRANCH_STACK) 291 evsel.core.attr.branch_sample_type |= PERF_SAMPLE_BRANCH_HW_INDEX; 292 293 for (i = 0; i < sizeof(regs); i++) 294 *(i + (u8 *)regs) = i & 0xfe; 295 296 if (read_format & PERF_FORMAT_GROUP) { 297 sample.read.group.nr = 4; 298 sample.read.group.values = values; 299 } else { 300 sample.read.one.value = 0x08789faeb786aa87ULL; 301 sample.read.one.id = 99; 302 sample.read.one.lost = 1; 303 } 304 305 sz = perf_event__sample_event_size(&sample, sample_type, read_format); 306 bufsz = sz + 4096; /* Add a bit for overrun checking */ 307 event = malloc(bufsz); 308 if (!event) { 309 pr_debug("malloc failed\n"); 310 return -1; 311 } 312 313 memset(event, 0xff, bufsz); 314 event->header.type = PERF_RECORD_SAMPLE; 315 event->header.misc = 0; 316 event->header.size = sz; 317 318 err = perf_event__synthesize_sample(event, sample_type, read_format, 319 &sample); 320 if (err) { 321 pr_debug("%s failed for sample_type %#"PRIx64", error %d\n", 322 "perf_event__synthesize_sample", sample_type, err); 323 goto out_free; 324 } 325 326 /* The data does not contain 0xff so we use that to check the size */ 327 for (i = bufsz; i > 0; i--) { 328 if (*(i - 1 + (u8 *)event) != 0xff) 329 break; 330 } 331 if (i != sz) { 332 pr_debug("Event size mismatch: actual %zu vs expected %zu\n", 333 i, sz); 334 goto out_free; 335 } 336 337 evsel.sample_size = __evsel__sample_size(sample_type); 338 339 err = evsel__parse_sample(&evsel, event, &sample_out); 340 if (err) { 341 pr_debug("%s failed for sample_type %#"PRIx64", error %d\n", 342 "evsel__parse_sample", sample_type, err); 343 goto out_free; 344 } 345 346 if (!samples_same(&sample, &sample_out, sample_type, read_format, evsel.needs_swap)) { 347 pr_debug("parsing failed for sample_type %#"PRIx64"\n", 348 sample_type); 349 goto out_free; 350 } 351 352 if (sample_type == PERF_SAMPLE_BRANCH_STACK) { 353 evsel.needs_swap = true; 354 evsel.sample_size = __evsel__sample_size(sample_type); 355 err = evsel__parse_sample(&evsel, event, &sample_out_endian); 356 if (err) { 357 pr_debug("%s failed for sample_type %#"PRIx64", error %d\n", 358 "evsel__parse_sample", sample_type, err); 359 goto out_free; 360 } 361 362 if (!samples_same(&sample, &sample_out_endian, sample_type, read_format, evsel.needs_swap)) { 363 pr_debug("parsing failed for sample_type %#"PRIx64"\n", 364 sample_type); 365 goto out_free; 366 } 367 } 368 369 ret = 0; 370 out_free: 371 free(event); 372 perf_sample__exit(&sample_out_endian); 373 perf_sample__exit(&sample_out); 374 if (ret && read_format) 375 pr_debug("read_format %#"PRIx64"\n", read_format); 376 return ret; 377 } 378 379 /** 380 * test__sample_parsing - test sample parsing. 381 * 382 * This function implements a test that synthesizes a sample event, parses it 383 * and then checks that the parsed sample matches the original sample. The test 384 * checks sample format bits separately and together. If the test passes %0 is 385 * returned, otherwise %-1 is returned. 386 */ 387 static int test__sample_parsing(struct test_suite *test __maybe_unused, int subtest __maybe_unused) 388 { 389 const u64 rf[] = {4, 5, 6, 7, 12, 13, 14, 15, 20, 21, 22, 28, 29, 30, 31}; 390 u64 sample_type; 391 u64 sample_regs; 392 size_t i; 393 int err; 394 395 /* 396 * Fail the test if it has not been updated when new sample format bits 397 * were added. Please actually update the test rather than just change 398 * the condition below. 399 */ 400 if (PERF_SAMPLE_MAX > PERF_SAMPLE_WEIGHT_STRUCT << 1) { 401 pr_debug("sample format has changed, some new PERF_SAMPLE_ bit was introduced - test needs updating\n"); 402 return -1; 403 } 404 405 /* Test each sample format bit separately */ 406 for (sample_type = 1; sample_type != PERF_SAMPLE_MAX; 407 sample_type <<= 1) { 408 /* Test read_format variations */ 409 if (sample_type == PERF_SAMPLE_READ) { 410 for (i = 0; i < ARRAY_SIZE(rf); i++) { 411 err = do_test(sample_type, 0, rf[i]); 412 if (err) 413 return err; 414 } 415 continue; 416 } 417 sample_regs = 0; 418 419 if (sample_type == PERF_SAMPLE_REGS_USER) 420 sample_regs = 0x3fff; 421 422 if (sample_type == PERF_SAMPLE_REGS_INTR) 423 sample_regs = 0xff0fff; 424 425 err = do_test(sample_type, sample_regs, 0); 426 if (err) 427 return err; 428 } 429 430 /* 431 * Test all sample format bits together 432 * Note: PERF_SAMPLE_WEIGHT and PERF_SAMPLE_WEIGHT_STRUCT cannot 433 * be set simultaneously. 434 */ 435 sample_type = (PERF_SAMPLE_MAX - 1) & ~PERF_SAMPLE_WEIGHT; 436 sample_regs = 0x3fff; /* shared yb intr and user regs */ 437 for (i = 0; i < ARRAY_SIZE(rf); i++) { 438 err = do_test(sample_type, sample_regs, rf[i]); 439 if (err) 440 return err; 441 } 442 443 return 0; 444 } 445 446 DEFINE_SUITE("Sample parsing", sample_parsing); 447