1 /* SPDX-License-Identifier: GPL-2.0-only */ 2 /* 3 * auxtrace.h: AUX area trace support 4 * Copyright (c) 2013-2015, Intel Corporation. 5 */ 6 7 #ifndef __PERF_AUXTRACE_H 8 #define __PERF_AUXTRACE_H 9 10 #include <sys/types.h> 11 #include <errno.h> 12 #include <stdbool.h> 13 #include <stddef.h> 14 #include <stdio.h> // FILE 15 #include <linux/list.h> 16 #include <linux/perf_event.h> 17 #include <linux/types.h> 18 #include <perf/cpumap.h> 19 #include <asm/bitsperlong.h> 20 #include <asm/barrier.h> 21 22 union perf_event; 23 struct perf_session; 24 struct evlist; 25 struct evsel; 26 struct perf_tool; 27 struct mmap; 28 struct perf_sample; 29 struct option; 30 struct record_opts; 31 struct perf_record_auxtrace_error; 32 struct perf_record_auxtrace_info; 33 struct events_stats; 34 struct perf_pmu; 35 36 enum auxtrace_error_type { 37 PERF_AUXTRACE_ERROR_ITRACE = 1, 38 PERF_AUXTRACE_ERROR_MAX 39 }; 40 41 /* Auxtrace records must have the same alignment as perf event records */ 42 #define PERF_AUXTRACE_RECORD_ALIGNMENT 8 43 44 enum auxtrace_type { 45 PERF_AUXTRACE_UNKNOWN, 46 PERF_AUXTRACE_INTEL_PT, 47 PERF_AUXTRACE_INTEL_BTS, 48 PERF_AUXTRACE_CS_ETM, 49 PERF_AUXTRACE_ARM_SPE, 50 PERF_AUXTRACE_S390_CPUMSF, 51 PERF_AUXTRACE_HISI_PTT, 52 }; 53 54 enum itrace_period_type { 55 PERF_ITRACE_PERIOD_INSTRUCTIONS, 56 PERF_ITRACE_PERIOD_TICKS, 57 PERF_ITRACE_PERIOD_NANOSECS, 58 }; 59 60 #define AUXTRACE_ERR_FLG_OVERFLOW (1 << ('o' - 'a')) 61 #define AUXTRACE_ERR_FLG_DATA_LOST (1 << ('l' - 'a')) 62 63 #define AUXTRACE_LOG_FLG_ALL_PERF_EVTS (1 << ('a' - 'a')) 64 #define AUXTRACE_LOG_FLG_ON_ERROR (1 << ('e' - 'a')) 65 #define AUXTRACE_LOG_FLG_USE_STDOUT (1 << ('o' - 'a')) 66 67 /** 68 * struct itrace_synth_opts - AUX area tracing synthesis options. 69 * @set: indicates whether or not options have been set 70 * @default_no_sample: Default to no sampling. 71 * @inject: indicates the event (not just the sample) must be fully synthesized 72 * because 'perf inject' will write it out 73 * @instructions: whether to synthesize 'instructions' events 74 * @cycles: whether to synthesize 'cycles' events 75 * (not fully accurate, since CYC packets are only emitted 76 * together with other events, such as branches) 77 * @branches: whether to synthesize 'branches' events 78 * (branch misses only for Arm SPE) 79 * @transactions: whether to synthesize events for transactions 80 * @ptwrites: whether to synthesize events for ptwrites 81 * @pwr_events: whether to synthesize power events 82 * @other_events: whether to synthesize other events recorded due to the use of 83 * aux_output 84 * @intr_events: whether to synthesize interrupt events 85 * @errors: whether to synthesize decoder error events 86 * @dont_decode: whether to skip decoding entirely 87 * @log: write a decoding log 88 * @calls: limit branch samples to calls (can be combined with @returns) 89 * @returns: limit branch samples to returns (can be combined with @calls) 90 * @callchain: add callchain to 'instructions' events 91 * @add_callchain: add callchain to existing event records 92 * @thread_stack: feed branches to the thread_stack 93 * @last_branch: add branch context to 'instruction' events 94 * @add_last_branch: add branch context to existing event records 95 * @approx_ipc: approximate IPC 96 * @flc: whether to synthesize first level cache events 97 * @llc: whether to synthesize last level cache events 98 * @tlb: whether to synthesize TLB events 99 * @remote_access: whether to synthesize remote access events 100 * @mem: whether to synthesize memory events 101 * @timeless_decoding: prefer "timeless" decoding i.e. ignore timestamps 102 * @use_timestamp: use the timestamp trace as kernel time 103 * @vm_time_correlation: perform VM Time Correlation 104 * @vm_tm_corr_dry_run: VM Time Correlation dry-run 105 * @vm_tm_corr_args: VM Time Correlation implementation-specific arguments 106 * @callchain_sz: maximum callchain size 107 * @last_branch_sz: branch context size 108 * @period: 'instructions' events period 109 * @period_type: 'instructions' events period type 110 * @initial_skip: skip N events at the beginning. 111 * @cpu_bitmap: CPUs for which to synthesize events, or NULL for all 112 * @ptime_range: time intervals to trace or NULL 113 * @range_num: number of time intervals to trace 114 * @error_plus_flags: flags to affect what errors are reported 115 * @error_minus_flags: flags to affect what errors are reported 116 * @log_plus_flags: flags to affect what is logged 117 * @log_minus_flags: flags to affect what is logged 118 * @quick: quicker (less detailed) decoding 119 * @log_on_error_size: size of log to keep for outputting log only on errors 120 */ 121 struct itrace_synth_opts { 122 bool set; 123 bool default_no_sample; 124 bool inject; 125 bool instructions; 126 bool cycles; 127 bool branches; 128 bool transactions; 129 bool ptwrites; 130 bool pwr_events; 131 bool other_events; 132 bool intr_events; 133 bool errors; 134 bool dont_decode; 135 bool log; 136 bool calls; 137 bool returns; 138 bool callchain; 139 bool add_callchain; 140 bool thread_stack; 141 bool last_branch; 142 bool add_last_branch; 143 bool approx_ipc; 144 bool flc; 145 bool llc; 146 bool tlb; 147 bool remote_access; 148 bool mem; 149 bool timeless_decoding; 150 bool use_timestamp; 151 bool vm_time_correlation; 152 bool vm_tm_corr_dry_run; 153 char *vm_tm_corr_args; 154 unsigned int callchain_sz; 155 unsigned int last_branch_sz; 156 unsigned long long period; 157 enum itrace_period_type period_type; 158 unsigned long initial_skip; 159 unsigned long *cpu_bitmap; 160 struct perf_time_interval *ptime_range; 161 int range_num; 162 unsigned int error_plus_flags; 163 unsigned int error_minus_flags; 164 unsigned int log_plus_flags; 165 unsigned int log_minus_flags; 166 unsigned int quick; 167 unsigned int log_on_error_size; 168 }; 169 170 /** 171 * struct auxtrace_index_entry - indexes a AUX area tracing event within a 172 * perf.data file. 173 * @file_offset: offset within the perf.data file 174 * @sz: size of the event 175 */ 176 struct auxtrace_index_entry { 177 u64 file_offset; 178 u64 sz; 179 }; 180 181 #define PERF_AUXTRACE_INDEX_ENTRY_COUNT 256 182 183 /** 184 * struct auxtrace_index - index of AUX area tracing events within a perf.data 185 * file. 186 * @list: linking a number of arrays of entries 187 * @nr: number of entries 188 * @entries: array of entries 189 */ 190 struct auxtrace_index { 191 struct list_head list; 192 size_t nr; 193 struct auxtrace_index_entry entries[PERF_AUXTRACE_INDEX_ENTRY_COUNT]; 194 }; 195 196 /** 197 * struct auxtrace - session callbacks to allow AUX area data decoding. 198 * @process_event: lets the decoder see all session events 199 * @process_auxtrace_event: process a PERF_RECORD_AUXTRACE event 200 * @queue_data: queue an AUX sample or PERF_RECORD_AUXTRACE event for later 201 * processing 202 * @dump_auxtrace_sample: dump AUX area sample data 203 * @flush_events: process any remaining data 204 * @free_events: free resources associated with event processing 205 * @free: free resources associated with the session 206 */ 207 struct auxtrace { 208 int (*process_event)(struct perf_session *session, 209 union perf_event *event, 210 struct perf_sample *sample, 211 const struct perf_tool *tool); 212 int (*process_auxtrace_event)(struct perf_session *session, 213 union perf_event *event, 214 const struct perf_tool *tool); 215 int (*queue_data)(struct perf_session *session, 216 struct perf_sample *sample, union perf_event *event, 217 u64 data_offset); 218 void (*dump_auxtrace_sample)(struct perf_session *session, 219 struct perf_sample *sample); 220 int (*flush_events)(struct perf_session *session, 221 const struct perf_tool *tool); 222 void (*free_events)(struct perf_session *session); 223 void (*free)(struct perf_session *session); 224 bool (*evsel_is_auxtrace)(struct perf_session *session, 225 struct evsel *evsel); 226 }; 227 228 /** 229 * struct auxtrace_buffer - a buffer containing AUX area tracing data. 230 * @list: buffers are queued in a list held by struct auxtrace_queue 231 * @size: size of the buffer in bytes 232 * @pid: in per-thread mode, the pid this buffer is associated with 233 * @tid: in per-thread mode, the tid this buffer is associated with 234 * @cpu: in per-cpu mode, the cpu this buffer is associated with 235 * @data: actual buffer data (can be null if the data has not been loaded) 236 * @data_offset: file offset at which the buffer can be read 237 * @mmap_addr: mmap address at which the buffer can be read 238 * @mmap_size: size of the mmap at @mmap_addr 239 * @data_needs_freeing: @data was malloc'd so free it when it is no longer 240 * needed 241 * @consecutive: the original data was split up and this buffer is consecutive 242 * to the previous buffer 243 * @offset: offset as determined by aux_head / aux_tail members of struct 244 * perf_event_mmap_page 245 * @reference: an implementation-specific reference determined when the data is 246 * recorded 247 * @buffer_nr: used to number each buffer 248 * @use_size: implementation actually only uses this number of bytes 249 * @use_data: implementation actually only uses data starting at this address 250 */ 251 struct auxtrace_buffer { 252 struct list_head list; 253 size_t size; 254 pid_t pid; 255 pid_t tid; 256 struct perf_cpu cpu; 257 void *data; 258 off_t data_offset; 259 void *mmap_addr; 260 size_t mmap_size; 261 bool data_needs_freeing; 262 bool consecutive; 263 u64 offset; 264 u64 reference; 265 u64 buffer_nr; 266 size_t use_size; 267 void *use_data; 268 }; 269 270 /** 271 * struct auxtrace_queue - a queue of AUX area tracing data buffers. 272 * @head: head of buffer list 273 * @tid: in per-thread mode, the tid this queue is associated with 274 * @cpu: in per-cpu mode, the cpu this queue is associated with 275 * @set: %true once this queue has been dedicated to a specific thread or cpu 276 * @priv: implementation-specific data 277 */ 278 struct auxtrace_queue { 279 struct list_head head; 280 pid_t tid; 281 int cpu; 282 bool set; 283 void *priv; 284 }; 285 286 /** 287 * struct auxtrace_queues - an array of AUX area tracing queues. 288 * @queue_array: array of queues 289 * @nr_queues: number of queues 290 * @new_data: set whenever new data is queued 291 * @populated: queues have been fully populated using the auxtrace_index 292 * @next_buffer_nr: used to number each buffer 293 */ 294 struct auxtrace_queues { 295 struct auxtrace_queue *queue_array; 296 unsigned int nr_queues; 297 bool new_data; 298 bool populated; 299 u64 next_buffer_nr; 300 }; 301 302 /** 303 * struct auxtrace_heap_item - element of struct auxtrace_heap. 304 * @queue_nr: queue number 305 * @ordinal: value used for sorting (lowest ordinal is top of the heap) expected 306 * to be a timestamp 307 */ 308 struct auxtrace_heap_item { 309 unsigned int queue_nr; 310 u64 ordinal; 311 }; 312 313 /** 314 * struct auxtrace_heap - a heap suitable for sorting AUX area tracing queues. 315 * @heap_array: the heap 316 * @heap_cnt: the number of elements in the heap 317 * @heap_sz: maximum number of elements (grows as needed) 318 */ 319 struct auxtrace_heap { 320 struct auxtrace_heap_item *heap_array; 321 unsigned int heap_cnt; 322 unsigned int heap_sz; 323 }; 324 325 /** 326 * struct auxtrace_mmap - records an mmap of the auxtrace buffer. 327 * @base: address of mapped area 328 * @userpg: pointer to buffer's perf_event_mmap_page 329 * @mask: %0 if @len is not a power of two, otherwise (@len - %1) 330 * @len: size of mapped area 331 * @prev: previous aux_head 332 * @idx: index of this mmap 333 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu 334 * mmap) otherwise %0 335 * @cpu: cpu number for a per-cpu mmap otherwise %-1 336 */ 337 struct auxtrace_mmap { 338 void *base; 339 void *userpg; 340 size_t mask; 341 size_t len; 342 u64 prev; 343 int idx; 344 pid_t tid; 345 int cpu; 346 }; 347 348 /** 349 * struct auxtrace_mmap_params - parameters to set up struct auxtrace_mmap. 350 * @mask: %0 if @len is not a power of two, otherwise (@len - %1) 351 * @offset: file offset of mapped area 352 * @len: size of mapped area 353 * @prot: mmap memory protection 354 * @idx: index of this mmap 355 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu 356 * mmap) otherwise %0 357 * @mmap_needed: set to %false for non-auxtrace events. This is needed because 358 * auxtrace mmapping is done in the same code path as non-auxtrace 359 * mmapping but not every evsel that needs non-auxtrace mmapping 360 * also needs auxtrace mmapping. 361 * @cpu: cpu number for a per-cpu mmap otherwise %-1 362 */ 363 struct auxtrace_mmap_params { 364 size_t mask; 365 off_t offset; 366 size_t len; 367 int prot; 368 int idx; 369 pid_t tid; 370 bool mmap_needed; 371 struct perf_cpu cpu; 372 }; 373 374 /** 375 * struct auxtrace_record - callbacks for recording AUX area data. 376 * @recording_options: validate and process recording options 377 * @info_priv_size: return the size of the private data in auxtrace_info_event 378 * @info_fill: fill-in the private data in auxtrace_info_event 379 * @free: free this auxtrace record structure 380 * @snapshot_start: starting a snapshot 381 * @snapshot_finish: finishing a snapshot 382 * @find_snapshot: find data to snapshot within auxtrace mmap 383 * @parse_snapshot_options: parse snapshot options 384 * @reference: provide a 64-bit reference number for auxtrace_event 385 * @read_finish: called after reading from an auxtrace mmap 386 * @alignment: alignment (if any) for AUX area data 387 * @default_aux_sample_size: default sample size for --aux sample option 388 * @pmu: associated pmu 389 * @evlist: selected events list 390 */ 391 struct auxtrace_record { 392 int (*recording_options)(struct auxtrace_record *itr, 393 struct evlist *evlist, 394 struct record_opts *opts); 395 size_t (*info_priv_size)(struct auxtrace_record *itr, 396 struct evlist *evlist); 397 int (*info_fill)(struct auxtrace_record *itr, 398 struct perf_session *session, 399 struct perf_record_auxtrace_info *auxtrace_info, 400 size_t priv_size); 401 void (*free)(struct auxtrace_record *itr); 402 int (*snapshot_start)(struct auxtrace_record *itr); 403 int (*snapshot_finish)(struct auxtrace_record *itr); 404 int (*find_snapshot)(struct auxtrace_record *itr, int idx, 405 struct auxtrace_mmap *mm, unsigned char *data, 406 u64 *head, u64 *old); 407 int (*parse_snapshot_options)(struct auxtrace_record *itr, 408 struct record_opts *opts, 409 const char *str); 410 u64 (*reference)(struct auxtrace_record *itr); 411 int (*read_finish)(struct auxtrace_record *itr, int idx); 412 unsigned int alignment; 413 unsigned int default_aux_sample_size; 414 struct evlist *evlist; 415 }; 416 417 /** 418 * struct addr_filter - address filter. 419 * @list: list node 420 * @range: true if it is a range filter 421 * @start: true if action is 'filter' or 'start' 422 * @action: 'filter', 'start' or 'stop' ('tracestop' is accepted but converted 423 * to 'stop') 424 * @sym_from: symbol name for the filter address 425 * @sym_to: symbol name that determines the filter size 426 * @sym_from_idx: selects n'th from symbols with the same name (0 means global 427 * and less than 0 means symbol must be unique) 428 * @sym_to_idx: same as @sym_from_idx but for @sym_to 429 * @addr: filter address 430 * @size: filter region size (for range filters) 431 * @filename: DSO file name or NULL for the kernel 432 * @str: allocated string that contains the other string members 433 */ 434 struct addr_filter { 435 struct list_head list; 436 bool range; 437 bool start; 438 const char *action; 439 const char *sym_from; 440 const char *sym_to; 441 int sym_from_idx; 442 int sym_to_idx; 443 u64 addr; 444 u64 size; 445 const char *filename; 446 char *str; 447 }; 448 449 /** 450 * struct addr_filters - list of address filters. 451 * @head: list of address filters 452 * @cnt: number of address filters 453 */ 454 struct addr_filters { 455 struct list_head head; 456 int cnt; 457 }; 458 459 struct auxtrace_cache; 460 461 #ifdef HAVE_AUXTRACE_SUPPORT 462 463 u64 compat_auxtrace_mmap__read_head(struct auxtrace_mmap *mm); 464 int compat_auxtrace_mmap__write_tail(struct auxtrace_mmap *mm, u64 tail); 465 466 static inline u64 auxtrace_mmap__read_head(struct auxtrace_mmap *mm, 467 int kernel_is_64_bit __maybe_unused) 468 { 469 struct perf_event_mmap_page *pc = mm->userpg; 470 u64 head; 471 472 #if BITS_PER_LONG == 32 473 if (kernel_is_64_bit) 474 return compat_auxtrace_mmap__read_head(mm); 475 #endif 476 head = READ_ONCE(pc->aux_head); 477 478 /* Ensure all reads are done after we read the head */ 479 smp_rmb(); 480 return head; 481 } 482 483 static inline int auxtrace_mmap__write_tail(struct auxtrace_mmap *mm, u64 tail, 484 int kernel_is_64_bit __maybe_unused) 485 { 486 struct perf_event_mmap_page *pc = mm->userpg; 487 488 #if BITS_PER_LONG == 32 489 if (kernel_is_64_bit) 490 return compat_auxtrace_mmap__write_tail(mm, tail); 491 #endif 492 /* Ensure all reads are done before we write the tail out */ 493 smp_mb(); 494 WRITE_ONCE(pc->aux_tail, tail); 495 return 0; 496 } 497 498 int auxtrace_mmap__mmap(struct auxtrace_mmap *mm, 499 struct auxtrace_mmap_params *mp, 500 void *userpg, int fd); 501 void auxtrace_mmap__munmap(struct auxtrace_mmap *mm); 502 void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp, 503 off_t auxtrace_offset, 504 unsigned int auxtrace_pages, 505 bool auxtrace_overwrite); 506 void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp, 507 struct evlist *evlist, 508 struct evsel *evsel, int idx); 509 510 typedef int (*process_auxtrace_t)(const struct perf_tool *tool, 511 struct mmap *map, 512 union perf_event *event, void *data1, 513 size_t len1, void *data2, size_t len2); 514 515 int auxtrace_mmap__read(struct mmap *map, struct auxtrace_record *itr, 516 const struct perf_tool *tool, process_auxtrace_t fn); 517 518 int auxtrace_mmap__read_snapshot(struct mmap *map, 519 struct auxtrace_record *itr, 520 const struct perf_tool *tool, process_auxtrace_t fn, 521 size_t snapshot_size); 522 523 int auxtrace_queues__init_nr(struct auxtrace_queues *queues, int nr_queues); 524 int auxtrace_queues__init(struct auxtrace_queues *queues); 525 int auxtrace_queues__add_event(struct auxtrace_queues *queues, 526 struct perf_session *session, 527 union perf_event *event, off_t data_offset, 528 struct auxtrace_buffer **buffer_ptr); 529 struct auxtrace_queue * 530 auxtrace_queues__sample_queue(struct auxtrace_queues *queues, 531 struct perf_sample *sample, 532 struct perf_session *session); 533 int auxtrace_queues__add_sample(struct auxtrace_queues *queues, 534 struct perf_session *session, 535 struct perf_sample *sample, u64 data_offset, 536 u64 reference); 537 void auxtrace_queues__free(struct auxtrace_queues *queues); 538 int auxtrace_queues__process_index(struct auxtrace_queues *queues, 539 struct perf_session *session); 540 int auxtrace_queue_data(struct perf_session *session, bool samples, 541 bool events); 542 struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue, 543 struct auxtrace_buffer *buffer); 544 void *auxtrace_buffer__get_data_rw(struct auxtrace_buffer *buffer, int fd, bool rw); 545 static inline void *auxtrace_buffer__get_data(struct auxtrace_buffer *buffer, int fd) 546 { 547 return auxtrace_buffer__get_data_rw(buffer, fd, false); 548 } 549 void auxtrace_buffer__put_data(struct auxtrace_buffer *buffer); 550 void auxtrace_buffer__drop_data(struct auxtrace_buffer *buffer); 551 void auxtrace_buffer__free(struct auxtrace_buffer *buffer); 552 553 int auxtrace_heap__add(struct auxtrace_heap *heap, unsigned int queue_nr, 554 u64 ordinal); 555 void auxtrace_heap__pop(struct auxtrace_heap *heap); 556 void auxtrace_heap__free(struct auxtrace_heap *heap); 557 558 struct auxtrace_cache_entry { 559 struct hlist_node hash; 560 u32 key; 561 }; 562 563 struct auxtrace_cache *auxtrace_cache__new(unsigned int bits, size_t entry_size, 564 unsigned int limit_percent); 565 void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache); 566 void *auxtrace_cache__alloc_entry(struct auxtrace_cache *c); 567 void auxtrace_cache__free_entry(struct auxtrace_cache *c, void *entry); 568 int auxtrace_cache__add(struct auxtrace_cache *c, u32 key, 569 struct auxtrace_cache_entry *entry); 570 void auxtrace_cache__remove(struct auxtrace_cache *c, u32 key); 571 void *auxtrace_cache__lookup(struct auxtrace_cache *c, u32 key); 572 573 struct auxtrace_record *auxtrace_record__init(struct evlist *evlist, 574 int *err); 575 576 int auxtrace_parse_snapshot_options(struct auxtrace_record *itr, 577 struct record_opts *opts, 578 const char *str); 579 int auxtrace_parse_sample_options(struct auxtrace_record *itr, 580 struct evlist *evlist, 581 struct record_opts *opts, const char *str); 582 void auxtrace_regroup_aux_output(struct evlist *evlist); 583 int auxtrace_record__options(struct auxtrace_record *itr, 584 struct evlist *evlist, 585 struct record_opts *opts); 586 size_t auxtrace_record__info_priv_size(struct auxtrace_record *itr, 587 struct evlist *evlist); 588 int auxtrace_record__info_fill(struct auxtrace_record *itr, 589 struct perf_session *session, 590 struct perf_record_auxtrace_info *auxtrace_info, 591 size_t priv_size); 592 void auxtrace_record__free(struct auxtrace_record *itr); 593 int auxtrace_record__snapshot_start(struct auxtrace_record *itr); 594 int auxtrace_record__snapshot_finish(struct auxtrace_record *itr, bool on_exit); 595 int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx, 596 struct auxtrace_mmap *mm, 597 unsigned char *data, u64 *head, u64 *old); 598 u64 auxtrace_record__reference(struct auxtrace_record *itr); 599 int auxtrace_record__read_finish(struct auxtrace_record *itr, int idx); 600 601 int auxtrace_index__auxtrace_event(struct list_head *head, union perf_event *event, 602 off_t file_offset); 603 int auxtrace_index__write(int fd, struct list_head *head); 604 int auxtrace_index__process(int fd, u64 size, struct perf_session *session, 605 bool needs_swap); 606 void auxtrace_index__free(struct list_head *head); 607 608 void auxtrace_synth_guest_error(struct perf_record_auxtrace_error *auxtrace_error, int type, 609 int code, int cpu, pid_t pid, pid_t tid, u64 ip, 610 const char *msg, u64 timestamp, 611 pid_t machine_pid, int vcpu); 612 void auxtrace_synth_error(struct perf_record_auxtrace_error *auxtrace_error, int type, 613 int code, int cpu, pid_t pid, pid_t tid, u64 ip, 614 const char *msg, u64 timestamp); 615 616 int perf_event__process_auxtrace_info(struct perf_session *session, 617 union perf_event *event); 618 s64 perf_event__process_auxtrace(struct perf_session *session, 619 union perf_event *event); 620 int perf_event__process_auxtrace_error(struct perf_session *session, 621 union perf_event *event); 622 int itrace_do_parse_synth_opts(struct itrace_synth_opts *synth_opts, 623 const char *str, int unset); 624 int itrace_parse_synth_opts(const struct option *opt, const char *str, 625 int unset); 626 void itrace_synth_opts__set_default(struct itrace_synth_opts *synth_opts, 627 bool no_sample); 628 629 size_t perf_event__fprintf_auxtrace_error(union perf_event *event, FILE *fp); 630 void perf_session__auxtrace_error_inc(struct perf_session *session, 631 union perf_event *event); 632 void events_stats__auxtrace_error_warn(const struct events_stats *stats); 633 634 void addr_filters__init(struct addr_filters *filts); 635 void addr_filters__exit(struct addr_filters *filts); 636 int addr_filters__parse_bare_filter(struct addr_filters *filts, 637 const char *filter); 638 int auxtrace_parse_filters(struct evlist *evlist); 639 640 int auxtrace__process_event(struct perf_session *session, union perf_event *event, 641 struct perf_sample *sample, const struct perf_tool *tool); 642 void auxtrace__dump_auxtrace_sample(struct perf_session *session, 643 struct perf_sample *sample); 644 int auxtrace__flush_events(struct perf_session *session, const struct perf_tool *tool); 645 void auxtrace__free_events(struct perf_session *session); 646 void auxtrace__free(struct perf_session *session); 647 bool auxtrace__evsel_is_auxtrace(struct perf_session *session, 648 struct evsel *evsel); 649 650 #define ITRACE_HELP \ 651 " i[period]: synthesize instructions events\n" \ 652 " y[period]: synthesize cycles events (same period as i)\n" \ 653 " b: synthesize branches events (branch misses for Arm SPE)\n" \ 654 " c: synthesize branches events (calls only)\n" \ 655 " r: synthesize branches events (returns only)\n" \ 656 " x: synthesize transactions events\n" \ 657 " w: synthesize ptwrite events\n" \ 658 " p: synthesize power events\n" \ 659 " o: synthesize other events recorded due to the use\n" \ 660 " of aux-output (refer to perf record)\n" \ 661 " I: synthesize interrupt or similar (asynchronous) events\n" \ 662 " (e.g. Intel PT Event Trace)\n" \ 663 " e[flags]: synthesize error events\n" \ 664 " each flag must be preceded by + or -\n" \ 665 " error flags are: o (overflow)\n" \ 666 " l (data lost)\n" \ 667 " d[flags]: create a debug log\n" \ 668 " each flag must be preceded by + or -\n" \ 669 " log flags are: a (all perf events)\n" \ 670 " o (output to stdout)\n" \ 671 " f: synthesize first level cache events\n" \ 672 " m: synthesize last level cache events\n" \ 673 " t: synthesize TLB events\n" \ 674 " a: synthesize remote access events\n" \ 675 " g[len]: synthesize a call chain (use with i or x)\n" \ 676 " G[len]: synthesize a call chain on existing event records\n" \ 677 " l[len]: synthesize last branch entries (use with i or x)\n" \ 678 " L[len]: synthesize last branch entries on existing event records\n" \ 679 " sNUMBER: skip initial number of events\n" \ 680 " q: quicker (less detailed) decoding\n" \ 681 " A: approximate IPC\n" \ 682 " Z: prefer to ignore timestamps (so-called \"timeless\" decoding)\n" \ 683 " T: use the timestamp trace as kernel time\n" \ 684 " PERIOD[ns|us|ms|i|t]: specify period to sample stream\n" \ 685 " concatenate multiple options. Default is iybxwpe or cewp\n" 686 687 static inline 688 void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts, 689 struct perf_time_interval *ptime_range, 690 int range_num) 691 { 692 opts->ptime_range = ptime_range; 693 opts->range_num = range_num; 694 } 695 696 static inline 697 void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts) 698 { 699 opts->ptime_range = NULL; 700 opts->range_num = 0; 701 } 702 703 #else 704 #include "debug.h" 705 706 static inline struct auxtrace_record * 707 auxtrace_record__init(struct evlist *evlist __maybe_unused, 708 int *err) 709 { 710 *err = 0; 711 return NULL; 712 } 713 714 static inline 715 void auxtrace_record__free(struct auxtrace_record *itr __maybe_unused) 716 { 717 } 718 719 static inline 720 int auxtrace_record__options(struct auxtrace_record *itr __maybe_unused, 721 struct evlist *evlist __maybe_unused, 722 struct record_opts *opts __maybe_unused) 723 { 724 return 0; 725 } 726 727 static inline 728 int perf_event__process_auxtrace_info(struct perf_session *session __maybe_unused, 729 union perf_event *event __maybe_unused) 730 { 731 return 0; 732 } 733 734 static inline 735 s64 perf_event__process_auxtrace(struct perf_session *session __maybe_unused, 736 union perf_event *event __maybe_unused) 737 { 738 return 0; 739 } 740 741 static inline 742 int perf_event__process_auxtrace_error(struct perf_session *session __maybe_unused, 743 union perf_event *event __maybe_unused) 744 { 745 return 0; 746 } 747 748 static inline 749 void perf_session__auxtrace_error_inc(struct perf_session *session 750 __maybe_unused, 751 union perf_event *event 752 __maybe_unused) 753 { 754 } 755 756 static inline 757 void events_stats__auxtrace_error_warn(const struct events_stats *stats 758 __maybe_unused) 759 { 760 } 761 762 static inline 763 int itrace_do_parse_synth_opts(struct itrace_synth_opts *synth_opts __maybe_unused, 764 const char *str __maybe_unused, int unset __maybe_unused) 765 { 766 pr_err("AUX area tracing not supported\n"); 767 return -EINVAL; 768 } 769 770 static inline 771 int itrace_parse_synth_opts(const struct option *opt __maybe_unused, 772 const char *str __maybe_unused, 773 int unset __maybe_unused) 774 { 775 pr_err("AUX area tracing not supported\n"); 776 return -EINVAL; 777 } 778 779 static inline 780 int auxtrace_parse_snapshot_options(struct auxtrace_record *itr __maybe_unused, 781 struct record_opts *opts __maybe_unused, 782 const char *str) 783 { 784 if (!str) 785 return 0; 786 pr_err("AUX area tracing not supported\n"); 787 return -EINVAL; 788 } 789 790 static inline 791 int auxtrace_parse_sample_options(struct auxtrace_record *itr __maybe_unused, 792 struct evlist *evlist __maybe_unused, 793 struct record_opts *opts __maybe_unused, 794 const char *str) 795 { 796 if (!str) 797 return 0; 798 pr_err("AUX area tracing not supported\n"); 799 return -EINVAL; 800 } 801 802 static inline 803 void auxtrace_regroup_aux_output(struct evlist *evlist __maybe_unused) 804 { 805 } 806 807 static inline 808 int auxtrace__process_event(struct perf_session *session __maybe_unused, 809 union perf_event *event __maybe_unused, 810 struct perf_sample *sample __maybe_unused, 811 const struct perf_tool *tool __maybe_unused) 812 { 813 return 0; 814 } 815 816 static inline 817 void auxtrace__dump_auxtrace_sample(struct perf_session *session __maybe_unused, 818 struct perf_sample *sample __maybe_unused) 819 { 820 } 821 822 static inline 823 int auxtrace__flush_events(struct perf_session *session __maybe_unused, 824 const struct perf_tool *tool __maybe_unused) 825 { 826 return 0; 827 } 828 829 static inline 830 void auxtrace__free_events(struct perf_session *session __maybe_unused) 831 { 832 } 833 834 static inline 835 void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache __maybe_unused) 836 { 837 } 838 839 static inline 840 void auxtrace__free(struct perf_session *session __maybe_unused) 841 { 842 } 843 844 static inline 845 int auxtrace_index__write(int fd __maybe_unused, 846 struct list_head *head __maybe_unused) 847 { 848 return -EINVAL; 849 } 850 851 static inline 852 int auxtrace_index__process(int fd __maybe_unused, 853 u64 size __maybe_unused, 854 struct perf_session *session __maybe_unused, 855 bool needs_swap __maybe_unused) 856 { 857 return -EINVAL; 858 } 859 860 static inline 861 void auxtrace_index__free(struct list_head *head __maybe_unused) 862 { 863 } 864 865 static inline 866 bool auxtrace__evsel_is_auxtrace(struct perf_session *session __maybe_unused, 867 struct evsel *evsel __maybe_unused) 868 { 869 return false; 870 } 871 872 static inline 873 int auxtrace_parse_filters(struct evlist *evlist __maybe_unused) 874 { 875 return 0; 876 } 877 878 int auxtrace_mmap__mmap(struct auxtrace_mmap *mm, 879 struct auxtrace_mmap_params *mp, 880 void *userpg, int fd); 881 void auxtrace_mmap__munmap(struct auxtrace_mmap *mm); 882 void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp, 883 off_t auxtrace_offset, 884 unsigned int auxtrace_pages, 885 bool auxtrace_overwrite); 886 void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp, 887 struct evlist *evlist, 888 struct evsel *evsel, int idx); 889 890 #define ITRACE_HELP "" 891 892 static inline 893 void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts 894 __maybe_unused, 895 struct perf_time_interval *ptime_range 896 __maybe_unused, 897 int range_num __maybe_unused) 898 { 899 } 900 901 static inline 902 void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts 903 __maybe_unused) 904 { 905 } 906 907 #endif 908 909 #endif 910