1 /* SPDX-License-Identifier: GPL-2.0-only */
2 /*
3 * auxtrace.h: AUX area trace support
4 * Copyright (c) 2013-2015, Intel Corporation.
5 */
6
7 #ifndef __PERF_AUXTRACE_H
8 #define __PERF_AUXTRACE_H
9
10 #include <sys/types.h>
11 #include <errno.h>
12 #include <stdbool.h>
13 #include <stddef.h>
14 #include <stdio.h> // FILE
15 #include <linux/list.h>
16 #include <linux/perf_event.h>
17 #include <linux/types.h>
18 #include <perf/cpumap.h>
19 #include <asm/bitsperlong.h>
20 #include <asm/barrier.h>
21
22 union perf_event;
23 struct perf_session;
24 struct evlist;
25 struct evsel;
26 struct perf_tool;
27 struct mmap;
28 struct perf_sample;
29 struct option;
30 struct record_opts;
31 struct perf_record_auxtrace_error;
32 struct perf_record_auxtrace_info;
33 struct events_stats;
34 struct perf_pmu;
35
36 enum auxtrace_error_type {
37 PERF_AUXTRACE_ERROR_ITRACE = 1,
38 PERF_AUXTRACE_ERROR_MAX
39 };
40
41 /* Auxtrace records must have the same alignment as perf event records */
42 #define PERF_AUXTRACE_RECORD_ALIGNMENT 8
43
44 enum auxtrace_type {
45 PERF_AUXTRACE_UNKNOWN,
46 PERF_AUXTRACE_INTEL_PT,
47 PERF_AUXTRACE_INTEL_BTS,
48 PERF_AUXTRACE_CS_ETM,
49 PERF_AUXTRACE_ARM_SPE,
50 PERF_AUXTRACE_S390_CPUMSF,
51 PERF_AUXTRACE_HISI_PTT,
52 };
53
54 enum itrace_period_type {
55 PERF_ITRACE_PERIOD_INSTRUCTIONS,
56 PERF_ITRACE_PERIOD_TICKS,
57 PERF_ITRACE_PERIOD_NANOSECS,
58 };
59
60 #define AUXTRACE_ERR_FLG_OVERFLOW (1 << ('o' - 'a'))
61 #define AUXTRACE_ERR_FLG_DATA_LOST (1 << ('l' - 'a'))
62
63 #define AUXTRACE_LOG_FLG_ALL_PERF_EVTS (1 << ('a' - 'a'))
64 #define AUXTRACE_LOG_FLG_ON_ERROR (1 << ('e' - 'a'))
65 #define AUXTRACE_LOG_FLG_USE_STDOUT (1 << ('o' - 'a'))
66
67 /**
68 * struct itrace_synth_opts - AUX area tracing synthesis options.
69 * @set: indicates whether or not options have been set
70 * @default_no_sample: Default to no sampling.
71 * @inject: indicates the event (not just the sample) must be fully synthesized
72 * because 'perf inject' will write it out
73 * @instructions: whether to synthesize 'instructions' events
74 * @cycles: whether to synthesize 'cycles' events
75 * (not fully accurate, since CYC packets are only emitted
76 * together with other events, such as branches)
77 * @branches: whether to synthesize 'branches' events
78 * (branch misses only for Arm SPE)
79 * @transactions: whether to synthesize events for transactions
80 * @ptwrites: whether to synthesize events for ptwrites
81 * @pwr_events: whether to synthesize power events
82 * @other_events: whether to synthesize other events recorded due to the use of
83 * aux_output
84 * @intr_events: whether to synthesize interrupt events
85 * @errors: whether to synthesize decoder error events
86 * @dont_decode: whether to skip decoding entirely
87 * @log: write a decoding log
88 * @calls: limit branch samples to calls (can be combined with @returns)
89 * @returns: limit branch samples to returns (can be combined with @calls)
90 * @callchain: add callchain to 'instructions' events
91 * @add_callchain: add callchain to existing event records
92 * @thread_stack: feed branches to the thread_stack
93 * @last_branch: add branch context to 'instruction' events
94 * @add_last_branch: add branch context to existing event records
95 * @approx_ipc: approximate IPC
96 * @flc: whether to synthesize first level cache events
97 * @llc: whether to synthesize last level cache events
98 * @tlb: whether to synthesize TLB events
99 * @remote_access: whether to synthesize remote access events
100 * @mem: whether to synthesize memory events
101 * @timeless_decoding: prefer "timeless" decoding i.e. ignore timestamps
102 * @use_timestamp: use the timestamp trace as kernel time
103 * @vm_time_correlation: perform VM Time Correlation
104 * @vm_tm_corr_dry_run: VM Time Correlation dry-run
105 * @vm_tm_corr_args: VM Time Correlation implementation-specific arguments
106 * @callchain_sz: maximum callchain size
107 * @last_branch_sz: branch context size
108 * @period: 'instructions' events period
109 * @period_type: 'instructions' events period type
110 * @initial_skip: skip N events at the beginning.
111 * @cpu_bitmap: CPUs for which to synthesize events, or NULL for all
112 * @ptime_range: time intervals to trace or NULL
113 * @range_num: number of time intervals to trace
114 * @error_plus_flags: flags to affect what errors are reported
115 * @error_minus_flags: flags to affect what errors are reported
116 * @log_plus_flags: flags to affect what is logged
117 * @log_minus_flags: flags to affect what is logged
118 * @quick: quicker (less detailed) decoding
119 * @log_on_error_size: size of log to keep for outputting log only on errors
120 */
121 struct itrace_synth_opts {
122 bool set;
123 bool default_no_sample;
124 bool inject;
125 bool instructions;
126 bool cycles;
127 bool branches;
128 bool transactions;
129 bool ptwrites;
130 bool pwr_events;
131 bool other_events;
132 bool intr_events;
133 bool errors;
134 bool dont_decode;
135 bool log;
136 bool calls;
137 bool returns;
138 bool callchain;
139 bool add_callchain;
140 bool thread_stack;
141 bool last_branch;
142 bool add_last_branch;
143 bool approx_ipc;
144 bool flc;
145 bool llc;
146 bool tlb;
147 bool remote_access;
148 bool mem;
149 bool timeless_decoding;
150 bool use_timestamp;
151 bool vm_time_correlation;
152 bool vm_tm_corr_dry_run;
153 char *vm_tm_corr_args;
154 unsigned int callchain_sz;
155 unsigned int last_branch_sz;
156 unsigned long long period;
157 enum itrace_period_type period_type;
158 unsigned long initial_skip;
159 unsigned long *cpu_bitmap;
160 struct perf_time_interval *ptime_range;
161 int range_num;
162 unsigned int error_plus_flags;
163 unsigned int error_minus_flags;
164 unsigned int log_plus_flags;
165 unsigned int log_minus_flags;
166 unsigned int quick;
167 unsigned int log_on_error_size;
168 };
169
170 /**
171 * struct auxtrace_index_entry - indexes a AUX area tracing event within a
172 * perf.data file.
173 * @file_offset: offset within the perf.data file
174 * @sz: size of the event
175 */
176 struct auxtrace_index_entry {
177 u64 file_offset;
178 u64 sz;
179 };
180
181 #define PERF_AUXTRACE_INDEX_ENTRY_COUNT 256
182
183 /**
184 * struct auxtrace_index - index of AUX area tracing events within a perf.data
185 * file.
186 * @list: linking a number of arrays of entries
187 * @nr: number of entries
188 * @entries: array of entries
189 */
190 struct auxtrace_index {
191 struct list_head list;
192 size_t nr;
193 struct auxtrace_index_entry entries[PERF_AUXTRACE_INDEX_ENTRY_COUNT];
194 };
195
196 /**
197 * struct auxtrace - session callbacks to allow AUX area data decoding.
198 * @process_event: lets the decoder see all session events
199 * @process_auxtrace_event: process a PERF_RECORD_AUXTRACE event
200 * @queue_data: queue an AUX sample or PERF_RECORD_AUXTRACE event for later
201 * processing
202 * @dump_auxtrace_sample: dump AUX area sample data
203 * @flush_events: process any remaining data
204 * @free_events: free resources associated with event processing
205 * @free: free resources associated with the session
206 */
207 struct auxtrace {
208 int (*process_event)(struct perf_session *session,
209 union perf_event *event,
210 struct perf_sample *sample,
211 const struct perf_tool *tool);
212 int (*process_auxtrace_event)(struct perf_session *session,
213 union perf_event *event,
214 const struct perf_tool *tool);
215 int (*queue_data)(struct perf_session *session,
216 struct perf_sample *sample, union perf_event *event,
217 u64 data_offset);
218 void (*dump_auxtrace_sample)(struct perf_session *session,
219 struct perf_sample *sample);
220 int (*flush_events)(struct perf_session *session,
221 const struct perf_tool *tool);
222 void (*free_events)(struct perf_session *session);
223 void (*free)(struct perf_session *session);
224 bool (*evsel_is_auxtrace)(struct perf_session *session,
225 struct evsel *evsel);
226 };
227
228 /**
229 * struct auxtrace_buffer - a buffer containing AUX area tracing data.
230 * @list: buffers are queued in a list held by struct auxtrace_queue
231 * @size: size of the buffer in bytes
232 * @pid: in per-thread mode, the pid this buffer is associated with
233 * @tid: in per-thread mode, the tid this buffer is associated with
234 * @cpu: in per-cpu mode, the cpu this buffer is associated with
235 * @data: actual buffer data (can be null if the data has not been loaded)
236 * @data_offset: file offset at which the buffer can be read
237 * @mmap_addr: mmap address at which the buffer can be read
238 * @mmap_size: size of the mmap at @mmap_addr
239 * @data_needs_freeing: @data was malloc'd so free it when it is no longer
240 * needed
241 * @consecutive: the original data was split up and this buffer is consecutive
242 * to the previous buffer
243 * @offset: offset as determined by aux_head / aux_tail members of struct
244 * perf_event_mmap_page
245 * @reference: an implementation-specific reference determined when the data is
246 * recorded
247 * @buffer_nr: used to number each buffer
248 * @use_size: implementation actually only uses this number of bytes
249 * @use_data: implementation actually only uses data starting at this address
250 */
251 struct auxtrace_buffer {
252 struct list_head list;
253 size_t size;
254 pid_t pid;
255 pid_t tid;
256 struct perf_cpu cpu;
257 void *data;
258 off_t data_offset;
259 void *mmap_addr;
260 size_t mmap_size;
261 bool data_needs_freeing;
262 bool consecutive;
263 u64 offset;
264 u64 reference;
265 u64 buffer_nr;
266 size_t use_size;
267 void *use_data;
268 };
269
270 /**
271 * struct auxtrace_queue - a queue of AUX area tracing data buffers.
272 * @head: head of buffer list
273 * @tid: in per-thread mode, the tid this queue is associated with
274 * @cpu: in per-cpu mode, the cpu this queue is associated with
275 * @set: %true once this queue has been dedicated to a specific thread or cpu
276 * @priv: implementation-specific data
277 */
278 struct auxtrace_queue {
279 struct list_head head;
280 pid_t tid;
281 int cpu;
282 bool set;
283 void *priv;
284 };
285
286 /**
287 * struct auxtrace_queues - an array of AUX area tracing queues.
288 * @queue_array: array of queues
289 * @nr_queues: number of queues
290 * @new_data: set whenever new data is queued
291 * @populated: queues have been fully populated using the auxtrace_index
292 * @next_buffer_nr: used to number each buffer
293 */
294 struct auxtrace_queues {
295 struct auxtrace_queue *queue_array;
296 unsigned int nr_queues;
297 bool new_data;
298 bool populated;
299 u64 next_buffer_nr;
300 };
301
302 /**
303 * struct auxtrace_heap_item - element of struct auxtrace_heap.
304 * @queue_nr: queue number
305 * @ordinal: value used for sorting (lowest ordinal is top of the heap) expected
306 * to be a timestamp
307 */
308 struct auxtrace_heap_item {
309 unsigned int queue_nr;
310 u64 ordinal;
311 };
312
313 /**
314 * struct auxtrace_heap - a heap suitable for sorting AUX area tracing queues.
315 * @heap_array: the heap
316 * @heap_cnt: the number of elements in the heap
317 * @heap_sz: maximum number of elements (grows as needed)
318 */
319 struct auxtrace_heap {
320 struct auxtrace_heap_item *heap_array;
321 unsigned int heap_cnt;
322 unsigned int heap_sz;
323 };
324
325 /**
326 * struct auxtrace_mmap - records an mmap of the auxtrace buffer.
327 * @base: address of mapped area
328 * @userpg: pointer to buffer's perf_event_mmap_page
329 * @mask: %0 if @len is not a power of two, otherwise (@len - %1)
330 * @len: size of mapped area
331 * @prev: previous aux_head
332 * @idx: index of this mmap
333 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu
334 * mmap) otherwise %0
335 * @cpu: cpu number for a per-cpu mmap otherwise %-1
336 */
337 struct auxtrace_mmap {
338 void *base;
339 void *userpg;
340 size_t mask;
341 size_t len;
342 u64 prev;
343 int idx;
344 pid_t tid;
345 int cpu;
346 };
347
348 /**
349 * struct auxtrace_mmap_params - parameters to set up struct auxtrace_mmap.
350 * @mask: %0 if @len is not a power of two, otherwise (@len - %1)
351 * @offset: file offset of mapped area
352 * @len: size of mapped area
353 * @prot: mmap memory protection
354 * @idx: index of this mmap
355 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu
356 * mmap) otherwise %0
357 * @mmap_needed: set to %false for non-auxtrace events. This is needed because
358 * auxtrace mmapping is done in the same code path as non-auxtrace
359 * mmapping but not every evsel that needs non-auxtrace mmapping
360 * also needs auxtrace mmapping.
361 * @cpu: cpu number for a per-cpu mmap otherwise %-1
362 */
363 struct auxtrace_mmap_params {
364 size_t mask;
365 off_t offset;
366 size_t len;
367 int prot;
368 int idx;
369 pid_t tid;
370 bool mmap_needed;
371 struct perf_cpu cpu;
372 };
373
374 /**
375 * struct auxtrace_record - callbacks for recording AUX area data.
376 * @recording_options: validate and process recording options
377 * @info_priv_size: return the size of the private data in auxtrace_info_event
378 * @info_fill: fill-in the private data in auxtrace_info_event
379 * @free: free this auxtrace record structure
380 * @snapshot_start: starting a snapshot
381 * @snapshot_finish: finishing a snapshot
382 * @find_snapshot: find data to snapshot within auxtrace mmap
383 * @parse_snapshot_options: parse snapshot options
384 * @reference: provide a 64-bit reference number for auxtrace_event
385 * @read_finish: called after reading from an auxtrace mmap
386 * @alignment: alignment (if any) for AUX area data
387 * @default_aux_sample_size: default sample size for --aux sample option
388 * @pmu: associated pmu
389 * @evlist: selected events list
390 */
391 struct auxtrace_record {
392 int (*recording_options)(struct auxtrace_record *itr,
393 struct evlist *evlist,
394 struct record_opts *opts);
395 size_t (*info_priv_size)(struct auxtrace_record *itr,
396 struct evlist *evlist);
397 int (*info_fill)(struct auxtrace_record *itr,
398 struct perf_session *session,
399 struct perf_record_auxtrace_info *auxtrace_info,
400 size_t priv_size);
401 void (*free)(struct auxtrace_record *itr);
402 int (*snapshot_start)(struct auxtrace_record *itr);
403 int (*snapshot_finish)(struct auxtrace_record *itr);
404 int (*find_snapshot)(struct auxtrace_record *itr, int idx,
405 struct auxtrace_mmap *mm, unsigned char *data,
406 u64 *head, u64 *old);
407 int (*parse_snapshot_options)(struct auxtrace_record *itr,
408 struct record_opts *opts,
409 const char *str);
410 u64 (*reference)(struct auxtrace_record *itr);
411 int (*read_finish)(struct auxtrace_record *itr, int idx);
412 unsigned int alignment;
413 unsigned int default_aux_sample_size;
414 struct evlist *evlist;
415 };
416
417 /**
418 * struct addr_filter - address filter.
419 * @list: list node
420 * @range: true if it is a range filter
421 * @start: true if action is 'filter' or 'start'
422 * @action: 'filter', 'start' or 'stop' ('tracestop' is accepted but converted
423 * to 'stop')
424 * @sym_from: symbol name for the filter address
425 * @sym_to: symbol name that determines the filter size
426 * @sym_from_idx: selects n'th from symbols with the same name (0 means global
427 * and less than 0 means symbol must be unique)
428 * @sym_to_idx: same as @sym_from_idx but for @sym_to
429 * @addr: filter address
430 * @size: filter region size (for range filters)
431 * @filename: DSO file name or NULL for the kernel
432 * @str: allocated string that contains the other string members
433 */
434 struct addr_filter {
435 struct list_head list;
436 bool range;
437 bool start;
438 const char *action;
439 const char *sym_from;
440 const char *sym_to;
441 int sym_from_idx;
442 int sym_to_idx;
443 u64 addr;
444 u64 size;
445 const char *filename;
446 char *str;
447 };
448
449 /**
450 * struct addr_filters - list of address filters.
451 * @head: list of address filters
452 * @cnt: number of address filters
453 */
454 struct addr_filters {
455 struct list_head head;
456 int cnt;
457 };
458
459 struct auxtrace_cache;
460
461 #ifdef HAVE_AUXTRACE_SUPPORT
462
463 u64 compat_auxtrace_mmap__read_head(struct auxtrace_mmap *mm);
464 int compat_auxtrace_mmap__write_tail(struct auxtrace_mmap *mm, u64 tail);
465
auxtrace_mmap__read_head(struct auxtrace_mmap * mm,int kernel_is_64_bit __maybe_unused)466 static inline u64 auxtrace_mmap__read_head(struct auxtrace_mmap *mm,
467 int kernel_is_64_bit __maybe_unused)
468 {
469 struct perf_event_mmap_page *pc = mm->userpg;
470 u64 head;
471
472 #if BITS_PER_LONG == 32
473 if (kernel_is_64_bit)
474 return compat_auxtrace_mmap__read_head(mm);
475 #endif
476 head = READ_ONCE(pc->aux_head);
477
478 /* Ensure all reads are done after we read the head */
479 smp_rmb();
480 return head;
481 }
482
auxtrace_mmap__write_tail(struct auxtrace_mmap * mm,u64 tail,int kernel_is_64_bit __maybe_unused)483 static inline int auxtrace_mmap__write_tail(struct auxtrace_mmap *mm, u64 tail,
484 int kernel_is_64_bit __maybe_unused)
485 {
486 struct perf_event_mmap_page *pc = mm->userpg;
487
488 #if BITS_PER_LONG == 32
489 if (kernel_is_64_bit)
490 return compat_auxtrace_mmap__write_tail(mm, tail);
491 #endif
492 /* Ensure all reads are done before we write the tail out */
493 smp_mb();
494 WRITE_ONCE(pc->aux_tail, tail);
495 return 0;
496 }
497
498 int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
499 struct auxtrace_mmap_params *mp,
500 void *userpg, int fd);
501 void auxtrace_mmap__munmap(struct auxtrace_mmap *mm);
502 void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
503 off_t auxtrace_offset,
504 unsigned int auxtrace_pages,
505 bool auxtrace_overwrite);
506 void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
507 struct evlist *evlist,
508 struct evsel *evsel, int idx);
509
510 typedef int (*process_auxtrace_t)(const struct perf_tool *tool,
511 struct mmap *map,
512 union perf_event *event, void *data1,
513 size_t len1, void *data2, size_t len2);
514
515 int auxtrace_mmap__read(struct mmap *map, struct auxtrace_record *itr,
516 const struct perf_tool *tool, process_auxtrace_t fn);
517
518 int auxtrace_mmap__read_snapshot(struct mmap *map,
519 struct auxtrace_record *itr,
520 const struct perf_tool *tool, process_auxtrace_t fn,
521 size_t snapshot_size);
522
523 int auxtrace_queues__init_nr(struct auxtrace_queues *queues, int nr_queues);
524 int auxtrace_queues__init(struct auxtrace_queues *queues);
525 int auxtrace_queues__add_event(struct auxtrace_queues *queues,
526 struct perf_session *session,
527 union perf_event *event, off_t data_offset,
528 struct auxtrace_buffer **buffer_ptr);
529 struct auxtrace_queue *
530 auxtrace_queues__sample_queue(struct auxtrace_queues *queues,
531 struct perf_sample *sample,
532 struct perf_session *session);
533 int auxtrace_queues__add_sample(struct auxtrace_queues *queues,
534 struct perf_session *session,
535 struct perf_sample *sample, u64 data_offset,
536 u64 reference);
537 void auxtrace_queues__free(struct auxtrace_queues *queues);
538 int auxtrace_queues__process_index(struct auxtrace_queues *queues,
539 struct perf_session *session);
540 int auxtrace_queue_data(struct perf_session *session, bool samples,
541 bool events);
542 struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue,
543 struct auxtrace_buffer *buffer);
544 void *auxtrace_buffer__get_data_rw(struct auxtrace_buffer *buffer, int fd, bool rw);
auxtrace_buffer__get_data(struct auxtrace_buffer * buffer,int fd)545 static inline void *auxtrace_buffer__get_data(struct auxtrace_buffer *buffer, int fd)
546 {
547 return auxtrace_buffer__get_data_rw(buffer, fd, false);
548 }
549 void auxtrace_buffer__put_data(struct auxtrace_buffer *buffer);
550 void auxtrace_buffer__drop_data(struct auxtrace_buffer *buffer);
551 void auxtrace_buffer__free(struct auxtrace_buffer *buffer);
552
553 int auxtrace_heap__add(struct auxtrace_heap *heap, unsigned int queue_nr,
554 u64 ordinal);
555 void auxtrace_heap__pop(struct auxtrace_heap *heap);
556 void auxtrace_heap__free(struct auxtrace_heap *heap);
557
558 struct auxtrace_cache_entry {
559 struct hlist_node hash;
560 u32 key;
561 };
562
563 struct auxtrace_cache *auxtrace_cache__new(unsigned int bits, size_t entry_size,
564 unsigned int limit_percent);
565 void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache);
566 void *auxtrace_cache__alloc_entry(struct auxtrace_cache *c);
567 void auxtrace_cache__free_entry(struct auxtrace_cache *c, void *entry);
568 int auxtrace_cache__add(struct auxtrace_cache *c, u32 key,
569 struct auxtrace_cache_entry *entry);
570 void auxtrace_cache__remove(struct auxtrace_cache *c, u32 key);
571 void *auxtrace_cache__lookup(struct auxtrace_cache *c, u32 key);
572
573 struct auxtrace_record *auxtrace_record__init(struct evlist *evlist,
574 int *err);
575
576 int auxtrace_parse_snapshot_options(struct auxtrace_record *itr,
577 struct record_opts *opts,
578 const char *str);
579 int auxtrace_parse_sample_options(struct auxtrace_record *itr,
580 struct evlist *evlist,
581 struct record_opts *opts, const char *str);
582 void auxtrace_regroup_aux_output(struct evlist *evlist);
583 int auxtrace_record__options(struct auxtrace_record *itr,
584 struct evlist *evlist,
585 struct record_opts *opts);
586 size_t auxtrace_record__info_priv_size(struct auxtrace_record *itr,
587 struct evlist *evlist);
588 int auxtrace_record__info_fill(struct auxtrace_record *itr,
589 struct perf_session *session,
590 struct perf_record_auxtrace_info *auxtrace_info,
591 size_t priv_size);
592 void auxtrace_record__free(struct auxtrace_record *itr);
593 int auxtrace_record__snapshot_start(struct auxtrace_record *itr);
594 int auxtrace_record__snapshot_finish(struct auxtrace_record *itr, bool on_exit);
595 int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx,
596 struct auxtrace_mmap *mm,
597 unsigned char *data, u64 *head, u64 *old);
598 u64 auxtrace_record__reference(struct auxtrace_record *itr);
599 int auxtrace_record__read_finish(struct auxtrace_record *itr, int idx);
600
601 int auxtrace_index__auxtrace_event(struct list_head *head, union perf_event *event,
602 off_t file_offset);
603 int auxtrace_index__write(int fd, struct list_head *head);
604 int auxtrace_index__process(int fd, u64 size, struct perf_session *session,
605 bool needs_swap);
606 void auxtrace_index__free(struct list_head *head);
607
608 void auxtrace_synth_guest_error(struct perf_record_auxtrace_error *auxtrace_error, int type,
609 int code, int cpu, pid_t pid, pid_t tid, u64 ip,
610 const char *msg, u64 timestamp,
611 pid_t machine_pid, int vcpu);
612 void auxtrace_synth_error(struct perf_record_auxtrace_error *auxtrace_error, int type,
613 int code, int cpu, pid_t pid, pid_t tid, u64 ip,
614 const char *msg, u64 timestamp);
615
616 int perf_event__process_auxtrace_info(struct perf_session *session,
617 union perf_event *event);
618 s64 perf_event__process_auxtrace(struct perf_session *session,
619 union perf_event *event);
620 int perf_event__process_auxtrace_error(struct perf_session *session,
621 union perf_event *event);
622 int itrace_do_parse_synth_opts(struct itrace_synth_opts *synth_opts,
623 const char *str, int unset);
624 int itrace_parse_synth_opts(const struct option *opt, const char *str,
625 int unset);
626 void itrace_synth_opts__set_default(struct itrace_synth_opts *synth_opts,
627 bool no_sample);
628
629 size_t perf_event__fprintf_auxtrace_error(union perf_event *event, FILE *fp);
630 void perf_session__auxtrace_error_inc(struct perf_session *session,
631 union perf_event *event);
632 void events_stats__auxtrace_error_warn(const struct events_stats *stats);
633
634 void addr_filters__init(struct addr_filters *filts);
635 void addr_filters__exit(struct addr_filters *filts);
636 int addr_filters__parse_bare_filter(struct addr_filters *filts,
637 const char *filter);
638 int auxtrace_parse_filters(struct evlist *evlist);
639
640 int auxtrace__process_event(struct perf_session *session, union perf_event *event,
641 struct perf_sample *sample, const struct perf_tool *tool);
642 void auxtrace__dump_auxtrace_sample(struct perf_session *session,
643 struct perf_sample *sample);
644 int auxtrace__flush_events(struct perf_session *session, const struct perf_tool *tool);
645 void auxtrace__free_events(struct perf_session *session);
646 void auxtrace__free(struct perf_session *session);
647 bool auxtrace__evsel_is_auxtrace(struct perf_session *session,
648 struct evsel *evsel);
649
650 #define ITRACE_HELP \
651 " i[period]: synthesize instructions events\n" \
652 " y[period]: synthesize cycles events (same period as i)\n" \
653 " b: synthesize branches events (branch misses for Arm SPE)\n" \
654 " c: synthesize branches events (calls only)\n" \
655 " r: synthesize branches events (returns only)\n" \
656 " x: synthesize transactions events\n" \
657 " w: synthesize ptwrite events\n" \
658 " p: synthesize power events\n" \
659 " o: synthesize other events recorded due to the use\n" \
660 " of aux-output (refer to perf record)\n" \
661 " I: synthesize interrupt or similar (asynchronous) events\n" \
662 " (e.g. Intel PT Event Trace)\n" \
663 " e[flags]: synthesize error events\n" \
664 " each flag must be preceded by + or -\n" \
665 " error flags are: o (overflow)\n" \
666 " l (data lost)\n" \
667 " d[flags]: create a debug log\n" \
668 " each flag must be preceded by + or -\n" \
669 " log flags are: a (all perf events)\n" \
670 " o (output to stdout)\n" \
671 " f: synthesize first level cache events\n" \
672 " m: synthesize last level cache events\n" \
673 " t: synthesize TLB events\n" \
674 " a: synthesize remote access events\n" \
675 " g[len]: synthesize a call chain (use with i or x)\n" \
676 " G[len]: synthesize a call chain on existing event records\n" \
677 " l[len]: synthesize last branch entries (use with i or x)\n" \
678 " L[len]: synthesize last branch entries on existing event records\n" \
679 " sNUMBER: skip initial number of events\n" \
680 " q: quicker (less detailed) decoding\n" \
681 " A: approximate IPC\n" \
682 " Z: prefer to ignore timestamps (so-called \"timeless\" decoding)\n" \
683 " T: use the timestamp trace as kernel time\n" \
684 " PERIOD[ns|us|ms|i|t]: specify period to sample stream\n" \
685 " concatenate multiple options. Default is iybxwpe or cewp\n"
686
687 static inline
itrace_synth_opts__set_time_range(struct itrace_synth_opts * opts,struct perf_time_interval * ptime_range,int range_num)688 void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts,
689 struct perf_time_interval *ptime_range,
690 int range_num)
691 {
692 opts->ptime_range = ptime_range;
693 opts->range_num = range_num;
694 }
695
696 static inline
itrace_synth_opts__clear_time_range(struct itrace_synth_opts * opts)697 void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts)
698 {
699 opts->ptime_range = NULL;
700 opts->range_num = 0;
701 }
702
703 #else
704 #include "debug.h"
705
706 static inline struct auxtrace_record *
auxtrace_record__init(struct evlist * evlist __maybe_unused,int * err)707 auxtrace_record__init(struct evlist *evlist __maybe_unused,
708 int *err)
709 {
710 *err = 0;
711 return NULL;
712 }
713
714 static inline
auxtrace_record__free(struct auxtrace_record * itr __maybe_unused)715 void auxtrace_record__free(struct auxtrace_record *itr __maybe_unused)
716 {
717 }
718
719 static inline
auxtrace_record__options(struct auxtrace_record * itr __maybe_unused,struct evlist * evlist __maybe_unused,struct record_opts * opts __maybe_unused)720 int auxtrace_record__options(struct auxtrace_record *itr __maybe_unused,
721 struct evlist *evlist __maybe_unused,
722 struct record_opts *opts __maybe_unused)
723 {
724 return 0;
725 }
726
727 static inline
perf_event__process_auxtrace_info(struct perf_session * session __maybe_unused,union perf_event * event __maybe_unused)728 int perf_event__process_auxtrace_info(struct perf_session *session __maybe_unused,
729 union perf_event *event __maybe_unused)
730 {
731 return 0;
732 }
733
734 static inline
perf_event__process_auxtrace(struct perf_session * session __maybe_unused,union perf_event * event __maybe_unused)735 s64 perf_event__process_auxtrace(struct perf_session *session __maybe_unused,
736 union perf_event *event __maybe_unused)
737 {
738 return 0;
739 }
740
741 static inline
perf_event__process_auxtrace_error(struct perf_session * session __maybe_unused,union perf_event * event __maybe_unused)742 int perf_event__process_auxtrace_error(struct perf_session *session __maybe_unused,
743 union perf_event *event __maybe_unused)
744 {
745 return 0;
746 }
747
748 static inline
perf_session__auxtrace_error_inc(struct perf_session * session __maybe_unused,union perf_event * event __maybe_unused)749 void perf_session__auxtrace_error_inc(struct perf_session *session
750 __maybe_unused,
751 union perf_event *event
752 __maybe_unused)
753 {
754 }
755
756 static inline
events_stats__auxtrace_error_warn(const struct events_stats * stats __maybe_unused)757 void events_stats__auxtrace_error_warn(const struct events_stats *stats
758 __maybe_unused)
759 {
760 }
761
762 static inline
itrace_do_parse_synth_opts(struct itrace_synth_opts * synth_opts __maybe_unused,const char * str __maybe_unused,int unset __maybe_unused)763 int itrace_do_parse_synth_opts(struct itrace_synth_opts *synth_opts __maybe_unused,
764 const char *str __maybe_unused, int unset __maybe_unused)
765 {
766 pr_err("AUX area tracing not supported\n");
767 return -EINVAL;
768 }
769
770 static inline
itrace_parse_synth_opts(const struct option * opt __maybe_unused,const char * str __maybe_unused,int unset __maybe_unused)771 int itrace_parse_synth_opts(const struct option *opt __maybe_unused,
772 const char *str __maybe_unused,
773 int unset __maybe_unused)
774 {
775 pr_err("AUX area tracing not supported\n");
776 return -EINVAL;
777 }
778
779 static inline
auxtrace_parse_snapshot_options(struct auxtrace_record * itr __maybe_unused,struct record_opts * opts __maybe_unused,const char * str)780 int auxtrace_parse_snapshot_options(struct auxtrace_record *itr __maybe_unused,
781 struct record_opts *opts __maybe_unused,
782 const char *str)
783 {
784 if (!str)
785 return 0;
786 pr_err("AUX area tracing not supported\n");
787 return -EINVAL;
788 }
789
790 static inline
auxtrace_parse_sample_options(struct auxtrace_record * itr __maybe_unused,struct evlist * evlist __maybe_unused,struct record_opts * opts __maybe_unused,const char * str)791 int auxtrace_parse_sample_options(struct auxtrace_record *itr __maybe_unused,
792 struct evlist *evlist __maybe_unused,
793 struct record_opts *opts __maybe_unused,
794 const char *str)
795 {
796 if (!str)
797 return 0;
798 pr_err("AUX area tracing not supported\n");
799 return -EINVAL;
800 }
801
802 static inline
auxtrace_regroup_aux_output(struct evlist * evlist __maybe_unused)803 void auxtrace_regroup_aux_output(struct evlist *evlist __maybe_unused)
804 {
805 }
806
807 static inline
auxtrace__process_event(struct perf_session * session __maybe_unused,union perf_event * event __maybe_unused,struct perf_sample * sample __maybe_unused,const struct perf_tool * tool __maybe_unused)808 int auxtrace__process_event(struct perf_session *session __maybe_unused,
809 union perf_event *event __maybe_unused,
810 struct perf_sample *sample __maybe_unused,
811 const struct perf_tool *tool __maybe_unused)
812 {
813 return 0;
814 }
815
816 static inline
auxtrace__dump_auxtrace_sample(struct perf_session * session __maybe_unused,struct perf_sample * sample __maybe_unused)817 void auxtrace__dump_auxtrace_sample(struct perf_session *session __maybe_unused,
818 struct perf_sample *sample __maybe_unused)
819 {
820 }
821
822 static inline
auxtrace__flush_events(struct perf_session * session __maybe_unused,const struct perf_tool * tool __maybe_unused)823 int auxtrace__flush_events(struct perf_session *session __maybe_unused,
824 const struct perf_tool *tool __maybe_unused)
825 {
826 return 0;
827 }
828
829 static inline
auxtrace__free_events(struct perf_session * session __maybe_unused)830 void auxtrace__free_events(struct perf_session *session __maybe_unused)
831 {
832 }
833
834 static inline
auxtrace_cache__free(struct auxtrace_cache * auxtrace_cache __maybe_unused)835 void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache __maybe_unused)
836 {
837 }
838
839 static inline
auxtrace__free(struct perf_session * session __maybe_unused)840 void auxtrace__free(struct perf_session *session __maybe_unused)
841 {
842 }
843
844 static inline
auxtrace_index__write(int fd __maybe_unused,struct list_head * head __maybe_unused)845 int auxtrace_index__write(int fd __maybe_unused,
846 struct list_head *head __maybe_unused)
847 {
848 return -EINVAL;
849 }
850
851 static inline
auxtrace_index__process(int fd __maybe_unused,u64 size __maybe_unused,struct perf_session * session __maybe_unused,bool needs_swap __maybe_unused)852 int auxtrace_index__process(int fd __maybe_unused,
853 u64 size __maybe_unused,
854 struct perf_session *session __maybe_unused,
855 bool needs_swap __maybe_unused)
856 {
857 return -EINVAL;
858 }
859
860 static inline
auxtrace_index__free(struct list_head * head __maybe_unused)861 void auxtrace_index__free(struct list_head *head __maybe_unused)
862 {
863 }
864
865 static inline
auxtrace__evsel_is_auxtrace(struct perf_session * session __maybe_unused,struct evsel * evsel __maybe_unused)866 bool auxtrace__evsel_is_auxtrace(struct perf_session *session __maybe_unused,
867 struct evsel *evsel __maybe_unused)
868 {
869 return false;
870 }
871
872 static inline
auxtrace_parse_filters(struct evlist * evlist __maybe_unused)873 int auxtrace_parse_filters(struct evlist *evlist __maybe_unused)
874 {
875 return 0;
876 }
877
878 int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
879 struct auxtrace_mmap_params *mp,
880 void *userpg, int fd);
881 void auxtrace_mmap__munmap(struct auxtrace_mmap *mm);
882 void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
883 off_t auxtrace_offset,
884 unsigned int auxtrace_pages,
885 bool auxtrace_overwrite);
886 void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
887 struct evlist *evlist,
888 struct evsel *evsel, int idx);
889
890 #define ITRACE_HELP ""
891
892 static inline
itrace_synth_opts__set_time_range(struct itrace_synth_opts * opts __maybe_unused,struct perf_time_interval * ptime_range __maybe_unused,int range_num __maybe_unused)893 void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts
894 __maybe_unused,
895 struct perf_time_interval *ptime_range
896 __maybe_unused,
897 int range_num __maybe_unused)
898 {
899 }
900
901 static inline
itrace_synth_opts__clear_time_range(struct itrace_synth_opts * opts __maybe_unused)902 void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts
903 __maybe_unused)
904 {
905 }
906
907 #endif
908
909 #endif
910