/linux/kernel/trace/ |
H A D | trace.c | 832 static u64 buffer_ftrace_now(struct array_buffer *buf, int cpu) in buffer_ftrace_now() 848 return buffer_ftrace_now(&global_trace.array_buffer, cpu); in ftrace_now() 1027 if (tr->array_buffer.buffer) in tracer_tracing_on() 1028 ring_buffer_record_on(tr->array_buffer.buffer); in tracer_tracing_on() 1093 buffer = tr->array_buffer.buffer; in __trace_array_puts() 1159 buffer = tr->array_buffer.buffer; in __trace_bputs() 1296 static int resize_buffer_duplicate_size(struct array_buffer *trace_buf, 1297 struct array_buffer *size_buf, int cpu_id); 1298 static void set_buffer_entries(struct array_buffer *buf, unsigned long val); 1308 order = ring_buffer_subbuf_order_get(tr->array_buffer.buffer); in tracing_alloc_snapshot_instance() [all …]
|
H A D | trace_sched_wakeup.c | 85 *data = per_cpu_ptr(tr->array_buffer.data, cpu); in func_prolog_preempt_disable() 399 struct trace_buffer *buffer = tr->array_buffer.buffer; in tracing_sched_switch_trace() 427 struct trace_buffer *buffer = tr->array_buffer.buffer; in tracing_sched_wakeup_trace() 476 disabled = atomic_inc_return(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled); in probe_wakeup_sched_switch() 490 data = per_cpu_ptr(wakeup_trace->array_buffer.data, wakeup_cpu); in probe_wakeup_sched_switch() 513 atomic_dec(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled); in probe_wakeup_sched_switch() 532 tracing_reset_online_cpus(&tr->array_buffer); in wakeup_reset() 568 disabled = atomic_inc_return(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled); in probe_wakeup() 600 data = per_cpu_ptr(wakeup_trace->array_buffer.data, wakeup_cpu); in probe_wakeup() 615 atomic_dec(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled); in probe_wakeup()
|
H A D | trace_mmiotrace.c | 33 tracing_reset_online_cpus(&tr->array_buffer); in mmio_reset_data() 123 unsigned long over = ring_buffer_overruns(iter->array_buffer->buffer); in count_overruns() 297 struct trace_buffer *buffer = tr->array_buffer.buffer; in __trace_mmiotrace_rw() 318 struct trace_array_cpu *data = per_cpu_ptr(tr->array_buffer.data, smp_processor_id()); in mmio_trace_rw() 326 struct trace_buffer *buffer = tr->array_buffer.buffer; in __trace_mmiotrace_map() 350 data = per_cpu_ptr(tr->array_buffer.data, smp_processor_id()); in mmio_trace_mapping()
|
H A D | trace_functions.c | 160 tr->array_buffer.cpu = raw_smp_processor_id(); in function_trace_init() 176 tracing_reset_online_cpus(&tr->array_buffer); in function_trace_start() 221 data = this_cpu_ptr(tr->array_buffer.data); in function_trace_call() 268 data = per_cpu_ptr(tr->array_buffer.data, cpu); in function_stack_trace_call() 293 ring_buffer_time_stamp(tr->array_buffer.buffer); in is_repeat_check() 334 data = this_cpu_ptr(tr->array_buffer.data); in function_no_repeats_trace_call() 381 data = per_cpu_ptr(tr->array_buffer.data, cpu); in function_stack_no_repeats_trace_call()
|
H A D | trace_irqsoff.c | 125 *data = per_cpu_ptr(tr->array_buffer.data, cpu); in func_prolog_dec() 173 tracing_reset_online_cpus(&irqsoff_trace->array_buffer); in irqsoff_display_graph() 405 data = per_cpu_ptr(tr->array_buffer.data, cpu); in start_critical_timing() 441 data = per_cpu_ptr(tr->array_buffer.data, cpu); in stop_critical_timing()
|
H A D | trace_functions_graph.c | 118 struct trace_buffer *buffer = tr->array_buffer.buffer; in __trace_graph_entry() 139 struct trace_buffer *buffer = tr->array_buffer.buffer; in __trace_graph_retaddr_entry() 240 data = per_cpu_ptr(tr->array_buffer.data, cpu); in trace_graph_entry() 289 struct trace_buffer *buffer = tr->array_buffer.buffer; in __trace_graph_return() 345 data = per_cpu_ptr(tr->array_buffer.data, cpu); in trace_graph_return() 567 ring_buffer_consume(iter->array_buffer->buffer, iter->cpu, in get_return_for_leaf() 569 event = ring_buffer_peek(iter->array_buffer->buffer, iter->cpu, in get_return_for_leaf() 629 usecs = iter->ts - iter->array_buffer->time_start; in print_graph_rel_time()
|
H A D | trace_branch.c | 57 data = this_cpu_ptr(tr->array_buffer.data); in probe_likely_condition() 62 buffer = tr->array_buffer.buffer; in probe_likely_condition()
|
H A D | trace.h | 211 struct array_buffer { struct 323 struct array_buffer array_buffer; member 336 struct array_buffer max_buffer; 641 void tracing_reset_online_cpus(struct array_buffer *buf); 1122 return this_cpu_read(tr->array_buffer.data->ftrace_ignore_pid) != in ftrace_trace_task()
|
H A D | trace_osnoise.c | 489 buffer = inst->tr->array_buffer.buffer; \ 532 buffer = inst->tr->array_buffer.buffer; in trace_osnoise_sample() 604 buffer = inst->tr->array_buffer.buffer; in trace_timerlat_sample() 691 buffer = inst->tr->array_buffer.buffer; in timerlat_dump_stack() 1198 trace_array_printk_buf(tr->array_buffer.buffer, _THIS_IP_, in osnoise_stop_exception() 1371 trace_array_printk_buf(tr->array_buffer.buffer, _THIS_IP_, in osnoise_stop_tracing()
|
H A D | trace_hwlat.c | 133 struct trace_buffer *buffer = tr->array_buffer.buffer; in trace_hwlat_sample() 189 trace_array_printk_buf(tr->array_buffer.buffer, _THIS_IP_, msg); \
|
H A D | trace_events.c | 628 data = this_cpu_ptr(tr->array_buffer.data); in trace_event_ignore_this_pid() 1062 this_cpu_write(tr->array_buffer.data->ignore_pid, ret || in event_filter_pid_sched_switch_probe_pre() 1080 this_cpu_write(tr->array_buffer.data->ignore_pid, in event_filter_pid_sched_switch_probe_post() 1092 if (!this_cpu_read(tr->array_buffer.data->ignore_pid)) in event_filter_pid_sched_wakeup_probe_pre() 1098 this_cpu_write(tr->array_buffer.data->ignore_pid, in event_filter_pid_sched_wakeup_probe_pre() 1110 if (this_cpu_read(tr->array_buffer.data->ignore_pid)) in event_filter_pid_sched_wakeup_probe_post() 1117 this_cpu_write(tr->array_buffer.data->ignore_pid, in event_filter_pid_sched_wakeup_probe_post() 1160 per_cpu_ptr(tr->array_buffer.data, cpu)->ignore_pid = false; in __ftrace_clear_event_pids() 2294 ring_buffer_print_page_header(tr->array_buffer.buffer, s); in show_header_page_file() 2342 this_cpu_write(tr->array_buffer.data->ignore_pid, in ignore_task_cpu()
|
H A D | blktrace.c | 81 buffer = blk_tr->array_buffer.buffer; in trace_note() 256 buffer = blk_tr->array_buffer.buffer; in __blk_add_trace()
|
H A D | trace_events_synth.c | 539 buffer = trace_file->tr->array_buffer.buffer; in trace_event_raw_event_synth() 1665 trace_state->buffer = file->tr->array_buffer.buffer; in __synth_event_trace_start()
|
H A D | fgraph.c | 1195 pid = this_cpu_read(tr->array_buffer.data->ftrace_ignore_pid); in fgraph_pid_func()
|
H A D | ftrace.c | 180 pid = this_cpu_read(tr->array_buffer.data->ftrace_ignore_pid); in ftrace_pid_func() 8013 this_cpu_write(tr->array_buffer.data->ftrace_ignore_pid, in ftrace_filter_pid_sched_switch_probe() 8016 this_cpu_write(tr->array_buffer.data->ftrace_ignore_pid, in ftrace_filter_pid_sched_switch_probe() 8082 per_cpu_ptr(tr->array_buffer.data, cpu)->ftrace_ignore_pid = FTRACE_PID_TRACE; in clear_ftrace_pids() 8281 this_cpu_write(tr->array_buffer.data->ftrace_ignore_pid, in ignore_task_cpu() 8284 this_cpu_write(tr->array_buffer.data->ftrace_ignore_pid, in ignore_task_cpu()
|
H A D | trace_output.c | 569 unsigned long long abs_ts = iter->ts - iter->array_buffer->time_start; in lat_print_timestamp()
|
/linux/include/linux/ |
H A D | trace_events.h | 14 struct array_buffer; 99 struct array_buffer *array_buffer; member
|