/linux/kernel/trace/ |
H A D | trace.c | 904 static u64 buffer_ftrace_now(struct array_buffer *buf, int cpu) in buffer_ftrace_now() 920 return buffer_ftrace_now(&global_trace.array_buffer, cpu); in ftrace_now() 1099 if (tr->array_buffer.buffer) in tracer_tracing_on() 1100 ring_buffer_record_on(tr->array_buffer.buffer); in tracer_tracing_on() 1165 buffer = tr->array_buffer.buffer; in __trace_array_puts() 1231 buffer = tr->array_buffer.buffer; in __trace_bputs() 1368 static int resize_buffer_duplicate_size(struct array_buffer *trace_buf, 1369 struct array_buffer *size_buf, int cpu_id); 1370 static void set_buffer_entries(struct array_buffer *buf, unsigned long val); 1380 order = ring_buffer_subbuf_order_get(tr->array_buffer.buffer); in tracing_alloc_snapshot_instance() [all …]
|
H A D | trace_selftest.c | 27 static int trace_test_buffer_cpu(struct array_buffer *buf, int cpu) in trace_test_buffer_cpu() 64 static int __maybe_unused trace_test_buffer(struct array_buffer *buf, unsigned long *count) in trace_test_buffer() 397 ret = trace_test_buffer(&tr->array_buffer, &count); in trace_selftest_startup_dynamic_tracing() 418 ret = trace_test_buffer(&tr->array_buffer, &count); in trace_selftest_startup_dynamic_tracing() 724 ret = trace_test_buffer(&tr->array_buffer, &count); in trace_selftest_startup_function() 1080 tracing_reset_online_cpus(&tr->array_buffer); in trace_selftest_startup_function_graph() 1102 ret = trace_test_buffer(&tr->array_buffer, &count); in trace_selftest_startup_function_graph() 1123 tracing_reset_online_cpus(&tr->array_buffer); in trace_selftest_startup_function_graph() 1158 ret = trace_test_buffer(&tr->array_buffer, &count); in trace_selftest_startup_function_graph() 1226 ret = trace_test_buffer(&tr->array_buffer, NULL); in trace_selftest_startup_irqsoff() [all …]
|
H A D | trace_sched_wakeup.c | 85 *data = per_cpu_ptr(tr->array_buffer.data, cpu); in func_prolog_preempt_disable() 397 struct trace_buffer *buffer = tr->array_buffer.buffer; in tracing_sched_switch_trace() 425 struct trace_buffer *buffer = tr->array_buffer.buffer; in tracing_sched_wakeup_trace() 474 disabled = local_inc_return(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled); in probe_wakeup_sched_switch() 488 data = per_cpu_ptr(wakeup_trace->array_buffer.data, wakeup_cpu); in probe_wakeup_sched_switch() 511 local_dec(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled); in probe_wakeup_sched_switch() 530 tracing_reset_online_cpus(&tr->array_buffer); in wakeup_reset() 566 disabled = local_inc_return(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled); in probe_wakeup() 598 data = per_cpu_ptr(wakeup_trace->array_buffer.data, wakeup_cpu); in probe_wakeup() 613 local_dec(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled); in probe_wakeup()
|
H A D | trace_mmiotrace.c | 33 tracing_reset_online_cpus(&tr->array_buffer); in mmio_reset_data() 123 unsigned long over = ring_buffer_overruns(iter->array_buffer->buffer); in count_overruns() 296 struct trace_buffer *buffer = tr->array_buffer.buffer; in __trace_mmiotrace_rw() 323 struct trace_buffer *buffer = tr->array_buffer.buffer; in __trace_mmiotrace_map()
|
H A D | trace_kdb.c | 46 ring_buffer_read_prepare(iter.array_buffer->buffer, in ftrace_dump_buf() 54 ring_buffer_read_prepare(iter.array_buffer->buffer, in ftrace_dump_buf()
|
H A D | trace_functions.c | 166 tr->array_buffer.cpu = raw_smp_processor_id(); in function_trace_init() 182 tracing_reset_online_cpus(&tr->array_buffer); in function_trace_start() 293 data = per_cpu_ptr(tr->array_buffer.data, cpu); in function_stack_trace_call() 318 ring_buffer_time_stamp(tr->array_buffer.buffer); in is_repeat_check() 404 data = per_cpu_ptr(tr->array_buffer.data, cpu); in function_stack_no_repeats_trace_call()
|
H A D | trace_irqsoff.c | 125 *data = per_cpu_ptr(tr->array_buffer.data, cpu); in func_prolog_dec() 173 tracing_reset_online_cpus(&irqsoff_trace->array_buffer); in irqsoff_display_graph() 410 data = per_cpu_ptr(tr->array_buffer.data, cpu); in start_critical_timing() 449 data = per_cpu_ptr(tr->array_buffer.data, cpu); in stop_critical_timing()
|
H A D | trace.h | 211 struct array_buffer { struct 328 struct array_buffer array_buffer; member 341 struct array_buffer max_buffer; 655 void tracing_reset_online_cpus(struct array_buffer *buf); 686 if (tr->array_buffer.buffer) in tracer_tracing_is_on_cpu() 687 return ring_buffer_record_is_on_cpu(tr->array_buffer.buffer, cpu); in tracer_tracing_is_on_cpu() 1159 return this_cpu_read(tr->array_buffer.data->ftrace_ignore_pid) != in ftrace_trace_task()
|
H A D | trace_functions_graph.c | 121 struct trace_buffer *buffer = tr->array_buffer.buffer; in __graph_entry() 161 struct trace_buffer *buffer = tr->array_buffer.buffer; in __trace_graph_retaddr_entry() 315 struct trace_buffer *buffer = tr->array_buffer.buffer; in __trace_graph_return() 614 ring_buffer_consume(iter->array_buffer->buffer, iter->cpu, in get_return_for_leaf() 616 event = ring_buffer_peek(iter->array_buffer->buffer, iter->cpu, in get_return_for_leaf() 676 usecs = iter->ts - iter->array_buffer->time_start; in print_graph_rel_time()
|
H A D | trace_osnoise.c | 462 buffer = inst->tr->array_buffer.buffer; \ 507 buffer = inst->tr->array_buffer.buffer; in record_osnoise_sample() 581 buffer = inst->tr->array_buffer.buffer; in record_timerlat_sample() 668 buffer = inst->tr->array_buffer.buffer; in timerlat_dump_stack() 1175 trace_array_printk_buf(tr->array_buffer.buffer, _THIS_IP_, in osnoise_stop_exception() 1348 trace_array_printk_buf(tr->array_buffer.buffer, _THIS_IP_, in osnoise_stop_tracing()
|
H A D | trace_hwlat.c | 133 struct trace_buffer *buffer = tr->array_buffer.buffer; in trace_hwlat_sample() 189 trace_array_printk_buf(tr->array_buffer.buffer, _THIS_IP_, msg); \
|
H A D | trace_events.c | 648 return this_cpu_read(tr->array_buffer.data->ignore_pid) != 0; in trace_event_ignore_this_pid() 1082 this_cpu_write(tr->array_buffer.data->ignore_pid, ret || in event_filter_pid_sched_switch_probe_pre() 1100 this_cpu_write(tr->array_buffer.data->ignore_pid, in event_filter_pid_sched_switch_probe_post() 1112 if (!this_cpu_read(tr->array_buffer.data->ignore_pid)) in event_filter_pid_sched_wakeup_probe_pre() 1118 this_cpu_write(tr->array_buffer.data->ignore_pid, in event_filter_pid_sched_wakeup_probe_pre() 1130 if (this_cpu_read(tr->array_buffer.data->ignore_pid)) in event_filter_pid_sched_wakeup_probe_post() 1137 this_cpu_write(tr->array_buffer.data->ignore_pid, in event_filter_pid_sched_wakeup_probe_post() 1180 per_cpu_ptr(tr->array_buffer.data, cpu)->ignore_pid = false; in __ftrace_clear_event_pids() 2330 ring_buffer_print_page_header(tr->array_buffer.buffer, s); in show_header_page_file() 2378 this_cpu_write(tr->array_buffer.data->ignore_pid, in ignore_task_cpu()
|
H A D | trace_branch.c | 60 buffer = tr->array_buffer.buffer; in probe_likely_condition()
|
H A D | blktrace.c | 81 buffer = blk_tr->array_buffer.buffer; in trace_note() 256 buffer = blk_tr->array_buffer.buffer; in __blk_add_trace()
|
H A D | trace_events_synth.c | 538 buffer = trace_file->tr->array_buffer.buffer; in trace_event_raw_event_synth() 1696 trace_state->buffer = file->tr->array_buffer.buffer; in __synth_event_trace_start()
|
H A D | fgraph.c | 1195 pid = this_cpu_read(tr->array_buffer.data->ftrace_ignore_pid); in fgraph_pid_func()
|
H A D | ftrace.c | 180 pid = this_cpu_read(tr->array_buffer.data->ftrace_ignore_pid); in ftrace_pid_func() 8162 this_cpu_write(tr->array_buffer.data->ftrace_ignore_pid, in ftrace_filter_pid_sched_switch_probe() 8165 this_cpu_write(tr->array_buffer.data->ftrace_ignore_pid, in ftrace_filter_pid_sched_switch_probe() 8231 per_cpu_ptr(tr->array_buffer.data, cpu)->ftrace_ignore_pid = FTRACE_PID_TRACE; in clear_ftrace_pids() 8430 this_cpu_write(tr->array_buffer.data->ftrace_ignore_pid, in ignore_task_cpu() 8433 this_cpu_write(tr->array_buffer.data->ftrace_ignore_pid, in ignore_task_cpu()
|
H A D | trace_output.c | 574 unsigned long long abs_ts = iter->ts - iter->array_buffer->time_start; in lat_print_timestamp()
|
/linux/include/linux/ |
H A D | trace_events.h | 14 struct array_buffer; 99 struct array_buffer *array_buffer; member
|