| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | test_stacktrace_build_id.c | 7 #ifndef PERF_MAX_STACK_DEPTH 8 #define PERF_MAX_STACK_DEPTH 127 macro 25 typedef struct bpf_stack_build_id stack_trace_t[PERF_MAX_STACK_DEPTH]; 46 * PERF_MAX_STACK_DEPTH; in oncpu()
|
| H A D | test_stacktrace_map.c | |
| H A D | stacktrace_ips.c | 8 #ifndef PERF_MAX_STACK_DEPTH 9 #define PERF_MAX_STACK_DEPTH 127 macro 12 typedef __u64 stack_trace_t[PERF_MAX_STACK_DEPTH];
|
| H A D | perf_event_stackmap.c | 6 #ifndef PERF_MAX_STACK_DEPTH 7 #define PERF_MAX_STACK_DEPTH 127 macro 10 typedef __u64 stack_trace_t[PERF_MAX_STACK_DEPTH];
|
| H A D | strobemeta.h | 19 #define PERF_MAX_STACK_DEPTH 127 macro 222 __uint(value_size, sizeof(uint64_t) * PERF_MAX_STACK_DEPTH); 229 __uint(value_size, sizeof(uint64_t) * PERF_MAX_STACK_DEPTH);
|
| /linux/samples/bpf/ |
| H A D | spintest.bpf.c | 12 #ifndef PERF_MAX_STACK_DEPTH 13 #define PERF_MAX_STACK_DEPTH 127 macro 32 __uint(value_size, PERF_MAX_STACK_DEPTH * sizeof(u64));
|
| H A D | offwaketime_user.c | 49 __u64 ip[PERF_MAX_STACK_DEPTH] = {}; in print_stack() 57 for (i = PERF_MAX_STACK_DEPTH - 1; i >= 0; i--) in print_stack() 64 for (i = 0; i < PERF_MAX_STACK_DEPTH; i++) in print_stack()
|
| H A D | offwaketime.bpf.c | 13 #ifndef PERF_MAX_STACK_DEPTH 14 #define PERF_MAX_STACK_DEPTH 127 macro 56 __uint(value_size, PERF_MAX_STACK_DEPTH * sizeof(u64));
|
| H A D | trace_event_user.c | 63 __u64 ip[PERF_MAX_STACK_DEPTH] = {}; in print_stack() 71 for (i = PERF_MAX_STACK_DEPTH - 1; i >= 0; i--) in print_stack() 78 for (i = PERF_MAX_STACK_DEPTH - 1; i >= 0; i--) in print_stack()
|
| H A D | trace_event_kern.c | 30 __uint(value_size, PERF_MAX_STACK_DEPTH * sizeof(u64));
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | stacktrace_build_id.c | 13 struct bpf_stack_build_id id_offs[PERF_MAX_STACK_DEPTH]; in test_stacktrace_build_id() 71 for (i = 0; i < PERF_MAX_STACK_DEPTH; ++i) in test_stacktrace_build_id() 95 stack_trace_len = PERF_MAX_STACK_DEPTH * in test_stacktrace_build_id()
|
| H A D | stacktrace_ips.c | 8 __u64 ips[PERF_MAX_STACK_DEPTH]; in check_stacktrace_ips() 19 if (!ASSERT_LT(cnt, PERF_MAX_STACK_DEPTH, "check_max")) in check_stacktrace_ips()
|
| H A D | stacktrace_build_id_nmi.c | 17 struct bpf_stack_build_id id_offs[PERF_MAX_STACK_DEPTH]; in test_stacktrace_build_id_nmi() 99 for (i = 0; i < PERF_MAX_STACK_DEPTH; ++i) in test_stacktrace_build_id_nmi()
|
| H A D | stacktrace_map.c | 11 __u64 stack[PERF_MAX_STACK_DEPTH]; in test_stacktrace_map() 46 stack_trace_len = PERF_MAX_STACK_DEPTH * sizeof(__u64); in test_stacktrace_map()
|
| /linux/kernel/events/ |
| H A D | callchain.c | 23 int sysctl_perf_event_max_stack __read_mostly = PERF_MAX_STACK_DEPTH;
|
| /linux/tools/include/uapi/linux/ |
| H A D | perf_event.h | 1285 #define PERF_MAX_STACK_DEPTH 127 macro
|
| /linux/include/uapi/linux/ |
| H A D | perf_event.h | 1285 #define PERF_MAX_STACK_DEPTH 127 macro
|
| /linux/tools/perf/util/ |
| H A D | db-export.c | 232 sample, NULL, NULL, PERF_MAX_STACK_DEPTH); in call_path_from_sample()
|
| H A D | build-id.c | 77 sample__for_each_callchain_node(thread, evsel, sample, PERF_MAX_STACK_DEPTH, in build_id__mark_dso_hit()
|
| /linux/tools/perf/ |
| H A D | builtin-report.c | 1321 .max_stack = PERF_MAX_STACK_DEPTH, in cmd_report() 1388 "Default: kernel.perf_event_max_stack or " __stringify(PERF_MAX_STACK_DEPTH)), in cmd_report()
|
| H A D | builtin-annotate.c | 228 ret = hist_entry_iter__add(&iter, &a, PERF_MAX_STACK_DEPTH, ann); in process_branch_callback()
|
| H A D | builtin-top.c | 1546 "Default: kernel.perf_event_max_stack or " __stringify(PERF_MAX_STACK_DEPTH)), in cmd_top()
|
| H A D | builtin-diff.c | 440 if (hist_entry_iter__add(&iter, &al, PERF_MAX_STACK_DEPTH, in diff__process_sample_event()
|
| H A D | builtin-inject.c | 938 sample__for_each_callchain_node(thread, evsel, sample, PERF_MAX_STACK_DEPTH, in perf_event__inject_buildid()
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | test_progs.c | 675 char val_buf[PERF_MAX_STACK_DEPTH * in compare_map_keys()
|