/linux/tools/testing/selftests/bpf/progs/ |
H A D | uretprobe_stack.c | 35 entry1_len = bpf_get_stack(ctx, &entry_stack1, SZ, BPF_F_USER_STACK); in BPF_UPROBE() 37 entry1_recur_len = bpf_get_stack(ctx, &entry_stack1_recur, SZ, BPF_F_USER_STACK); in BPF_UPROBE() 51 exit1_recur_len = bpf_get_stack(ctx, &exit_stack1_recur, SZ, BPF_F_USER_STACK); in BPF_URETPROBE() 53 exit1_len = bpf_get_stack(ctx, &exit_stack1, SZ, BPF_F_USER_STACK); in BPF_URETPROBE() 62 entry2_len = bpf_get_stack(ctx, &entry_stack2, SZ, BPF_F_USER_STACK); in BPF_UPROBE() 71 entry3_len = bpf_get_stack(ctx, &entry_stack3, SZ, BPF_F_USER_STACK); in BPF_UPROBE() 80 entry4_len = bpf_get_stack(ctx, &entry_stack4, SZ, BPF_F_USER_STACK); in BPF_UPROBE() 87 exit4_len = bpf_get_stack(ctx, &exit_stack4, SZ, BPF_F_USER_STACK); in BPF_URETPROBE() 94 usdt_len = bpf_get_stack(ctx, &usdt_stack, SZ, BPF_F_USER_STACK); in BPF_USDT()
|
H A D | test_global_func_ctx_args.c | 19 return bpf_get_stack(ctx, &stack, sizeof(stack), 0); in kprobe_typedef_ctx_subprog() 50 return bpf_get_stack((void *)ctx, &stack, sizeof(stack), 0); in kprobe_struct_ctx_subprog() 67 return bpf_get_stack(ctx, &stack, sizeof(stack), 0); in kprobe_workaround_ctx_subprog() 83 return bpf_get_stack(ctx, &stack, sizeof(stack), 0); in raw_tp_ctx_subprog() 99 return bpf_get_stack(ctx, &stack, sizeof(stack), 0); in raw_tp_writable_ctx_subprog() 115 return bpf_get_stack(ctx, &stack, sizeof(stack), 0); in perf_event_ctx_subprog() 130 return bpf_get_stack(ctx, stack, sizeof(stack), 0); in subprog_ctx_tag() 142 return bpf_get_stack(ctx1, stack, sizeof(stack), 0) + in subprog_multi_ctx_tags() 144 bpf_get_stack(ctx2, stack, sizeof(stack), 0); in subprog_multi_ctx_tags()
|
H A D | test_get_stack_rawtp.c | 73 data->kern_stack_size = bpf_get_stack(ctx, data->kern_stack, in bpf_prog1() 75 data->user_stack_size = bpf_get_stack(ctx, data->user_stack, max_len, in bpf_prog1() 77 data->user_stack_buildid_size = bpf_get_stack( in bpf_prog1() 87 usize = bpf_get_stack(ctx, raw_data, max_len, BPF_F_USER_STACK); in bpf_prog1() 91 ksize = bpf_get_stack(ctx, raw_data + usize, max_len - usize, 0); in bpf_prog1()
|
H A D | test_build_id.c | 16 res_nofault = bpf_get_stack(ctx, stack_nofault, sizeof(stack_nofault), in uprobe_nofault() 25 res_sleepable = bpf_get_stack(ctx, stack_sleepable, sizeof(stack_sleepable), in uprobe_sleepable()
|
H A D | perf_event_stackmap.c | 48 val = bpf_get_stack(ctx, trace, sizeof(stack_trace_t), 0); in oncpu() 52 val = bpf_get_stack(ctx, trace, sizeof(stack_trace_t), BPF_F_USER_STACK); in oncpu()
|
H A D | test_get_stack_rawtp_err.c | 15 error = bpf_get_stack(ctx, stack, 0, -1); in bpf_prog2()
|
H A D | verifier_bpf_get_stack.c | 73 : __imm(bpf_get_stack), in stack_return_r0_within_range()
|
H A D | stacktrace_map_skip.c | 55 bpf_get_stack(ctx, stack_p, max_len, TEST_STACK_DEPTH); in oncpu()
|
H A D | test_stacktrace_build_id.c | 60 bpf_get_stack(args, stack_p, max_len, in oncpu()
|
H A D | test_stacktrace_map.c | 70 bpf_get_stack(ctx, stack_p, max_len, 0); in oncpu()
|
H A D | verifier_global_subprogs.c | 176 return bpf_get_stack(ctx, stack, sizeof(stack), 0); in subprog_ctx_tag()
|
/linux/kernel/bpf/ |
H A D | stackmap.c | 496 BPF_CALL_4(bpf_get_stack, struct pt_regs *, regs, void *, buf, u32, size, in BPF_CALL_4() argument 503 .func = bpf_get_stack,
|
/linux/kernel/trace/ |
H A D | bpf_trace.c | 85 u64 bpf_get_stack(u64 r1, u64 r2, u64 r3, u64 r4, u64 r5); 1711 return bpf_get_stack((unsigned long) regs, (unsigned long) buf, in BPF_CALL_4() 1948 ret = bpf_get_stack((unsigned long) regs, (unsigned long) buf, in BPF_CALL_4()
|