/linux/fs/xfs/scrub/ |
H A D | xfarray.c | 764 if (si->stack_depth >= si->max_stack_depth - 1) { in xfarray_qsort_push() 765 ASSERT(si->stack_depth < si->max_stack_depth - 1); in xfarray_qsort_push() 770 si->stack_depth + 2); in xfarray_qsort_push() 772 si_lo[si->stack_depth + 1] = lo + 1; in xfarray_qsort_push() 773 si_hi[si->stack_depth + 1] = si_hi[si->stack_depth]; in xfarray_qsort_push() 774 si_hi[si->stack_depth++] = lo - 1; in xfarray_qsort_push() 780 if (si_hi[si->stack_depth] - si_lo[si->stack_depth] > in xfarray_qsort_push() 781 si_hi[si->stack_depth - 1] - si_lo[si->stack_depth - 1]) { in xfarray_qsort_push() 782 swap(si_lo[si->stack_depth], si_lo[si->stack_depth - 1]); in xfarray_qsort_push() 783 swap(si_hi[si->stack_depth], si_hi[si->stack_depth - 1]); in xfarray_qsort_push() [all …]
|
H A D | trace.h | 1121 __field(int, stack_depth) 1128 __entry->stack_depth = si->stack_depth; 1136 __entry->stack_depth,
|
/linux/lib/ |
H A D | test_bpf.c | 84 int stack_depth; /* for eBPF only, since tests don't call verifier */ member 452 self->stack_depth = 40; in __bpf_fill_stxdw() 8159 .stack_depth = 8, 8179 .stack_depth = 8, 8196 .stack_depth = 0, 8212 .stack_depth = 0, 8228 .stack_depth = 0, 8248 .stack_depth = 8, 8268 .stack_depth = 8, 8285 .stack_depth = 0, [all …]
|
/linux/arch/s390/net/ |
H A D | bpf_jit_comp.c | 424 static void restore_regs(struct bpf_jit *jit, u32 rs, u32 re, u32 stack_depth) in restore_regs() argument 429 off += STK_OFF + stack_depth; in restore_regs() 473 static void save_restore_regs(struct bpf_jit *jit, int op, u32 stack_depth, in save_restore_regs() argument 497 restore_regs(jit, rs, re, stack_depth); in save_restore_regs() 557 u32 stack_depth) in bpf_jit_prologue() argument 590 save_restore_regs(jit, REGS_SAVE, stack_depth, in bpf_jit_prologue() 614 EMIT4_IMM(0xa70b0000, REG_15, -(STK_OFF + stack_depth)); in bpf_jit_prologue() 662 static void bpf_jit_epilogue(struct bpf_jit *jit, u32 stack_depth) in bpf_jit_epilogue() argument 668 save_restore_regs(jit, REGS_RESTORE, stack_depth, 0); in bpf_jit_epilogue() 859 int i, bool extra_pass, u32 stack_depth) in bpf_jit_insn() argument [all …]
|
/linux/arch/x86/net/ |
H A D | bpf_jit_comp.c | 508 static void emit_prologue(u8 **pprog, u32 stack_depth, bool ebpf_from_cbpf, in emit_prologue() argument 550 if (stack_depth) in emit_prologue() 551 EMIT3_off32(0x48, 0x81, 0xEC, round_up(stack_depth, 8)); in emit_prologue() 707 u32 stack_depth, u8 *ip, in emit_bpf_tail_call_indirect() argument 710 int tcc_ptr_off = BPF_TAIL_CALL_CNT_PTR_STACK_OFF(stack_depth); in emit_bpf_tail_call_indirect() 772 if (stack_depth) in emit_bpf_tail_call_indirect() 774 round_up(stack_depth, 8)); in emit_bpf_tail_call_indirect() 796 bool *callee_regs_used, u32 stack_depth, in emit_bpf_tail_call_direct() argument 799 int tcc_ptr_off = BPF_TAIL_CALL_CNT_PTR_STACK_OFF(stack_depth); in emit_bpf_tail_call_direct() 839 if (stack_depth) in emit_bpf_tail_call_direct() [all …]
|
H A D | bpf_jit_comp32.c | 181 #define _STACK_SIZE (stack_depth + SCRATCH_SIZE) 1200 static void emit_prologue(u8 **pprog, u32 stack_depth) in emit_prologue() argument 1245 static void emit_epilogue(u8 **pprog, u32 stack_depth) in emit_epilogue() argument 1670 emit_prologue(&prog, bpf_prog->aux->stack_depth); in do_jit() 2472 emit_epilogue(&prog, bpf_prog->aux->stack_depth); in do_jit()
|
/linux/drivers/firewire/ |
H A D | core-topology.c | 107 int phy_id, stack_depth; in build_tree() local 114 stack_depth = 0; in build_tree() 168 if (child_port_count > stack_depth) { in build_tree() 236 stack_depth += 1 - child_port_count; in build_tree()
|
/linux/drivers/net/ethernet/netronome/nfp/bpf/ |
H A D | verifier.c | 714 frame_depths[frame] = nfp_prog->subprog[idx].stack_depth; in nfp_bpf_get_stack_usage() 782 nfp_prog->subprog[i].stack_depth = info[i].stack_depth; in nfp_bpf_finalize() 788 nfp_prog->subprog[i].stack_depth += REG_WIDTH; in nfp_bpf_finalize() 791 nfp_prog->subprog[i].stack_depth += BPF_REG_SIZE * 4; in nfp_bpf_finalize()
|
H A D | main.h | 484 u16 stack_depth; member
|
H A D | jit.c | 3264 u32 ret_tgt, stack_depth, offset_br; in bpf_to_bpf_call() local 3267 stack_depth = round_up(nfp_prog->stack_frame_depth, STACK_FRAME_ALIGN); in bpf_to_bpf_call() 3271 if (stack_depth) { in bpf_to_bpf_call() 3272 tmp_reg = ur_load_imm_any(nfp_prog, stack_depth, in bpf_to_bpf_call() 3331 if (stack_depth) { in bpf_to_bpf_call() 3332 tmp_reg = ur_load_imm_any(nfp_prog, stack_depth, in bpf_to_bpf_call() 3649 unsigned int depth = nfp_prog->subprog[meta->subprog_idx].stack_depth; in nfp_start_subprog() 3837 depth = nfp_prog->subprog[0].stack_depth; in nfp_translate()
|
/linux/include/linux/ |
H A D | kcsan-checks.h | 138 int stack_depth; member
|
H A D | bpf_verifier.h | 652 u16 stack_depth; /* max. stack depth used by this function */ member
|
H A D | bpf.h | 1500 u32 stack_depth; member 2590 void bpf_patch_call_args(struct bpf_insn *insn, u32 stack_depth);
|
/linux/kernel/kcsan/ |
H A D | core.c | 437 reorder_access->stack_depth = get_kcsan_stack_depth(); in set_reorder_access() 1122 if (get_kcsan_stack_depth() <= reorder_access->stack_depth) { in __tsan_func_exit() 1133 reorder_access->stack_depth = INT_MIN; in __tsan_func_exit()
|
/linux/kernel/bpf/ |
H A D | core.c | 2284 void bpf_patch_call_args(struct bpf_insn *insn, u32 stack_depth) in bpf_patch_call_args() argument 2286 stack_depth = max_t(u32, stack_depth, 1); in bpf_patch_call_args() 2288 insn->imm = interpreters_args[(round_up(stack_depth, 32) / 32) - 1] - in bpf_patch_call_args() 2382 u32 stack_depth = max_t(u32, fp->aux->stack_depth, 1); in bpf_prog_select_func() local 2384 fp->bpf_func = interpreters[(round_up(stack_depth, 32) / 32) - 1]; in bpf_prog_select_func()
|
H A D | verifier.c | 1460 if (env->subprog_info[state->subprogno].stack_depth < size) in grow_stack_state() 1461 env->subprog_info[state->subprogno].stack_depth = size; in grow_stack_state() 6306 static int round_up_stack_depth(struct bpf_verifier_env *env, int stack_depth) in round_up_stack_depth() argument 6309 return round_up(stack_depth, 16); in round_up_stack_depth() 6314 return round_up(max_t(u32, stack_depth, 1), 32); in round_up_stack_depth() 6364 subprog_depth = round_up_stack_depth(env, subprog[idx].stack_depth); in check_max_stack_depth_subprog() 6480 depth -= round_up_stack_depth(env, subprog[idx].stack_depth); in check_max_stack_depth_subprog() 6543 return env->subprog_info[subprog].stack_depth; in get_callee_stack_depth() 20350 -(subprogs[0].stack_depth + 8)); in convert_ctx_accesses() 20357 subprogs[0].stack_depth += 8; in convert_ctx_accesses() [all …]
|
/linux/fs/overlayfs/ |
H A D | super.c | 398 struct ovl_fs *ofs, int *stack_depth) in ovl_lower_dir() argument 407 *stack_depth = max(*stack_depth, path->mnt->mnt_sb->s_stack_depth); in ovl_lower_dir()
|
/linux/arch/sparc/net/ |
H A D | bpf_jit_comp_64.c | 803 u32 stack_depth; in build_prologue() local 805 stack_depth = prog->aux->stack_depth; in build_prologue() 806 stack_needed += round_up(stack_depth, 16); in build_prologue()
|
/linux/Documentation/bpf/ |
H A D | drgn.rst | 132 .stack_depth = (u32)8,
|
/linux/arch/mips/net/ |
H A D | bpf_jit_comp64.c | 587 locals = ALIGN(ctx->program->aux->stack_depth, MIPS_STACK_ALIGNMENT); in build_prologue()
|
/linux/security/ |
H A D | Kconfig.hardening | 214 the /proc file system. In particular, /proc/<pid>/stack_depth
|
/linux/arch/parisc/net/ |
H A D | bpf_jit_comp64.c | 1113 bpf_stack_adjust = ctx->prog->aux->stack_depth; in bpf_jit_build_prologue()
|
H A D | bpf_jit_comp32.c | 1483 bpf_stack_adjust = ctx->prog->aux->stack_depth; in bpf_jit_build_prologue()
|
/linux/arch/loongarch/net/ |
H A D | bpf_jit.c | 93 bpf_stack_adjust = round_up(ctx->prog->aux->stack_depth, 16); in build_prologue()
|
/linux/arch/riscv/net/ |
H A D | bpf_jit_comp32.c | 1309 round_up(ctx->prog->aux->stack_depth, STACK_ALIGN); in bpf_jit_build_prologue()
|