Lines Matching refs:allocated_stack

612        int allocated_slots = state->allocated_stack / BPF_REG_SIZE;  in is_spi_bounds_valid()
856 for (i = 1; i < state->allocated_stack / BPF_REG_SIZE; i++) { in unmark_stack_slots_dynptr()
1446 size_t n = src->allocated_stack / BPF_REG_SIZE; in copy_stack_state()
1453 dst->allocated_stack = src->allocated_stack; in copy_stack_state()
1473 size_t old_n = state->allocated_stack / BPF_REG_SIZE, n; in grow_stack_state()
1486 state->allocated_stack = size; in grow_stack_state()
4635 for (j = 0; j < func->allocated_stack / BPF_REG_SIZE; j++) { in mark_all_scalars_precise()
4665 for (j = 0; j < func->allocated_stack / BPF_REG_SIZE; j++) { in mark_all_scalars_imprecise()
4899 if (verifier_bug_if(i >= func->allocated_stack / BPF_REG_SIZE, in __mark_chain_precision()
4901 i, func->allocated_stack / BPF_REG_SIZE)) in __mark_chain_precision()
7446 min_valid_off = -state->allocated_stack; in check_stack_slot_within_bounds()
8100 if (state->allocated_stack <= stack_off) in check_stack_range_initialized()
8117 if (state->allocated_stack <= slot) { in check_stack_range_initialized()
8914 for (i = 0; i < fold->allocated_stack / BPF_REG_SIZE; i++) { in widen_imprecise_scalars()
16552 for (j = 0; j < func->allocated_stack / BPF_REG_SIZE; j++) { in collect_linked_regs()
18468 for (i = 0; i < st->allocated_stack / BPF_REG_SIZE; i++) { in clean_func_state()
18735 for (i = 0; i < old->allocated_stack; i++) { in stacksafe()
18741 (i >= cur->allocated_stack || in stacksafe()
18763 if (i >= cur->allocated_stack) in stacksafe()
19066 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE && in propagate_liveness()
19067 i < parent->allocated_stack / BPF_REG_SIZE; i++) { in propagate_liveness()
19112 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) { in propagate_precision()
19270 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) { in iter_active_depths_differ()
19674 for (i = 0; i < frame->allocated_stack / BPF_REG_SIZE; i++) { in is_state_visited()