Lines Matching refs:curframe
1693 for (i = 0; i <= state->curframe; i++) { in free_verifier_state()
1760 for (i = src->curframe + 1; i <= dst_state->curframe; i++) { in copy_verifier_state()
1770 dst_state->curframe = src->curframe; in copy_verifier_state()
1779 for (i = 0; i <= src->curframe; i++) { in copy_verifier_state()
1802 struct bpf_func_state *state = cur->frame[cur->curframe]; in explored_state()
1811 if (a->curframe != b->curframe) in same_callsites()
1814 for (fr = a->curframe; fr >= 0; fr--) in same_callsites()
1824 return frame == st->curframe in frame_insn_idx()
1849 for (i = 0; i <= st->curframe; i++) { in compute_scc_callchain()
1855 } else if (i < st->curframe) { in compute_scc_callchain()
3970 struct bpf_func_state *state = vstate->frame[vstate->curframe]; in check_reg_arg()
4737 st->curframe); in mark_all_scalars_precise()
4747 for (i = 0; i <= st->curframe; i++) { in mark_all_scalars_precise()
4781 for (i = 0; i <= st->curframe; i++) { in mark_all_scalars_imprecise()
4907 bt_init(bt, starting_state->curframe); in __mark_chain_precision()
4942 if (st->curframe == 0 && in __mark_chain_precision()
5249 cur = env->cur_state->frame[env->cur_state->curframe]; in check_stack_write_fixed_off()
5389 cur = env->cur_state->frame[env->cur_state->curframe]; in check_stack_write_var_off()
5496 struct bpf_func_state *state = vstate->frame[vstate->curframe]; in mark_reg_stack_read()
5536 struct bpf_func_state *state = vstate->frame[vstate->curframe]; in check_stack_read_fixed_off()
5874 struct bpf_func_state *state = vstate->frame[vstate->curframe]; in check_mem_region_access()
6219 struct bpf_func_state *state = vstate->frame[vstate->curframe]; in check_map_access()
9093 for (fr = old->curframe; fr >= 0; fr--) { in widen_imprecise_scalars()
9208 struct bpf_func_state *cur_fr = cur_st->frame[cur_st->curframe], *queued_fr; in process_iter_next_call()
9247 queued_fr = queued_st->frame[queued_st->curframe]; in process_iter_next_call()
10528 struct bpf_func_state *state = vstate->frame[vstate->curframe]; in mark_pkt_end()
10625 if (state->curframe + 1 >= MAX_CALL_FRAMES) { in setup_func_entry()
10627 state->curframe + 2); in setup_func_entry()
10631 if (state->frame[state->curframe + 1]) { in setup_func_entry()
10632 verifier_bug(env, "Frame %d already allocated", state->curframe + 1); in setup_func_entry()
10636 caller = state->frame[state->curframe]; in setup_func_entry()
10640 state->frame[state->curframe + 1] = callee; in setup_func_entry()
10649 state->curframe + 1 /* frameno within this callchain */, in setup_func_entry()
10656 state->curframe++; in setup_func_entry()
10662 state->frame[state->curframe + 1] = NULL; in setup_func_entry()
10803 caller = state->frame[state->curframe]; in push_callback_call()
10859 callback_state->frame[callback_state->curframe - 1]->callback_depth++; in push_callback_call()
10877 caller = state->frame[state->curframe]; in check_func_call()
10940 print_verifier_state(env, state, state->curframe, true); in check_func_call()
11191 if (!state->curframe) in in_rbtree_lock_required_cb()
11194 callee = state->frame[state->curframe]; in in_rbtree_lock_required_cb()
11224 callee = state->frame[state->curframe]; in prepare_func_exit()
11237 caller = state->frame[state->curframe - 1]; in prepare_func_exit()
11284 state->frame[state->curframe--] = NULL; in prepare_func_exit()
11818 env->cur_state->curframe); in check_helper_call()
12131 if (env->cur_state->curframe) { in check_helper_call()
14129 regs = branch->frame[branch->curframe]->regs; in check_kfunc_call()
14612 regs = branch->frame[branch->curframe]->regs; in sanitize_speculative_path()
14847 struct bpf_func_state *state = vstate->frame[vstate->curframe]; in adjust_ptr_min_max_vals()
15999 regs = branch->frame[branch->curframe]->regs; in maybe_fork_scalars()
16176 struct bpf_func_state *state = vstate->frame[vstate->curframe]; in adjust_reg_min_max_vals()
16257 print_verifier_state(env, vstate, vstate->curframe, true); in adjust_reg_min_max_vals()
16262 print_verifier_state(env, vstate, vstate->curframe, true); in adjust_reg_min_max_vals()
17218 struct bpf_func_state *state = vstate->frame[vstate->curframe]; in mark_ptr_or_null_regs()
17374 for (i = vstate->curframe; i >= 0; i--) { in collect_linked_regs()
17454 struct bpf_reg_state *regs = this_branch->frame[this_branch->curframe]->regs; in check_cond_jmp_op()
17569 print_insn_state(env, this_branch, this_branch->curframe); in check_cond_jmp_op()
17584 print_insn_state(env, this_branch, this_branch->curframe); in check_cond_jmp_op()
17606 other_branch_regs = other_branch->frame[other_branch->curframe]->regs; in check_cond_jmp_op()
17702 print_insn_state(env, this_branch, this_branch->curframe); in check_cond_jmp_op()
19640 for (i = 0; i <= st->curframe; i++) { in clean_verifier_state()
20202 if (old->curframe != cur->curframe) in states_equal()
20222 for (i = 0; i <= old->curframe; i++) { in states_equal()
20245 for (fr = old->curframe; fr >= 0; fr--) { in propagate_precision()
20332 int i, fr = cur->curframe; in states_maybe_looping()
20334 if (old->curframe != fr) in states_maybe_looping()
20415 for (fr = old->curframe; fr >= 0; fr--) { in iter_active_depths_differ()
20470 struct bpf_func_state *frame = sl->state.frame[sl->state.curframe]; in is_state_visited()
20473 frame->async_entry_cnt != cur->frame[cur->curframe]->async_entry_cnt) { in is_state_visited()
20529 cur_frame = cur->frame[cur->curframe]; in is_state_visited()
20570 print_verifier_state(env, cur, cur->curframe, true); in is_state_visited()
20572 print_verifier_state(env, &sl->state, cur->curframe, true); in is_state_visited()
20907 !env->cur_state->curframe, in process_bpf_exit_full()
20924 if (env->cur_state->curframe) { in process_bpf_exit_full()
21273 print_verifier_state(env, state, state->curframe, true); in do_check()
21279 print_insn_state(env, state, state->curframe); in do_check()
24543 state->curframe = 0; in do_check_common()
24560 regs = state->frame[state->curframe]->regs; in do_check_common()