Lines Matching refs:allocated_stack

351 	for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) {  in print_verifier_state()
370 if (WARN_ON_ONCE(dst->allocated_stack < src->allocated_stack)) { in copy_stack_state()
376 sizeof(*src->stack) * (src->allocated_stack / BPF_REG_SIZE)); in copy_stack_state()
390 u32 old_size = state->allocated_stack; in realloc_func_state()
397 state->allocated_stack = slot * BPF_REG_SIZE; in realloc_func_state()
415 state->allocated_stack = slot * BPF_REG_SIZE; in realloc_func_state()
450 err = realloc_func_state(dst, src->allocated_stack, false); in copy_func_state()
453 memcpy(dst, src, offsetof(struct bpf_func_state, allocated_stack)); in copy_func_state()
1145 if (parent->frame[frameno]->allocated_stack <= slot * BPF_REG_SIZE) in mark_stack_slot_read()
1173 if (reg_state->allocated_stack <= slot) { in check_stack_read()
1893 if (state->allocated_stack <= slot) in check_stack_boundary()
2308 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) { in __clear_all_pkt_pointers()
3458 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) { in find_good_pkt_pointers()
3715 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) { in mark_map_regs()
4484 if (old->allocated_stack > cur->allocated_stack) in stacksafe()
4491 for (i = 0; i < old->allocated_stack; i++) { in stacksafe()
4645 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE && in propagate_liveness()
4646 i < parent->allocated_stack / BPF_REG_SIZE; i++) { in propagate_liveness()
4725 for (i = 0; i < frame->allocated_stack / BPF_REG_SIZE; i++) in is_state_visited()