Lines Matching refs:BPF_REG_SIZE

501 	for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) {  in print_verifier_state()
502 char types_buf[BPF_REG_SIZE + 1]; in print_verifier_state()
506 for (j = 0; j < BPF_REG_SIZE; j++) { in print_verifier_state()
512 types_buf[BPF_REG_SIZE] = 0; in print_verifier_state()
515 verbose(env, " fp%d", (-i - 1) * BPF_REG_SIZE); in print_verifier_state()
556 COPY_STATE_FN(stack, allocated_stack, stack, BPF_REG_SIZE) in COPY_STATE_FN() argument
596 REALLOC_STATE_FN(stack, allocated_stack, stack, BPF_REG_SIZE)
1516 spi = (-insn->off - 1) / BPF_REG_SIZE; in backtrack_insn()
1535 spi = (-insn->off - 1) / BPF_REG_SIZE; in backtrack_insn()
1649 for (j = 0; j < func->allocated_stack / BPF_REG_SIZE; j++) { in mark_all_scalars_precise()
1774 if (i >= func->allocated_stack / BPF_REG_SIZE) { in __mark_chain_precision()
1877 for (i = 0; i < BPF_REG_SIZE; i++) in save_register_state()
1889 int i, slot = -off - 1, spi = slot / BPF_REG_SIZE, err; in check_stack_write()
1893 err = realloc_func_state(state, round_up(slot + 1, BPF_REG_SIZE), in check_stack_write()
1902 size != BPF_REG_SIZE) { in check_stack_write()
1911 if (reg && size == BPF_REG_SIZE && register_is_const(reg) && in check_stack_write()
1927 if (size != BPF_REG_SIZE) { in check_stack_write()
1944 for (i = 0; i < BPF_REG_SIZE; i++) in check_stack_write()
1951 int soff = (-spi - 1) * BPF_REG_SIZE; in check_stack_write()
1981 for (i = 0; i < BPF_REG_SIZE; i++) in check_stack_write()
1992 if (size == BPF_REG_SIZE) in check_stack_write()
2006 state->stack[spi].slot_type[(slot - i) % BPF_REG_SIZE] = in check_stack_write()
2018 int i, slot = -off - 1, spi = slot / BPF_REG_SIZE; in check_stack_read()
2031 if (size != BPF_REG_SIZE) { in check_stack_read()
2044 for (i = 1; i < BPF_REG_SIZE; i++) { in check_stack_read()
2045 if (stype[(slot - i) % BPF_REG_SIZE] != STACK_SPILL) { in check_stack_read()
2065 if (stype[(slot - i) % BPF_REG_SIZE] == STACK_MISC) in check_stack_read()
2067 if (stype[(slot - i) % BPF_REG_SIZE] == STACK_ZERO) { in check_stack_read()
2879 if (!err && size < BPF_REG_SIZE && value_regno >= 0 && t == BPF_READ && in check_mem_access()
3045 spi = slot / BPF_REG_SIZE; in check_stack_boundary()
3048 stype = &state->stack[spi].slot_type[slot % BPF_REG_SIZE]; in check_stack_boundary()
3059 for (j = 0; j < BPF_REG_SIZE; j++) in check_stack_boundary()
6681 #define ID_MAP_SIZE (MAX_BPF_REG + MAX_BPF_STACK / BPF_REG_SIZE)
6733 for (i = 0; i < st->allocated_stack / BPF_REG_SIZE; i++) { in clean_func_state()
6739 for (j = 0; j < BPF_REG_SIZE; j++) in clean_func_state()
6940 spi = i / BPF_REG_SIZE; in stacksafe()
6943 i += BPF_REG_SIZE - 1; in stacksafe()
6948 if (old->stack[spi].slot_type[i % BPF_REG_SIZE] == STACK_INVALID) in stacksafe()
6961 if (old->stack[spi].slot_type[i % BPF_REG_SIZE] == STACK_MISC && in stacksafe()
6962 cur->stack[spi].slot_type[i % BPF_REG_SIZE] == STACK_ZERO) in stacksafe()
6964 if (old->stack[spi].slot_type[i % BPF_REG_SIZE] != in stacksafe()
6965 cur->stack[spi].slot_type[i % BPF_REG_SIZE]) in stacksafe()
6972 if (i % BPF_REG_SIZE) in stacksafe()
7153 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE && in propagate_liveness()
7154 i < parent->allocated_stack / BPF_REG_SIZE; i++) { in propagate_liveness()
7189 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) { in propagate_precision()
7198 (-i - 1) * BPF_REG_SIZE); in propagate_precision()
7422 for (i = 0; i < frame->allocated_stack / BPF_REG_SIZE; i++) { in is_state_visited()