Searched refs:BPF_REG_SIZE (Results 1 – 4 of 4) sorted by relevance
177 #define BPF_REG_SIZE 8 /* size of eBPF register in bytes */ macro181 u8 slot_type[BPF_REG_SIZE];238 #define BPF_ID_MAP_SIZE (MAX_BPF_REG + MAX_BPF_STACK / BPF_REG_SIZE)308 (((slot < frame->allocated_stack / BPF_REG_SIZE) && \315 iter < frame->allocated_stack / BPF_REG_SIZE; \
20 #define BPF_REG_SIZE 8 macro156 ins(BPF_ALU64_IMM(BPF_MOV, BPF_REG_ARG2, BPF_REG_SIZE), pos); in gen_read_mem()314 BPF_REG_FP, -BPF_REG_SIZE * (i + 1)), pos); in gen_prologue_slowpath()
704 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) { in print_verifier_state()705 char types_buf[BPF_REG_SIZE + 1]; in print_verifier_state()709 for (j = 0; j < BPF_REG_SIZE; j++) { in print_verifier_state()715 types_buf[BPF_REG_SIZE] = 0; in print_verifier_state()718 verbose(env, " fp%d", (-i - 1) * BPF_REG_SIZE); in print_verifier_state()808 size_t n = src->allocated_stack / BPF_REG_SIZE; in copy_stack_state()832 size_t old_n = state->allocated_stack / BPF_REG_SIZE, n = size / BPF_REG_SIZE; in grow_stack_state()2242 spi = (-insn->off - 1) / BPF_REG_SIZE; in backtrack_insn()2261 spi = (-insn->off - 1) / BPF_REG_SIZE; in backtrack_insn()2375 for (j = 0; j < func->allocated_stack / BPF_REG_SIZE; j++) { in mark_all_scalars_precise()[all …]
108 stack_entry = &state->stack[soff / BPF_REG_SIZE]; in nfp_bpf_map_update_value_ok()109 if (stack_entry->slot_type[soff % BPF_REG_SIZE] == STACK_ZERO) in nfp_bpf_map_update_value_ok()791 nfp_prog->subprog[i].stack_depth += BPF_REG_SIZE * 4; in nfp_bpf_finalize()