Lines Matching refs:src_reg

249 	       insn->src_reg == 0;  in bpf_helper_call()
255 insn->src_reg == BPF_PSEUDO_CALL; in bpf_pseudo_call()
261 insn->src_reg == BPF_PSEUDO_KFUNC_CALL; in bpf_pseudo_kfunc_call()
2849 insn[i].src_reg == 0 && in check_subprogs()
3008 if (insn->src_reg == BPF_PSEUDO_CALL) in is_reg64()
3088 return insn->src_reg; in insn_def_regno()
3223 if (insn->src_reg != BPF_PSEUDO_KFUNC_CALL) in disasm_kfunc_name()
3411 u32 sreg = insn->src_reg; in backtrack_insn()
3468 if (insn->src_reg != BPF_REG_FP) in backtrack_insn()
3588 if (insn->src_reg == BPF_PSEUDO_KFUNC_CALL && insn->imm == 0) in backtrack_insn()
6707 err = check_reg_arg(env, insn->src_reg, SRC_OP); in check_atomic()
6730 if (is_pointer_value(env, insn->src_reg)) { in check_atomic()
6731 verbose(env, "R%d leaks addr into mem\n", insn->src_reg); in check_atomic()
6749 load_reg = insn->src_reg; in check_atomic()
8917 insn->src_reg == 0 && in __check_func_call()
11801 mark_reg_unknown(env, regs, insn->src_reg); in sanitize_speculative_path()
11923 u32 dst = insn->dst_reg, src = insn->src_reg; in sanitize_err()
12253 struct bpf_reg_state *src_reg) in scalar32_min_max_add() argument
12255 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_add()
12256 s32 smax_val = src_reg->s32_max_value; in scalar32_min_max_add()
12257 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_add()
12258 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_add()
12279 struct bpf_reg_state *src_reg) in scalar_min_max_add() argument
12281 s64 smin_val = src_reg->smin_value; in scalar_min_max_add()
12282 s64 smax_val = src_reg->smax_value; in scalar_min_max_add()
12283 u64 umin_val = src_reg->umin_value; in scalar_min_max_add()
12284 u64 umax_val = src_reg->umax_value; in scalar_min_max_add()
12305 struct bpf_reg_state *src_reg) in scalar32_min_max_sub() argument
12307 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_sub()
12308 s32 smax_val = src_reg->s32_max_value; in scalar32_min_max_sub()
12309 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_sub()
12310 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_sub()
12333 struct bpf_reg_state *src_reg) in scalar_min_max_sub() argument
12335 s64 smin_val = src_reg->smin_value; in scalar_min_max_sub()
12336 s64 smax_val = src_reg->smax_value; in scalar_min_max_sub()
12337 u64 umin_val = src_reg->umin_value; in scalar_min_max_sub()
12338 u64 umax_val = src_reg->umax_value; in scalar_min_max_sub()
12361 struct bpf_reg_state *src_reg) in scalar32_min_max_mul() argument
12363 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_mul()
12364 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_mul()
12365 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_mul()
12393 struct bpf_reg_state *src_reg) in scalar_min_max_mul() argument
12395 s64 smin_val = src_reg->smin_value; in scalar_min_max_mul()
12396 u64 umin_val = src_reg->umin_value; in scalar_min_max_mul()
12397 u64 umax_val = src_reg->umax_value; in scalar_min_max_mul()
12425 struct bpf_reg_state *src_reg) in scalar32_min_max_and() argument
12427 bool src_known = tnum_subreg_is_const(src_reg->var_off); in scalar32_min_max_and()
12430 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_and()
12431 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_and()
12459 struct bpf_reg_state *src_reg) in scalar_min_max_and() argument
12461 bool src_known = tnum_is_const(src_reg->var_off); in scalar_min_max_and()
12463 s64 smin_val = src_reg->smin_value; in scalar_min_max_and()
12464 u64 umax_val = src_reg->umax_value; in scalar_min_max_and()
12494 struct bpf_reg_state *src_reg) in scalar32_min_max_or() argument
12496 bool src_known = tnum_subreg_is_const(src_reg->var_off); in scalar32_min_max_or()
12499 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_or()
12500 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_or()
12528 struct bpf_reg_state *src_reg) in scalar_min_max_or() argument
12530 bool src_known = tnum_is_const(src_reg->var_off); in scalar_min_max_or()
12532 s64 smin_val = src_reg->smin_value; in scalar_min_max_or()
12533 u64 umin_val = src_reg->umin_value; in scalar_min_max_or()
12563 struct bpf_reg_state *src_reg) in scalar32_min_max_xor() argument
12565 bool src_known = tnum_subreg_is_const(src_reg->var_off); in scalar32_min_max_xor()
12568 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_xor()
12592 struct bpf_reg_state *src_reg) in scalar_min_max_xor() argument
12594 bool src_known = tnum_is_const(src_reg->var_off); in scalar_min_max_xor()
12596 s64 smin_val = src_reg->smin_value; in scalar_min_max_xor()
12641 struct bpf_reg_state *src_reg) in scalar32_min_max_lsh() argument
12643 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_lsh()
12644 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_lsh()
12689 struct bpf_reg_state *src_reg) in scalar_min_max_lsh() argument
12691 u64 umax_val = src_reg->umax_value; in scalar_min_max_lsh()
12692 u64 umin_val = src_reg->umin_value; in scalar_min_max_lsh()
12704 struct bpf_reg_state *src_reg) in scalar32_min_max_rsh() argument
12707 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_rsh()
12708 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_rsh()
12736 struct bpf_reg_state *src_reg) in scalar_min_max_rsh() argument
12738 u64 umax_val = src_reg->umax_value; in scalar_min_max_rsh()
12739 u64 umin_val = src_reg->umin_value; in scalar_min_max_rsh()
12770 struct bpf_reg_state *src_reg) in scalar32_min_max_arsh() argument
12772 u64 umin_val = src_reg->u32_min_value; in scalar32_min_max_arsh()
12793 struct bpf_reg_state *src_reg) in scalar_min_max_arsh() argument
12795 u64 umin_val = src_reg->umin_value; in scalar_min_max_arsh()
12826 struct bpf_reg_state src_reg) in adjust_scalar_min_max_vals() argument
12839 smin_val = src_reg.smin_value; in adjust_scalar_min_max_vals()
12840 smax_val = src_reg.smax_value; in adjust_scalar_min_max_vals()
12841 umin_val = src_reg.umin_value; in adjust_scalar_min_max_vals()
12842 umax_val = src_reg.umax_value; in adjust_scalar_min_max_vals()
12844 s32_min_val = src_reg.s32_min_value; in adjust_scalar_min_max_vals()
12845 s32_max_val = src_reg.s32_max_value; in adjust_scalar_min_max_vals()
12846 u32_min_val = src_reg.u32_min_value; in adjust_scalar_min_max_vals()
12847 u32_max_val = src_reg.u32_max_value; in adjust_scalar_min_max_vals()
12850 src_known = tnum_subreg_is_const(src_reg.var_off); in adjust_scalar_min_max_vals()
12861 src_known = tnum_is_const(src_reg.var_off); in adjust_scalar_min_max_vals()
12901 scalar32_min_max_add(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12902 scalar_min_max_add(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12903 dst_reg->var_off = tnum_add(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
12906 scalar32_min_max_sub(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12907 scalar_min_max_sub(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12908 dst_reg->var_off = tnum_sub(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
12911 dst_reg->var_off = tnum_mul(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
12912 scalar32_min_max_mul(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12913 scalar_min_max_mul(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12916 dst_reg->var_off = tnum_and(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
12917 scalar32_min_max_and(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12918 scalar_min_max_and(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12921 dst_reg->var_off = tnum_or(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
12922 scalar32_min_max_or(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12923 scalar_min_max_or(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12926 dst_reg->var_off = tnum_xor(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
12927 scalar32_min_max_xor(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12928 scalar_min_max_xor(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12939 scalar32_min_max_lsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12941 scalar_min_max_lsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12952 scalar32_min_max_rsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12954 scalar_min_max_rsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12965 scalar32_min_max_arsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12967 scalar_min_max_arsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
12989 struct bpf_reg_state *regs = state->regs, *dst_reg, *src_reg; in adjust_reg_min_max_vals() local
12995 src_reg = NULL; in adjust_reg_min_max_vals()
13004 src_reg = &regs[insn->src_reg]; in adjust_reg_min_max_vals()
13005 if (src_reg->type != SCALAR_VALUE) { in adjust_reg_min_max_vals()
13028 src_reg, dst_reg); in adjust_reg_min_max_vals()
13032 err = mark_chain_precision(env, insn->src_reg); in adjust_reg_min_max_vals()
13036 dst_reg, src_reg); in adjust_reg_min_max_vals()
13039 err = mark_chain_precision(env, insn->src_reg); in adjust_reg_min_max_vals()
13049 src_reg = &off_reg; in adjust_reg_min_max_vals()
13052 ptr_reg, src_reg); in adjust_reg_min_max_vals()
13061 if (WARN_ON(!src_reg)) { in adjust_reg_min_max_vals()
13066 return adjust_scalar_min_max_vals(env, insn, dst_reg, *src_reg); in adjust_reg_min_max_vals()
13079 insn->src_reg != BPF_REG_0 || in check_alu_op()
13085 if (insn->src_reg != BPF_REG_0 || insn->off != 0 || in check_alu_op()
13132 err = check_reg_arg(env, insn->src_reg, SRC_OP); in check_alu_op()
13136 if (insn->src_reg != BPF_REG_0 || insn->off != 0) { in check_alu_op()
13148 struct bpf_reg_state *src_reg = regs + insn->src_reg; in check_alu_op() local
13150 bool need_id = src_reg->type == SCALAR_VALUE && !src_reg->id && in check_alu_op()
13151 !tnum_is_const(src_reg->var_off); in check_alu_op()
13163 src_reg->id = ++env->id_gen; in check_alu_op()
13164 copy_register_state(dst_reg, src_reg); in check_alu_op()
13169 if (is_pointer_value(env, insn->src_reg)) { in check_alu_op()
13172 insn->src_reg); in check_alu_op()
13174 } else if (src_reg->type == SCALAR_VALUE) { in check_alu_op()
13177 no_sext = src_reg->umax_value < (1ULL << (insn->off - 1)); in check_alu_op()
13179 src_reg->id = ++env->id_gen; in check_alu_op()
13180 copy_register_state(dst_reg, src_reg); in check_alu_op()
13192 if (is_pointer_value(env, insn->src_reg)) { in check_alu_op()
13195 insn->src_reg); in check_alu_op()
13197 } else if (src_reg->type == SCALAR_VALUE) { in check_alu_op()
13199 bool is_src_reg_u32 = src_reg->umax_value <= U32_MAX; in check_alu_op()
13202 src_reg->id = ++env->id_gen; in check_alu_op()
13203 copy_register_state(dst_reg, src_reg); in check_alu_op()
13214 bool no_sext = src_reg->umax_value < (1ULL << (insn->off - 1)); in check_alu_op()
13217 src_reg->id = ++env->id_gen; in check_alu_op()
13218 copy_register_state(dst_reg, src_reg); in check_alu_op()
13261 err = check_reg_arg(env, insn->src_reg, SRC_OP); in check_alu_op()
13265 if (insn->src_reg != BPF_REG_0 || insn->off > 1 || in check_alu_op()
13594 struct bpf_reg_state *src_reg, in is_pkt_ptr_branch_taken() argument
13599 if (src_reg->type == PTR_TO_PACKET_END) { in is_pkt_ptr_branch_taken()
13602 pkt = src_reg; in is_pkt_ptr_branch_taken()
13812 static void __reg_combine_min_max(struct bpf_reg_state *src_reg, in __reg_combine_min_max() argument
13815 src_reg->umin_value = dst_reg->umin_value = max(src_reg->umin_value, in __reg_combine_min_max()
13817 src_reg->umax_value = dst_reg->umax_value = min(src_reg->umax_value, in __reg_combine_min_max()
13819 src_reg->smin_value = dst_reg->smin_value = max(src_reg->smin_value, in __reg_combine_min_max()
13821 src_reg->smax_value = dst_reg->smax_value = min(src_reg->smax_value, in __reg_combine_min_max()
13823 src_reg->var_off = dst_reg->var_off = tnum_intersect(src_reg->var_off, in __reg_combine_min_max()
13825 reg_bounds_sync(src_reg); in __reg_combine_min_max()
13917 struct bpf_reg_state *src_reg, in try_match_pkt_pointers() argument
13931 src_reg->type == PTR_TO_PACKET_END) || in try_match_pkt_pointers()
13933 reg_is_init_pkt_pointer(src_reg, PTR_TO_PACKET))) { in try_match_pkt_pointers()
13939 src_reg->type == PTR_TO_PACKET) || in try_match_pkt_pointers()
13941 src_reg->type == PTR_TO_PACKET_META)) { in try_match_pkt_pointers()
13943 find_good_pkt_pointers(other_branch, src_reg, in try_match_pkt_pointers()
13944 src_reg->type, true); in try_match_pkt_pointers()
13945 mark_pkt_end(this_branch, insn->src_reg, false); in try_match_pkt_pointers()
13952 src_reg->type == PTR_TO_PACKET_END) || in try_match_pkt_pointers()
13954 reg_is_init_pkt_pointer(src_reg, PTR_TO_PACKET))) { in try_match_pkt_pointers()
13960 src_reg->type == PTR_TO_PACKET) || in try_match_pkt_pointers()
13962 src_reg->type == PTR_TO_PACKET_META)) { in try_match_pkt_pointers()
13964 find_good_pkt_pointers(this_branch, src_reg, in try_match_pkt_pointers()
13965 src_reg->type, false); in try_match_pkt_pointers()
13966 mark_pkt_end(other_branch, insn->src_reg, true); in try_match_pkt_pointers()
13973 src_reg->type == PTR_TO_PACKET_END) || in try_match_pkt_pointers()
13975 reg_is_init_pkt_pointer(src_reg, PTR_TO_PACKET))) { in try_match_pkt_pointers()
13981 src_reg->type == PTR_TO_PACKET) || in try_match_pkt_pointers()
13983 src_reg->type == PTR_TO_PACKET_META)) { in try_match_pkt_pointers()
13985 find_good_pkt_pointers(other_branch, src_reg, in try_match_pkt_pointers()
13986 src_reg->type, false); in try_match_pkt_pointers()
13987 mark_pkt_end(this_branch, insn->src_reg, true); in try_match_pkt_pointers()
13994 src_reg->type == PTR_TO_PACKET_END) || in try_match_pkt_pointers()
13996 reg_is_init_pkt_pointer(src_reg, PTR_TO_PACKET))) { in try_match_pkt_pointers()
14002 src_reg->type == PTR_TO_PACKET) || in try_match_pkt_pointers()
14004 src_reg->type == PTR_TO_PACKET_META)) { in try_match_pkt_pointers()
14006 find_good_pkt_pointers(this_branch, src_reg, in try_match_pkt_pointers()
14007 src_reg->type, true); in try_match_pkt_pointers()
14008 mark_pkt_end(other_branch, insn->src_reg, false); in try_match_pkt_pointers()
14038 struct bpf_reg_state *dst_reg, *other_branch_regs, *src_reg = NULL; in check_cond_jmp_op() local
14064 err = check_reg_arg(env, insn->src_reg, SRC_OP); in check_cond_jmp_op()
14068 src_reg = &regs[insn->src_reg]; in check_cond_jmp_op()
14069 if (!(reg_is_pkt_pointer_any(dst_reg) && reg_is_pkt_pointer_any(src_reg)) && in check_cond_jmp_op()
14070 is_pointer_value(env, insn->src_reg)) { in check_cond_jmp_op()
14072 insn->src_reg); in check_cond_jmp_op()
14076 if (insn->src_reg != BPF_REG_0) { in check_cond_jmp_op()
14086 } else if (src_reg->type == SCALAR_VALUE && in check_cond_jmp_op()
14087 is_jmp32 && tnum_is_const(tnum_subreg(src_reg->var_off))) { in check_cond_jmp_op()
14089 tnum_subreg(src_reg->var_off).value, in check_cond_jmp_op()
14092 } else if (src_reg->type == SCALAR_VALUE && in check_cond_jmp_op()
14093 !is_jmp32 && tnum_is_const(src_reg->var_off)) { in check_cond_jmp_op()
14095 src_reg->var_off.value, in check_cond_jmp_op()
14100 pred = is_branch_taken(src_reg, in check_cond_jmp_op()
14106 pred = is_branch_taken(src_reg, in check_cond_jmp_op()
14111 reg_is_pkt_pointer_any(src_reg) && in check_cond_jmp_op()
14113 pred = is_pkt_ptr_branch_taken(dst_reg, src_reg, opcode); in check_cond_jmp_op()
14123 !__is_pointer_value(false, src_reg)) in check_cond_jmp_op()
14124 err = mark_chain_precision(env, insn->src_reg); in check_cond_jmp_op()
14167 struct bpf_reg_state *src_reg = &regs[insn->src_reg]; in check_cond_jmp_op() local
14170 src_reg->type == SCALAR_VALUE) { in check_cond_jmp_op()
14171 if (tnum_is_const(src_reg->var_off) || in check_cond_jmp_op()
14173 tnum_is_const(tnum_subreg(src_reg->var_off)))) in check_cond_jmp_op()
14176 src_reg->var_off.value, in check_cond_jmp_op()
14177 tnum_subreg(src_reg->var_off).value, in check_cond_jmp_op()
14182 reg_set_min_max_inv(&other_branch_regs[insn->src_reg], in check_cond_jmp_op()
14183 src_reg, in check_cond_jmp_op()
14190 reg_combine_min_max(&other_branch_regs[insn->src_reg], in check_cond_jmp_op()
14192 src_reg, dst_reg, opcode); in check_cond_jmp_op()
14193 if (src_reg->id && in check_cond_jmp_op()
14194 !WARN_ON_ONCE(src_reg->id != other_branch_regs[insn->src_reg].id)) { in check_cond_jmp_op()
14195 find_equal_scalars(this_branch, src_reg); in check_cond_jmp_op()
14196 find_equal_scalars(other_branch, &other_branch_regs[insn->src_reg]); in check_cond_jmp_op()
14225 __is_pointer_value(false, src_reg) && __is_pointer_value(false, dst_reg) && in check_cond_jmp_op()
14226 type_may_be_null(src_reg->type) != type_may_be_null(dst_reg->type) && in check_cond_jmp_op()
14227 base_type(src_reg->type) != PTR_TO_BTF_ID && in check_cond_jmp_op()
14242 if (type_may_be_null(src_reg->type)) in check_cond_jmp_op()
14243 mark_ptr_not_null_reg(&eq_branch_regs[insn->src_reg]); in check_cond_jmp_op()
14263 } else if (!try_match_pkt_pointers(insn, dst_reg, &regs[insn->src_reg], in check_cond_jmp_op()
14298 if (insn->src_reg == 0) { in check_ld_imm()
14312 if (insn->src_reg == BPF_PSEUDO_BTF_ID) { in check_ld_imm()
14329 if (insn->src_reg == BPF_PSEUDO_FUNC) { in check_ld_imm()
14351 if (insn->src_reg == BPF_PSEUDO_MAP_VALUE || in check_ld_imm()
14352 insn->src_reg == BPF_PSEUDO_MAP_IDX_VALUE) { in check_ld_imm()
14357 } else if (insn->src_reg == BPF_PSEUDO_MAP_FD || in check_ld_imm()
14358 insn->src_reg == BPF_PSEUDO_MAP_IDX) { in check_ld_imm()
14414 (mode == BPF_ABS && insn->src_reg != BPF_REG_0)) { in check_ld_abs()
14452 err = check_reg_arg(env, insn->src_reg, SRC_OP); in check_ld_abs()
14826 if (insn->src_reg == 0 && insn->imm == BPF_FUNC_timer_set_callback) in visit_insn()
14833 if (insn->src_reg == BPF_PSEUDO_KFUNC_CALL) { in visit_insn()
14853 return visit_func_call_insn(t, insns, env, insn->src_reg == BPF_PSEUDO_CALL); in visit_insn()
16544 err = check_reg_arg(env, insn->src_reg, SRC_OP); in do_check()
16552 src_reg_type = regs[insn->src_reg].type; in do_check()
16557 err = check_mem_access(env, env->insn_idx, insn->src_reg, in do_check()
16584 err = check_reg_arg(env, insn->src_reg, SRC_OP); in do_check()
16597 BPF_WRITE, insn->src_reg, false, false); in do_check()
16608 insn->src_reg != BPF_REG_0) { in do_check()
16635 (insn->src_reg != BPF_PSEUDO_KFUNC_CALL in do_check()
16637 (insn->src_reg != BPF_REG_0 && in do_check()
16638 insn->src_reg != BPF_PSEUDO_CALL && in do_check()
16639 insn->src_reg != BPF_PSEUDO_KFUNC_CALL) || in do_check()
16647 if ((insn->src_reg == BPF_REG_0 && insn->imm != BPF_FUNC_spin_unlock) || in do_check()
16648 (insn->src_reg == BPF_PSEUDO_CALL) || in do_check()
16649 (insn->src_reg == BPF_PSEUDO_KFUNC_CALL && in do_check()
16655 if (insn->src_reg == BPF_PSEUDO_CALL) in do_check()
16657 else if (insn->src_reg == BPF_PSEUDO_KFUNC_CALL) in do_check()
16667 insn->src_reg != BPF_REG_0 || in do_check()
16684 insn->src_reg != BPF_REG_0 || in do_check()
17063 insn[1].dst_reg != 0 || insn[1].src_reg != 0 || in resolve_pseudo_ldimm64()
17069 if (insn[0].src_reg == 0) in resolve_pseudo_ldimm64()
17073 if (insn[0].src_reg == BPF_PSEUDO_BTF_ID) { in resolve_pseudo_ldimm64()
17081 if (insn[0].src_reg == BPF_PSEUDO_FUNC) { in resolve_pseudo_ldimm64()
17090 switch (insn[0].src_reg) { in resolve_pseudo_ldimm64()
17104 switch (insn[0].src_reg) { in resolve_pseudo_ldimm64()
17136 if (insn[0].src_reg == BPF_PSEUDO_MAP_FD || in resolve_pseudo_ldimm64()
17137 insn[0].src_reg == BPF_PSEUDO_MAP_IDX) { in resolve_pseudo_ldimm64()
17245 if (insn->src_reg == BPF_PSEUDO_FUNC) in convert_pseudo_ld_imm64()
17247 insn->src_reg = 0; in convert_pseudo_ld_imm64()
17685 zext_patch[1].src_reg = load_reg; in opt_subreg_zext_lo32_rnd_hi32()
18352 BPF_JNE | BPF_K, insn->src_reg, in do_misc_fixups()
18361 BPF_JEQ | BPF_K, insn->src_reg, in do_misc_fixups()
18421 off_reg = issrc ? insn->src_reg : insn->dst_reg; in do_misc_fixups()
18435 *patch++ = BPF_MOV64_REG(insn->dst_reg, insn->src_reg); in do_misc_fixups()
18436 insn->src_reg = BPF_REG_AX; in do_misc_fixups()
18457 if (insn->src_reg == BPF_PSEUDO_CALL) in do_misc_fixups()
18459 if (insn->src_reg == BPF_PSEUDO_KFUNC_CALL) { in do_misc_fixups()
18930 insn->src_reg == 0 && in is_bpf_loop_call()