Lines Matching refs:smin_value
664 if (reg->smin_value != reg->umin_value && in print_verifier_state()
665 reg->smin_value != S64_MIN) in print_verifier_state()
667 (long long)reg->smin_value); in print_verifier_state()
684 if (reg->s32_min_value != reg->smin_value && in print_verifier_state()
1076 reg->smin_value = (s64)imm; in ___mark_reg_known()
1211 reg->smin_value = S64_MIN; in __mark_reg_unbounded()
1224 reg->smin_value = S64_MIN; in __mark_reg64_unbounded()
1256 reg->smin_value = max_t(s64, reg->smin_value, in __update_reg64_bounds()
1314 if (reg->smin_value >= 0 || reg->smax_value < 0) { in __reg64_deduce_bounds()
1315 reg->smin_value = reg->umin_value = max_t(u64, reg->smin_value, in __reg64_deduce_bounds()
1328 reg->smin_value = reg->umin_value; in __reg64_deduce_bounds()
1335 reg->smin_value = reg->umin_value = max_t(u64, reg->smin_value, in __reg64_deduce_bounds()
1373 reg->smin_value = reg->s32_min_value; in __reg_assign_32_into_64()
1375 reg->smin_value = 0; in __reg_assign_32_into_64()
1421 if (__reg64_bound_s32(reg->smin_value) && __reg64_bound_s32(reg->smax_value)) { in __reg_combine_64_into_32()
1422 reg->s32_min_value = (s32)reg->smin_value; in __reg_combine_64_into_32()
2608 reg->smin_value == S64_MIN && reg->smax_value == S64_MAX && in __is_scalar_unbounded()
2785 min_off = ptr_reg->smin_value + off; in check_stack_write_var_off()
3032 min_off = reg->smin_value + off; in check_stack_read_var_off()
3212 if (reg->smin_value < 0 && in check_mem_region_access()
3213 (reg->smin_value == S64_MIN || in check_mem_region_access()
3214 (off + reg->smin_value != (s64)(s32)(off + reg->smin_value)) || in check_mem_region_access()
3215 reg->smin_value + off < 0)) { in check_mem_region_access()
3220 err = __check_mem_access(env, regno, reg->smin_value + off, size, in check_mem_region_access()
3271 if (reg->smin_value + off < lock + sizeof(struct bpf_spin_lock) && in check_map_access()
3280 if (reg->smin_value + off < t + sizeof(struct bpf_timer) && in check_map_access()
3353 if (reg->smin_value < 0) { in check_packet_access()
3438 if (reg->smin_value < 0) { in check_sock_access()
3873 reg->smin_value = reg->umin_value; in coerce_reg_to_size()
4080 reg->smin_value <= -BPF_MAX_VAR_OFF) { in check_stack_access_within_bounds()
4085 min_off = reg->smin_value + off; in check_stack_access_within_bounds()
4489 min_off = reg->smin_value + off; in check_stack_range_initialized()
5164 if (reg->smin_value < 0) { in check_func_arg()
6006 ret_reg->smin_value = -MAX_ERRNO; in do_refine_retval_range()
6596 s64 smin = reg->smin_value; in check_reg_sane_offset()
6656 ptr_reg->smin_value : in retrieve_ptr_limit()
6745 bool off_is_neg = off_reg->smin_value < 0; in sanitize_ptr_alu()
6765 (off_reg->smin_value < 0) != (off_reg->smax_value < 0)) in sanitize_ptr_alu()
6962 s64 smin_val = off_reg->smin_value, smax_val = off_reg->smax_value, in adjust_ptr_min_max_vals()
6963 smin_ptr = ptr_reg->smin_value, smax_ptr = ptr_reg->smax_value; in adjust_ptr_min_max_vals()
7048 dst_reg->smin_value = smin_ptr; in adjust_ptr_min_max_vals()
7068 dst_reg->smin_value = S64_MIN; in adjust_ptr_min_max_vals()
7071 dst_reg->smin_value = smin_ptr + smin_val; in adjust_ptr_min_max_vals()
7110 dst_reg->smin_value = smin_ptr; in adjust_ptr_min_max_vals()
7126 dst_reg->smin_value = S64_MIN; in adjust_ptr_min_max_vals()
7129 dst_reg->smin_value = smin_ptr - smax_val; in adjust_ptr_min_max_vals()
7213 s64 smin_val = src_reg->smin_value; in scalar_min_max_add()
7218 if (signed_add_overflows(dst_reg->smin_value, smin_val) || in scalar_min_max_add()
7220 dst_reg->smin_value = S64_MIN; in scalar_min_max_add()
7223 dst_reg->smin_value += smin_val; in scalar_min_max_add()
7267 s64 smin_val = src_reg->smin_value; in scalar_min_max_sub()
7272 if (signed_sub_overflows(dst_reg->smin_value, smax_val) || in scalar_min_max_sub()
7275 dst_reg->smin_value = S64_MIN; in scalar_min_max_sub()
7278 dst_reg->smin_value -= smax_val; in scalar_min_max_sub()
7327 s64 smin_val = src_reg->smin_value; in scalar_min_max_mul()
7331 if (smin_val < 0 || dst_reg->smin_value < 0) { in scalar_min_max_mul()
7348 dst_reg->smin_value = S64_MIN; in scalar_min_max_mul()
7351 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_mul()
7395 s64 smin_val = src_reg->smin_value; in scalar_min_max_and()
7408 if (dst_reg->smin_value < 0 || smin_val < 0) { in scalar_min_max_and()
7412 dst_reg->smin_value = S64_MIN; in scalar_min_max_and()
7418 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_and()
7464 s64 smin_val = src_reg->smin_value; in scalar_min_max_or()
7477 if (dst_reg->smin_value < 0 || smin_val < 0) { in scalar_min_max_or()
7481 dst_reg->smin_value = S64_MIN; in scalar_min_max_or()
7487 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_or()
7528 s64 smin_val = src_reg->smin_value; in scalar_min_max_xor()
7540 if (dst_reg->smin_value >= 0 && smin_val >= 0) { in scalar_min_max_xor()
7544 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_xor()
7547 dst_reg->smin_value = S64_MIN; in scalar_min_max_xor()
7606 dst_reg->smin_value = (s64)dst_reg->s32_min_value << 32; in __scalar64_min_max_lsh()
7608 dst_reg->smin_value = S64_MIN; in __scalar64_min_max_lsh()
7687 dst_reg->smin_value = S64_MIN; in scalar_min_max_rsh()
7732 dst_reg->smin_value >>= umin_val; in scalar_min_max_arsh()
7771 smin_val = src_reg.smin_value; in adjust_scalar_min_max_vals()
8369 if (reg->smin_value > sval) in is_branch64_taken()
8383 else if (reg->smin_value >= sval) in is_branch64_taken()
8393 if (reg->smin_value >= sval) in is_branch64_taken()
8407 else if (reg->smin_value > sval) in is_branch64_taken()
8605 true_reg->smin_value = max(true_reg->smin_value, true_smin); in reg_set_min_max()
8642 false_reg->smin_value = max(false_reg->smin_value, false_smin); in reg_set_min_max()
8690 src_reg->smin_value = dst_reg->smin_value = max(src_reg->smin_value, in __reg_combine_min_max()
8691 dst_reg->smin_value); in __reg_combine_min_max()
8739 if (WARN_ON_ONCE(reg->smin_value || reg->smax_value || in mark_ptr_or_null_reg()
10064 old->smin_value <= cur->smin_value && in range_within()