Lines Matching refs:smin_value

672 				if (reg->smin_value != reg->umin_value &&  in print_verifier_state()
673 reg->smin_value != S64_MIN) in print_verifier_state()
675 (long long)reg->smin_value); in print_verifier_state()
692 if (reg->s32_min_value != reg->smin_value && in print_verifier_state()
1084 reg->smin_value = (s64)imm; in ___mark_reg_known()
1220 reg->smin_value = S64_MIN; in __mark_reg_unbounded()
1233 reg->smin_value = S64_MIN; in __mark_reg64_unbounded()
1265 reg->smin_value = max_t(s64, reg->smin_value, in __update_reg64_bounds()
1323 if (reg->smin_value >= 0 || reg->smax_value < 0) { in __reg64_deduce_bounds()
1324 reg->smin_value = reg->umin_value = max_t(u64, reg->smin_value, in __reg64_deduce_bounds()
1337 reg->smin_value = reg->umin_value; in __reg64_deduce_bounds()
1344 reg->smin_value = reg->umin_value = max_t(u64, reg->smin_value, in __reg64_deduce_bounds()
1385 reg->smin_value = reg->s32_min_value; in __reg_assign_32_into_64()
1388 reg->smin_value = 0; in __reg_assign_32_into_64()
1436 if (__reg64_bound_s32(reg->smin_value) && __reg64_bound_s32(reg->smax_value)) { in __reg_combine_64_into_32()
1437 reg->s32_min_value = (s32)reg->smin_value; in __reg_combine_64_into_32()
2757 reg->smin_value == S64_MIN && reg->smax_value == S64_MAX && in __is_scalar_unbounded()
2940 min_off = ptr_reg->smin_value + off; in check_stack_write_var_off()
3206 min_off = reg->smin_value + off; in check_stack_read_var_off()
3386 if (reg->smin_value < 0 && in check_mem_region_access()
3387 (reg->smin_value == S64_MIN || in check_mem_region_access()
3388 (off + reg->smin_value != (s64)(s32)(off + reg->smin_value)) || in check_mem_region_access()
3389 reg->smin_value + off < 0)) { in check_mem_region_access()
3394 err = __check_mem_access(env, regno, reg->smin_value + off, size, in check_mem_region_access()
3445 if (reg->smin_value + off < lock + sizeof(struct bpf_spin_lock) && in check_map_access()
3454 if (reg->smin_value + off < t + sizeof(struct bpf_timer) && in check_map_access()
3527 if (reg->smin_value < 0) { in check_packet_access()
3612 if (reg->smin_value < 0) { in check_sock_access()
4047 reg->smin_value = reg->umin_value; in coerce_reg_to_size()
4269 reg->smin_value <= -BPF_MAX_VAR_OFF) { in check_stack_access_within_bounds()
4274 min_off = reg->smin_value + off; in check_stack_access_within_bounds()
4691 min_off = reg->smin_value + off; in check_stack_range_initialized()
5369 if (reg->smin_value < 0) { in check_func_arg()
6221 ret_reg->smin_value = -MAX_ERRNO; in do_refine_retval_range()
6821 s64 smin = reg->smin_value; in check_reg_sane_offset()
6881 ptr_reg->smin_value : in retrieve_ptr_limit()
6970 bool off_is_neg = off_reg->smin_value < 0; in sanitize_ptr_alu()
6990 (off_reg->smin_value < 0) != (off_reg->smax_value < 0)) in sanitize_ptr_alu()
7187 s64 smin_val = off_reg->smin_value, smax_val = off_reg->smax_value, in adjust_ptr_min_max_vals()
7188 smin_ptr = ptr_reg->smin_value, smax_ptr = ptr_reg->smax_value; in adjust_ptr_min_max_vals()
7273 dst_reg->smin_value = smin_ptr; in adjust_ptr_min_max_vals()
7293 dst_reg->smin_value = S64_MIN; in adjust_ptr_min_max_vals()
7296 dst_reg->smin_value = smin_ptr + smin_val; in adjust_ptr_min_max_vals()
7335 dst_reg->smin_value = smin_ptr; in adjust_ptr_min_max_vals()
7351 dst_reg->smin_value = S64_MIN; in adjust_ptr_min_max_vals()
7354 dst_reg->smin_value = smin_ptr - smax_val; in adjust_ptr_min_max_vals()
7438 s64 smin_val = src_reg->smin_value; in scalar_min_max_add()
7443 if (signed_add_overflows(dst_reg->smin_value, smin_val) || in scalar_min_max_add()
7445 dst_reg->smin_value = S64_MIN; in scalar_min_max_add()
7448 dst_reg->smin_value += smin_val; in scalar_min_max_add()
7492 s64 smin_val = src_reg->smin_value; in scalar_min_max_sub()
7497 if (signed_sub_overflows(dst_reg->smin_value, smax_val) || in scalar_min_max_sub()
7500 dst_reg->smin_value = S64_MIN; in scalar_min_max_sub()
7503 dst_reg->smin_value -= smax_val; in scalar_min_max_sub()
7552 s64 smin_val = src_reg->smin_value; in scalar_min_max_mul()
7556 if (smin_val < 0 || dst_reg->smin_value < 0) { in scalar_min_max_mul()
7573 dst_reg->smin_value = S64_MIN; in scalar_min_max_mul()
7576 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_mul()
7620 s64 smin_val = src_reg->smin_value; in scalar_min_max_and()
7633 if (dst_reg->smin_value < 0 || smin_val < 0) { in scalar_min_max_and()
7637 dst_reg->smin_value = S64_MIN; in scalar_min_max_and()
7643 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_and()
7689 s64 smin_val = src_reg->smin_value; in scalar_min_max_or()
7702 if (dst_reg->smin_value < 0 || smin_val < 0) { in scalar_min_max_or()
7706 dst_reg->smin_value = S64_MIN; in scalar_min_max_or()
7712 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_or()
7753 s64 smin_val = src_reg->smin_value; in scalar_min_max_xor()
7765 if (dst_reg->smin_value >= 0 && smin_val >= 0) { in scalar_min_max_xor()
7769 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_xor()
7772 dst_reg->smin_value = S64_MIN; in scalar_min_max_xor()
7831 dst_reg->smin_value = (s64)dst_reg->s32_min_value << 32; in __scalar64_min_max_lsh()
7833 dst_reg->smin_value = S64_MIN; in __scalar64_min_max_lsh()
7912 dst_reg->smin_value = S64_MIN; in scalar_min_max_rsh()
7957 dst_reg->smin_value >>= umin_val; in scalar_min_max_arsh()
7996 smin_val = src_reg.smin_value; in adjust_scalar_min_max_vals()
8598 if (reg->smin_value > sval) in is_branch64_taken()
8612 else if (reg->smin_value >= sval) in is_branch64_taken()
8622 if (reg->smin_value >= sval) in is_branch64_taken()
8636 else if (reg->smin_value > sval) in is_branch64_taken()
8834 true_reg->smin_value = max(true_reg->smin_value, true_smin); in reg_set_min_max()
8871 false_reg->smin_value = max(false_reg->smin_value, false_smin); in reg_set_min_max()
8919 src_reg->smin_value = dst_reg->smin_value = max(src_reg->smin_value, in __reg_combine_min_max()
8920 dst_reg->smin_value); in __reg_combine_min_max()
8968 if (WARN_ON_ONCE(reg->smin_value || reg->smax_value || in mark_ptr_or_null_reg()
10294 old->smin_value <= cur->smin_value && in range_within()