Lines Matching refs:umin_value
1759 reg->umin_value = imm; in ___mark_reg_known()
1911 reg->umin_value = 0; in __mark_reg_unbounded()
1924 reg->umin_value = 0; in __mark_reg64_unbounded()
1959 reg->umin_value = max(reg->umin_value, reg->var_off.value); in __update_reg64_bounds()
1994 if ((reg->umin_value >> 32) == (reg->umax_value >> 32)) { in __reg32_deduce_bounds()
1998 reg->u32_min_value = max_t(u32, reg->u32_min_value, (u32)reg->umin_value); in __reg32_deduce_bounds()
2001 if ((s32)reg->umin_value <= (s32)reg->umax_value) { in __reg32_deduce_bounds()
2002 reg->s32_min_value = max_t(s32, reg->s32_min_value, (s32)reg->umin_value); in __reg32_deduce_bounds()
2031 if ((u32)(reg->umin_value >> 32) + 1 == (u32)(reg->umax_value >> 32) && in __reg32_deduce_bounds()
2032 (s32)reg->umin_value < 0 && (s32)reg->umax_value >= 0) { in __reg32_deduce_bounds()
2033 reg->s32_min_value = max_t(s32, reg->s32_min_value, (s32)reg->umin_value); in __reg32_deduce_bounds()
2127 if ((s64)reg->umin_value <= (s64)reg->umax_value) { in __reg64_deduce_bounds()
2128 reg->smin_value = max_t(s64, reg->smin_value, reg->umin_value); in __reg64_deduce_bounds()
2136 reg->umin_value = max_t(u64, reg->smin_value, reg->umin_value); in __reg64_deduce_bounds()
2159 new_umin = (reg->umin_value & ~0xffffffffULL) | reg->u32_min_value; in __reg_deduce_mixed_bounds()
2161 reg->umin_value = max_t(u64, reg->umin_value, new_umin); in __reg_deduce_mixed_bounds()
2172 new_umin = (reg->umin_value & ~0xffffffffULL) | (u32)reg->s32_min_value; in __reg_deduce_mixed_bounds()
2174 reg->umin_value = max_t(u64, reg->umin_value, new_umin); in __reg_deduce_mixed_bounds()
2215 reg->umin_value = reg->s32_min_value; in __reg_deduce_mixed_bounds()
2233 tnum_range(reg->umin_value, in __reg_bound_offset()
2263 if (reg->umin_value > reg->umax_value || in reg_bounds_sanity_check()
2275 if (reg->umin_value != uval || reg->umax_value != uval || in reg_bounds_sanity_check()
2297 ctx, msg, reg->umin_value, reg->umax_value, in reg_bounds_sanity_check()
2315 reg->umin_value = reg->u32_min_value; in __reg_assign_32_into_64()
6256 if ((reg->umin_value & ~mask) == (reg->umax_value & ~mask)) { in coerce_reg_to_size()
6257 reg->umin_value &= mask; in coerce_reg_to_size()
6260 reg->umin_value = 0; in coerce_reg_to_size()
6263 reg->smin_value = reg->umin_value; in coerce_reg_to_size()
6289 reg->umin_value = reg->u32_min_value = 0; in set_sext64_default_val()
6313 reg->umax_value = reg->umin_value = u64_cval; in coerce_reg_to_size_sx()
6344 reg->u32_min_value = reg->umin_value = s64_min; in coerce_reg_to_size_sx()
7559 if (reg->umin_value == 0 && !zero_size_allowed) { in check_mem_size_reg()
7561 regno, reg->umin_value, reg->umax_value); in check_mem_size_reg()
10205 ret_reg->umin_value = 0; in do_refine_retval_range()
13264 u64 umin_val = off_reg->umin_value, umax_val = off_reg->umax_value, in adjust_ptr_min_max_vals()
13265 umin_ptr = ptr_reg->umin_value, umax_ptr = ptr_reg->umax_value; in adjust_ptr_min_max_vals()
13360 dst_reg->umin_value = umin_ptr; in adjust_ptr_min_max_vals()
13381 if (check_add_overflow(umin_ptr, umin_val, &dst_reg->umin_value) || in adjust_ptr_min_max_vals()
13383 dst_reg->umin_value = 0; in adjust_ptr_min_max_vals()
13416 dst_reg->umin_value = umin_ptr; in adjust_ptr_min_max_vals()
13435 dst_reg->umin_value = 0; in adjust_ptr_min_max_vals()
13439 dst_reg->umin_value = umin_ptr - umax_val; in adjust_ptr_min_max_vals()
13506 u64 *dst_umin = &dst_reg->umin_value; in scalar_min_max_add()
13514 if (check_add_overflow(*dst_umin, src_reg->umin_value, dst_umin) || in scalar_min_max_add()
13551 u64 umin_val = src_reg->umin_value; in scalar_min_max_sub()
13560 if (dst_reg->umin_value < umax_val) { in scalar_min_max_sub()
13562 dst_reg->umin_value = 0; in scalar_min_max_sub()
13566 dst_reg->umin_value -= umax_val; in scalar_min_max_sub()
13607 u64 umin_val = src_reg->umin_value; in scalar_min_max_mul()
13623 dst_reg->umin_value *= umin_val; in scalar_min_max_mul()
13630 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_mul()
13681 dst_reg->umin_value = dst_reg->var_off.value; in scalar_min_max_and()
13687 if ((s64)dst_reg->umin_value <= (s64)dst_reg->umax_value) { in scalar_min_max_and()
13688 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_and()
13734 u64 umin_val = src_reg->umin_value; in scalar_min_max_or()
13744 dst_reg->umin_value = max(dst_reg->umin_value, umin_val); in scalar_min_max_or()
13750 if ((s64)dst_reg->umin_value <= (s64)dst_reg->umax_value) { in scalar_min_max_or()
13751 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_or()
13802 dst_reg->umin_value = dst_reg->var_off.value; in scalar_min_max_xor()
13808 if ((s64)dst_reg->umin_value <= (s64)dst_reg->umax_value) { in scalar_min_max_xor()
13809 dst_reg->smin_value = dst_reg->umin_value; in scalar_min_max_xor()
13877 dst_reg->umin_value = 0; in __scalar64_min_max_lsh()
13880 dst_reg->umin_value <<= umin_val; in __scalar64_min_max_lsh()
13889 u64 umin_val = src_reg->umin_value; in scalar_min_max_lsh()
13936 u64 umin_val = src_reg->umin_value; in scalar_min_max_rsh()
13955 dst_reg->umin_value >>= umax_val; in scalar_min_max_rsh()
13992 u64 umin_val = src_reg->umin_value; in scalar_min_max_arsh()
14005 dst_reg->umin_value = 0; in scalar_min_max_arsh()
14030 && src_reg->umin_value == src_reg->umax_value) in is_safe_to_compute_dst_reg_range()
14618 u64 umin1 = is_jmp32 ? (u64)reg1->u32_min_value : reg1->umin_value; in is_scalar_branch_taken()
14622 u64 umin2 = is_jmp32 ? (u64)reg2->u32_min_value : reg2->umin_value; in is_scalar_branch_taken()
14912 reg1->umin_value = max(reg1->umin_value, reg2->umin_value); in regs_refine_cond_op()
14916 reg2->umin_value = reg1->umin_value; in regs_refine_cond_op()
14954 if (reg1->umin_value == (u64)val) in regs_refine_cond_op()
14955 reg1->umin_value++; in regs_refine_cond_op()
15008 reg2->umin_value = max(reg1->umin_value, reg2->umin_value); in regs_refine_cond_op()
15017 reg2->umin_value = max(reg1->umin_value + 1, reg2->umin_value); in regs_refine_cond_op()
17048 return old->umin_value <= cur->umin_value && in range_within()