Lines Matching refs:umax_value

1760 	reg->umax_value = imm;  in ___mark_reg_known()
1912 reg->umax_value = U64_MAX; in __mark_reg_unbounded()
1925 reg->umax_value = U64_MAX; in __mark_reg64_unbounded()
1960 reg->umax_value = min(reg->umax_value, in __update_reg64_bounds()
1994 if ((reg->umin_value >> 32) == (reg->umax_value >> 32)) { in __reg32_deduce_bounds()
1999 reg->u32_max_value = min_t(u32, reg->u32_max_value, (u32)reg->umax_value); in __reg32_deduce_bounds()
2001 if ((s32)reg->umin_value <= (s32)reg->umax_value) { in __reg32_deduce_bounds()
2003 reg->s32_max_value = min_t(s32, reg->s32_max_value, (s32)reg->umax_value); in __reg32_deduce_bounds()
2031 if ((u32)(reg->umin_value >> 32) + 1 == (u32)(reg->umax_value >> 32) && in __reg32_deduce_bounds()
2032 (s32)reg->umin_value < 0 && (s32)reg->umax_value >= 0) { in __reg32_deduce_bounds()
2034 reg->s32_max_value = min_t(s32, reg->s32_max_value, (s32)reg->umax_value); in __reg32_deduce_bounds()
2127 if ((s64)reg->umin_value <= (s64)reg->umax_value) { in __reg64_deduce_bounds()
2129 reg->smax_value = min_t(s64, reg->smax_value, reg->umax_value); in __reg64_deduce_bounds()
2137 reg->umax_value = min_t(u64, reg->smax_value, reg->umax_value); in __reg64_deduce_bounds()
2160 new_umax = (reg->umax_value & ~0xffffffffULL) | reg->u32_max_value; in __reg_deduce_mixed_bounds()
2162 reg->umax_value = min_t(u64, reg->umax_value, new_umax); in __reg_deduce_mixed_bounds()
2173 new_umax = (reg->umax_value & ~0xffffffffULL) | (u32)reg->s32_max_value; in __reg_deduce_mixed_bounds()
2175 reg->umax_value = min_t(u64, reg->umax_value, new_umax); in __reg_deduce_mixed_bounds()
2216 reg->umax_value = reg->s32_max_value; in __reg_deduce_mixed_bounds()
2234 reg->umax_value)); in __reg_bound_offset()
2263 if (reg->umin_value > reg->umax_value || in reg_bounds_sanity_check()
2275 if (reg->umin_value != uval || reg->umax_value != uval || in reg_bounds_sanity_check()
2297 ctx, msg, reg->umin_value, reg->umax_value, in reg_bounds_sanity_check()
2316 reg->umax_value = reg->u32_max_value; in __reg_assign_32_into_64()
4582 return fls64(reg->umax_value); in get_reg_width()
5284 if (reg->umax_value >= BPF_MAX_VAR_OFF) { in check_mem_region_access()
5289 err = __check_mem_access(env, regno, reg->umax_value + off, size, in check_mem_region_access()
5577 p < reg->umax_value + off + size) { in check_map_access()
5691 off + reg->umax_value + size - 1); in check_packet_access()
6256 if ((reg->umin_value & ~mask) == (reg->umax_value & ~mask)) { in coerce_reg_to_size()
6258 reg->umax_value &= mask; in coerce_reg_to_size()
6261 reg->umax_value = mask; in coerce_reg_to_size()
6264 reg->smax_value = reg->umax_value; in coerce_reg_to_size()
6290 reg->umax_value = U64_MAX; in set_sext64_default_val()
6313 reg->umax_value = reg->umin_value = u64_cval; in coerce_reg_to_size_sx()
6345 reg->u32_max_value = reg->umax_value = s64_max; in coerce_reg_to_size_sx()
7543 meta->msize_max_value = reg->umax_value; in check_mem_size_reg()
7561 regno, reg->umin_value, reg->umax_value); in check_mem_size_reg()
7565 if (reg->umax_value >= BPF_MAX_VAR_SIZ) { in check_mem_size_reg()
7570 err = check_helper_mem_access(env, regno - 1, reg->umax_value, in check_mem_size_reg()
10201 ret_reg->umax_value = nr_cpu_ids - 1; in do_refine_retval_range()
10624 if (cur_func(env)->callback_depth < regs[BPF_REG_1].umax_value) { in check_helper_call()
12957 ptr_reg->umax_value) + ptr_reg->off; in retrieve_ptr_limit()
13264 u64 umin_val = off_reg->umin_value, umax_val = off_reg->umax_value, in adjust_ptr_min_max_vals()
13265 umin_ptr = ptr_reg->umin_value, umax_ptr = ptr_reg->umax_value; in adjust_ptr_min_max_vals()
13361 dst_reg->umax_value = umax_ptr; in adjust_ptr_min_max_vals()
13382 check_add_overflow(umax_ptr, umax_val, &dst_reg->umax_value)) { in adjust_ptr_min_max_vals()
13384 dst_reg->umax_value = U64_MAX; in adjust_ptr_min_max_vals()
13417 dst_reg->umax_value = umax_ptr; in adjust_ptr_min_max_vals()
13436 dst_reg->umax_value = U64_MAX; in adjust_ptr_min_max_vals()
13440 dst_reg->umax_value = umax_ptr - umin_val; in adjust_ptr_min_max_vals()
13507 u64 *dst_umax = &dst_reg->umax_value; in scalar_min_max_add()
13515 check_add_overflow(*dst_umax, src_reg->umax_value, dst_umax)) { in scalar_min_max_add()
13552 u64 umax_val = src_reg->umax_value; in scalar_min_max_sub()
13563 dst_reg->umax_value = U64_MAX; in scalar_min_max_sub()
13567 dst_reg->umax_value -= umin_val; in scalar_min_max_sub()
13608 u64 umax_val = src_reg->umax_value; in scalar_min_max_mul()
13618 if (umax_val > U32_MAX || dst_reg->umax_value > U32_MAX) { in scalar_min_max_mul()
13624 dst_reg->umax_value *= umax_val; in scalar_min_max_mul()
13625 if (dst_reg->umax_value > S64_MAX) { in scalar_min_max_mul()
13631 dst_reg->smax_value = dst_reg->umax_value; in scalar_min_max_mul()
13671 u64 umax_val = src_reg->umax_value; in scalar_min_max_and()
13682 dst_reg->umax_value = min(dst_reg->umax_value, umax_val); in scalar_min_max_and()
13687 if ((s64)dst_reg->umin_value <= (s64)dst_reg->umax_value) { in scalar_min_max_and()
13689 dst_reg->smax_value = dst_reg->umax_value; in scalar_min_max_and()
13745 dst_reg->umax_value = dst_reg->var_off.value | dst_reg->var_off.mask; in scalar_min_max_or()
13750 if ((s64)dst_reg->umin_value <= (s64)dst_reg->umax_value) { in scalar_min_max_or()
13752 dst_reg->smax_value = dst_reg->umax_value; in scalar_min_max_or()
13803 dst_reg->umax_value = dst_reg->var_off.value | dst_reg->var_off.mask; in scalar_min_max_xor()
13808 if ((s64)dst_reg->umin_value <= (s64)dst_reg->umax_value) { in scalar_min_max_xor()
13810 dst_reg->smax_value = dst_reg->umax_value; in scalar_min_max_xor()
13876 if (dst_reg->umax_value > 1ULL << (63 - umax_val)) { in __scalar64_min_max_lsh()
13878 dst_reg->umax_value = U64_MAX; in __scalar64_min_max_lsh()
13881 dst_reg->umax_value <<= umax_val; in __scalar64_min_max_lsh()
13888 u64 umax_val = src_reg->umax_value; in scalar_min_max_lsh()
13935 u64 umax_val = src_reg->umax_value; in scalar_min_max_rsh()
13956 dst_reg->umax_value >>= umin_val; in scalar_min_max_rsh()
14006 dst_reg->umax_value = U64_MAX; in scalar_min_max_arsh()
14030 && src_reg->umin_value == src_reg->umax_value) in is_safe_to_compute_dst_reg_range()
14050 return (src_is_const && src_reg->umax_value < insn_bitness); in is_safe_to_compute_dst_reg_range()
14405 no_sext = src_reg->umax_value < (1ULL << (insn->off - 1)); in check_alu_op()
14442 bool no_sext = src_reg->umax_value < (1ULL << (insn->off - 1)); in check_alu_op()
14545 if (dst_reg->umax_value > MAX_PACKET_OFF || in find_good_pkt_pointers()
14546 dst_reg->umax_value + dst_reg->off > MAX_PACKET_OFF) in find_good_pkt_pointers()
14619 u64 umax1 = is_jmp32 ? (u64)reg1->u32_max_value : reg1->umax_value; in is_scalar_branch_taken()
14623 u64 umax2 = is_jmp32 ? (u64)reg2->u32_max_value : reg2->umax_value; in is_scalar_branch_taken()
14913 reg1->umax_value = min(reg1->umax_value, reg2->umax_value); in regs_refine_cond_op()
14917 reg2->umax_value = reg1->umax_value; in regs_refine_cond_op()
14956 if (reg1->umax_value == (u64)val) in regs_refine_cond_op()
14957 reg1->umax_value--; in regs_refine_cond_op()
15007 reg1->umax_value = min(reg1->umax_value, reg2->umax_value); in regs_refine_cond_op()
15016 reg1->umax_value = min(reg1->umax_value, reg2->umax_value - 1); in regs_refine_cond_op()
17049 old->umax_value >= cur->umax_value && in range_within()