Lines Matching refs:umax_value

1107 				if (reg->smax_value != reg->umax_value &&  in print_verifier_state()
1112 if (reg->umax_value != U64_MAX) in print_verifier_state()
1113 verbose_a("umax=%llu", (unsigned long long)reg->umax_value); in print_verifier_state()
1129 if (reg->u32_max_value != reg->umax_value && in print_verifier_state()
1547 reg->umax_value = imm; in ___mark_reg_known()
1687 reg->umax_value = U64_MAX; in __mark_reg_unbounded()
1700 reg->umax_value = U64_MAX; in __mark_reg64_unbounded()
1735 reg->umax_value = min(reg->umax_value, in __update_reg64_bounds()
1790 reg->smax_value = reg->umax_value = min_t(u64, reg->smax_value, in __reg64_deduce_bounds()
1791 reg->umax_value); in __reg64_deduce_bounds()
1797 if ((s64)reg->umax_value >= 0) { in __reg64_deduce_bounds()
1802 reg->smax_value = reg->umax_value = min_t(u64, reg->smax_value, in __reg64_deduce_bounds()
1803 reg->umax_value); in __reg64_deduce_bounds()
1810 reg->smax_value = reg->umax_value; in __reg64_deduce_bounds()
1825 reg->umax_value)); in __reg_bound_offset()
1856 reg->umax_value = reg->u32_max_value; in __reg_assign_32_into_64()
1910 if (__reg64_bound_u32(reg->umin_value) && __reg64_bound_u32(reg->umax_value)) { in __reg_combine_64_into_32()
1912 reg->u32_max_value = (u32)reg->umax_value; in __reg_combine_64_into_32()
3428 reg->umin_value == 0 && reg->umax_value == U64_MAX && in __is_scalar_unbounded()
4119 if (reg->umax_value >= BPF_MAX_VAR_OFF) { in check_mem_region_access()
4124 err = __check_mem_access(env, regno, reg->umax_value + off, size, in check_mem_region_access()
4329 p < reg->umax_value + off + size) { in check_map_access()
4442 off + reg->umax_value + size - 1); in check_packet_access()
4933 if ((reg->umin_value & ~mask) == (reg->umax_value & ~mask)) { in coerce_reg_to_size()
4935 reg->umax_value &= mask; in coerce_reg_to_size()
4938 reg->umax_value = mask; in coerce_reg_to_size()
4941 reg->smax_value = reg->umax_value; in coerce_reg_to_size()
5909 meta->msize_max_value = reg->umax_value; in check_mem_size_reg()
5936 if (reg->umax_value >= BPF_MAX_VAR_SIZ) { in check_mem_size_reg()
5942 reg->umax_value, in check_mem_size_reg()
10171 ptr_reg->umax_value) + ptr_reg->off; in retrieve_ptr_limit()
10478 u64 umin_val = off_reg->umin_value, umax_val = off_reg->umax_value, in adjust_ptr_min_max_vals()
10479 umin_ptr = ptr_reg->umin_value, umax_ptr = ptr_reg->umax_value; in adjust_ptr_min_max_vals()
10564 dst_reg->umax_value = umax_ptr; in adjust_ptr_min_max_vals()
10590 dst_reg->umax_value = U64_MAX; in adjust_ptr_min_max_vals()
10593 dst_reg->umax_value = umax_ptr + umax_val; in adjust_ptr_min_max_vals()
10626 dst_reg->umax_value = umax_ptr; in adjust_ptr_min_max_vals()
10648 dst_reg->umax_value = U64_MAX; in adjust_ptr_min_max_vals()
10652 dst_reg->umax_value = umax_ptr - umin_val; in adjust_ptr_min_max_vals()
10725 u64 umax_val = src_reg->umax_value; in scalar_min_max_add()
10736 dst_reg->umax_value + umax_val < umax_val) { in scalar_min_max_add()
10738 dst_reg->umax_value = U64_MAX; in scalar_min_max_add()
10741 dst_reg->umax_value += umax_val; in scalar_min_max_add()
10779 u64 umax_val = src_reg->umax_value; in scalar_min_max_sub()
10793 dst_reg->umax_value = U64_MAX; in scalar_min_max_sub()
10797 dst_reg->umax_value -= umin_val; in scalar_min_max_sub()
10838 u64 umax_val = src_reg->umax_value; in scalar_min_max_mul()
10848 if (umax_val > U32_MAX || dst_reg->umax_value > U32_MAX) { in scalar_min_max_mul()
10854 dst_reg->umax_value *= umax_val; in scalar_min_max_mul()
10855 if (dst_reg->umax_value > S64_MAX) { in scalar_min_max_mul()
10861 dst_reg->smax_value = dst_reg->umax_value; in scalar_min_max_mul()
10905 u64 umax_val = src_reg->umax_value; in scalar_min_max_and()
10916 dst_reg->umax_value = min(dst_reg->umax_value, umax_val); in scalar_min_max_and()
10928 dst_reg->smax_value = dst_reg->umax_value; in scalar_min_max_and()
10985 dst_reg->umax_value = dst_reg->var_off.value | dst_reg->var_off.mask; in scalar_min_max_or()
10997 dst_reg->smax_value = dst_reg->umax_value; in scalar_min_max_or()
11047 dst_reg->umax_value = dst_reg->var_off.value | dst_reg->var_off.mask; in scalar_min_max_xor()
11054 dst_reg->smax_value = dst_reg->umax_value; in scalar_min_max_xor()
11120 if (dst_reg->umax_value > 1ULL << (63 - umax_val)) { in __scalar64_min_max_lsh()
11122 dst_reg->umax_value = U64_MAX; in __scalar64_min_max_lsh()
11125 dst_reg->umax_value <<= umax_val; in __scalar64_min_max_lsh()
11132 u64 umax_val = src_reg->umax_value; in scalar_min_max_lsh()
11179 u64 umax_val = src_reg->umax_value; in scalar_min_max_rsh()
11200 dst_reg->umax_value >>= umin_val; in scalar_min_max_rsh()
11250 dst_reg->umax_value = U64_MAX; in scalar_min_max_arsh()
11283 umax_val = src_reg.umax_value; in adjust_scalar_min_max_vals()
11698 if (dst_reg->umax_value > MAX_PACKET_OFF || in find_good_pkt_pointers()
11699 dst_reg->umax_value + dst_reg->off > MAX_PACKET_OFF) in find_good_pkt_pointers()
11859 else if (reg->umax_value <= val) in is_branch64_taken()
11869 if (reg->umax_value < val) in is_branch64_taken()
11883 else if (reg->umax_value < val) in is_branch64_taken()
11893 if (reg->umax_value <= val) in is_branch64_taken()
12087 false_reg->umax_value = min(false_reg->umax_value, false_umax); in reg_set_min_max()
12126 true_reg->umax_value = min(true_reg->umax_value, true_umax); in reg_set_min_max()
12189 src_reg->umax_value = dst_reg->umax_value = min(src_reg->umax_value, in __reg_combine_min_max()
12190 dst_reg->umax_value); in __reg_combine_min_max()
13680 old->umax_value >= cur->umax_value && in range_within()