Lines Matching refs:base

208 		unsigned long base = round_down(index, 1UL << order);  in check_xa_mark_1()  local
209 unsigned long next = base + (1UL << order); in check_xa_mark_1()
219 for (i = base; i < next; i++) { in check_xa_mark_1()
822 unsigned long index, base, next_index, next_next_index; in check_xa_multi_store_adv() local
826 base = round_down(index, nrpages); in check_xa_multi_store_adv()
827 next_index = round_down(base + nrpages, nrpages); in check_xa_multi_store_adv()
830 check_xa_multi_store_adv_add(xa, base, order, &some_val); in check_xa_multi_store_adv()
833 XA_BUG_ON(xa, test_get_entry(xa, base + i) != &some_val); in check_xa_multi_store_adv()
851 check_xa_multi_store_adv_delete(xa, base, order); in check_xa_multi_store_adv()
862 XA_BUG_ON(xa, test_get_entry(xa, base + i) != NULL); in check_xa_multi_store_adv()
878 XA_BUG_ON(xa, test_get_entry(xa, base + i) != NULL); in check_xa_multi_store_adv()
910 static noinline void check_xa_alloc_1(struct xarray *xa, unsigned int base) in check_xa_alloc_1() argument
917 xa_alloc_index(xa, base, GFP_KERNEL); in check_xa_alloc_1()
920 xa_erase_index(xa, base); in check_xa_alloc_1()
924 xa_alloc_index(xa, base, GFP_KERNEL); in check_xa_alloc_1()
927 for (i = base + 1; i < 2 * XA_CHUNK_SIZE; i++) in check_xa_alloc_1()
929 for (i = base; i < 2 * XA_CHUNK_SIZE; i++) in check_xa_alloc_1()
931 xa_alloc_index(xa, base, GFP_KERNEL); in check_xa_alloc_1()
937 xa_alloc_index(xa, base, GFP_KERNEL); in check_xa_alloc_1()
940 xa_alloc_index(xa, base + 1, GFP_KERNEL); in check_xa_alloc_1()
941 xa_erase_index(xa, base + 1); in check_xa_alloc_1()
944 xa_store_index(xa, base + 1, GFP_KERNEL); in check_xa_alloc_1()
945 xa_alloc_index(xa, base + 2, GFP_KERNEL); in check_xa_alloc_1()
948 xa_erase_index(xa, base); in check_xa_alloc_1()
949 xa_alloc_index(xa, base, GFP_KERNEL); in check_xa_alloc_1()
951 xa_erase_index(xa, base + 1); in check_xa_alloc_1()
952 xa_erase_index(xa, base + 2); in check_xa_alloc_1()
955 xa_alloc_index(xa, base + i, GFP_KERNEL); in check_xa_alloc_1()
985 static noinline void check_xa_alloc_2(struct xarray *xa, unsigned int base) in check_xa_alloc_2() argument
994 XA_BUG_ON(xa, id != base); in check_xa_alloc_2()
1002 XA_BUG_ON(xa, id != base); in check_xa_alloc_2()
1007 for (i = base; i < base + 10; i++) { in check_xa_alloc_2()
1024 for (i = base; i < base + 9; i++) { in check_xa_alloc_2()
1030 XA_BUG_ON(xa, xa_erase(xa, base + 9) != NULL); in check_xa_alloc_2()
1036 static noinline void check_xa_alloc_3(struct xarray *xa, unsigned int base) in check_xa_alloc_3() argument
1068 if (base != 0) in check_xa_alloc_3()
1069 xa_erase_index(xa, base); in check_xa_alloc_3()
1070 xa_erase_index(xa, base + 1); in check_xa_alloc_3()
1075 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id, xa_mk_index(base), in check_xa_alloc_3()
1077 XA_BUG_ON(xa, id != base); in check_xa_alloc_3()
1078 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id, xa_mk_index(base + 1), in check_xa_alloc_3()
1080 XA_BUG_ON(xa, id != base + 1); in check_xa_alloc_3()
1091 ret = __xa_alloc_cyclic(xa, &id, xa_mk_index(base), xa_limit_32b, in check_xa_alloc_3()
1720 unsigned long base = xas.xa_index; in check_create_range_4() local
1730 void *old = xas_store(&xas, xa_mk_index(base + i)); in check_create_range_4()
1732 XA_BUG_ON(xa, old != xa_mk_index(base + i)); in check_create_range_4()
1743 for (i = base; i < base + (1UL << order); i++) in check_create_range_4()