| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| A D | mlx5hws_buddy.c | 11 buddy->max_order = max_order; in hws_buddy_init() 13 buddy->bitmap = kcalloc(buddy->max_order + 1, in hws_buddy_init() 19 buddy->num_free = kcalloc(buddy->max_order + 1, in hws_buddy_init() 27 for (i = 0; i <= (int)buddy->max_order; ++i) { in hws_buddy_init() 28 s = 1 << (buddy->max_order - i); in hws_buddy_init() 38 buddy->num_free[buddy->max_order] = 1; in hws_buddy_init() 43 for (i = 0; i <= (int)buddy->max_order; ++i) in hws_buddy_init() 61 if (hws_buddy_init(buddy, max_order)) in mlx5hws_buddy_create() 75 for (i = 0; i <= (int)buddy->max_order; ++i) in mlx5hws_buddy_cleanup() 94 m = 1 << (buddy->max_order - order_iter); in hws_buddy_find_free_seg() [all …]
|
| A D | mlx5hws_buddy.h | 10 u32 max_order; member 13 struct mlx5hws_buddy_mem *mlx5hws_buddy_create(u32 max_order);
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
| A D | dr_buddy.c | 11 unsigned int max_order) in mlx5dr_buddy_init() argument 15 buddy->max_order = max_order; in mlx5dr_buddy_init() 19 buddy->bitmap = kcalloc(buddy->max_order + 1, in mlx5dr_buddy_init() 22 buddy->num_free = kcalloc(buddy->max_order + 1, in mlx5dr_buddy_init() 31 for (i = 0; i <= buddy->max_order; ++i) { in mlx5dr_buddy_init() 32 unsigned int size = 1 << (buddy->max_order - i); in mlx5dr_buddy_init() 45 buddy->num_free[buddy->max_order] = 1; in mlx5dr_buddy_init() 50 for (i = 0; i <= buddy->max_order; ++i) in mlx5dr_buddy_init() 65 for (i = 0; i <= buddy->max_order; ++i) in mlx5dr_buddy_cleanup() 84 m = 1 << (buddy->max_order - order_iter); in dr_buddy_find_free_seg() [all …]
|
| A D | mlx5dr.h | 173 u32 max_order; member 188 unsigned int max_order);
|
| /linux/drivers/gpu/drm/tests/ |
| A D | drm_buddy_test.c | 267 const int max_order = 12; in drm_test_buddy_alloc_clear() local 275 mm_size = SZ_4K << max_order; in drm_test_buddy_alloc_clear() 278 KUNIT_EXPECT_EQ(test, mm.max_order, max_order); in drm_test_buddy_alloc_clear() 493 const int max_order = 3; in drm_test_buddy_alloc_pathological() local 508 mm_size = SZ_4K << max_order; in drm_test_buddy_alloc_pathological() 512 KUNIT_EXPECT_EQ(test, mm.max_order, max_order); in drm_test_buddy_alloc_pathological() 551 top, max_order); in drm_test_buddy_alloc_pathological() 591 KUNIT_EXPECT_EQ(test, mm.max_order, max_order); in drm_test_buddy_alloc_pessimistic() 654 max_order); in drm_test_buddy_alloc_pessimistic() 670 const int max_order = 16; in drm_test_buddy_alloc_optimistic() local [all …]
|
| /linux/drivers/gpu/drm/i915/gem/ |
| A D | i915_gem_internal.c | 39 int max_order = MAX_PAGE_ORDER; in i915_gem_object_get_pages_internal() local 48 max_order = min(max_order, get_order(max_segment)); in i915_gem_object_get_pages_internal() 71 int order = min(fls(npages) - 1, max_order); in i915_gem_object_get_pages_internal() 83 max_order = order; in i915_gem_object_get_pages_internal() 102 max_order = 0; in i915_gem_object_get_pages_internal()
|
| /linux/virt/kvm/ |
| A D | guest_memfd.c | 557 int *max_order) in __kvm_gmem_get_pfn() argument 585 if (max_order) in __kvm_gmem_get_pfn() 586 *max_order = 0; in __kvm_gmem_get_pfn() 593 gfn_t gfn, kvm_pfn_t *pfn, int *max_order) in kvm_gmem_get_pfn() argument 630 int ret = 0, max_order; in kvm_gmem_populate() local 648 for (i = 0; i < npages; i += (1 << max_order)) { in kvm_gmem_populate() 673 WARN_ON(!IS_ALIGNED(gfn, 1 << max_order) || in kvm_gmem_populate() 674 (npages - i) < (1 << max_order)); in kvm_gmem_populate() 680 if (!max_order) in kvm_gmem_populate() 682 max_order--; in kvm_gmem_populate() [all …]
|
| /linux/drivers/infiniband/hw/mthca/ |
| A D | mthca_mr.c | 92 for (o = order; o <= buddy->max_order; ++o) in mthca_buddy_alloc() 94 m = 1 << (buddy->max_order - o); in mthca_buddy_alloc() 144 buddy->max_order = max_order; in mthca_buddy_init() 147 buddy->bits = kcalloc(buddy->max_order + 1, sizeof(long *), in mthca_buddy_init() 154 for (i = 0; i <= buddy->max_order; ++i) { in mthca_buddy_init() 155 buddy->bits[i] = bitmap_zalloc(1 << (buddy->max_order - i), in mthca_buddy_init() 161 __set_bit(0, buddy->bits[buddy->max_order]); in mthca_buddy_init() 162 buddy->num_free[buddy->max_order] = 1; in mthca_buddy_init() 167 for (i = 0; i <= buddy->max_order; ++i) in mthca_buddy_init() 181 for (i = 0; i <= buddy->max_order; ++i) in mthca_buddy_cleanup() [all …]
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| A D | page_alloc.c | 113 for (; (order + 1) <= pool->max_order; order++) { in __hyp_attach_page() 206 while (i <= pool->max_order && list_empty(&pool->free_area[i])) in hyp_alloc_pages() 208 if (i > pool->max_order) { in hyp_alloc_pages() 231 pool->max_order = min(MAX_PAGE_ORDER, in hyp_pool_init() 233 for (i = 0; i <= pool->max_order; i++) in hyp_pool_init()
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| A D | mr.c | 54 for (o = order; o <= buddy->max_order; ++o) in mlx4_buddy_alloc() 56 m = 1 << (buddy->max_order - o); in mlx4_buddy_alloc() 106 buddy->max_order = max_order; in mlx4_buddy_init() 109 buddy->bits = kcalloc(buddy->max_order + 1, sizeof(long *), in mlx4_buddy_init() 116 for (i = 0; i <= buddy->max_order; ++i) { in mlx4_buddy_init() 117 s = BITS_TO_LONGS(1UL << (buddy->max_order - i)); in mlx4_buddy_init() 123 set_bit(0, buddy->bits[buddy->max_order]); in mlx4_buddy_init() 124 buddy->num_free[buddy->max_order] = 1; in mlx4_buddy_init() 129 for (i = 0; i <= buddy->max_order; ++i) in mlx4_buddy_init() 143 for (i = 0; i <= buddy->max_order; ++i) in mlx4_buddy_cleanup() [all …]
|
| /linux/tools/testing/radix-tree/ |
| A D | iteration_check.c | 19 static int max_order; variable 29 for (order = max_order; order >= 0; order--) { in my_item_insert() 172 max_order = order; in iteration_test()
|
| /linux/drivers/gpu/drm/ |
| A D | drm_buddy.c | 175 if (min_order > mm->max_order) in __force_merge() 254 mm->max_order = ilog2(size) - ilog2(chunk_size); in drm_buddy_init() 256 BUG_ON(mm->max_order > DRM_BUDDY_MAX_ORDER); in drm_buddy_init() 258 mm->free_list = kmalloc_array(mm->max_order + 1, in drm_buddy_init() 264 for (i = 0; i <= mm->max_order; ++i) in drm_buddy_init() 296 BUG_ON(i > mm->max_order); in drm_buddy_init() 592 for (i = order; i <= mm->max_order; ++i) { in get_maxblock() 635 for (tmp = order; tmp <= mm->max_order; ++tmp) { in alloc_from_freelist() 653 for (tmp = order; tmp <= mm->max_order; ++tmp) { in alloc_from_freelist() 1034 BUG_ON(order > mm->max_order); in drm_buddy_alloc_blocks() [all …]
|
| /linux/drivers/dma-buf/heaps/ |
| A D | system_heap.c | 315 unsigned int max_order) in alloc_largest_available() argument 323 if (max_order < orders[i]) in alloc_largest_available() 342 unsigned int max_order = orders[0]; in system_heap_allocate() local 371 page = alloc_largest_available(size_remaining, max_order); in system_heap_allocate() 377 max_order = compound_order(page); in system_heap_allocate()
|
| /linux/mm/ |
| A D | workingset.c | 802 unsigned int max_order; in workingset_init() local 814 max_order = fls_long(totalram_pages() - 1); in workingset_init() 815 if (max_order > timestamp_bits) in workingset_init() 816 bucket_order = max_order - timestamp_bits; in workingset_init() 818 timestamp_bits, max_order, bucket_order); in workingset_init()
|
| /linux/lib/ |
| A D | test_xarray.c | 207 for (order = 2; order < max_order; order++) { in check_xa_mark_1() 353 for (order = 0; order < max_order; order++) { in check_xa_shrink() 698 for (i = 0; i < max_order; i++) { in check_multi_store() 699 for (j = 0; j < max_order; j++) { in check_multi_store() 703 for (k = 0; k < max_order; k++) { in check_multi_store() 902 for (i = 0; i < max_order; i++) { in check_multi_store_advanced() 1134 for (i = 0; i < max_order; i++) { in check_store_iter() 1205 for (i = 0; i < max_order; i++) { in check_multi_find_2() 1723 for (order = 0; order < max_order; order++) { in check_create_range() 1993 for (order = 0; order < max_order; order++) { in check_get_order() [all …]
|
| /linux/arch/arm64/kvm/hyp/include/nvhe/ |
| A D | gfp.h | 22 unsigned short max_order; member
|
| /linux/kernel/events/ |
| A D | ring_buffer.c | 682 int ret = -ENOMEM, max_order; in rb_alloc_aux() local 704 max_order = get_order(watermark); in rb_alloc_aux() 710 max_order = ilog2(nr_pages); in rb_alloc_aux() 730 order = min(max_order, ilog2(nr_pages - rb->aux_nr_pages)); in rb_alloc_aux() 750 if (page_private(page) != max_order) in rb_alloc_aux()
|
| /linux/drivers/gpu/drm/i915/selftests/ |
| A D | scatterlist.c | 281 const unsigned long max_order = 20; /* approximating a 4GiB object */ in igt_sg_alloc() local 286 for_each_prime_number(prime, max_order) { in igt_sg_alloc()
|
| /linux/include/drm/ |
| A D | drm_buddy.h | 87 unsigned int max_order; member
|
| /linux/drivers/pci/ |
| A D | setup-bus.c | 954 int max_order) in calculate_mem_align() argument 960 for (order = 0; order <= max_order; order++) { in calculate_mem_align() 1064 int order, max_order; in pbus_size_mem() local 1079 max_order = 0; in pbus_size_mem() 1130 if (order > max_order) in pbus_size_mem() 1131 max_order = order; in pbus_size_mem() 1142 min_align = calculate_mem_align(aligns, max_order); in pbus_size_mem() 1150 min_align = 1ULL << (max_order + __ffs(SZ_1M)); in pbus_size_mem()
|
| /linux/arch/x86/kvm/svm/ |
| A D | svm.h | 763 int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order); 787 static inline int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order) in sev_gmem_prepare() argument
|
| /linux/drivers/accel/qaic/ |
| A D | qaic_data.c | 447 int max_order; in create_sgt() local 460 max_order = min(MAX_PAGE_ORDER, get_order(size)); in create_sgt() 465 max_order = 0; in create_sgt() 482 order = min(get_order(nr_pages * PAGE_SIZE), max_order); in create_sgt() 496 max_order = order; in create_sgt()
|
| /linux/tools/testing/kunit/test_data/ |
| A D | test_is_test_passed-no_tests_run_no_header.log | 31 workingset: timestamp_bits=62 max_order=13 bucket_order=0
|
| A D | test_output_isolated_correctly.log | 56 workingset: timestamp_bits=62 max_order=16 bucket_order=0
|
| /linux/include/linux/ |
| A D | kvm_host.h | 2464 gfn_t gfn, kvm_pfn_t *pfn, int *max_order); 2468 kvm_pfn_t *pfn, int *max_order) in kvm_gmem_get_pfn() argument 2476 int kvm_arch_gmem_prepare(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn, int max_order);
|