Lines Matching refs:size

75 static void nested_table_free(union nested_table *ntbl, unsigned int size)  in nested_table_free()  argument
85 if (size > len) { in nested_table_free()
86 size >>= shift; in nested_table_free()
88 nested_table_free(ntbl + i, size); in nested_table_free()
96 unsigned int size = tbl->size >> tbl->nest; in nested_bucket_table_free() local
104 nested_table_free(ntbl + i, size); in nested_bucket_table_free()
154 size_t size; in nested_bucket_table_alloc() local
159 size = sizeof(*tbl) + sizeof(tbl->buckets[0]); in nested_bucket_table_alloc()
162 kmalloc_noprof(size, gfp|__GFP_ZERO)); in nested_bucket_table_alloc()
182 size_t size; in bucket_table_alloc() local
190 size = nbuckets; in bucket_table_alloc()
202 tbl->size = size; in bucket_table_alloc()
331 for (old_hash = 0; old_hash < old_tbl->size; old_hash++) { in rhashtable_rehash_table()
360 unsigned int size) in rhashtable_rehash_alloc() argument
367 new_tbl = bucket_table_alloc(ht, size, GFP_KERNEL); in rhashtable_rehash_alloc()
398 unsigned int size = 0; in rhashtable_shrink() local
401 size = roundup_pow_of_two(nelems * 3 / 2); in rhashtable_shrink()
402 if (size < ht->p.min_size) in rhashtable_shrink()
403 size = ht->p.min_size; in rhashtable_shrink()
405 if (old_tbl->size <= size) in rhashtable_shrink()
411 return rhashtable_rehash_alloc(ht, old_tbl, size); in rhashtable_shrink()
427 err = rhashtable_rehash_alloc(ht, tbl, tbl->size * 2); in rht_deferred_worker()
431 err = rhashtable_rehash_alloc(ht, tbl, tbl->size); in rht_deferred_worker()
451 unsigned int size; in rhashtable_insert_rehash() local
456 size = tbl->size; in rhashtable_insert_rehash()
461 size *= 2; in rhashtable_insert_rehash()
468 new_tbl = bucket_table_alloc(ht, size, GFP_ATOMIC | __GFP_NOWARN); in rhashtable_insert_rehash()
810 for (; iter->slot < tbl->size; iter->slot++) { in __rhashtable_walk_find_next()
1031 size_t size; in rhashtable_init_noprof() local
1058 size = rounded_hashtable_size(&ht->p); in rhashtable_init_noprof()
1075 tbl = bucket_table_alloc(ht, size, GFP_KERNEL); in rhashtable_init_noprof()
1077 size = max_t(u16, ht->p.min_size, HASH_MIN_SIZE); in rhashtable_init_noprof()
1078 tbl = bucket_table_alloc(ht, size, GFP_KERNEL | __GFP_NOFAIL); in rhashtable_init_noprof()
1157 for (i = 0; i < tbl->size; i++) { in rhashtable_free_and_destroy()
1193 unsigned int size = tbl->size >> tbl->nest; in __rht_bucket_nested() local
1201 while (ntbl && size > (1 << shift)) { in __rht_bucket_nested()
1205 size >>= shift; in __rht_bucket_nested()
1233 unsigned int size = tbl->size >> tbl->nest; in rht_bucket_nested_insert() local
1239 size <= (1 << shift)); in rht_bucket_nested_insert()
1241 while (ntbl && size > (1 << shift)) { in rht_bucket_nested_insert()
1243 size >>= shift; in rht_bucket_nested_insert()
1246 size <= (1 << shift)); in rht_bucket_nested_insert()