| /mm/ |
| A D | mempool.c | 139 BUG_ON(pool->min_nr != 0 && pool->curr_nr >= pool->min_nr); in add_element() 142 pool->elements[pool->curr_nr++] = element; in add_element() 147 void *element = pool->elements[--pool->curr_nr]; in remove_element() 170 pool->free(element, pool->pool_data); in mempool_exit() 218 while (pool->curr_nr < max(1, pool->min_nr)) { in mempool_init_node() 221 element = pool->alloc(gfp_mask, pool->pool_data); in mempool_init_node() 326 pool->free(element, pool->pool_data); in mempool_resize() 353 while (pool->curr_nr < pool->min_nr) { in mempool_resize() 359 if (pool->curr_nr < pool->min_nr) { in mempool_resize() 544 if (likely(pool->curr_nr < pool->min_nr)) { in mempool_free() [all …]
|
| A D | dmapool.c | 84 pool->name, pool->nr_active, in pools_show() 85 pool->nr_blocks, pool->size, in pools_show() 188 pool->nr_active++; in pool_block_pop() 309 while (offset + pool->size <= pool->allocation) { in pool_initialise_page() 334 pool->nr_pages++; in pool_initialise_page() 345 page->vaddr = dma_alloc_coherent(pool->dev, pool->allocation, in pool_alloc_page() 381 dev_err(pool->dev, "%s %s busy\n", __func__, pool->name); in dma_pool_destroy() 387 dma_free_coherent(pool->dev, pool->allocation, in dma_pool_destroy() 393 kfree(pool); in dma_pool_destroy() 506 if (pool) in dmam_pool_create() [all …]
|
| A D | zsmalloc.c | 694 debugfs_create_file("classes", S_IFREG | 0444, pool->stat_dentry, pool, in zs_pool_stat_create() 1050 zspage->pool = pool; in alloc_zspage() 1753 pool = zspage->pool; in zs_page_migrate() 2054 pool->shrinker = shrinker_alloc(0, "mm-zspool:%s", pool->name); in zs_register_shrinker() 2061 pool->shrinker->private_data = pool; in zs_register_shrinker() 2105 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in zs_create_pool() 2106 if (!pool) in zs_create_pool() 2114 if (!pool->name) in zs_create_pool() 2205 return pool; in zs_create_pool() 2242 kfree(pool->name); in zs_destroy_pool() [all …]
|
| A D | zswap.c | 269 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in zswap_pool_create() 270 if (!pool) in zswap_pool_create() 284 pool->acomp_ctx = alloc_percpu(*pool->acomp_ctx); in zswap_pool_create() 309 return pool; in zswap_pool_create() 396 pool = container_of(ref, typeof(*pool), ref); in __zswap_pool_empty() 412 if (!pool) in zswap_pool_tryget() 583 if (pool) { in __zswap_param_set() 591 if (!pool) in __zswap_param_set() 605 if (pool) in __zswap_param_set() 1538 entry->pool = pool; in zswap_store_page() [all …]
|
| A D | zpool.c | 23 void *pool; member 174 zpool->pool = driver->create(name, gfp); in zpool_create_pool() 176 if (!zpool->pool) { in zpool_create_pool() 203 zpool->driver->destroy(zpool->pool); in zpool_destroy_pool() 244 return zpool->driver->malloc(zpool->pool, size, gfp, handle, nid); in zpool_malloc() 263 zpool->driver->free(zpool->pool, handle); in zpool_free() 283 return zpool->driver->obj_read_begin(zpool->pool, handle, local_copy); in zpool_obj_read_begin() 297 zpool->driver->obj_read_end(zpool->pool, handle, handle_mem); in zpool_obj_read_end() 311 zpool->driver->obj_write(zpool->pool, handle, handle_mem, mem_len); in zpool_obj_write() 324 return zpool->driver->total_pages(zpool->pool); in zpool_get_total_pages()
|
| A D | dmapool_test.c | 31 static struct dma_pool *pool; variable 45 p[i].v = dma_pool_alloc(pool, GFP_KERNEL, in dmapool_test_alloc() 52 dma_pool_free(pool, p[i].v, p[i].dma); in dmapool_test_alloc() 58 dma_pool_free(pool, p[i].v, p[i].dma); in dmapool_test_alloc() 73 pool = dma_pool_create("test pool", &test_dev, parms->size, in dmapool_test_block() 75 if (!pool) { in dmapool_test_block() 95 dma_pool_destroy(pool); in dmapool_test_block()
|
| A D | page_io.c | 334 mempool_t *pool = mempool_create_kmalloc_pool( in sio_pool_init() local 336 if (cmpxchg(&sio_pool, NULL, pool)) in sio_pool_init() 337 mempool_destroy(pool); in sio_pool_init()
|
| A D | vmalloc.c | 918 struct vmap_pool pool[MAX_VA_SIZE_PAGES]; member 1903 return &vn->pool[idx]; in size_to_va_pool() 2197 if (list_empty(&vn->pool[i].head)) in decay_va_pool_node() 2202 list_replace_init(&vn->pool[i].head, &tmp_list); in decay_va_pool_node() 2205 pool_len = n_decay = vn->pool[i].len; in decay_va_pool_node() 2206 WRITE_ONCE(vn->pool[i].len, 0); in decay_va_pool_node() 2229 list_replace_init(&tmp_list, &vn->pool[i].head); in decay_va_pool_node() 2230 WRITE_ONCE(vn->pool[i].len, pool_len); in decay_va_pool_node() 5202 INIT_LIST_HEAD(&vn->pool[i].head); in vmap_init_nodes() 5203 WRITE_ONCE(vn->pool[i].len, 0); in vmap_init_nodes() [all …]
|
| A D | Kconfig.debug | 266 int "Kmemleak memory pool size" 273 freed before kmemleak is fully initialised, use a static pool 275 fully initialised, this memory pool acts as an emergency one
|
| A D | Kconfig | 33 compress them into a dynamically allocated RAM-based memory pool. 49 bool "Shrink the zswap pool on memory pressure" 54 stored in the zswap pool will become available for reclaim (i.e 57 This means that zswap writeback could happen even if the pool is 59 reducing the chance that cold pages will reside in the zswap pool 183 initialization of the pool. 188 the pool, as zsmalloc merges size classes with similar
|
| /mm/kasan/ |
| A D | kasan_test_c.c | 1144 memset(pool, 0, sizeof(*pool)); in mempool_prepare_kmalloc() 1169 memset(pool, 0, sizeof(*pool)); in mempool_prepare_slab() 1187 memset(pool, 0, sizeof(*pool)); in mempool_prepare_page() 1218 mempool_t pool; in mempool_kmalloc_oob_right() local 1232 mempool_t pool; in mempool_kmalloc_large_oob_right() local 1246 mempool_t pool; in mempool_slab_oob_right() local 1281 mempool_t pool; in mempool_kmalloc_uaf() local 1295 mempool_t pool; in mempool_kmalloc_large_uaf() local 1309 mempool_t pool; in mempool_slab_uaf() local 1323 mempool_t pool; in mempool_page_alloc_uaf() local [all …]
|