Home
last modified time | relevance | path

Searched refs:pool (Results 1 – 25 of 51) sorted by relevance

123

/include/net/
A Dxdp_sock_drv.h42 return pool->chunk_size; in xsk_pool_get_chunk_size()
47 return xsk_pool_get_chunk_size(pool) - xsk_pool_get_headroom(pool); in xsk_pool_get_rx_frame_size()
53 xp_set_rxq_info(pool, rxq); in xsk_pool_set_rxq_info()
59 xp_fill_cb(pool, desc); in xsk_pool_fill_cb()
65 xp_dma_unmap(pool, attrs); in xsk_pool_dma_unmap()
71 struct xdp_umem *umem = pool->umem; in xsk_pool_dma_map()
92 return xp_alloc(pool); in xsk_buff_alloc()
108 return xp_can_alloc(pool, count); in xsk_buff_can_alloc()
191 return xp_raw_get_dma(pool, addr); in xsk_buff_raw_get_dma()
233 if (!pool->tx_metadata_len) in __xsk_buff_get_metadata()
[all …]
A Dxsk_buff_pool.h30 struct xsk_buff_pool *pool; member
124 xskb->xdp.data_hard_start = pool->addrs + addr + pool->headroom; in xp_init_xskb_addr()
167 xskb->pool->frame_len, in xp_dma_sync_for_cpu()
192 return pool->dma_pages && in xp_desc_crosses_non_contig_pg()
203 return addr & pool->chunk_mask; in xp_aligned_extract_addr()
224 return xp_aligned_extract_addr(pool, addr) >> pool->chunk_shift; in xp_aligned_extract_idx()
229 if (xskb->pool->unaligned) in xp_release()
230 xskb->pool->free_heads[xskb->pool->free_heads_cnt++] = xskb; in xp_release()
234 struct xsk_buff_pool *pool) in xp_get_handle() argument
239 if (!pool->unaligned) in xp_get_handle()
[all …]
A Dnetdev_rx_queue.h27 struct xsk_buff_pool *pool; member
/include/net/page_pool/
A Dhelpers.h96 return page_pool_alloc_pages(pool, gfp); in page_pool_dev_alloc_pages()
128 return page_pool_alloc_netmems(pool, gfp); in page_pool_alloc_netmem()
141 pool->frag_offset = max_size; in page_pool_alloc_netmem()
160 return page_pool_alloc_netmems(pool, gfp); in page_pool_dev_alloc_netmems()
234 return pool->p.dma_dir; in page_pool_get_dma_dir()
456 page_pool_get_dma_dir(pool)); in __page_pool_dma_sync_for_cpu()
484 if (!pool->dma_sync_for_cpu) in page_pool_dma_sync_netmem_for_cpu()
487 __page_pool_dma_sync_for_cpu(pool, in page_pool_dma_sync_netmem_for_cpu()
499 if (unlikely(pool->p.nid != new_nid)) in page_pool_nid_changed()
500 page_pool_update_nid(pool, new_nid); in page_pool_nid_changed()
[all …]
A Dmemory_provider.h13 netmem_ref (*alloc_netmems)(struct page_pool *pool, gfp_t gfp);
14 bool (*release_netmem)(struct page_pool *pool, netmem_ref netmem);
15 int (*init)(struct page_pool *pool);
16 void (*destroy)(struct page_pool *pool);
23 void net_mp_niov_set_page_pool(struct page_pool *pool, struct net_iov *niov);
45 static inline void net_mp_netmem_place_in_cache(struct page_pool *pool, in net_mp_netmem_place_in_cache() argument
48 pool->alloc.cache[pool->alloc.count++] = netmem; in net_mp_netmem_place_in_cache()
A Dtypes.h187 void (*disconnect)(void *pool);
254 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp);
255 netmem_ref page_pool_alloc_netmems(struct page_pool *pool, gfp_t gfp);
258 netmem_ref page_pool_alloc_frag_netmem(struct page_pool *pool,
268 void page_pool_enable_direct_recycling(struct page_pool *pool,
270 void page_pool_disable_direct_recycling(struct page_pool *pool);
271 void page_pool_destroy(struct page_pool *pool);
276 static inline void page_pool_destroy(struct page_pool *pool) in page_pool_destroy() argument
280 static inline void page_pool_use_xdp_mem(struct page_pool *pool, in page_pool_use_xdp_mem() argument
294 void page_pool_put_unrefed_page(struct page_pool *pool, struct page *page,
[all …]
/include/trace/events/
A Dpage_pool.h16 TP_PROTO(const struct page_pool *pool,
19 TP_ARGS(pool, inflight, hold, release),
30 __entry->pool = pool;
34 __entry->cnt = pool->destroy_cnt;
47 TP_ARGS(pool, netmem, release),
57 __entry->pool = pool;
73 TP_ARGS(pool, netmem, hold),
83 __entry->pool = pool;
98 TP_ARGS(pool, new_nid),
107 __entry->pool = pool;
[all …]
/include/linux/
A Dgenalloc.h52 void *data, struct gen_pool *pool,
104 return gen_pool_add_owner(pool, addr, phys, size, nid, NULL); in gen_pool_add_virt()
122 return gen_pool_add_virt(pool, addr, -1, size, nid); in gen_pool_add()
131 return gen_pool_alloc_algo_owner(pool, size, pool->algo, pool->data, in gen_pool_alloc_owner()
153 return gen_pool_alloc_algo(pool, size, pool->algo, pool->data); in gen_pool_alloc()
172 gen_pool_free_owner(pool, addr, size, NULL); in gen_pool_free()
185 struct gen_pool *pool, unsigned long start_addr);
189 void *data, struct gen_pool *pool, unsigned long start_addr);
193 void *data, struct gen_pool *pool, unsigned long start_addr);
198 void *data, struct gen_pool *pool, unsigned long start_addr);
[all …]
A Dzsmalloc.h27 void zs_destroy_pool(struct zs_pool *pool);
29 unsigned long zs_malloc(struct zs_pool *pool, size_t size, gfp_t flags,
31 void zs_free(struct zs_pool *pool, unsigned long obj);
33 size_t zs_huge_class_size(struct zs_pool *pool);
35 unsigned long zs_get_total_pages(struct zs_pool *pool);
36 unsigned long zs_compact(struct zs_pool *pool);
38 unsigned int zs_lookup_class_index(struct zs_pool *pool, unsigned int size);
40 void zs_pool_stats(struct zs_pool *pool, struct zs_pool_stats *stats);
42 void *zs_obj_read_begin(struct zs_pool *pool, unsigned long handle,
44 void zs_obj_read_end(struct zs_pool *pool, unsigned long handle,
[all …]
A Dzpool.h20 const char *zpool_get_type(struct zpool *pool);
22 void zpool_destroy_pool(struct zpool *pool);
27 void zpool_free(struct zpool *pool, unsigned long handle);
38 u64 zpool_get_total_pages(struct zpool *pool);
64 void (*destroy)(void *pool);
66 int (*malloc)(void *pool, size_t size, gfp_t gfp,
68 void (*free)(void *pool, unsigned long handle);
72 void (*obj_read_end)(void *pool, unsigned long handle,
74 void (*obj_write)(void *pool, unsigned long handle,
77 u64 (*total_pages)(void *pool);
[all …]
A Dobjpool.h128 struct objpool_slot *slot = pool->cpu_slots[cpu]; in __objpool_try_get_slot()
147 if (READ_ONCE(slot->last) - head - 1 >= pool->nr_objs) { in __objpool_try_get_slot()
169 static inline void *objpool_pop(struct objpool_head *pool) in objpool_pop() argument
180 obj = __objpool_try_get_slot(pool, cpu); in objpool_pop()
193 struct objpool_slot *slot = pool->cpu_slots[cpu]; in __objpool_try_add_slot()
202 WARN_ON_ONCE(tail - head > pool->nr_objs); in __objpool_try_add_slot()
221 static inline int objpool_push(void *obj, struct objpool_head *pool) in objpool_push() argument
228 rc = __objpool_try_add_slot(obj, pool, raw_smp_processor_id()); in objpool_push()
251 int objpool_drop(void *obj, struct objpool_head *pool);
257 void objpool_free(struct objpool_head *pool);
[all …]
A Ddmapool.h25 void dma_pool_destroy(struct dma_pool *pool);
27 void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags,
29 void dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t addr);
36 void dmam_pool_destroy(struct dma_pool *pool);
45 static inline void dma_pool_destroy(struct dma_pool *pool) { } in dma_pool_destroy() argument
46 static inline void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, in dma_pool_alloc() argument
48 static inline void dma_pool_free(struct dma_pool *pool, void *vaddr, in dma_pool_free() argument
53 static inline void dmam_pool_destroy(struct dma_pool *pool) { } in dmam_pool_destroy() argument
71 static inline void *dma_pool_zalloc(struct dma_pool *pool, gfp_t mem_flags, in dma_pool_zalloc() argument
74 return dma_pool_alloc(pool, mem_flags | __GFP_ZERO, handle); in dma_pool_zalloc()
A Dmempool.h30 static inline bool mempool_initialized(mempool_t *pool) in mempool_initialized() argument
32 return pool->elements != NULL; in mempool_initialized()
35 static inline bool mempool_is_saturated(mempool_t *pool) in mempool_is_saturated() argument
37 return READ_ONCE(pool->curr_nr) >= pool->min_nr; in mempool_is_saturated()
40 void mempool_exit(mempool_t *pool);
41 int mempool_init_node(mempool_t *pool, int min_nr, mempool_alloc_t *alloc_fn,
63 extern int mempool_resize(mempool_t *pool, int new_min_nr);
64 extern void mempool_destroy(mempool_t *pool);
66 extern void *mempool_alloc_noprof(mempool_t *pool, gfp_t gfp_mask) __malloc;
70 extern void *mempool_alloc_preallocated(mempool_t *pool) __malloc;
[all …]
A Dswiotlb.h247 unsigned long attrs, struct io_tlb_pool *pool);
252 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_tbl_unmap_single() local
254 if (unlikely(pool)) in swiotlb_tbl_unmap_single()
260 struct io_tlb_pool *pool);
264 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_sync_single_for_device() local
266 if (unlikely(pool)) in swiotlb_sync_single_for_device()
267 __swiotlb_sync_single_for_device(dev, addr, size, dir, pool); in swiotlb_sync_single_for_device()
272 struct io_tlb_pool *pool);
276 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_sync_single_for_cpu() local
278 if (unlikely(pool)) in swiotlb_sync_single_for_cpu()
[all …]
A Dcgroup_dmem.h23 void dmem_cgroup_uncharge(struct dmem_cgroup_pool_state *pool, u64 size);
28 void dmem_cgroup_pool_state_put(struct dmem_cgroup_pool_state *pool);
51 static inline void dmem_cgroup_uncharge(struct dmem_cgroup_pool_state *pool, u64 size) in dmem_cgroup_uncharge() argument
62 static inline void dmem_cgroup_pool_state_put(struct dmem_cgroup_pool_state *pool) in dmem_cgroup_pool_state_put() argument
A Dtee_core.h62 struct tee_shm_pool *pool; member
133 struct tee_shm_pool *pool,
202 int (*alloc)(struct tee_shm_pool *pool, struct tee_shm *shm,
204 void (*free)(struct tee_shm_pool *pool, struct tee_shm *shm);
205 void (*destroy_pool)(struct tee_shm_pool *pool);
227 static inline void tee_shm_pool_free(struct tee_shm_pool *pool) in tee_shm_pool_free() argument
229 pool->ops->destroy_pool(pool); in tee_shm_pool_free()
A Dsram.h9 void *sram_exec_copy(struct gen_pool *pool, void *dst, void *src, size_t size);
11 static inline void *sram_exec_copy(struct gen_pool *pool, void *dst, void *src, in sram_exec_copy() argument
A Drethook.h37 struct objpool_head pool; member
/include/drm/ttm/
A Dttm_pool.h52 struct ttm_pool *pool; member
83 int ttm_pool_alloc(struct ttm_pool *pool, struct ttm_tt *tt,
85 void ttm_pool_free(struct ttm_pool *pool, struct ttm_tt *tt);
87 void ttm_pool_init(struct ttm_pool *pool, struct device *dev,
89 void ttm_pool_fini(struct ttm_pool *pool);
91 int ttm_pool_debugfs(struct ttm_pool *pool, struct seq_file *m);
95 long ttm_pool_backup(struct ttm_pool *pool, struct ttm_tt *ttm,
97 int ttm_pool_restore_and_alloc(struct ttm_pool *pool, struct ttm_tt *tt,
/include/linux/ceph/
A Dmsgpool.h13 mempool_t *pool; member
19 int ceph_msgpool_init(struct ceph_msgpool *pool, int type,
22 extern void ceph_msgpool_destroy(struct ceph_msgpool *pool);
23 struct ceph_msg *ceph_msgpool_get(struct ceph_msgpool *pool, int front_len,
A Dosdmap.h23 uint64_t pool; member
63 static inline bool ceph_can_shift_osds(struct ceph_pg_pool_info *pool) in ceph_can_shift_osds() argument
65 switch (pool->type) { in ceph_can_shift_osds()
76 s64 pool; member
82 oloc->pool = -1; in ceph_oloc_init()
88 return oloc->pool == -1; in ceph_oloc_empty()
246 pgid->pool = ceph_decode_64(p); in ceph_decode_pgid()
/include/soc/fsl/
A Dbman.h94 void bman_free_pool(struct bman_pool *pool);
103 int bman_get_bpid(const struct bman_pool *pool);
114 int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num);
127 int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num);
/include/linux/firmware/qcom/
A Dqcom_tzmem.h44 void qcom_tzmem_pool_free(struct qcom_tzmem_pool *pool);
49 void *qcom_tzmem_alloc(struct qcom_tzmem_pool *pool, size_t size, gfp_t gfp);
/include/net/libeth/
A Dxsk.h122 xsk_buff_raw_dma_sync_for_device(sq->pool, desc.addr, desc.len); in libeth_xsk_tx_fill_buf()
199 ctx = xsk_buff_raw_get_ctx(sq->pool, xdesc->addr); in __libeth_xsk_xmit_fill_buf_md()
228 .addr = xsk_buff_raw_get_dma(sq->pool, xdesc->addr), in __libeth_xsk_xmit_fill_buf()
259 xsk_buff_raw_dma_sync_for_device(sq->pool, desc.addr, desc.len); in libeth_xsk_xmit_fill_buf()
296 wake = xsk_uses_need_wakeup(pool); in libeth_xsk_xmit_do_bulk()
298 xsk_clear_tx_need_wakeup(pool); in libeth_xsk_xmit_do_bulk()
300 n = xsk_tx_peek_release_desc_batch(pool, budget); in libeth_xsk_xmit_do_bulk()
301 bulk = container_of(&pool->tx_descs[0], typeof(*bulk), desc); in libeth_xsk_xmit_do_bulk()
309 xsk_set_tx_need_wakeup(pool); in libeth_xsk_xmit_do_bulk()
604 struct xsk_buff_pool *pool;
[all …]
/include/uapi/linux/netfilter_bridge/
A Debt_among.h43 struct ebt_mac_wormhash_tuple pool[]; member

Completed in 37 milliseconds

123