| /include/net/ |
| A D | xdp_sock_drv.h | 42 return pool->chunk_size; in xsk_pool_get_chunk_size() 47 return xsk_pool_get_chunk_size(pool) - xsk_pool_get_headroom(pool); in xsk_pool_get_rx_frame_size() 53 xp_set_rxq_info(pool, rxq); in xsk_pool_set_rxq_info() 59 xp_fill_cb(pool, desc); in xsk_pool_fill_cb() 65 xp_dma_unmap(pool, attrs); in xsk_pool_dma_unmap() 71 struct xdp_umem *umem = pool->umem; in xsk_pool_dma_map() 92 return xp_alloc(pool); in xsk_buff_alloc() 108 return xp_can_alloc(pool, count); in xsk_buff_can_alloc() 191 return xp_raw_get_dma(pool, addr); in xsk_buff_raw_get_dma() 233 if (!pool->tx_metadata_len) in __xsk_buff_get_metadata() [all …]
|
| A D | xsk_buff_pool.h | 30 struct xsk_buff_pool *pool; member 124 xskb->xdp.data_hard_start = pool->addrs + addr + pool->headroom; in xp_init_xskb_addr() 167 xskb->pool->frame_len, in xp_dma_sync_for_cpu() 192 return pool->dma_pages && in xp_desc_crosses_non_contig_pg() 203 return addr & pool->chunk_mask; in xp_aligned_extract_addr() 224 return xp_aligned_extract_addr(pool, addr) >> pool->chunk_shift; in xp_aligned_extract_idx() 229 if (xskb->pool->unaligned) in xp_release() 230 xskb->pool->free_heads[xskb->pool->free_heads_cnt++] = xskb; in xp_release() 234 struct xsk_buff_pool *pool) in xp_get_handle() argument 239 if (!pool->unaligned) in xp_get_handle() [all …]
|
| A D | netdev_rx_queue.h | 27 struct xsk_buff_pool *pool; member
|
| /include/net/page_pool/ |
| A D | helpers.h | 96 return page_pool_alloc_pages(pool, gfp); in page_pool_dev_alloc_pages() 128 return page_pool_alloc_netmems(pool, gfp); in page_pool_alloc_netmem() 141 pool->frag_offset = max_size; in page_pool_alloc_netmem() 160 return page_pool_alloc_netmems(pool, gfp); in page_pool_dev_alloc_netmems() 234 return pool->p.dma_dir; in page_pool_get_dma_dir() 456 page_pool_get_dma_dir(pool)); in __page_pool_dma_sync_for_cpu() 484 if (!pool->dma_sync_for_cpu) in page_pool_dma_sync_netmem_for_cpu() 487 __page_pool_dma_sync_for_cpu(pool, in page_pool_dma_sync_netmem_for_cpu() 499 if (unlikely(pool->p.nid != new_nid)) in page_pool_nid_changed() 500 page_pool_update_nid(pool, new_nid); in page_pool_nid_changed() [all …]
|
| A D | memory_provider.h | 13 netmem_ref (*alloc_netmems)(struct page_pool *pool, gfp_t gfp); 14 bool (*release_netmem)(struct page_pool *pool, netmem_ref netmem); 15 int (*init)(struct page_pool *pool); 16 void (*destroy)(struct page_pool *pool); 23 void net_mp_niov_set_page_pool(struct page_pool *pool, struct net_iov *niov); 45 static inline void net_mp_netmem_place_in_cache(struct page_pool *pool, in net_mp_netmem_place_in_cache() argument 48 pool->alloc.cache[pool->alloc.count++] = netmem; in net_mp_netmem_place_in_cache()
|
| A D | types.h | 187 void (*disconnect)(void *pool); 254 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp); 255 netmem_ref page_pool_alloc_netmems(struct page_pool *pool, gfp_t gfp); 258 netmem_ref page_pool_alloc_frag_netmem(struct page_pool *pool, 268 void page_pool_enable_direct_recycling(struct page_pool *pool, 270 void page_pool_disable_direct_recycling(struct page_pool *pool); 271 void page_pool_destroy(struct page_pool *pool); 276 static inline void page_pool_destroy(struct page_pool *pool) in page_pool_destroy() argument 280 static inline void page_pool_use_xdp_mem(struct page_pool *pool, in page_pool_use_xdp_mem() argument 294 void page_pool_put_unrefed_page(struct page_pool *pool, struct page *page, [all …]
|
| /include/trace/events/ |
| A D | page_pool.h | 16 TP_PROTO(const struct page_pool *pool, 19 TP_ARGS(pool, inflight, hold, release), 30 __entry->pool = pool; 34 __entry->cnt = pool->destroy_cnt; 47 TP_ARGS(pool, netmem, release), 57 __entry->pool = pool; 73 TP_ARGS(pool, netmem, hold), 83 __entry->pool = pool; 98 TP_ARGS(pool, new_nid), 107 __entry->pool = pool; [all …]
|
| /include/linux/ |
| A D | genalloc.h | 52 void *data, struct gen_pool *pool, 104 return gen_pool_add_owner(pool, addr, phys, size, nid, NULL); in gen_pool_add_virt() 122 return gen_pool_add_virt(pool, addr, -1, size, nid); in gen_pool_add() 131 return gen_pool_alloc_algo_owner(pool, size, pool->algo, pool->data, in gen_pool_alloc_owner() 153 return gen_pool_alloc_algo(pool, size, pool->algo, pool->data); in gen_pool_alloc() 172 gen_pool_free_owner(pool, addr, size, NULL); in gen_pool_free() 185 struct gen_pool *pool, unsigned long start_addr); 189 void *data, struct gen_pool *pool, unsigned long start_addr); 193 void *data, struct gen_pool *pool, unsigned long start_addr); 198 void *data, struct gen_pool *pool, unsigned long start_addr); [all …]
|
| A D | zsmalloc.h | 27 void zs_destroy_pool(struct zs_pool *pool); 29 unsigned long zs_malloc(struct zs_pool *pool, size_t size, gfp_t flags, 31 void zs_free(struct zs_pool *pool, unsigned long obj); 33 size_t zs_huge_class_size(struct zs_pool *pool); 35 unsigned long zs_get_total_pages(struct zs_pool *pool); 36 unsigned long zs_compact(struct zs_pool *pool); 38 unsigned int zs_lookup_class_index(struct zs_pool *pool, unsigned int size); 40 void zs_pool_stats(struct zs_pool *pool, struct zs_pool_stats *stats); 42 void *zs_obj_read_begin(struct zs_pool *pool, unsigned long handle, 44 void zs_obj_read_end(struct zs_pool *pool, unsigned long handle, [all …]
|
| A D | zpool.h | 20 const char *zpool_get_type(struct zpool *pool); 22 void zpool_destroy_pool(struct zpool *pool); 27 void zpool_free(struct zpool *pool, unsigned long handle); 38 u64 zpool_get_total_pages(struct zpool *pool); 64 void (*destroy)(void *pool); 66 int (*malloc)(void *pool, size_t size, gfp_t gfp, 68 void (*free)(void *pool, unsigned long handle); 72 void (*obj_read_end)(void *pool, unsigned long handle, 74 void (*obj_write)(void *pool, unsigned long handle, 77 u64 (*total_pages)(void *pool); [all …]
|
| A D | objpool.h | 128 struct objpool_slot *slot = pool->cpu_slots[cpu]; in __objpool_try_get_slot() 147 if (READ_ONCE(slot->last) - head - 1 >= pool->nr_objs) { in __objpool_try_get_slot() 169 static inline void *objpool_pop(struct objpool_head *pool) in objpool_pop() argument 180 obj = __objpool_try_get_slot(pool, cpu); in objpool_pop() 193 struct objpool_slot *slot = pool->cpu_slots[cpu]; in __objpool_try_add_slot() 202 WARN_ON_ONCE(tail - head > pool->nr_objs); in __objpool_try_add_slot() 221 static inline int objpool_push(void *obj, struct objpool_head *pool) in objpool_push() argument 228 rc = __objpool_try_add_slot(obj, pool, raw_smp_processor_id()); in objpool_push() 251 int objpool_drop(void *obj, struct objpool_head *pool); 257 void objpool_free(struct objpool_head *pool); [all …]
|
| A D | dmapool.h | 25 void dma_pool_destroy(struct dma_pool *pool); 27 void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, 29 void dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t addr); 36 void dmam_pool_destroy(struct dma_pool *pool); 45 static inline void dma_pool_destroy(struct dma_pool *pool) { } in dma_pool_destroy() argument 46 static inline void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, in dma_pool_alloc() argument 48 static inline void dma_pool_free(struct dma_pool *pool, void *vaddr, in dma_pool_free() argument 53 static inline void dmam_pool_destroy(struct dma_pool *pool) { } in dmam_pool_destroy() argument 71 static inline void *dma_pool_zalloc(struct dma_pool *pool, gfp_t mem_flags, in dma_pool_zalloc() argument 74 return dma_pool_alloc(pool, mem_flags | __GFP_ZERO, handle); in dma_pool_zalloc()
|
| A D | mempool.h | 30 static inline bool mempool_initialized(mempool_t *pool) in mempool_initialized() argument 32 return pool->elements != NULL; in mempool_initialized() 35 static inline bool mempool_is_saturated(mempool_t *pool) in mempool_is_saturated() argument 37 return READ_ONCE(pool->curr_nr) >= pool->min_nr; in mempool_is_saturated() 40 void mempool_exit(mempool_t *pool); 41 int mempool_init_node(mempool_t *pool, int min_nr, mempool_alloc_t *alloc_fn, 63 extern int mempool_resize(mempool_t *pool, int new_min_nr); 64 extern void mempool_destroy(mempool_t *pool); 66 extern void *mempool_alloc_noprof(mempool_t *pool, gfp_t gfp_mask) __malloc; 70 extern void *mempool_alloc_preallocated(mempool_t *pool) __malloc; [all …]
|
| A D | swiotlb.h | 247 unsigned long attrs, struct io_tlb_pool *pool); 252 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_tbl_unmap_single() local 254 if (unlikely(pool)) in swiotlb_tbl_unmap_single() 260 struct io_tlb_pool *pool); 264 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_sync_single_for_device() local 266 if (unlikely(pool)) in swiotlb_sync_single_for_device() 267 __swiotlb_sync_single_for_device(dev, addr, size, dir, pool); in swiotlb_sync_single_for_device() 272 struct io_tlb_pool *pool); 276 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_sync_single_for_cpu() local 278 if (unlikely(pool)) in swiotlb_sync_single_for_cpu() [all …]
|
| A D | cgroup_dmem.h | 23 void dmem_cgroup_uncharge(struct dmem_cgroup_pool_state *pool, u64 size); 28 void dmem_cgroup_pool_state_put(struct dmem_cgroup_pool_state *pool); 51 static inline void dmem_cgroup_uncharge(struct dmem_cgroup_pool_state *pool, u64 size) in dmem_cgroup_uncharge() argument 62 static inline void dmem_cgroup_pool_state_put(struct dmem_cgroup_pool_state *pool) in dmem_cgroup_pool_state_put() argument
|
| A D | tee_core.h | 62 struct tee_shm_pool *pool; member 133 struct tee_shm_pool *pool, 202 int (*alloc)(struct tee_shm_pool *pool, struct tee_shm *shm, 204 void (*free)(struct tee_shm_pool *pool, struct tee_shm *shm); 205 void (*destroy_pool)(struct tee_shm_pool *pool); 227 static inline void tee_shm_pool_free(struct tee_shm_pool *pool) in tee_shm_pool_free() argument 229 pool->ops->destroy_pool(pool); in tee_shm_pool_free()
|
| A D | sram.h | 9 void *sram_exec_copy(struct gen_pool *pool, void *dst, void *src, size_t size); 11 static inline void *sram_exec_copy(struct gen_pool *pool, void *dst, void *src, in sram_exec_copy() argument
|
| A D | rethook.h | 37 struct objpool_head pool; member
|
| /include/drm/ttm/ |
| A D | ttm_pool.h | 52 struct ttm_pool *pool; member 83 int ttm_pool_alloc(struct ttm_pool *pool, struct ttm_tt *tt, 85 void ttm_pool_free(struct ttm_pool *pool, struct ttm_tt *tt); 87 void ttm_pool_init(struct ttm_pool *pool, struct device *dev, 89 void ttm_pool_fini(struct ttm_pool *pool); 91 int ttm_pool_debugfs(struct ttm_pool *pool, struct seq_file *m); 95 long ttm_pool_backup(struct ttm_pool *pool, struct ttm_tt *ttm, 97 int ttm_pool_restore_and_alloc(struct ttm_pool *pool, struct ttm_tt *tt,
|
| /include/linux/ceph/ |
| A D | msgpool.h | 13 mempool_t *pool; member 19 int ceph_msgpool_init(struct ceph_msgpool *pool, int type, 22 extern void ceph_msgpool_destroy(struct ceph_msgpool *pool); 23 struct ceph_msg *ceph_msgpool_get(struct ceph_msgpool *pool, int front_len,
|
| A D | osdmap.h | 23 uint64_t pool; member 63 static inline bool ceph_can_shift_osds(struct ceph_pg_pool_info *pool) in ceph_can_shift_osds() argument 65 switch (pool->type) { in ceph_can_shift_osds() 76 s64 pool; member 82 oloc->pool = -1; in ceph_oloc_init() 88 return oloc->pool == -1; in ceph_oloc_empty() 246 pgid->pool = ceph_decode_64(p); in ceph_decode_pgid()
|
| /include/soc/fsl/ |
| A D | bman.h | 94 void bman_free_pool(struct bman_pool *pool); 103 int bman_get_bpid(const struct bman_pool *pool); 114 int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num); 127 int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num);
|
| /include/linux/firmware/qcom/ |
| A D | qcom_tzmem.h | 44 void qcom_tzmem_pool_free(struct qcom_tzmem_pool *pool); 49 void *qcom_tzmem_alloc(struct qcom_tzmem_pool *pool, size_t size, gfp_t gfp);
|
| /include/net/libeth/ |
| A D | xsk.h | 122 xsk_buff_raw_dma_sync_for_device(sq->pool, desc.addr, desc.len); in libeth_xsk_tx_fill_buf() 199 ctx = xsk_buff_raw_get_ctx(sq->pool, xdesc->addr); in __libeth_xsk_xmit_fill_buf_md() 228 .addr = xsk_buff_raw_get_dma(sq->pool, xdesc->addr), in __libeth_xsk_xmit_fill_buf() 259 xsk_buff_raw_dma_sync_for_device(sq->pool, desc.addr, desc.len); in libeth_xsk_xmit_fill_buf() 296 wake = xsk_uses_need_wakeup(pool); in libeth_xsk_xmit_do_bulk() 298 xsk_clear_tx_need_wakeup(pool); in libeth_xsk_xmit_do_bulk() 300 n = xsk_tx_peek_release_desc_batch(pool, budget); in libeth_xsk_xmit_do_bulk() 301 bulk = container_of(&pool->tx_descs[0], typeof(*bulk), desc); in libeth_xsk_xmit_do_bulk() 309 xsk_set_tx_need_wakeup(pool); in libeth_xsk_xmit_do_bulk() 604 struct xsk_buff_pool *pool; [all …]
|
| /include/uapi/linux/netfilter_bridge/ |
| A D | ebt_among.h | 43 struct ebt_mac_wormhash_tuple pool[]; member
|