| /linux/tools/testing/selftests/drivers/net/mlxsw/ |
| A D | sharedbuffer_configuration.py | 28 def __init__(self, pools): argument 30 for pool in pools: 187 pools = PoolList() 191 return pools 194 def do_check_pools(dlname, pools, vp): argument 195 for pool in pools: 219 def check_pools(dlname, pools): argument 224 do_check_pools(dlname, pools, RandomValuePicker(pools)) 409 pools = get_pools(dlname) 411 check_pools(dlname, pools) [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
| A D | pool.c | 25 if (!xsk->pools) { in mlx5e_xsk_get_pools() 26 xsk->pools = kcalloc(MLX5E_MAX_NUM_CHANNELS, in mlx5e_xsk_get_pools() 27 sizeof(*xsk->pools), GFP_KERNEL); in mlx5e_xsk_get_pools() 28 if (unlikely(!xsk->pools)) in mlx5e_xsk_get_pools() 41 kfree(xsk->pools); in mlx5e_xsk_put_pools() 42 xsk->pools = NULL; in mlx5e_xsk_put_pools() 54 xsk->pools[ix] = pool; in mlx5e_xsk_add_pool() 60 xsk->pools[ix] = NULL; in mlx5e_xsk_remove_pool()
|
| A D | pool.h | 12 if (!xsk || !xsk->pools) in mlx5e_xsk_get_pool() 18 return xsk->pools[ix]; in mlx5e_xsk_get_pool()
|
| /linux/arch/sparc/kernel/ |
| A D | iommu-common.c | 82 spin_lock_init(&(iommu->pools[i].lock)); in iommu_tbl_pool_init() 83 iommu->pools[i].start = start; in iommu_tbl_pool_init() 84 iommu->pools[i].hint = start; in iommu_tbl_pool_init() 86 iommu->pools[i].end = start - 1; in iommu_tbl_pool_init() 131 pool = &(iommu->pools[pool_nr]); in iommu_tbl_range_alloc() 161 pool = &(iommu->pools[0]); in iommu_tbl_range_alloc() 193 pool = &(iommu->pools[pool_nr]); in iommu_tbl_range_alloc() 237 p = &tbl->pools[pool_nr]; in get_pool()
|
| /linux/mm/ |
| A D | dmapool.c | 51 struct list_head pools; member 81 list_for_each_entry(pool, &dev->dma_pools, pools) { in pools_show() 105 static DEVICE_ATTR_RO(pools); 169 INIT_LIST_HEAD(&retval->pools); in dma_pool_create() 183 list_add(&retval->pools, &dev->dma_pools); in dma_pool_create() 191 list_del(&retval->pools); in dma_pool_create() 277 list_del(&pool->pools); in dma_pool_destroy()
|
| /linux/drivers/net/ethernet/chelsio/libcxgb/ |
| A D | libcxgb_ppm.c | 348 struct cxgbi_ppm_pool *pools; in ppm_alloc_cpu_pool() local 350 unsigned int max = (PCPU_MIN_UNIT_SIZE - sizeof(*pools)) << 3; in ppm_alloc_cpu_pool() 367 alloc_sz = sizeof(*pools) + sizeof(unsigned long) * bmap; in ppm_alloc_cpu_pool() 368 pools = __alloc_percpu(alloc_sz, __alignof__(struct cxgbi_ppm_pool)); in ppm_alloc_cpu_pool() 370 if (!pools) in ppm_alloc_cpu_pool() 374 struct cxgbi_ppm_pool *ppool = per_cpu_ptr(pools, cpu); in ppm_alloc_cpu_pool() 384 return pools; in ppm_alloc_cpu_pool()
|
| /linux/drivers/atm/ |
| A D | zatm.c | 1038 unsigned long pools; in zatm_int() local 1041 pools = zin(RQA); in zatm_int() 1044 if (pools & 1) { in zatm_int() 1048 pools >>= 1; in zatm_int() 1052 unsigned long pools; in zatm_int() local 1054 pools = zin(RQU); in zatm_int() 1056 dev->number,pools); in zatm_int() 1059 if (pools & 1) { in zatm_int() 1063 pools >>= 1; in zatm_int() 1268 int pools,vccs,rx; in zatm_start() local [all …]
|
| /linux/Documentation/core-api/ |
| A D | workqueue.rst | 60 * Use per-CPU unified worker pools shared by all wq to provide 83 called worker-pools. 87 which manages worker-pools and processes the queued work items. 89 There are two worker-pools, one for normal work items and the other 91 worker-pools to serve work items queued on unbound workqueues - the 92 number of these backing pools is dynamic. 132 For unbound workqueues, the number of backing pools is dynamic. 169 worker-pools which host workers which are not bound to any 172 worker-pools try to start execution of work items as soon as 196 worker-pool of the target cpu. Highpri worker-pools are [all …]
|
| A D | mm-api.rst | 76 Memory pools 82 DMA pools
|
| /linux/arch/powerpc/kernel/ |
| A D | iommu.c | 243 pool = &(tbl->pools[pool_nr]); in iommu_range_alloc() 271 pool = &(tbl->pools[0]); in iommu_range_alloc() 293 pool = &tbl->pools[pool_nr]; in iommu_range_alloc() 420 p = &tbl->pools[pool_nr]; in get_pool() 746 p = &tbl->pools[i]; in iommu_init_table() 1104 spin_lock_nest_lock(&tbl->pools[i].lock, &tbl->large_pool.lock); in iommu_take_ownership() 1114 spin_unlock(&tbl->pools[i].lock); in iommu_take_ownership() 1127 spin_lock_nest_lock(&tbl->pools[i].lock, &tbl->large_pool.lock); in iommu_release_ownership() 1135 spin_unlock(&tbl->pools[i].lock); in iommu_release_ownership()
|
| /linux/drivers/soc/ti/ |
| A D | knav_qmss.h | 203 struct list_head pools; member 304 struct list_head pools; member 363 list_for_each_entry(pool, &kdev->pools, list)
|
| A D | knav_qmss_queue.c | 820 node = ®ion->pools; in knav_pool_create() 821 list_for_each_entry(pi, ®ion->pools, region_inst) { in knav_pool_create() 835 list_add_tail(&pool->list, &kdev->pools); in knav_pool_create() 1037 list_add(&pool->region_inst, ®ion->pools); in knav_queue_setup_region() 1121 INIT_LIST_HEAD(®ion->pools); in knav_queue_setup_regions() 1359 list_for_each_entry_safe(pool, tmp, ®ion->pools, region_inst) in knav_queue_free_regions() 1783 INIT_LIST_HEAD(&kdev->pools); in knav_queue_probe()
|
| /linux/Documentation/devicetree/bindings/soc/ti/ |
| A D | keystone-navigator-qmss.txt | 6 processors(PDSP), linking RAM, descriptor pools and infrastructure 49 - queue-pools : child node classifying the queue ranges into pools. 50 Queue ranges are grouped into 3 type of pools: 151 queue-pools {
|
| /linux/drivers/md/ |
| A D | dm.c | 2804 struct dm_md_mempools *pools = kzalloc_node(sizeof(*pools), GFP_KERNEL, md->numa_node_id); in dm_alloc_md_mempools() local 2809 if (!pools) in dm_alloc_md_mempools() 2818 ret = bioset_init(&pools->io_bs, pool_size, io_front_pad, 0); in dm_alloc_md_mempools() 2833 ret = bioset_init(&pools->bs, pool_size, front_pad, 0); in dm_alloc_md_mempools() 2840 return pools; in dm_alloc_md_mempools() 2843 dm_free_md_mempools(pools); in dm_alloc_md_mempools() 2848 void dm_free_md_mempools(struct dm_md_mempools *pools) in dm_free_md_mempools() argument 2850 if (!pools) in dm_free_md_mempools() 2853 bioset_exit(&pools->bs); in dm_free_md_mempools() 2854 bioset_exit(&pools->io_bs); in dm_free_md_mempools() [all …]
|
| A D | dm.h | 230 void dm_free_md_mempools(struct dm_md_mempools *pools);
|
| /linux/Documentation/networking/device_drivers/ethernet/freescale/dpaa2/ |
| A D | ethernet-driver.rst | 26 - buffer pools 69 DPBPs represent hardware buffer pools. Packet I/O is performed in the context 124 The role of hardware buffer pools is storage of ingress frame data. Each network
|
| A D | overview.rst | 25 The MC uses DPAA2 hardware resources such as queues, buffer pools, and 56 | -buffer pools -DPMCP | 360 - DPBPs for network buffer pools
|
| /linux/arch/sparc/include/asm/ |
| A D | iommu-common.h | 26 struct iommu_pool pools[IOMMU_NR_POOLS]; member
|
| /linux/Documentation/arm/keystone/ |
| A D | knav-qmss.rst | 12 processors(PDSP), linking RAM, descriptor pools and infrastructure 25 allocate descriptor pools, map the descriptors, push/pop to queues etc. For
|
| /linux/Documentation/devicetree/bindings/misc/ |
| A D | fsl,qoriq-mc.txt | 6 block is enabled, pools of hardware resources are available, such as 7 queues, buffer pools, I/O interfaces. These resources are building
|
| /linux/drivers/soc/fsl/qbman/ |
| A D | qman_priv.h | 177 u32 pools; member
|
| A D | qman_portal.c | 248 pcfg->pools = qm_get_pools_sdqcr(); in qman_portal_probe()
|
| /linux/arch/arm/boot/dts/ |
| A D | keystone-k2g-netcp.dtsi | 36 queue-pools {
|
| A D | keystone-k2l-netcp.dtsi | 35 queue-pools {
|
| /linux/arch/powerpc/include/asm/ |
| A D | iommu.h | 107 struct iommu_pool pools[IOMMU_NR_POOLS]; member
|