| /drivers/net/ethernet/mellanox/mlxsw/ |
| A D | spectrum_cnt.c | 22 u64 pool_size; member 139 &pool->pool_size); in mlxsw_sp_counter_pool_init() 145 pool->usage = bitmap_zalloc(pool->pool_size, GFP_KERNEL); in mlxsw_sp_counter_pool_init() 173 WARN_ON(find_first_bit(pool->usage, pool->pool_size) != in mlxsw_sp_counter_pool_fini() 174 pool->pool_size); in mlxsw_sp_counter_pool_fini() 231 if (WARN_ON(counter_index >= pool->pool_size)) in mlxsw_sp_counter_free() 250 u64 pool_size; in mlxsw_sp_counter_resources_register() local 263 pool_size, bank_size, in mlxsw_sp_counter_resources_register() 267 pool_size, in mlxsw_sp_counter_resources_register() 283 if (base_index + sub_pool_size > pool_size) in mlxsw_sp_counter_resources_register() [all …]
|
| /drivers/net/ethernet/mscc/ |
| A D | ocelot_devlink.c | 359 buf_shr_i = ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_ING] - in ocelot_setup_sharing_watermarks() 361 buf_shr_e = ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_EGR] - in ocelot_setup_sharing_watermarks() 363 ref_shr_i = ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_ING] - in ocelot_setup_sharing_watermarks() 365 ref_shr_e = ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_EGR] - in ocelot_setup_sharing_watermarks() 391 if (buf_rsrv_i > ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_ING]) { in ocelot_watermark_validate() 396 if (buf_rsrv_e > ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_EGR]) { in ocelot_watermark_validate() 401 if (ref_rsrv_i > ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_ING]) { in ocelot_watermark_validate() 546 pool_info->size = ocelot->pool_size[sb_index][pool_index]; in ocelot_sb_pool_get() 585 old_pool_size = ocelot->pool_size[sb_index][pool_index]; in ocelot_sb_pool_set() 586 ocelot->pool_size[sb_index][pool_index] = size; in ocelot_sb_pool_set() [all …]
|
| /drivers/md/dm-vdo/ |
| A D | vio.c | 321 int make_vio_pool(struct vdo *vdo, size_t pool_size, size_t block_count, thread_id_t thread_id, in make_vio_pool() argument 330 result = vdo_allocate_extended(struct vio_pool, pool_size, struct pooled_vio, in make_vio_pool() 339 result = vdo_allocate(pool_size * per_vio_size, char, in make_vio_pool() 347 for (pool->size = 0; pool->size < pool_size; pool->size++, ptr += per_vio_size) { in make_vio_pool()
|
| A D | data-vio.c | 842 int make_data_vio_pool(struct vdo *vdo, data_vio_count_t pool_size, in make_data_vio_pool() argument 849 result = vdo_allocate_extended(struct data_vio_pool, pool_size, struct data_vio, in make_data_vio_pool() 854 VDO_ASSERT_LOG_ONLY((discard_limit <= pool_size), in make_data_vio_pool() 859 initialize_limiter(&pool->limiter, pool, assign_data_vio_to_waiter, pool_size); in make_data_vio_pool() 875 for (i = 0; i < pool_size; i++) { in make_data_vio_pool()
|
| A D | vio.h | 195 int __must_check make_vio_pool(struct vdo *vdo, size_t pool_size, size_t block_count,
|
| A D | data-vio.h | 331 int make_data_vio_pool(struct vdo *vdo, data_vio_count_t pool_size,
|
| /drivers/mtd/ubi/ |
| A D | fastmap.c | 401 __be32 *pebs, int pool_size, unsigned long long *max_sqnum, in scan_pool() argument 422 dbg_bld("scanning fastmap pool: size = %i", pool_size); in scan_pool() 428 for (i = 0; i < pool_size; i++) { in scan_pool() 561 int ret, i, j, pool_size, wl_pool_size; in ubi_attach_fastmap() local 607 pool_size = be16_to_cpu(fmpl->size); in ubi_attach_fastmap() 612 if (pool_size > UBI_FM_MAX_POOL_SIZE || pool_size < 0) { in ubi_attach_fastmap() 613 ubi_err(ubi, "bad pool size: %i", pool_size); in ubi_attach_fastmap() 765 ret = scan_pool(ubi, ai, fmpl->pebs, pool_size, &max_sqnum, &free); in ubi_attach_fastmap()
|
| /drivers/net/ethernet/intel/i40e/ |
| A D | i40e_dcb.c | 1567 u32 pool_size[I40E_MAX_TRAFFIC_CLASS]; in i40e_dcb_hw_calculate_pool_sizes() local 1599 pool_size[i] = high_wm[i]; in i40e_dcb_hw_calculate_pool_sizes() 1600 pool_size[i] += I40E_BT2B(I40E_STD_DV_TC(mfs_max, in i40e_dcb_hw_calculate_pool_sizes() 1604 pool_size[i] = (I40E_DCB_WATERMARK_START_FACTOR * in i40e_dcb_hw_calculate_pool_sizes() 1606 high_wm[i] = pool_size[i]; in i40e_dcb_hw_calculate_pool_sizes() 1608 total_pool_size += pool_size[i]; in i40e_dcb_hw_calculate_pool_sizes() 1619 pb_cfg->tc_pool_size[i] = pool_size[i]; in i40e_dcb_hw_calculate_pool_sizes()
|
| /drivers/dma/ppc4xx/ |
| A D | adma.h | 67 size_t pool_size; member
|
| A D | adma.c | 1783 for (; i < (ppc440spe_chan->device->pool_size / db_sz); i++) { in ppc440spe_adma_alloc_chan_resources() 4015 u32 id, pool_size; in ppc440spe_adma_probe() local 4023 pool_size = PAGE_SIZE << 1; in ppc440spe_adma_probe() 4045 pool_size = (id == PPC440SPE_DMA0_ID) ? in ppc440spe_adma_probe() 4047 pool_size <<= 2; in ppc440spe_adma_probe() 4075 adev->pool_size = pool_size; in ppc440spe_adma_probe() 4078 adev->pool_size, &adev->dma_desc_pool, in ppc440spe_adma_probe() 4083 adev->pool_size); in ppc440spe_adma_probe() 4216 dma_free_coherent(adev->dev, adev->pool_size, in ppc440spe_adma_probe() 4272 dma_free_coherent(adev->dev, adev->pool_size, in ppc440spe_adma_remove()
|
| /drivers/misc/ |
| A D | ibmvmc.h | 111 __be16 pool_size; /* Maximum number of buffers supported per HMC member
|
| /drivers/dma/ |
| A D | mv_xor.h | 117 size_t pool_size; member
|
| /drivers/net/ethernet/ibm/ |
| A D | ibmveth.h | 125 static int pool_size[] = { 512, 1024 * 2, 1024 * 16, 1024 * 32, 1024 * 64 }; variable
|
| A D | ibmveth.c | 152 u32 pool_index, u32 pool_size, in ibmveth_init_buffer_pool() argument 155 pool->size = pool_size; in ibmveth_init_buffer_pool() 158 pool->threshold = pool_size * 7 / 8; in ibmveth_init_buffer_pool() 1883 pool_count[i], pool_size[i], in ibmveth_probe() 2195 pool_count[i], pool_size[i], in ibmveth_remove_buffer_from_pool_test() 2248 pool_count[i], pool_size[i], in ibmveth_rxq_get_buffer_test()
|
| A D | ibmvnic.c | 1064 u64 pool_size; /* # of buffers in one pool */ in init_rx_pools() local 1098 i, pool_size, buff_size); in init_rx_pools() 1100 rx_pool->size = pool_size; in init_rx_pools() 1122 adapter->prev_rx_pool_size = pool_size; in init_rx_pools() 1223 int pool_size, int buf_size) in init_one_tx_pool() argument 1227 tx_pool->tx_buff = kcalloc(pool_size, in init_one_tx_pool() 1240 for (i = 0; i < pool_size; i++) in init_one_tx_pool() 1245 tx_pool->num_buffers = pool_size; in init_one_tx_pool() 1304 u64 pool_size; /* # of buffers in pool */ in init_tx_pools() local 1354 pool_size, buff_size); in init_tx_pools() [all …]
|
| /drivers/net/ethernet/intel/libeth/ |
| A D | rx.c | 157 .pool_size = fq->count, in libeth_rx_fq_create()
|
| /drivers/md/ |
| A D | dm-table.c | 1057 unsigned int min_pool_size = 0, pool_size; in dm_table_alloc_md_mempools() local 1071 pool_size = dm_get_reserved_rq_based_ios(); in dm_table_alloc_md_mempools() 1085 pool_size = max(dm_get_reserved_bio_based_ios(), min_pool_size); in dm_table_alloc_md_mempools() 1091 if (bioset_init(&pools->io_bs, pool_size, io_front_pad, bioset_flags)) in dm_table_alloc_md_mempools() 1094 if (bioset_init(&pools->bs, pool_size, front_pad, 0)) in dm_table_alloc_md_mempools()
|
| A D | raid5.h | 670 int pool_size; /* number of disks in stripeheads in pool */ member
|
| A D | raid5.c | 524 int num = sh->raid_conf->pool_size; in shrink_buffers() 547 int num = sh->raid_conf->pool_size; in grow_buffers() 2422 conf->pool_size = devs; in grow_stripes() 2588 for(i=0; i<conf->pool_size; i++) { in resize_stripes() 2611 for (i = 0; i < conf->pool_size; i++) in resize_stripes() 2614 for (i = conf->pool_size; i < newsize; i++) { in resize_stripes() 2621 for (i = conf->pool_size; i < newsize; i++) in resize_stripes() 2672 conf->pool_size = newsize; in resize_stripes() 7348 for (i = 0; i < conf->pool_size; i++) in free_conf() 8167 if (unlikely(number >= conf->pool_size)) in raid5_remove_disk() [all …]
|
| /drivers/net/ethernet/ti/ |
| A D | cpsw_priv.c | 1135 pp_params.pool_size = size; in cpsw_create_page_pool() 1150 int ret = 0, pool_size; in cpsw_create_rx_pool() local 1152 pool_size = cpdma_chan_get_rx_buf_num(cpsw->rxv[ch].ch); in cpsw_create_rx_pool() 1153 pool = cpsw_create_page_pool(cpsw, pool_size); in cpsw_create_rx_pool()
|
| /drivers/net/ethernet/google/gve/ |
| A D | gve_buffer_mgmt_dqo.c | 253 .pool_size = GVE_PAGE_POOL_SIZE_MULTIPLIER * priv->rx_desc_cnt, in gve_rx_create_page_pool()
|
| /drivers/gpu/drm/amd/pm/swsmu/ |
| A D | amdgpu_smu.c | 1072 uint64_t pool_size = smu->pool_size; in smu_alloc_memory_pool() local 1075 if (pool_size == SMU_MEMORY_POOL_SIZE_ZERO) in smu_alloc_memory_pool() 1078 memory_pool->size = pool_size; in smu_alloc_memory_pool() 1085 switch (pool_size) { in smu_alloc_memory_pool() 1324 smu->pool_size = adev->pm.smu_prv_buffer_size; in smu_sw_init()
|
| /drivers/net/ethernet/microchip/sparx5/lan969x/ |
| A D | lan969x_fdma.c | 159 .pool_size = fdma->n_dcbs * fdma->n_dbs, in lan969x_fdma_rx_alloc()
|
| /drivers/infiniband/hw/efa/ |
| A D | efa_com.c | 357 size_t pool_size = aq->depth * sizeof(*aq->comp_ctx_pool); in efa_com_init_comp_ctxt() local 363 aq->comp_ctx_pool = devm_kzalloc(aq->dmadev, pool_size, GFP_KERNEL); in efa_com_init_comp_ctxt()
|
| /drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | en_main.c | 791 u32 *pool_size, in mlx5_rq_shampo_alloc() argument 831 pp_params.pool_size = hd_pool_size; in mlx5_rq_shampo_alloc() 847 *pool_size += hd_pool_size; in mlx5_rq_shampo_alloc() 893 u32 pool_size; in mlx5e_alloc_rq() local 908 pool_size = 1 << params->log_rq_mtu_frames; in mlx5e_alloc_rq() 939 pool_size = rq->mpwqe.pages_per_wqe << in mlx5e_alloc_rq() 943 pool_size *= 2; /* additional page per packet for the linear part */ in mlx5e_alloc_rq() 960 err = mlx5_rq_shampo_alloc(mdev, params, rqp, rq, &pool_size, node); in mlx5e_alloc_rq() 995 pp_params.pool_size = pool_size; in mlx5e_alloc_rq()
|