Home
last modified time | relevance | path

Searched refs:cq_table (Results 1 – 20 of 20) sorted by relevance

/drivers/infiniband/hw/hns/
A Dhns_roce_cq.c60 struct hns_roce_cq_table *cq_table = &hr_dev->cq_table; in alloc_cqn() local
65 mutex_lock(&cq_table->bank_mutex); in alloc_cqn()
67 bank = &cq_table->bank[bankid]; in alloc_cqn()
78 mutex_unlock(&cq_table->bank_mutex); in alloc_cqn()
91 struct hns_roce_cq_table *cq_table = &hr_dev->cq_table; in free_cqn() local
98 mutex_lock(&cq_table->bank_mutex); in free_cqn()
133 struct hns_roce_cq_table *cq_table = &hr_dev->cq_table; in alloc_cqc() local
175 struct hns_roce_cq_table *cq_table = &hr_dev->cq_table; in free_cqc() local
488 xa_lock(&hr_dev->cq_table.array); in hns_roce_cq_event()
513 struct hns_roce_cq_table *cq_table = &hr_dev->cq_table; in hns_roce_init_cq_table() local
[all …]
A Dhns_roce_main.c829 ret = hns_roce_init_hem_table(hr_dev, &hr_dev->cq_table.table, in hns_roce_init_hem()
918 hns_roce_cleanup_hem_table(hr_dev, &hr_dev->cq_table.table); in hns_roce_init_hem()
A Dhns_roce_device.h993 struct hns_roce_cq_table cq_table; member
A Dhns_roce_hem.c896 hns_roce_cleanup_hem_table(hr_dev, &hr_dev->cq_table.table); in hns_roce_cleanup_hem()
/drivers/net/ethernet/mellanox/mlx4/
A Dcq.c128 struct mlx4_cq_table *cq_table = &mlx4_priv(dev)->cq_table; in mlx4_cq_event() local
220 struct mlx4_cq_table *cq_table = &priv->cq_table; in __mlx4_cq_alloc_icm() local
267 struct mlx4_cq_table *cq_table = &priv->cq_table; in __mlx4_cq_free_icm() local
349 struct mlx4_cq_table *cq_table = &priv->cq_table; in mlx4_cq_alloc() local
364 spin_lock(&cq_table->lock); in mlx4_cq_alloc()
366 spin_unlock(&cq_table->lock); in mlx4_cq_alloc()
425 spin_lock(&cq_table->lock); in mlx4_cq_alloc()
427 spin_unlock(&cq_table->lock); in mlx4_cq_alloc()
439 struct mlx4_cq_table *cq_table = &priv->cq_table; in mlx4_cq_free() local
446 spin_lock(&cq_table->lock); in mlx4_cq_free()
[all …]
A Dmain.c1624 err = mlx4_init_icm_table(dev, &priv->cq_table.cmpt_table, in mlx4_init_cmpt_table()
1645 mlx4_cleanup_icm_table(dev, &priv->cq_table.cmpt_table); in mlx4_init_cmpt_table()
1779 err = mlx4_init_icm_table(dev, &priv->cq_table.table, in mlx4_init_icm()
1823 mlx4_cleanup_icm_table(dev, &priv->cq_table.table); in mlx4_init_icm()
1848 mlx4_cleanup_icm_table(dev, &priv->cq_table.cmpt_table); in mlx4_init_icm()
1867 mlx4_cleanup_icm_table(dev, &priv->cq_table.table); in mlx4_free_icms()
1876 mlx4_cleanup_icm_table(dev, &priv->cq_table.cmpt_table); in mlx4_free_icms()
A Dmlx4.h903 struct mlx4_cq_table cq_table; member
/drivers/infiniband/hw/mthca/
A Dmthca_cq.c239 spin_lock(&dev->cq_table.lock); in mthca_cq_event()
245 spin_unlock(&dev->cq_table.lock); in mthca_cq_event()
258 spin_lock(&dev->cq_table.lock); in mthca_cq_event()
261 spin_unlock(&dev->cq_table.lock); in mthca_cq_event()
850 spin_lock_irq(&dev->cq_table.lock); in mthca_init_cq()
857 spin_unlock_irq(&dev->cq_table.lock); in mthca_init_cq()
893 spin_lock_irq(&dev->cq_table.lock); in get_cq_refcount()
895 spin_unlock_irq(&dev->cq_table.lock); in get_cq_refcount()
927 spin_lock_irq(&dev->cq_table.lock); in mthca_free_cq()
928 mthca_array_clear(&dev->cq_table.cq, in mthca_free_cq()
[all …]
A Dmthca_main.c476 mdev->cq_table.table = mthca_alloc_icm_table(mdev, init_hca->cqc_base, in mthca_init_icm()
481 if (!mdev->cq_table.table) { in mthca_init_icm()
527 mthca_free_icm_table(mdev, mdev->cq_table.table); in mthca_init_icm()
562 mthca_free_icm_table(mdev, mdev->cq_table.table); in mthca_free_icms()
A Dmthca_dev.h342 struct mthca_cq_table cq_table; member
/drivers/infiniband/hw/mana/
A Dcq.c144 WARN_ON(gc->cq_table[cq->queue.id]); in mana_ib_install_cq_cb()
156 gc->cq_table[cq->queue.id] = gdma_cq; in mana_ib_install_cq_cb()
171 kfree(gc->cq_table[cq->queue.id]); in mana_ib_remove_cq_cb()
172 gc->cq_table[cq->queue.id] = NULL; in mana_ib_remove_cq_cb()
/drivers/net/ethernet/mellanox/mlx5/core/
A Deq.c96 struct mlx5_cq_table *table = &eq->cq_table; in mlx5_eq_cq_get()
260 struct mlx5_cq_table *cq_table = &eq->cq_table; in create_map_eq() local
273 memset(cq_table, 0, sizeof(*cq_table)); in create_map_eq()
274 spin_lock_init(&cq_table->lock); in create_map_eq()
275 INIT_RADIX_TREE(&cq_table->tree, GFP_ATOMIC); in create_map_eq()
400 struct mlx5_cq_table *table = &eq->cq_table; in mlx5_eq_add_cq()
412 struct mlx5_cq_table *table = &eq->cq_table; in mlx5_eq_del_cq()
/drivers/net/ethernet/microsoft/mana/
A Dhw_channel.c680 gc->cq_table = vcalloc(gc->max_num_cqs, sizeof(struct gdma_queue *)); in mana_hwc_establish_channel()
681 if (!gc->cq_table) in mana_hwc_establish_channel()
684 gc->cq_table[cq->id] = cq; in mana_hwc_establish_channel()
830 vfree(gc->cq_table); in mana_hwc_destroy_channel()
831 gc->cq_table = NULL; in mana_hwc_destroy_channel()
A Dgdma_main.c512 cq = gc->cq_table[cq_id]; in mana_gd_process_eqe()
827 if (!gc->cq_table[id]) in mana_gd_destroy_cq()
830 gc->cq_table[id] = NULL; in mana_gd_destroy_cq()
A Dmana_en.c2254 gc->cq_table[cq->gdma_id] = cq->gdma_cq; in mana_create_txq()
2550 gc->cq_table[cq->gdma_id] = cq->gdma_cq; in mana_create_rxq()
/drivers/net/ethernet/mellanox/mlx5/core/lib/
A Deq.h28 struct mlx5_cq_table cq_table; member
/drivers/infiniband/hw/irdma/
A Dmain.h317 struct irdma_cq **cq_table; member
A Dhw.c344 iwcq = rf->cq_table[info->qp_cq_id]; in irdma_process_aeq()
1970 rf->cq_table = (struct irdma_cq **)(&rf->qp_table[rf->max_qp]); in irdma_set_hw_rsrc()
A Dutils.c780 iwdev->rf->cq_table[iwcq->cq_num] = NULL; in irdma_cq_rem_ref()
A Dverbs.c2228 rf->cq_table[cq_num] = iwcq; in irdma_create_cq()

Completed in 735 milliseconds