Home
last modified time | relevance | path

Searched refs:mcq (Results 1 – 25 of 44) sorted by relevance

12

/drivers/infiniband/hw/mlx4/
A Dcq.c98 return mlx4_cq_modify(dev->dev, &mcq->mcq, cq_count, cq_period); in mlx4_ib_modify_cq()
236 *cq->mcq.set_ci_db = 0; in mlx4_ib_create_cq()
237 *cq->mcq.arm_db = 0; in mlx4_ib_create_cq()
253 &cq->mcq, vector, 0, in mlx4_ib_create_cq()
352 i = cq->mcq.cons_index; in mlx4_ib_get_outstanding_cqes()
366 i = cq->mcq.cons_index; in mlx4_ib_cq_resize_copy_cqes()
381 ++cq->mcq.cons_index; in mlx4_ib_cq_resize_copy_cqes()
487 mlx4_cq_free(dev->dev, &mcq->mcq); in mlx4_ib_destroy_cq()
496 &mcq->db); in mlx4_ib_destroy_cq()
685 ++cq->mcq.cons_index; in mlx4_ib_poll_one()
[all …]
A Dqp.c869 struct mlx4_ib_cq *mcq; in create_rq() local
973 mcq = to_mcq(init_attr->send_cq); in create_rq()
975 mcq = to_mcq(init_attr->recv_cq); in create_rq()
1005 struct mlx4_ib_cq *mcq; in create_qp_common() local
1251 mcq = to_mcq(init_attr->send_cq); in create_qp_common()
1253 mcq = to_mcq(init_attr->recv_cq); in create_qp_common()
1316 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx4_ib_lock_cqs()
1331 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx4_ib_unlock_cqs()
4424 if (!mcq->mcq.reset_notify_added) in handle_drain_completion()
4425 mcq->mcq.reset_notify_added = 1; in handle_drain_completion()
[all …]
A Dmlx4_ib.h118 struct mlx4_cq mcq; member
698 static inline struct mlx4_ib_cq *to_mibcq(struct mlx4_cq *mcq) in to_mibcq() argument
700 return container_of(mcq, struct mlx4_ib_cq, mcq); in to_mibcq()
/drivers/net/ethernet/mellanox/mlx4/
A Den_cq.c98 cq->mcq.set_ci_db = cq->wqres.db.db; in mlx4_en_activate_cq()
99 cq->mcq.arm_db = cq->wqres.db.db + 1; in mlx4_en_activate_cq()
100 *cq->mcq.set_ci_db = 0; in mlx4_en_activate_cq()
101 *cq->mcq.arm_db = 0; in mlx4_en_activate_cq()
139 cq->mcq.usage = MLX4_RES_USAGE_DRIVER; in mlx4_en_activate_cq()
141 &mdev->priv_uar, cq->wqres.db.dma, &cq->mcq, in mlx4_en_activate_cq()
147 cq->mcq.event = mlx4_en_cq_event; in mlx4_en_activate_cq()
151 cq->mcq.comp = mlx4_en_tx_irq; in mlx4_en_activate_cq()
158 cq->mcq.comp = mlx4_en_rx_irq; in mlx4_en_activate_cq()
211 mlx4_cq_free(priv->mdev->dev, &cq->mcq); in mlx4_en_deactivate_cq()
[all …]
A Dcq.c63 struct mlx4_cq *mcq, *temp; in mlx4_cq_tasklet_cb() local
69 list_for_each_entry_safe(mcq, temp, &ctx->process_list, tasklet_ctx.list) { in mlx4_cq_tasklet_cb()
70 list_del_init(&mcq->tasklet_ctx.list); in mlx4_cq_tasklet_cb()
71 mcq->tasklet_ctx.comp(mcq); in mlx4_cq_tasklet_cb()
72 if (refcount_dec_and_test(&mcq->refcount)) in mlx4_cq_tasklet_cb()
73 complete(&mcq->free); in mlx4_cq_tasklet_cb()
A Den_rx.c335 ring->cqn = priv->rx_cq[ring_ind]->mcq.cqn; in mlx4_en_activate_rx_rings()
709 index = cq->mcq.cons_index & ring->size_mask; in mlx4_en_process_rx_cq()
714 cq->mcq.cons_index & cq->size)) { in mlx4_en_process_rx_cq()
930 ++cq->mcq.cons_index; in mlx4_en_process_rx_cq()
931 index = (cq->mcq.cons_index) & ring->size_mask; in mlx4_en_process_rx_cq()
946 mlx4_cq_set_ci(&cq->mcq); in mlx4_en_process_rx_cq()
948 ring->cons = cq->mcq.cons_index; in mlx4_en_process_rx_cq()
957 void mlx4_en_rx_irq(struct mlx4_cq *mcq) in mlx4_en_rx_irq() argument
959 struct mlx4_en_cq *cq = container_of(mcq, struct mlx4_en_cq, mcq); in mlx4_en_rx_irq()
A Den_tx.c428 struct mlx4_cq *mcq = &cq->mcq; in mlx4_en_process_tx_cq() local
434 u32 cons_index = mcq->cons_index; in mlx4_en_process_tx_cq()
514 mcq->cons_index = cons_index; in mlx4_en_process_tx_cq()
515 mlx4_cq_set_ci(mcq); in mlx4_en_process_tx_cq()
538 void mlx4_en_tx_irq(struct mlx4_cq *mcq) in mlx4_en_tx_irq() argument
540 struct mlx4_en_cq *cq = container_of(mcq, struct mlx4_en_cq, mcq); in mlx4_en_tx_irq()
/drivers/infiniband/hw/mlx5/
A Dcq.c53 struct mlx5_ib_cq *cq = container_of(mcq, struct mlx5_ib_cq, mcq); in mlx5_ib_cq_event()
60 type, mcq->cqn); in mlx5_ib_cq_event()
472 ++cq->mcq.cons_index; in mlx5_poll_one()
596 cq->mcq.cqn); in poll_soft_wc()
663 mlx5_cq_arm(&cq->mcq, in mlx5_ib_arm_cq()
1063 return mlx5_core_destroy_cq(dev->mdev, &mcq->mcq); in mlx5_ib_pre_destroy_cq()
1164 err = mlx5_core_modify_cq_moderation(dev->mdev, &mcq->mcq, in mlx5_ib_modify_cq()
1167 mlx5_ib_warn(dev, "modify cq 0x%x failed\n", mcq->mcq.cqn); in mlx5_ib_modify_cq()
1247 i = cq->mcq.cons_index; in copy_resize_cqes()
1274 cq->mcq.cqn); in copy_resize_cqes()
[all …]
A Dqp.c2177 to_mcq(init_attr->send_cq)->mcq.cqn); in create_dci()
2181 to_mcq(init_attr->recv_cq)->mcq.cqn); in create_dci()
2572 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs()
2576 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs()
2602 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_unlock_cqs()
2605 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { in mlx5_ib_unlock_cqs()
4285 MLX5_SET(qpc, qpc, cqn_snd, send_cq->mcq.cqn); in __mlx5_ib_modify_qp()
5736 struct mlx5_ib_cq *mcq = to_mcq(cq); in handle_drain_completion() local
5742 if (!mcq->mcq.reset_notify_added) in handle_drain_completion()
5743 mcq->mcq.reset_notify_added = 1; in handle_drain_completion()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/
A Den_dim.c38 struct mlx5_core_dev *mdev, struct mlx5_core_cq *mcq) in mlx5e_complete_dim_work() argument
40 mlx5e_modify_cq_moderation(mdev, mcq, moder.usec, moder.pkts, in mlx5e_complete_dim_work()
52 mlx5e_complete_dim_work(dim, cur_moder, rq->mdev, &rq->cq.mcq); in mlx5e_rx_dim_work()
62 mlx5e_complete_dim_work(dim, cur_moder, sq->cq.mdev, &sq->cq.mcq); in mlx5e_tx_dim_work()
67 u8 cq_period_mode, struct mlx5_core_cq *mcq, in mlx5e_dim_enable() argument
82 err = mlx5e_modify_cq_period_mode(mdev, mcq, dim->mode); in mlx5e_dim_enable()
107 c->rx_cq_moder.cq_period_mode, &rq->cq.mcq, rq); in mlx5e_dim_rx_change()
134 c->tx_cq_moder.cq_period_mode, &sq->cq.mcq, sq); in mlx5e_dim_tx_change()
A Dwc.c20 struct mlx5_core_cq mcq; member
46 struct mlx5_core_cq *mcq = &cq->mcq; in mlx5_wc_create_cqwq() local
55 mcq->cqe_sz = 64; in mlx5_wc_create_cqwq()
56 mcq->set_ci_db = cq->wq_ctrl.db.db; in mlx5_wc_create_cqwq()
57 mcq->arm_db = cq->wq_ctrl.db.db + 1; in mlx5_wc_create_cqwq()
74 struct mlx5_core_cq *mcq = &cq->mcq; in create_wc_cq() local
102 err = mlx5_core_create_cq(mdev, mcq, in, inlen, out, sizeof(out)); in create_wc_cq()
147 mlx5_core_destroy_cq(cq->mdev, &cq->mcq); in mlx5_wc_destroy_cq()
168 MLX5_SET(sqc, sqc, cqn, sq->cq.mcq.cqn); in create_wc_sq()
A Den_txrx.c260 void mlx5e_completion_event(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe) in mlx5e_completion_event() argument
262 struct mlx5e_cq *cq = container_of(mcq, struct mlx5e_cq, mcq); in mlx5e_completion_event()
269 void mlx5e_cq_error_event(struct mlx5_core_cq *mcq, enum mlx5_event event) in mlx5e_cq_error_event() argument
271 struct mlx5e_cq *cq = container_of(mcq, struct mlx5e_cq, mcq); in mlx5e_cq_error_event()
275 __func__, mcq->cqn, event); in mlx5e_cq_error_event()
A Dcq.c49 struct mlx5_core_cq *mcq; in mlx5_cq_tasklet_cb() local
56 list_for_each_entry_safe(mcq, temp, &ctx->process_list, in mlx5_cq_tasklet_cb()
58 list_del_init(&mcq->tasklet_ctx.list); in mlx5_cq_tasklet_cb()
59 mcq->tasklet_ctx.comp(mcq, NULL); in mlx5_cq_tasklet_cb()
60 mlx5_cq_put(mcq); in mlx5_cq_tasklet_cb()
/drivers/net/ethernet/mellanox/mlx5/core/lib/
A Daso.c15 struct mlx5_core_cq mcq; member
48 struct mlx5_core_cq *mcq = &cq->mcq; in mlx5_aso_alloc_cq() local
60 mcq->cqe_sz = 64; in mlx5_aso_alloc_cq()
61 mcq->set_ci_db = cq->wq_ctrl.db.db; in mlx5_aso_alloc_cq()
62 mcq->arm_db = cq->wq_ctrl.db.db + 1; in mlx5_aso_alloc_cq()
79 struct mlx5_core_cq *mcq = &cq->mcq; in create_aso_cq() local
108 err = mlx5_core_create_cq(mdev, mcq, in, inlen, out, sizeof(out)); in create_aso_cq()
117 mlx5_core_destroy_cq(cq->mdev, &cq->mcq); in mlx5_aso_destroy_cq()
195 MLX5_SET(sqc, sqc, cqn, sq->cq.mcq.cqn); in create_aso_sq()
/drivers/net/ethernet/mellanox/mlx5/core/fpga/
A Dconn.c361 mlx5_cq_arm(&conn->cq.mcq, MLX5_CQ_DB_REQ_NOT, in mlx5_fpga_conn_arm_cq()
405 conn = container_of(mcq, struct mlx5_fpga_conn, cq.mcq); in mlx5_fpga_conn_cq_complete()
471 conn->cq.mcq.cqe_sz = 64; in mlx5_fpga_conn_create_cq()
474 *conn->cq.mcq.set_ci_db = 0; in mlx5_fpga_conn_create_cq()
475 *conn->cq.mcq.arm_db = 0; in mlx5_fpga_conn_create_cq()
476 conn->cq.mcq.vector = 0; in mlx5_fpga_conn_create_cq()
478 conn->cq.mcq.uar = fdev->conn_res.uar; in mlx5_fpga_conn_create_cq()
574 MLX5_SET(qpc, qpc, cqn_snd, conn->cq.mcq.cqn); in mlx5_fpga_conn_create_qp()
575 MLX5_SET(qpc, qpc, cqn_rcv, conn->cq.mcq.cqn); in mlx5_fpga_conn_create_qp()
686 MLX5_SET(qpc, qpc, cqn_snd, conn->cq.mcq.cqn); in mlx5_fpga_conn_init_qp()
[all …]
A Dconn.h58 struct mlx5_core_cq mcq; member
/drivers/net/ethernet/mellanox/mlx5/core/en/
A Dtxrx.h83 void mlx5e_completion_event(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe);
84 void mlx5e_cq_error_event(struct mlx5_core_cq *mcq, enum mlx5_event event);
312 struct mlx5_core_cq *mcq; in mlx5e_cq_arm() local
314 mcq = &cq->mcq; in mlx5e_cq_arm()
315 mlx5_cq_arm(mcq, MLX5_CQ_DB_REQ_NOT, mcq->uar->map, cq->wq.cc); in mlx5e_cq_arm()
410 cq->mcq.cqn, ci, qn, in mlx5e_dump_error_cqe()
A Dreporter_tx.c137 eq = sq->cq.mcq.eq; in mlx5e_tx_reporter_timeout_recover()
239 mlx5e_health_eq_diag_fmsg(sq->cq.mcq.eq, fmsg); in mlx5e_tx_reporter_build_diagnose_output_sq_common()
516 sq->ch_ix, sq->sqn, sq->cq.mcq.cqn, sq->cc, sq->pc, in mlx5e_reporter_tx_timeout()
537 txqsq->ch_ix, txqsq->sqn, ts_cq->mcq.cqn, map->undelivered_counter, map->capacity); in mlx5e_reporter_tx_ptpsq_unhealthy()
A Dhealth.c26 mlx5_core_query_cq(cq->mdev, &cq->mcq, out); in mlx5e_health_cq_diag_fmsg()
31 devlink_fmsg_u32_pair_put(fmsg, "cqn", cq->mcq.cqn); in mlx5e_health_cq_diag_fmsg()
A Dreporter_rx.c182 eq = rq->cq.mcq.eq; in mlx5e_rx_reporter_timeout_recover()
220 devlink_fmsg_u32_pair_put(fmsg, "cqn", icosq->cq.mcq.cqn); in mlx5e_reporter_icosq_diagnose()
268 mlx5e_health_eq_diag_fmsg(rq->cq.mcq.eq, fmsg); in mlx5e_rx_reporter_build_diagnose_output_rq_common()
611 rq->ix, icosq_str, rq->rqn, rq->cq.mcq.cqn); in mlx5e_reporter_rx_timeout()
/drivers/vfio/pci/mlx5/
A Dcmd.c1100 tracker.cq.mcq)); in mlx5vf_cq_event()
1148 tracker.cq.mcq); in mlx5vf_cq_complete()
1172 cq->mcq.set_ci_db = cq->db.db; in mlx5vf_create_cq()
1173 cq->mcq.arm_db = cq->db.db + 1; in mlx5vf_create_cq()
1174 cq->mcq.cqe_sz = cqe_size; in mlx5vf_create_cq()
1203 cq->mcq.comp = mlx5vf_cq_complete; in mlx5vf_create_cq()
1204 cq->mcq.event = mlx5vf_cq_event; in mlx5vf_create_cq()
1210 cq->mcq.cons_index); in mlx5vf_create_cq()
1707 ++cq->mcq.cons_index; in mlx5vf_cq_poll_one()
1763 cq->mcq.cons_index); in mlx5vf_tracker_read_and_clear()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
A Dsend.c756 MLX5_SET(sqc, sqc, cqn, cq->mcq.cqn); in hws_send_ring_create_sq()
876 static void hws_cq_complete(struct mlx5_core_cq *mcq, in hws_cq_complete() argument
879 pr_err("CQ completion CQ: #%u\n", mcq->cqn); in hws_cq_complete()
888 struct mlx5_core_cq *mcq = &cq->mcq; in hws_send_ring_alloc_cq() local
901 mcq->cqe_sz = 64; in hws_send_ring_alloc_cq()
902 mcq->set_ci_db = cq->wq_ctrl.db.db; in hws_send_ring_alloc_cq()
903 mcq->arm_db = cq->wq_ctrl.db.db + 1; in hws_send_ring_alloc_cq()
904 mcq->comp = hws_cq_complete; in hws_send_ring_alloc_cq()
922 struct mlx5_core_cq *mcq = &cq->mcq; in hws_send_ring_create_cq() local
947 err = mlx5_core_create_cq(mdev, mcq, in, inlen, out, sizeof(out)); in hws_send_ring_create_cq()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/
A Ddr_send.c1055 pr_err("CQ completion CQ: #%u\n", mcq->cqn); in dr_cq_complete()
1115 cq->mcq.comp = dr_cq_complete; in dr_create_cq()
1123 cq->mcq.cqe_sz = 64; in dr_create_cq()
1124 cq->mcq.set_ci_db = cq->wq_ctrl.db.db; in dr_create_cq()
1125 cq->mcq.arm_db = cq->wq_ctrl.db.db + 1; in dr_create_cq()
1126 *cq->mcq.set_ci_db = 0; in dr_create_cq()
1131 *cq->mcq.arm_db = cpu_to_be32(2 << 28); in dr_create_cq()
1133 cq->mcq.vector = 0; in dr_create_cq()
1134 cq->mcq.uar = uar; in dr_create_cq()
1148 mlx5_core_destroy_cq(mdev, &cq->mcq); in dr_destroy_cq()
[all …]
/drivers/ufs/core/
A DMakefile4 ufshcd-core-y += ufshcd.o ufs-sysfs.o ufs-mcq.o
/drivers/vdpa/mlx5/net/
A Dmlx5_vnet.c71 struct mlx5_core_cq mcq; member
507 vcq->mcq.cons_index++; in mlx5_vdpa_poll_one()
517 mlx5_cq_set_ci(&mvq->cq.mcq); in mlx5_vdpa_handle_completions()
530 struct mlx5_vdpa_virtqueue *mvq = container_of(mcq, struct mlx5_vdpa_virtqueue, cq.mcq); in mlx5_vdpa_cq_comp()
552 mlx5_cq_arm(&mvq->cq.mcq, MLX5_CQ_DB_REQ_NOT, uar_page, mvq->cq.mcq.cons_index); in mlx5_vdpa_cq_comp()
573 vcq->mcq.set_ci_db = vcq->db.db; in cq_create()
574 vcq->mcq.arm_db = vcq->db.db + 1; in cq_create()
575 vcq->mcq.cqe_sz = 64; in cq_create()
617 vcq->mcq.set_ci_db = vcq->db.db; in cq_create()
618 vcq->mcq.arm_db = vcq->db.db + 1; in cq_create()
[all …]

Completed in 846 milliseconds

12