/linux-6.3-rc2/drivers/infiniband/core/ |
A D | uverbs_std_types_qp.c | 95 struct ib_cq *send_cq = NULL; in UVERBS_HANDLER() local 168 send_cq = uverbs_attr_get_obj(attrs, in UVERBS_HANDLER() 170 if (IS_ERR(send_cq)) in UVERBS_HANDLER() 171 return PTR_ERR(send_cq); in UVERBS_HANDLER() 175 send_cq = uverbs_attr_get_obj(attrs, in UVERBS_HANDLER() 177 if (IS_ERR(send_cq)) in UVERBS_HANDLER() 178 return PTR_ERR(send_cq); in UVERBS_HANDLER() 234 attr.send_cq = send_cq; in UVERBS_HANDLER()
|
A D | verbs.c | 1175 qp->send_cq = qp->recv_cq = NULL; in create_xrc_qp_user() 1227 qp->send_cq = attr->send_cq; in create_qp() 1241 qp->send_cq = attr->send_cq; in create_qp() 1301 if (qp->send_cq) in ib_qp_usecnt_inc() 1302 atomic_inc(&qp->send_cq->usecnt); in ib_qp_usecnt_inc() 1320 if (qp->send_cq) in ib_qp_usecnt_dec() 1321 atomic_dec(&qp->send_cq->usecnt); in ib_qp_usecnt_dec() 2747 struct ib_cq *cq = qp->send_cq; in __ib_drain_sq() 2840 trace_cq_drain_complete(qp->send_cq); in ib_drain_sq()
|
/linux-6.3-rc2/drivers/infiniband/hw/hns/ |
A D | hns_roce_qp.c | 250 hr_send_cq = send_cq ? to_hr_cq(send_cq) : NULL; in add_qp_to_list() 1361 __acquire(&send_cq->lock); in hns_roce_lock_cqs() 1364 spin_lock_irq(&send_cq->lock); in hns_roce_lock_cqs() 1368 __acquire(&send_cq->lock); in hns_roce_lock_cqs() 1369 } else if (send_cq == recv_cq) { in hns_roce_lock_cqs() 1370 spin_lock_irq(&send_cq->lock); in hns_roce_lock_cqs() 1373 spin_lock_irq(&send_cq->lock); in hns_roce_lock_cqs() 1387 __release(&send_cq->lock); in hns_roce_unlock_cqs() 1390 spin_unlock(&send_cq->lock); in hns_roce_unlock_cqs() 1392 __release(&send_cq->lock); in hns_roce_unlock_cqs() [all …]
|
A D | hns_roce_hw_v2.c | 2737 free_mr->rsv_qp[i]->ibqp.send_cq = cq; in free_mr_init_qp() 2793 qp_init_attr.send_cq = cq; in free_mr_alloc_res() 5241 if (ibqp->send_cq) in clear_qp() 5242 hns_roce_v2_cq_clean(to_hr_cq(ibqp->send_cq), in clear_qp() 5484 qp_init_attr->send_cq = ibqp->send_cq; in hns_roce_v2_query_qp() 5508 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v2_destroy_qp_common() local 5522 send_cq = hr_qp->ibqp.send_cq ? to_hr_cq(hr_qp->ibqp.send_cq) : NULL; in hns_roce_v2_destroy_qp_common() 5526 hns_roce_lock_cqs(send_cq, recv_cq); in hns_roce_v2_destroy_qp_common() 5535 if (send_cq && send_cq != recv_cq) in hns_roce_v2_destroy_qp_common() 5536 __hns_roce_v2_cq_clean(send_cq, hr_qp->qpn, NULL); in hns_roce_v2_destroy_qp_common() [all …]
|
/linux-6.3-rc2/drivers/infiniband/ulp/ipoib/ |
A D | ipoib_verbs.c | 187 priv->send_cq = ib_create_cq(priv->ca, ipoib_ib_tx_completion, NULL, in ipoib_transport_dev_init() 189 if (IS_ERR(priv->send_cq)) { in ipoib_transport_dev_init() 197 init_attr.send_cq = priv->send_cq; in ipoib_transport_dev_init() 218 if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP)) in ipoib_transport_dev_init() 244 ib_destroy_cq(priv->send_cq); in ipoib_transport_dev_init() 266 ib_destroy_cq(priv->send_cq); in ipoib_transport_dev_cleanup()
|
A D | ipoib_ib.c | 439 n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); in poll_tx() 507 n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); in ipoib_tx_poll() 519 if (unlikely(ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | in ipoib_tx_poll() 649 if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | in ipoib_send()
|
/linux-6.3-rc2/drivers/infiniband/hw/mthca/ |
A D | mthca_qp.c | 1164 struct mthca_cq *send_cq, in mthca_alloc_qp_common() argument 1293 struct mthca_cq *send_cq, in mthca_alloc_qp() argument 1339 if (send_cq == recv_cq) { in mthca_lock_cqs() 1340 spin_lock_irq(&send_cq->lock); in mthca_lock_cqs() 1343 spin_lock_irq(&send_cq->lock); in mthca_lock_cqs() 1354 if (send_cq == recv_cq) { in mthca_unlock_cqs() 1361 spin_unlock(&send_cq->lock); in mthca_unlock_cqs() 1368 struct mthca_cq *send_cq, in mthca_alloc_sqp() argument 1448 struct mthca_cq *send_cq; in mthca_free_qp() local 1451 send_cq = to_mcq(qp->ibqp.send_cq); in mthca_free_qp() [all …]
|
A D | mthca_dev.h | 535 struct mthca_cq *send_cq, 544 struct mthca_cq *send_cq,
|
/linux-6.3-rc2/drivers/infiniband/hw/mlx5/ |
A D | qp.c | 1309 send_cq ? get_ts_format(dev, send_cq, fr_sup, rt_sup) : in get_qp_ts_format() 2208 if (send_cq) in create_dci() 2411 if (send_cq) in create_user_qp() 2540 if (send_cq) in create_kernel_qp() 2557 if (send_cq) { in mlx5_ib_lock_cqs() 2587 if (send_cq) { in mlx5_ib_unlock_cqs() 2618 *send_cq = NULL; in get_cqs() 2682 get_cqs(qp->type, qp->ibqp.send_cq, qp->ibqp.recv_cq, &send_cq, in destroy_qp_common() 2689 if (send_cq) in destroy_qp_common() 4223 if (send_cq) in __mlx5_ib_modify_qp() [all …]
|
A D | gsi.c | 50 struct ib_cq *gsi_cq = mqp->ibqp.send_cq; in generate_completions() 141 hw_init_attr.send_cq = gsi->cq; in mlx5_ib_create_gsi() 205 .send_cq = gsi->cq, in create_gsi_ud_qp()
|
A D | mem.c | 246 qp_init_attr.send_cq = cq; in mlx5_ib_test_wc()
|
/linux-6.3-rc2/include/rdma/ |
A D | rdmavt_qp.h | 805 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.send_cq); in rvt_send_cq() 920 static inline u32 ib_cq_tail(struct ib_cq *send_cq) in ib_cq_tail() argument 922 struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); in ib_cq_tail() 924 return ibcq_to_rvtcq(send_cq)->ip ? in ib_cq_tail() 926 ibcq_to_rvtcq(send_cq)->kqueue->tail; in ib_cq_tail() 936 static inline u32 ib_cq_head(struct ib_cq *send_cq) in ib_cq_head() argument 938 struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); in ib_cq_head() 940 return ibcq_to_rvtcq(send_cq)->ip ? in ib_cq_head() 942 ibcq_to_rvtcq(send_cq)->kqueue->head; in ib_cq_head()
|
/linux-6.3-rc2/drivers/infiniband/hw/mana/ |
A D | qp.c | 261 struct mana_ib_cq *send_cq = in mana_ib_create_qp_raw() local 262 container_of(attr->send_cq, struct mana_ib_cq, ibcq); in mana_ib_create_qp_raw() 350 cq_spec.gdma_region = send_cq->gdma_region; in mana_ib_create_qp_raw() 351 cq_spec.queue_size = send_cq->cqe * COMP_ENTRY_SIZE; in mana_ib_create_qp_raw() 366 send_cq->gdma_region = GDMA_INVALID_DMA_REGION; in mana_ib_create_qp_raw() 369 send_cq->id = cq_spec.queue_index; in mana_ib_create_qp_raw() 373 qp->tx_object, qp->sq_id, send_cq->id); in mana_ib_create_qp_raw() 376 resp.cqid = send_cq->id; in mana_ib_create_qp_raw()
|
/linux-6.3-rc2/drivers/infiniband/hw/mlx4/ |
A D | qp.c | 1301 if (send_cq == recv_cq) { in mlx4_ib_lock_cqs() 1302 spin_lock(&send_cq->lock); in mlx4_ib_lock_cqs() 1305 spin_lock(&send_cq->lock); in mlx4_ib_lock_cqs() 1316 if (send_cq == recv_cq) { in mlx4_ib_unlock_cqs() 1352 *recv_cq = *send_cq; in get_cqs() 1355 *send_cq = to_mcq(qp->ibqp.send_cq); in get_cqs() 1356 *recv_cq = *send_cq; in get_cqs() 1361 *send_cq = (src == MLX4_IB_QP_SRC) ? to_mcq(qp->ibqp.send_cq) : in get_cqs() 1443 if (send_cq != recv_cq) in destroy_qp_common() 2355 recv_cq = send_cq; in __mlx4_ib_modify_qp() [all …]
|
/linux-6.3-rc2/Documentation/translations/zh_CN/infiniband/ |
A D | tag_matching.rst | 39 1. Eager协议--当发送方处理完发送时,完整的信息就会被发送。在send_cq中会收到
|
/linux-6.3-rc2/net/sunrpc/xprtrdma/ |
A D | verbs.c | 341 if (ep->re_attr.send_cq) in rpcrdma_ep_destroy() 342 ib_free_cq(ep->re_attr.send_cq); in rpcrdma_ep_destroy() 343 ep->re_attr.send_cq = NULL; in rpcrdma_ep_destroy() 410 ep->re_attr.send_cq = ib_alloc_cq_any(device, r_xprt, in rpcrdma_ep_create() 413 if (IS_ERR(ep->re_attr.send_cq)) { in rpcrdma_ep_create() 414 rc = PTR_ERR(ep->re_attr.send_cq); in rpcrdma_ep_create() 415 ep->re_attr.send_cq = NULL; in rpcrdma_ep_create() 613 sc->sc_cid.ci_queue_id = ep->re_attr.send_cq->res.id; in rpcrdma_sendctx_create()
|
/linux-6.3-rc2/fs/ksmbd/ |
A D | transport_rdma.c | 104 struct ib_cq *send_cq; member 446 if (t->send_cq) in free_transport() 447 ib_free_cq(t->send_cq); in free_transport() 1875 t->send_cq = ib_alloc_cq(t->cm_id->device, t, in smb_direct_create_qpair() 1878 if (IS_ERR(t->send_cq)) { in smb_direct_create_qpair() 1880 ret = PTR_ERR(t->send_cq); in smb_direct_create_qpair() 1881 t->send_cq = NULL; in smb_direct_create_qpair() 1900 qp_attr.send_cq = t->send_cq; in smb_direct_create_qpair() 1935 if (t->send_cq) { in smb_direct_create_qpair() 1936 ib_destroy_cq(t->send_cq); in smb_direct_create_qpair() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/ibm/ehea/ |
A D | ehea_main.c | 803 struct ehea_cq *send_cq = pr->send_cq; in ehea_proc_cqes() local 812 cqe = ehea_poll_cq(send_cq); in ehea_proc_cqes() 814 ehea_inc_cq(send_cq); in ehea_proc_cqes() 854 cqe = ehea_poll_cq(send_cq); in ehea_proc_cqes() 892 ehea_reset_cq_ep(pr->send_cq); in ehea_poll() 894 ehea_reset_cq_n1(pr->send_cq); in ehea_poll() 897 cqe_skb = ehea_poll_cq(pr->send_cq); in ehea_poll() 1479 if (!pr->send_cq) { in ehea_init_port_res() 1486 pr->send_cq->attr.act_nr_of_cqes, in ehea_init_port_res() 1561 ehea_destroy_cq(pr->send_cq); in ehea_init_port_res() [all …]
|
A D | ehea.h | 351 struct ehea_cq *send_cq; member
|
/linux-6.3-rc2/drivers/infiniband/hw/vmw_pvrdma/ |
A D | pvrdma_qp.c | 58 static inline void get_cqs(struct pvrdma_qp *qp, struct pvrdma_cq **send_cq, in get_cqs() argument 61 *send_cq = to_vcq(qp->ibqp.send_cq); in get_cqs() 359 cmd->send_cq_handle = to_vcq(init_attr->send_cq)->cq_handle; in pvrdma_create_qp() 1042 init_attr->send_cq = qp->ibqp.send_cq; in pvrdma_query_qp()
|
/linux-6.3-rc2/drivers/infiniband/sw/rxe/ |
A D | rxe_qp.c | 75 if (!init->recv_cq || !init->send_cq) { in rxe_qp_chk_init() 304 struct rxe_cq *scq = to_rcq(init->send_cq); in rxe_qp_from_init() 362 init->send_cq = qp->ibqp.send_cq; in rxe_qp_to_init()
|
/linux-6.3-rc2/fs/cifs/ |
A D | smbdirect.c | 1357 ib_free_cq(info->send_cq); in smbd_destroy() 1543 info->send_cq = NULL; in _smbd_get_connection() 1545 info->send_cq = in _smbd_get_connection() 1548 if (IS_ERR(info->send_cq)) { in _smbd_get_connection() 1549 info->send_cq = NULL; in _smbd_get_connection() 1571 qp_attr.send_cq = info->send_cq; in _smbd_get_connection() 1688 if (info->send_cq) in _smbd_get_connection() 1689 ib_free_cq(info->send_cq); in _smbd_get_connection()
|
A D | smbdirect.h | 59 struct ib_cq *send_cq, *recv_cq; member
|
/linux-6.3-rc2/drivers/infiniband/ulp/srp/ |
A D | ib_srp.h | 157 struct ib_cq *send_cq; member
|
A D | ib_srp.c | 521 ib_process_cq_direct(ch->send_cq, -1); in srp_destroy_qp() 534 struct ib_cq *recv_cq, *send_cq; in srp_create_ch_ib() local 554 if (IS_ERR(send_cq)) { in srp_create_ch_ib() 555 ret = PTR_ERR(send_cq); in srp_create_ch_ib() 566 init_attr->send_cq = send_cq; in srp_create_ch_ib() 604 if (ch->send_cq) in srp_create_ch_ib() 605 ib_free_cq(ch->send_cq); in srp_create_ch_ib() 609 ch->send_cq = send_cq; in srp_create_ch_ib() 627 ib_free_cq(send_cq); in srp_create_ch_ib() 672 ib_free_cq(ch->send_cq); in srp_free_ch_ib() [all …]
|