| /linux/drivers/infiniband/hw/erdma/ |
| A D | erdma_qp.c | 220 while (i < send_wr->num_sge) { in fill_inline_data() 221 bytes += send_wr->sg_list[i].length; in fill_inline_data() 267 while (i < send_wr->num_sge) { in fill_sgl() 273 bytes += send_wr->sg_list[i].length; in fill_sgl() 290 enum ib_wr_opcode op = send_wr->opcode; in erdma_push_one_sqe() 311 flags = send_wr->send_flags; in erdma_push_one_sqe() 347 if (unlikely(send_wr->num_sge != 1)) in erdma_push_one_sqe() 397 mr = to_emr(reg_wr(send_wr)->mr); in erdma_push_one_sqe() 476 send_wr->num_sge); in erdma_push_one_sqe() 508 if (!send_wr) in erdma_post_send() [all …]
|
| A D | erdma_verbs.h | 358 int erdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *send_wr,
|
| /linux/drivers/infiniband/ulp/isert/ |
| A D | ib_isert.c | 783 struct ib_send_wr send_wr; in isert_login_post_send() local 791 send_wr.next = NULL; in isert_login_post_send() 795 send_wr.opcode = IB_WR_SEND; in isert_login_post_send() 864 struct ib_send_wr *send_wr) in isert_init_send_wr() argument 875 send_wr->opcode = IB_WR_SEND; in isert_init_send_wr() 1763 struct ib_send_wr *send_wr = &isert_cmd->tx_desc.send_wr; in isert_put_response() local 1851 struct ib_send_wr *send_wr = &isert_cmd->tx_desc.send_wr; in isert_put_nopin() local 1870 struct ib_send_wr *send_wr = &isert_cmd->tx_desc.send_wr; in isert_put_logout_rsp() local 1888 struct ib_send_wr *send_wr = &isert_cmd->tx_desc.send_wr; in isert_put_tm_rsp() local 1906 struct ib_send_wr *send_wr = &isert_cmd->tx_desc.send_wr; in isert_put_reject() local [all …]
|
| A D | ib_isert.h | 132 struct ib_send_wr send_wr; member
|
| /linux/drivers/infiniband/core/ |
| A D | mad.c | 62 struct ib_ud_wr *wr = &mad_send_wr->send_wr; in create_mad_addr_info() 607 struct ib_ud_wr *send_wr = &mad_send_wr->send_wr; in handle_outgoing_dr_smp() local 616 port_num = send_wr->port_num; in handle_outgoing_dr_smp() 689 send_wr->wr.wr_cqe, drslid, in handle_outgoing_dr_smp() 690 send_wr->pkey_index, in handle_outgoing_dr_smp() 691 send_wr->port_num, &mad_wc); in handle_outgoing_dr_smp() 799 pad = send_wr->pad; in alloc_send_rmpp_list() 805 free_send_rmpp_list(send_wr); in alloc_send_rmpp_list() 821 send_wr->cur_seg = container_of(send_wr->rmpp_list.next, in alloc_send_rmpp_list() 823 send_wr->last_ack_seg = send_wr->cur_seg; in alloc_send_rmpp_list() [all …]
|
| A D | cma_trace.h | 186 __field(u32, send_wr) 198 __entry->send_wr = qp_init_attr->cap.max_send_wr; 218 rdma_show_qp_type(__entry->qp_type), __entry->send_wr,
|
| A D | mad_priv.h | 128 struct ib_ud_wr send_wr; member
|
| A D | agent.c | 138 mad_send_wr->send_wr.port_num = port_num; in agent_send_response()
|
| /linux/net/sunrpc/xprtrdma/ |
| A D | svc_rdma_sendto.c | 342 struct ib_send_wr *send_wr = &ctxt->sc_send_wr; in svc_rdma_post_send() local 351 send_wr->sg_list[0].addr, in svc_rdma_post_send() 352 send_wr->sg_list[0].length, in svc_rdma_post_send() 885 struct ib_send_wr *send_wr = &sctxt->sc_send_wr; in svc_rdma_send_reply_msg() local 899 send_wr->opcode = IB_WR_SEND_WITH_INV; in svc_rdma_send_reply_msg() 900 send_wr->ex.invalidate_rkey = rctxt->rc_inv_rkey; in svc_rdma_send_reply_msg() 902 send_wr->opcode = IB_WR_SEND; in svc_rdma_send_reply_msg()
|
| A D | frwr_ops.c | 386 struct ib_send_wr *post_wr, *send_wr = &req->rl_wr; in frwr_send() local 393 post_wr = send_wr; in frwr_send() 408 send_wr->send_flags |= IB_SEND_SIGNALED; in frwr_send() 412 send_wr->send_flags &= ~IB_SEND_SIGNALED; in frwr_send()
|
| /linux/drivers/infiniband/ulp/iser/ |
| A D | iscsi_iser.h | 144 #define ISER_GET_MAX_XMIT_CMDS(send_wr) ((send_wr \ argument 247 struct ib_send_wr send_wr; member
|
| A D | iser_memory.c | 281 wr->wr.next = &tx_desc->send_wr; in iser_reg_sig_mr() 328 wr->wr.next = &tx_desc->send_wr; in iser_fast_reg_mr()
|
| A D | iser_verbs.c | 859 struct ib_send_wr *wr = &tx_desc->send_wr; in iser_post_send()
|
| /linux/fs/smb/client/ |
| A D | smbdirect.c | 686 struct ib_send_wr send_wr; in smbd_post_send_negotiate_req() local 725 send_wr.next = NULL; in smbd_post_send_negotiate_req() 726 send_wr.wr_cqe = &request->cqe; in smbd_post_send_negotiate_req() 727 send_wr.sg_list = request->sge; in smbd_post_send_negotiate_req() 728 send_wr.num_sge = request->num_sge; in smbd_post_send_negotiate_req() 729 send_wr.opcode = IB_WR_SEND; in smbd_post_send_negotiate_req() 796 struct ib_send_wr send_wr; in smbd_post_send() local 812 send_wr.next = NULL; in smbd_post_send() 813 send_wr.wr_cqe = &request->cqe; in smbd_post_send() 814 send_wr.sg_list = request->sge; in smbd_post_send() [all …]
|
| /linux/drivers/nvme/target/ |
| A D | rdma.c | 63 struct ib_send_wr send_wr; member 420 r->send_wr.wr_cqe = &r->send_cqe; in nvmet_rdma_alloc_rsp() 421 r->send_wr.sg_list = &r->send_sge; in nvmet_rdma_alloc_rsp() 422 r->send_wr.num_sge = 1; in nvmet_rdma_alloc_rsp() 423 r->send_wr.send_flags = IB_SEND_SIGNALED; in nvmet_rdma_alloc_rsp() 717 rsp->send_wr.opcode = IB_WR_SEND_WITH_INV; in nvmet_rdma_queue_response() 718 rsp->send_wr.ex.invalidate_rkey = rsp->invalidate_rkey; in nvmet_rdma_queue_response() 720 rsp->send_wr.opcode = IB_WR_SEND; in nvmet_rdma_queue_response() 729 cm_id->port_num, NULL, &rsp->send_wr); in nvmet_rdma_queue_response() 731 first_wr = &rsp->send_wr; in nvmet_rdma_queue_response() [all …]
|
| /linux/drivers/infiniband/hw/mlx5/ |
| A D | wr.c | 416 static int set_sig_data_segment(const struct ib_send_wr *send_wr, in set_sig_data_segment() argument 557 static int set_pi_umr_wr(const struct ib_send_wr *send_wr, in set_pi_umr_wr() argument 561 const struct ib_reg_wr *wr = reg_wr(send_wr); in set_pi_umr_wr() 569 if (unlikely(send_wr->num_sge != 0) || in set_pi_umr_wr() 599 ret = set_sig_data_segment(send_wr, wr->mr, sig_attrs, qp, seg, size, in set_pi_umr_wr()
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| A D | ib_verbs.h | 235 int bnxt_re_post_send(struct ib_qp *qp, const struct ib_send_wr *send_wr,
|
| /linux/drivers/infiniband/ulp/srpt/ |
| A D | ib_srpt.c | 2875 struct ib_send_wr send_wr, *first_wr = &send_wr; in srpt_queue_response() local 2938 send_wr.next = NULL; in srpt_queue_response() 2939 send_wr.wr_cqe = &ioctx->ioctx.cqe; in srpt_queue_response() 2940 send_wr.sg_list = &sge; in srpt_queue_response() 2941 send_wr.num_sge = 1; in srpt_queue_response() 2942 send_wr.opcode = IB_WR_SEND; in srpt_queue_response() 2943 send_wr.send_flags = IB_SEND_SIGNALED; in srpt_queue_response()
|
| /linux/include/uapi/rdma/ |
| A D | ib_user_verbs.h | 843 struct ib_uverbs_send_wr send_wr[]; member
|
| /linux/include/rdma/ |
| A D | ib_verbs.h | 2335 int (*post_send)(struct ib_qp *qp, const struct ib_send_wr *send_wr, 3855 const struct ib_send_wr *send_wr, in ib_post_send() argument 3860 return qp->device->ops.post_send(qp, send_wr, bad_send_wr ? : &dummy); in ib_post_send()
|
| /linux/drivers/infiniband/hw/hns/ |
| A D | hns_roce_hw_v2.c | 3416 struct ib_send_wr *send_wr; in free_mr_post_send_lp_wqe() local 3419 send_wr = &rdma_wr.wr; in free_mr_post_send_lp_wqe() 3420 send_wr->opcode = IB_WR_RDMA_WRITE; in free_mr_post_send_lp_wqe() 3422 ret = hns_roce_v2_post_send(&hr_qp->ibqp, send_wr, &bad_wr); in free_mr_post_send_lp_wqe()
|