| /linux/net/9p/ |
| A D | trans_rdma.c | 260 if (rdma) in p9_cm_event_handler() 364 if (!rdma) in rdma_destroy_trans() 367 if (rdma->qp && !IS_ERR(rdma->qp)) in rdma_destroy_trans() 370 if (rdma->pd && !IS_ERR(rdma->pd)) in rdma_destroy_trans() 373 if (rdma->cq && !IS_ERR(rdma->cq)) in rdma_destroy_trans() 376 if (rdma->cm_id && !IS_ERR(rdma->cm_id)) in rdma_destroy_trans() 559 if (!rdma) in rdma_close() 586 sema_init(&rdma->sq_sem, rdma->sq_depth); in alloc_rdma() 587 sema_init(&rdma->rq_sem, rdma->rq_depth); in alloc_rdma() 704 rdma->pd = ib_alloc_pd(rdma->cm_id->device, 0); in rdma_create_trans() [all …]
|
| /linux/drivers/gpu/drm/mediatek/ |
| A D | mtk_disp_rdma.c | 52 #define RDMA_FIFO_SIZE(rdma) ((rdma)->data->fifo_size) argument 130 rdma->vblank_cb = NULL; in mtk_rdma_unregister_vblank_cb() 193 mtk_ddp_write_mask(cmdq_pkt, width, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_config() 195 mtk_ddp_write_mask(cmdq_pkt, height, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_config() 198 if (rdma->fifo_size) in mtk_rdma_config() 277 &rdma->cmdq_reg, rdma->regs, DISP_REG_RDMA_SIZE_CON_0, in mtk_rdma_layer_config() 280 mtk_ddp_write_mask(cmdq_pkt, 0, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_layer_config() 284 mtk_ddp_write_relaxed(cmdq_pkt, addr, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_layer_config() 286 mtk_ddp_write_relaxed(cmdq_pkt, pitch, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_layer_config() 288 mtk_ddp_write(cmdq_pkt, RDMA_MEM_GMC, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_layer_config() [all …]
|
| /linux/drivers/gpu/drm/meson/ |
| A D | meson_rdma.c | 26 if (!priv->rdma.addr) { in meson_rdma_init() 28 priv->rdma.addr = in meson_rdma_init() 32 if (!priv->rdma.addr) in meson_rdma_init() 36 priv->rdma.offset = 0; in meson_rdma_init() 50 if (!priv->rdma.addr && !priv->rdma.addr_dma) in meson_rdma_free() 56 priv->rdma.addr, priv->rdma.addr_dma); in meson_rdma_free() 58 priv->rdma.addr = NULL; in meson_rdma_free() 88 priv->rdma.offset = 0; in meson_rdma_reset() 99 priv->rdma.addr[priv->rdma.offset++] = reg; in meson_rdma_writel() 100 priv->rdma.addr[priv->rdma.offset++] = val; in meson_rdma_writel() [all …]
|
| /linux/net/sunrpc/xprtrdma/ |
| A D | svc_rdma_sendto.c | 139 ctxt->sc_rdma = rdma; in svc_rdma_send_ctxt_alloc() 147 rdma->sc_max_req_size); in svc_rdma_send_ctxt_alloc() 176 rdma->sc_max_req_size, in svc_rdma_send_ctxts_destroy() 195 spin_lock(&rdma->sc_send_lock); in svc_rdma_send_ctxt_get() 197 spin_unlock(&rdma->sc_send_lock); in svc_rdma_send_ctxt_get() 284 wake_up(&rdma->sc_send_wait); in svc_rdma_wake_send_waiters() 368 wait_event(rdma->sc_send_wait, in svc_rdma_post_send() 825 .md_rdma = rdma, in svc_rdma_map_reply_msg() 942 *p++ = rdma->sc_fc_credits; in svc_rdma_send_error_msg() 992 struct svcxprt_rdma *rdma = in svc_rdma_sendto() local [all …]
|
| A D | svc_rdma_recvfrom.c | 244 struct svcxprt_rdma *rdma = in svc_rdma_release_ctxt() local 271 rdma->sc_pending_recvs++; in svc_rdma_refresh_recvs() 309 total = (rdma->sc_max_requests * 2) + rdma->sc_recv_batch; in svc_rdma_post_recvs() 319 return svc_rdma_refresh_recvs(rdma, rdma->sc_max_requests); in svc_rdma_post_recvs() 334 rdma->sc_pending_recvs--; in svc_rdma_wc_receive() 352 if (rdma->sc_pending_recvs < rdma->sc_max_requests) in svc_rdma_wc_receive() 353 if (!svc_rdma_refresh_recvs(rdma, rdma->sc_recv_batch)) in svc_rdma_wc_receive() 359 spin_lock(&rdma->sc_rq_dto_lock); in svc_rdma_wc_receive() 363 spin_unlock(&rdma->sc_rq_dto_lock); in svc_rdma_wc_receive() 365 svc_xprt_enqueue(&rdma->sc_xprt); in svc_rdma_wc_receive() [all …]
|
| A D | svc_rdma_rw.c | 62 spin_lock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 64 spin_unlock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 138 ret = rdma_rw_ctx_init(&ctxt->rw_ctx, rdma->sc_qp, rdma->sc_port_num, in svc_rdma_rw_ctx_init() 144 svc_rdma_put_rw_ctxt(rdma, ctxt); in svc_rdma_rw_ctx_init() 160 svc_rdma_send_cid_init(rdma, cid); in svc_rdma_cc_init() 211 info->wi_rdma = rdma; in svc_rdma_write_info_alloc() 333 spin_lock(&rdma->sc_rq_dto_lock); in svc_rdma_wc_read_done() 338 svc_xprt_enqueue(&rdma->sc_xprt); in svc_rdma_wc_read_done() 400 wait_event(rdma->sc_send_wait, in svc_rdma_post_chunk_ctxt() 413 wake_up(&rdma->sc_send_wait); in svc_rdma_post_chunk_ctxt() [all …]
|
| A D | svc_rdma_transport.c | 585 struct svcxprt_rdma *rdma = in svc_rdma_detach() local 593 struct svcxprt_rdma *rdma = in __svc_rdma_free() local 598 if (rdma->sc_qp && !IS_ERR(rdma->sc_qp)) in __svc_rdma_free() 599 ib_drain_qp(rdma->sc_qp); in __svc_rdma_free() 609 if (rdma->sc_qp && !IS_ERR(rdma->sc_qp)) in __svc_rdma_free() 612 if (rdma->sc_sq_cq && !IS_ERR(rdma->sc_sq_cq)) in __svc_rdma_free() 615 if (rdma->sc_rq_cq && !IS_ERR(rdma->sc_rq_cq)) in __svc_rdma_free() 618 if (rdma->sc_pd && !IS_ERR(rdma->sc_pd)) in __svc_rdma_free() 625 kfree(rdma); in __svc_rdma_free() 630 struct svcxprt_rdma *rdma = in svc_rdma_free() local [all …]
|
| A D | svc_rdma_backchannel.c | 75 static int svc_rdma_bc_sendto(struct svcxprt_rdma *rdma, in svc_rdma_bc_sendto() argument 83 ret = svc_rdma_map_reply_msg(rdma, sctxt, &empty_pcl, &empty_pcl, in svc_rdma_bc_sendto() 93 return svc_rdma_post_send(rdma, sctxt); in svc_rdma_bc_sendto() 136 rpcrdma_bc_send_request(struct svcxprt_rdma *rdma, struct rpc_rqst *rqst) in rpcrdma_bc_send_request() argument 144 ctxt = svc_rdma_send_ctxt_get(rdma); in rpcrdma_bc_send_request() 160 rc = svc_rdma_bc_sendto(rdma, rqst, ctxt); in rpcrdma_bc_send_request() 166 svc_rdma_send_ctxt_put(rdma, ctxt); in rpcrdma_bc_send_request() 183 struct svcxprt_rdma *rdma = in xprt_rdma_bc_send_request() local 190 ret = rpcrdma_bc_send_request(rdma, rqst); in xprt_rdma_bc_send_request()
|
| /linux/include/linux/sunrpc/ |
| A D | svc_rdma.h | 148 cid->ci_queue_id = rdma->sc_rq_cq->res.id; in svc_rdma_recv_cid_init() 160 cid->ci_queue_id = rdma->sc_sq_cq->res.id; in svc_rdma_send_cid_init() 257 extern bool svc_rdma_post_recvs(struct svcxprt_rdma *rdma); 259 svc_rdma_recv_ctxt_get(struct svcxprt_rdma *rdma); 260 extern void svc_rdma_recv_ctxt_put(struct svcxprt_rdma *rdma, 267 extern void svc_rdma_cc_init(struct svcxprt_rdma *rdma, 270 extern void svc_rdma_cc_init(struct svcxprt_rdma *rdma, 272 extern void svc_rdma_cc_release(struct svcxprt_rdma *rdma, 292 svc_rdma_send_ctxt_get(struct svcxprt_rdma *rdma); 295 extern int svc_rdma_post_send(struct svcxprt_rdma *rdma, [all …]
|
| /linux/Documentation/devicetree/bindings/display/mediatek/ |
| A D | mediatek,rdma.yaml | 27 - mediatek,mt2701-disp-rdma 28 - mediatek,mt8173-disp-rdma 29 - mediatek,mt8183-disp-rdma 30 - mediatek,mt8195-disp-rdma 33 - mediatek,mt8188-disp-rdma 37 - mediatek,mt7623-disp-rdma 38 - mediatek,mt2712-disp-rdma 42 - mediatek,mt6795-disp-rdma 71 mediatek,rdma-fifo-size: 74 corresponding rdma. [all …]
|
| /linux/Documentation/admin-guide/cgroup-v1/ |
| A D | rdma.rst | 28 Currently user space applications can easily take away all the rdma verb 31 rdma resources. This can lead to service unavailability. 34 of processes can be limited. Through this controller different rdma 43 by rdma cgroup, which can be extended later if required. 60 Whenever RDMA resource charging occurs, owner rdma cgroup is returned to 86 Following resources can be accounted by rdma controller. 98 echo mlx4_0 hca_handle=2 hca_object=2000 > /sys/fs/cgroup/rdma/1/rdma.max 99 echo ocrdma1 hca_handle=3 > /sys/fs/cgroup/rdma/2/rdma.max 103 cat /sys/fs/cgroup/rdma/2/rdma.max 110 cat /sys/fs/cgroup/rdma/2/rdma.current [all …]
|
| /linux/Documentation/devicetree/bindings/media/ |
| A D | mediatek,mdp3-rdma.yaml | 4 $id: http://devicetree.org/schemas/media/mediatek,mdp3-rdma.yaml# 25 - mediatek,mt8183-mdp3-rdma 26 - mediatek,mt8188-mdp3-rdma 27 - mediatek,mt8195-mdp3-rdma 28 - mediatek,mt8195-vdo1-rdma 30 - const: mediatek,mt8188-vdo1-rdma 31 - const: mediatek,mt8195-vdo1-rdma 104 const: mediatek,mt8183-mdp3-rdma 122 const: mediatek,mt8195-mdp3-rdma 139 const: mediatek,mt8195-vdo1-rdma [all …]
|
| A D | mediatek-mdp.txt | 12 "mediatek,mt8173-mdp-rdma" - read DMA 24 "mediatek,mt8173-mdp-rdma" 32 mdp_rdma0: rdma@14001000 { 33 compatible = "mediatek,mt8173-mdp-rdma"; 43 mdp_rdma1: rdma@14002000 { 44 compatible = "mediatek,mt8173-mdp-rdma";
|
| /linux/Documentation/ABI/stable/ |
| A D | sysfs-driver-ib_srp | 4 Contact: linux-rdma@vger.kernel.org 77 Contact: linux-rdma@vger.kernel.org 83 Contact: linux-rdma@vger.kernel.org 89 Contact: linux-rdma@vger.kernel.org 97 Contact: linux-rdma@vger.kernel.org 104 Contact: linux-rdma@vger.kernel.org 111 Contact: linux-rdma@vger.kernel.org 117 Contact: linux-rdma@vger.kernel.org 124 Contact: linux-rdma@vger.kernel.org 131 Contact: linux-rdma@vger.kernel.org [all …]
|
| A D | sysfs-class-infiniband | 9 Contact: linux-rdma@vger.kernel.org 24 Contact: linux-rdma@vger.kernel.org 34 Contact: linux-rdma@vger.kernel.org 49 Contact: linux-rdma@vger.kernel.org 80 Contact: linux-rdma@vger.kernel.org 108 Contact: linux-rdma@vger.kernel.org 203 Contact: linux-rdma@vger.kernel.org 222 Contact: linux-rdma@vger.kernel.org 229 Contact: linux-rdma@vger.kernel.org 570 linux-rdma@vger.kernel.org [all …]
|
| A D | sysfs-transport-srp | 4 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 11 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 20 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 36 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 50 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org
|
| /linux/drivers/net/ethernet/seeq/ |
| A D | sgiseeq.c | 69 volatile struct hpc_dma_desc rdma; member 213 sp->rx_desc[i].rdma.pbuf = dma_addr; in seeq_init_ring() 263 i, (&r[i]), r[i].rdma.pbuf, r[i].rdma.cntinfo, in sgiseeq_dump_rings() 264 r[i].rdma.pnext); in sgiseeq_dump_rings() 267 i, (&r[i]), r[i].rdma.pbuf, r[i].rdma.cntinfo, in sgiseeq_dump_rings() 268 r[i].rdma.pnext); in sgiseeq_dump_rings() 355 while (!(rd->rdma.cntinfo & HPCDMA_OWN)) { in sgiseeq_rx() 404 rd->rdma.cntinfo = RCNTINFO_INIT; in sgiseeq_rx() 711 buf[i].rdma.pbuf = 0; in setup_rx_ring() 715 buf[i].rdma.pbuf = 0; in setup_rx_ring() [all …]
|
| /linux/drivers/infiniband/ulp/rtrs/ |
| A D | rtrs-clt-stats.c | 35 this_cpu_inc(stats->pcpu_stats->rdma.failover_cnt); in rtrs_clt_inc_failover_cnt() 90 r = &per_cpu_ptr(stats->pcpu_stats, cpu)->rdma; in rtrs_clt_stats_rdma_to_str() 120 memset(&s->rdma, 0, sizeof(s->rdma)); in rtrs_clt_reset_rdma_stats() 168 this_cpu_inc(stats->pcpu_stats->rdma.dir[d].cnt); in rtrs_clt_update_rdma_stats() 169 this_cpu_add(stats->pcpu_stats->rdma.dir[d].size_total, size); in rtrs_clt_update_rdma_stats()
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| A D | cxgb3_offload.c | 281 struct rdma_info *rdma = data; in cxgb_rdma_ctl() local 286 rdma->tpt_base = in cxgb_rdma_ctl() 289 rdma->pbl_base = in cxgb_rdma_ctl() 295 rdma->pdev = pdev; in cxgb_rdma_ctl() 300 struct rdma_cq_op *rdma = data; in cxgb_rdma_ctl() local 304 ret = t3_sge_cqcntxt_op(adapter, rdma->id, rdma->op, in cxgb_rdma_ctl() 305 rdma->credits); in cxgb_rdma_ctl() 337 rdma->base_addr, rdma->size, in cxgb_rdma_ctl() 339 rdma->ovfl_mode, rdma->credits, in cxgb_rdma_ctl() 340 rdma->credit_thres); in cxgb_rdma_ctl() [all …]
|
| /linux/drivers/media/platform/mediatek/mdp3/ |
| A D | mtk-mdp3-comp.c | 116 reg = CFG_COMP(MT8183, ctx->param, rdma.src_ctrl); in config_rdma_frame() 118 reg = CFG_COMP(MT8195, ctx->param, rdma.src_ctrl); in config_rdma_frame() 154 reg = CFG_COMP(MT8183, ctx->param, rdma.control); in config_rdma_frame() 157 reg = CFG_COMP(MT8195, ctx->param, rdma.control); in config_rdma_frame() 165 reg = CFG_COMP(MT8183, ctx->param, rdma.iova[0]); in config_rdma_frame() 167 reg = CFG_COMP(MT8195, ctx->param, rdma.iova[0]); in config_rdma_frame() 172 reg = CFG_COMP(MT8183, ctx->param, rdma.iova[1]); in config_rdma_frame() 174 reg = CFG_COMP(MT8195, ctx->param, rdma.iova[1]); in config_rdma_frame() 179 reg = CFG_COMP(MT8183, ctx->param, rdma.iova[2]); in config_rdma_frame() 181 reg = CFG_COMP(MT8195, ctx->param, rdma.iova[2]); in config_rdma_frame() [all …]
|
| /linux/drivers/nvme/target/ |
| A D | Makefile | 7 obj-$(CONFIG_NVME_TARGET_RDMA) += nvmet-rdma.o 19 nvmet-rdma-y += rdma.o
|
| /linux/drivers/nvme/host/ |
| A D | Makefile | 8 obj-$(CONFIG_NVME_RDMA) += nvme-rdma.o 26 nvme-rdma-y += rdma.o
|
| /linux/drivers/infiniband/ |
| A D | Kconfig | 25 need libibumad from rdma-core 26 <https://github.com/linux-rdma/rdma-core>. 38 rdma-core <https://github.com/linux-rdma/rdma-core>.
|
| /linux/drivers/macintosh/ |
| A D | rack-meter.c | 146 struct rackmeter_dma *rdma = rm->dma_buf_v; in rackmeter_do_pause() local 155 memset(rdma->buf1, 0, sizeof(rdma->buf1)); in rackmeter_do_pause() 156 memset(rdma->buf2, 0, sizeof(rdma->buf2)); in rackmeter_do_pause() 371 struct resource ri2s, rdma; in rackmeter_probe() local 431 of_address_to_resource(i2s, 1, &rdma)) { in rackmeter_probe() 441 pr_debug(" dma @0x%08x\n", (unsigned int)rdma.start); in rackmeter_probe() 475 rm->dma_regs = ioremap(rdma.start, 0x100); in rackmeter_probe()
|
| /linux/fs/smb/client/ |
| A D | smbdirect.c | 51 struct smb_extract_to_rdma *rdma); 2415 struct ib_sge *sge = &rdma->sge[rdma->nr_sge]; in smb_set_sge() 2426 rdma->nr_sge++; in smb_set_sge() 2460 if (rdma->nr_sge >= rdma->max_sge || maxsize <= 0) in smb_extract_bvec_to_rdma() 2516 } while (len > 0 && rdma->nr_sge < rdma->max_sge); in smb_extract_kvec_to_rdma() 2518 if (rdma->nr_sge >= rdma->max_sge || maxsize <= 0) in smb_extract_kvec_to_rdma() 2576 } while (rdma->nr_sge < rdma->max_sge || maxsize > 0); in smb_extract_folioq_to_rdma() 2600 int before = rdma->nr_sge; in smb_extract_iter_to_rdma() 2618 while (rdma->nr_sge > before) { in smb_extract_iter_to_rdma() 2619 struct ib_sge *sge = &rdma->sge[rdma->nr_sge--]; in smb_extract_iter_to_rdma() [all …]
|