Home
last modified time | relevance | path

Searched refs:mw (Results 1 – 25 of 29) sorted by relevance

12

/drivers/infiniband/sw/rxe/
A Drxe_mw.c32 mw->state = (mw->ibmw.type == IB_MW_TYPE_2) ? in rxe_alloc_mw()
124 ((wqe->wr.wr.mw.addr + wqe->wr.wr.mw.length) > in rxe_check_bind_mw()
140 mw->rkey = (mw->rkey & ~0xff) | key; in rxe_do_bind_mw()
143 mw->addr = wqe->wr.wr.mw.addr; in rxe_do_bind_mw()
144 mw->length = wqe->wr.wr.mw.length; in rxe_do_bind_mw()
219 rxe_put(mw); in rxe_bind_mw()
242 qp = mw->qp; in rxe_do_invalidate_mw()
247 mr = mw->mr; in rxe_do_invalidate_mw()
265 if (!mw) { in rxe_invalidate_mw()
298 if (!mw) in rxe_lookup_mw()
[all …]
A Drxe.h57 #define rxe_dbg_mw(mw, fmt, ...) ibdev_dbg((mw)->ibmw.device, \ argument
58 "mw#%d %s: " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__)
78 #define rxe_err_mw(mw, fmt, ...) ibdev_err_ratelimited((mw)->ibmw.device, \ argument
79 "mw#%d %s: " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__)
99 #define rxe_info_mw(mw, fmt, ...) ibdev_info_ratelimited((mw)->ibmw.device, \ argument
100 "mw#%d %s: " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__)
A Drxe_resp.c487 if (!mw) { in check_rkey()
493 mr = mw->mr; in check_rkey()
504 rxe_put(mw); in check_rkey()
505 mw = NULL; in check_rkey()
559 if (mw) in check_rkey()
560 rxe_put(mw); in check_rkey()
852 if (!mw) in rxe_recheck_mr()
855 mr = mw->mr; in rxe_recheck_mr()
856 if (mw->rkey != rkey || mw->state != RXE_MW_STATE_VALID || in rxe_recheck_mr()
858 rxe_put(mw); in rxe_recheck_mr()
[all …]
A Drxe_verbs.h496 static inline struct rxe_mw *to_rmw(struct ib_mw *mw) in to_rmw() argument
498 return mw ? container_of(mw, struct rxe_mw, ibmw) : NULL; in to_rmw()
511 static inline struct rxe_pd *rxe_mw_pd(struct rxe_mw *mw) in rxe_mw_pd() argument
513 return to_rpd(mw->ibmw.pd); in rxe_mw_pd()
/drivers/ntb/
A Dntb_transport.c633 if (!mw->virt_addr) in ntb_transport_setup_qp_mw()
798 if (!mw->virt_addr) in ntb_free_mw()
803 mw->alloc_addr, mw->dma_addr); in ntb_free_mw()
804 mw->xlat_size = 0; in ntb_free_mw()
805 mw->buff_size = 0; in ntb_free_mw()
835 mw->alloc_size); in ntb_alloc_mw_buffer()
847 if (mw->alloc_size > mw->buff_size) { in ntb_alloc_mw_buffer()
893 if (mw->buff_size) in ntb_set_mw()
1339 mw->vbase = ioremap_wc(mw->phys_addr, mw->phys_size); in ntb_transport_probe()
1340 if (!mw->vbase) { in ntb_transport_probe()
[all …]
/drivers/net/ethernet/mellanox/mlx4/
A Dmr.c815 struct mlx4_mw *mw) in mlx4_mw_alloc() argument
830 mw->pd = pd; in mlx4_mw_alloc()
831 mw->type = type; in mlx4_mw_alloc()
832 mw->enabled = MLX4_MPT_DISABLED; in mlx4_mw_alloc()
860 if (mw->type == MLX4_MW_TYPE_2) { in mlx4_mw_enable()
867 key_to_hw_index(mw->key) & in mlx4_mw_enable()
873 mw->enabled = MLX4_MPT_EN_HW; in mlx4_mw_enable()
892 if (mw->enabled == MLX4_MPT_EN_HW) { in mlx4_mw_free()
894 key_to_hw_index(mw->key) & in mlx4_mw_free()
899 mw->enabled = MLX4_MPT_EN_SW; in mlx4_mw_free()
[all …]
/drivers/net/ethernet/marvell/octeontx2/af/
A Drvu.c2216 mbox = &mw->mbox; in __rvu_mbox_handler()
2323 mbox = &mw->mbox_up; in __rvu_mbox_up_handler()
2536 if (!mw->mbox_wq) { in rvu_mbox_init()
2543 if (!mw->mbox_wrk) { in rvu_mbox_init()
2550 if (!mw->mbox_wrk_up) { in rvu_mbox_init()
2605 if (mw->mbox_wq) { in rvu_mbox_destroy()
2607 mw->mbox_wq = NULL; in rvu_mbox_destroy()
2633 mbox = &mw->mbox; in rvu_queue_work()
2648 queue_work(mw->mbox_wq, &mw->mbox_wrk[i].work); in rvu_queue_work()
2650 mbox = &mw->mbox_up; in rvu_queue_work()
[all …]
A Drvu.h479 void (*rvu_queue_work_hdlr)(struct mbox_wq_info *mw, int first,
481 void (*afvf_queue_work_hdlr)(struct mbox_wq_info *mw, int first,
961 void rvu_queue_work(struct mbox_wq_info *mw, int first,
/drivers/clk/rockchip/
A Dclk.h717 .mux_width = mw, \
738 .mux_width = mw, \
786 #define COMPOSITE_NODIV(_id, cname, pnames, f, mo, ms, mw, mf, \ argument
797 .mux_width = mw, \
815 .mux_width = mw, \
824 mw, mf, ds, dw, df, dt) \ argument
834 .mux_width = mw, \
894 #define COMPOSITE_DDRCLK(_id, cname, pnames, f, mo, ms, mw, \ argument
905 .mux_width = mw, \
1103 .mux_width = mw, \
[all …]
/drivers/net/ethernet/marvell/octeontx2/nic/
A Dcn20k.c28 struct mbox *mw = &pf->mbox; in cn20k_pfaf_mbox_intr_handler() local
40 mbox = &mw->mbox_up; in cn20k_pfaf_mbox_intr_handler()
46 queue_work(pf->mbox_wq, &mw->mbox_up_wrk); in cn20k_pfaf_mbox_intr_handler()
53 mbox = &mw->mbox; in cn20k_pfaf_mbox_intr_handler()
59 queue_work(pf->mbox_wq, &mw->mbox_wrk); in cn20k_pfaf_mbox_intr_handler()
A Dotx2_pf.c312 mbox = &mw->mbox; in otx2_queue_vf_work()
323 mw[i].num_msgs = hdr->num_msgs; in otx2_queue_vf_work()
325 queue_work(mbox_wq, &mw[i].mbox_wrk); in otx2_queue_vf_work()
328 mbox = &mw->mbox_up; in otx2_queue_vf_work()
332 mw[i].up_num_msgs = hdr->num_msgs; in otx2_queue_vf_work()
334 queue_work(mbox_wq, &mw[i].mbox_up_wrk); in otx2_queue_vf_work()
1047 struct mbox *mw = &pf->mbox; in otx2_pfaf_mbox_intr_handler() local
1063 mbox = &mw->mbox_up; in otx2_pfaf_mbox_intr_handler()
1069 queue_work(pf->mbox_wq, &mw->mbox_up_wrk); in otx2_pfaf_mbox_intr_handler()
1082 mbox = &mw->mbox; in otx2_pfaf_mbox_intr_handler()
[all …]
/drivers/infiniband/hw/mlx4/
A Dmr.c358 struct mlx4_ib_mw *mw = to_mmw(ibmw); in mlx4_ib_alloc_mw() local
362 to_mlx4_type(ibmw->type), &mw->mmw); in mlx4_ib_alloc_mw()
366 err = mlx4_mw_enable(dev->dev, &mw->mmw); in mlx4_ib_alloc_mw()
370 ibmw->rkey = mw->mmw.key; in mlx4_ib_alloc_mw()
374 mlx4_mw_free(dev->dev, &mw->mmw); in mlx4_ib_alloc_mw()
380 struct mlx4_ib_mw *mw = to_mmw(ibmw); in mlx4_ib_dealloc_mw() local
382 mlx4_mw_free(to_mdev(ibmw->device)->dev, &mw->mmw); in mlx4_ib_dealloc_mw()
A Dmlx4_ib.h765 int mlx4_ib_alloc_mw(struct ib_mw *mw, struct ib_udata *udata);
766 int mlx4_ib_dealloc_mw(struct ib_mw *mw);
/drivers/mtd/maps/
A Dphysmap-core.c147 map_word mw; in physmap_addr_gpios_read() local
155 mw.x[0] = word; in physmap_addr_gpios_read()
156 return mw; in physmap_addr_gpios_read()
181 static void physmap_addr_gpios_write(struct map_info *map, map_word mw, in physmap_addr_gpios_write() argument
192 word = mw.x[0]; in physmap_addr_gpios_write()
/drivers/pci/endpoint/functions/
A Dpci-epf-ntb.c236 enum pci_epc_interface_type type, u32 mw) in epf_ntb_configure_mw() argument
252 peer_barno = peer_ntb_epc->epf_ntb_bar[mw + NTB_MW_OFFSET]; in epf_ntb_configure_mw()
259 if (mw + NTB_MW_OFFSET == BAR_DB_MW1) in epf_ntb_configure_mw()
262 if (size > ntb->mws_size[mw]) { in epf_ntb_configure_mw()
265 pci_epc_interface_string(type), mw, size, in epf_ntb_configure_mw()
266 ntb->mws_size[mw]); in epf_ntb_configure_mw()
278 pci_epc_interface_string(type), mw); in epf_ntb_configure_mw()
295 enum pci_epc_interface_type type, u32 mw) in epf_ntb_teardown_mw() argument
309 peer_barno = peer_ntb_epc->epf_ntb_bar[mw + NTB_MW_OFFSET]; in epf_ntb_teardown_mw()
314 if (mw + NTB_MW_OFFSET == BAR_DB_MW1) in epf_ntb_teardown_mw()
A Dpci-epf-vntb.c204 static int epf_ntb_configure_mw(struct epf_ntb *ntb, u32 mw) in epf_ntb_configure_mw() argument
211 phys_addr = ntb->vpci_mw_phy[mw]; in epf_ntb_configure_mw()
221 "Failed to map memory window %d address\n", mw); in epf_ntb_configure_mw()
233 static void epf_ntb_teardown_mw(struct epf_ntb *ntb, u32 mw) in epf_ntb_teardown_mw() argument
238 ntb->vpci_mw_phy[mw]); in epf_ntb_teardown_mw()
/drivers/infiniband/hw/bnxt_re/
A Dib_verbs.c531 if (fence->mw) { in bnxt_re_destroy_fence_mr()
533 fence->mw = NULL; in bnxt_re_destroy_fence_mr()
559 struct ib_mw *mw; in bnxt_re_create_fence_mr() local
611 if (IS_ERR(mw)) { in bnxt_re_create_fence_mr()
614 rc = PTR_ERR(mw); in bnxt_re_create_fence_mr()
617 fence->mw = mw; in bnxt_re_create_fence_mr()
4119 mw = kzalloc(sizeof(*mw), GFP_KERNEL); in bnxt_re_alloc_mw()
4120 if (!mw) in bnxt_re_alloc_mw()
4133 mw->ib_mw.rkey = mw->qplib_mw.rkey; in bnxt_re_alloc_mw()
4141 kfree(mw); in bnxt_re_alloc_mw()
[all …]
A Dib_verbs.h53 struct ib_mw *mw; member
258 int bnxt_re_dealloc_mw(struct ib_mw *mw);
/drivers/media/platform/nxp/dw100/
A Ddw100.c382 u32 sw, sh, mw, mh, idx; in dw100_ctrl_dewarping_map_init() local
389 mw = ctrl->dims[0]; in dw100_ctrl_dewarping_map_init()
394 qdx = qsw / (mw - 1); in dw100_ctrl_dewarping_map_init()
397 ctx->map_width = mw; in dw100_ctrl_dewarping_map_init()
399 ctx->map_size = mh * mw * sizeof(u32); in dw100_ctrl_dewarping_map_init()
402 qy = min_t(u32, (idx / mw) * qdy, qsh); in dw100_ctrl_dewarping_map_init()
403 qx = min_t(u32, (idx % mw) * qdx, qsw); in dw100_ctrl_dewarping_map_init()
/drivers/infiniband/core/
A Duverbs_main.c118 int uverbs_dealloc_mw(struct ib_mw *mw) in uverbs_dealloc_mw() argument
120 struct ib_pd *pd = mw->pd; in uverbs_dealloc_mw()
123 ret = mw->device->ops.dealloc_mw(mw); in uverbs_dealloc_mw()
128 kfree(mw); in uverbs_dealloc_mw()
A Duverbs.h216 int uverbs_dealloc_mw(struct ib_mw *mw);
A Duverbs_cmd.c926 struct ib_mw *mw; in ib_uverbs_alloc_mw() local
949 mw = rdma_zalloc_drv_obj(ib_dev, ib_mw); in ib_uverbs_alloc_mw()
950 if (!mw) { in ib_uverbs_alloc_mw()
955 mw->device = ib_dev; in ib_uverbs_alloc_mw()
956 mw->pd = pd; in ib_uverbs_alloc_mw()
957 mw->uobject = uobj; in ib_uverbs_alloc_mw()
958 mw->type = cmd.mw_type; in ib_uverbs_alloc_mw()
960 ret = pd->device->ops.alloc_mw(mw, &attrs->driver_udata); in ib_uverbs_alloc_mw()
966 uobj->object = mw; in ib_uverbs_alloc_mw()
970 resp.rkey = mw->rkey; in ib_uverbs_alloc_mw()
[all …]
/drivers/infiniband/hw/mlx5/
A Dmr.c2543 struct mlx5_ib_mw *mw = to_mmw(ibmw); in mlx5_ib_alloc_mw() local
2583 err = mlx5_ib_create_mkey(dev, &mw->mmkey, in, inlen); in mlx5_ib_alloc_mw()
2587 mw->mmkey.type = MLX5_MKEY_MW; in mlx5_ib_alloc_mw()
2588 ibmw->rkey = mw->mmkey.key; in mlx5_ib_alloc_mw()
2589 mw->mmkey.ndescs = ndescs; in mlx5_ib_alloc_mw()
2600 err = mlx5r_store_odp_mkey(dev, &mw->mmkey); in mlx5_ib_alloc_mw()
2609 mlx5_core_destroy_mkey(dev->mdev, mw->mmkey.key); in mlx5_ib_alloc_mw()
2615 int mlx5_ib_dealloc_mw(struct ib_mw *mw) in mlx5_ib_dealloc_mw() argument
2617 struct mlx5_ib_dev *dev = to_mdev(mw->device); in mlx5_ib_dealloc_mw()
2618 struct mlx5_ib_mw *mmw = to_mmw(mw); in mlx5_ib_dealloc_mw()
/drivers/ntb/test/
A Dntb_tool.c227 struct tool_mw *mw; member
869 struct tool_mw *outmw = outmw_wrap->mw; in tool_peer_mw_trans_read()
927 struct tool_mw *outmw = outmw_wrap->mw; in tool_peer_mw_trans_write()
999 tc->peers[pidx].outmws[widx].mw = &tc->outmws[widx]; in tool_init_mws()
/drivers/infiniband/hw/ocrdma/
A Docrdma_stats.c161 (u64)rsrc_stats->mw); in ocrdma_resource_stats()
215 (u64)rsrc_stats->mw); in ocrdma_resource_stats()

Completed in 142 milliseconds

12