Home
last modified time | relevance | path

Searched refs:ibmr (Results 1 – 25 of 59) sorted by relevance

123

/drivers/infiniband/hw/vmw_pvrdma/
A Dpvrdma_mr.c94 mr->ibmr.lkey = resp->lkey; in pvrdma_get_dma_mr()
95 mr->ibmr.rkey = resp->rkey; in pvrdma_get_dma_mr()
97 return &mr->ibmr; in pvrdma_get_dma_mr()
187 mr->ibmr.lkey = resp->lkey; in pvrdma_reg_user_mr()
188 mr->ibmr.rkey = resp->rkey; in pvrdma_reg_user_mr()
190 return &mr->ibmr; in pvrdma_reg_user_mr()
259 mr->ibmr.lkey = resp->lkey; in pvrdma_alloc_mr()
260 mr->ibmr.rkey = resp->rkey; in pvrdma_alloc_mr()
264 return &mr->ibmr; in pvrdma_alloc_mr()
284 struct pvrdma_user_mr *mr = to_vmr(ibmr); in pvrdma_dereg_mr()
[all …]
A Dpvrdma.h141 struct ib_mr ibmr; member
284 static inline struct pvrdma_user_mr *to_vmr(struct ib_mr *ibmr) in to_vmr() argument
286 return container_of(ibmr, struct pvrdma_user_mr, ibmr); in to_vmr()
/drivers/infiniband/hw/mlx4/
A Dmr.c76 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_get_dma_mr()
79 return &mr->ibmr; in mlx4_ib_get_dma_mr()
183 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_reg_user_mr()
184 mr->ibmr.page_size = 1U << shift; in mlx4_ib_reg_user_mr()
186 return &mr->ibmr; in mlx4_ib_reg_user_mr()
340 struct mlx4_ib_mr *mr = to_mmr(ibmr); in mlx4_ib_dereg_mr()
415 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_alloc_mr()
418 return &mr->ibmr; in mlx4_ib_alloc_mr()
421 mr->ibmr.device = pd->device; in mlx4_ib_alloc_mr()
432 struct mlx4_ib_mr *mr = to_mmr(ibmr); in mlx4_set_page()
[all …]
/drivers/infiniband/hw/mlx5/
A Dmr.c1090 return &mr->ibmr; in mlx5_ib_get_dma_mr()
1187 mr->ibmr.pd = pd; in alloc_cacheable_mr()
1240 mr->ibmr.pd = pd; in reg_create_crossing_vhca_mr()
1244 return &mr->ibmr; in reg_create_crossing_vhca_mr()
1279 mr->ibmr.pd = pd; in reg_create()
1398 return &mr->ibmr; in mlx5_ib_get_dm_mr()
1514 return &mr->ibmr; in create_real_mr()
1571 return &mr->ibmr; in create_user_odp_mr()
2813 pi_mr->ibmr.page_size = ibmr->page_size; in mlx5_ib_map_mtt_mr_sg_pi()
2889 ibmr->length = pi_mr->ibmr.length; in mlx5_ib_map_klm_mr_sg_pi()
[all …]
A Drestrack.c81 static int fill_stat_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in fill_stat_mr_entry() argument
83 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_stat_mr_entry()
124 static int fill_res_mr_entry_raw(struct sk_buff *msg, struct ib_mr *ibmr) in fill_res_mr_entry_raw() argument
126 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_res_mr_entry_raw()
132 static int fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in fill_res_mr_entry() argument
134 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_res_mr_entry()
A Dwr.c206 seg->start_addr = cpu_to_be64(mr->ibmr.iova); in set_reg_mkey_seg()
207 seg->len = cpu_to_be64(mr->ibmr.length); in set_reg_mkey_seg()
436 data_key = pi_mr->ibmr.lkey; in set_sig_data_segment()
440 prot_key = pi_mr->ibmr.lkey; in set_sig_data_segment()
576 region_len = pi_mr->ibmr.length; in set_pi_umr_wr()
867 reg_pi_wr.mr = &pi_mr->ibmr; in handle_reg_mr_integrity()
869 reg_pi_wr.key = pi_mr->ibmr.rkey; in handle_reg_mr_integrity()
889 pa_pi_mr.ibmr.lkey = mr->ibmr.pd->local_dma_lkey; in handle_reg_mr_integrity()
899 pa_pi_mr.ibmr.length = mr->ibmr.length; in handle_reg_mr_integrity()
902 (*ctrl)->imm = cpu_to_be32(mr->ibmr.rkey); in handle_reg_mr_integrity()
[all …]
A Dodp.c232 mlx5_ib_dereg_mr(&mr->ibmr, NULL); in free_implicit_child_mr_work()
493 mr->ibmr.pd = imr->ibmr.pd; in implicit_get_child_mr()
496 mr->ibmr.lkey = mr->mmkey.key; in implicit_get_child_mr()
497 mr->ibmr.rkey = mr->mmkey.key; in implicit_get_child_mr()
629 imr->ibmr.pd = &pd->ibpd; in mlx5_ib_alloc_implicit_mr()
630 imr->ibmr.iova = 0; in mlx5_ib_alloc_implicit_mr()
632 imr->ibmr.lkey = imr->mmkey.key; in mlx5_ib_alloc_implicit_mr()
633 imr->ibmr.rkey = imr->mmkey.key; in mlx5_ib_alloc_implicit_mr()
634 imr->ibmr.device = &dev->ib_dev; in mlx5_ib_alloc_implicit_mr()
912 u64 offset = io_virt < mr->ibmr.iova ? 0 : io_virt - mr->ibmr.iova; in pagefault_mr()
[all …]
/drivers/infiniband/sw/rxe/
A Drxe_mr.c29 switch (mr->ibmr.type) { in mr_check_range()
35 if (iova < mr->ibmr.iova || in mr_check_range()
36 iova + length > mr->ibmr.iova + mr->ibmr.length) { in mr_check_range()
57 mr->lkey = mr->ibmr.lkey = key; in rxe_mr_init()
58 mr->rkey = mr->ibmr.rkey = key; in rxe_mr_init()
61 mr->ibmr.page_size = PAGE_SIZE; in rxe_mr_init()
72 mr->ibmr.type = IB_MR_TYPE_DMA; in rxe_mr_init_dma()
153 mr->ibmr.type = IB_MR_TYPE_USER; in rxe_mr_init_user()
200 mr->ibmr.type = IB_MR_TYPE_MEM_REG; in rxe_mr_init_fast()
210 struct rxe_mr *mr = to_rmr(ibmr); in rxe_set_page()
[all …]
A Drxe_verbs.c1259 mr->ibmr.pd = ibpd; in rxe_get_dma_mr()
1260 mr->ibmr.device = ibpd->device; in rxe_get_dma_mr()
1264 return &mr->ibmr; in rxe_get_dma_mr()
1302 mr->ibmr.pd = ibpd; in rxe_reg_user_mr()
1303 mr->ibmr.device = ibpd->device; in rxe_reg_user_mr()
1315 return &mr->ibmr; in rxe_reg_user_mr()
1332 struct rxe_mr *mr = to_rmr(ibmr); in rxe_rereg_user_mr()
1347 mr->ibmr.pd = ibpd; in rxe_rereg_user_mr()
1385 mr->ibmr.pd = ibpd; in rxe_alloc_mr()
1386 mr->ibmr.device = ibpd->device; in rxe_alloc_mr()
[all …]
A Drxe_mw.c117 if (unlikely(wqe->wr.wr.mw.length > mr->ibmr.length)) { in rxe_check_bind_mw()
123 if (unlikely((wqe->wr.wr.mw.addr < mr->ibmr.iova) || in rxe_check_bind_mw()
125 (mr->ibmr.iova + mr->ibmr.length)))) { in rxe_check_bind_mw()
A Drxe_verbs.h340 struct ib_mr ibmr; member
362 return mr ? mr->ibmr.page_size : PAGE_SIZE; in mr_page_size()
493 return mr ? container_of(mr, struct rxe_mr, ibmr) : NULL; in to_rmr()
508 return to_rpd(mr->ibmr.pd); in mr_pd()
A Drxe.h55 #define rxe_dbg_mr(mr, fmt, ...) ibdev_dbg((mr)->ibmr.device, \
76 #define rxe_err_mr(mr, fmt, ...) ibdev_err_ratelimited((mr)->ibmr.device, \
97 #define rxe_info_mr(mr, fmt, ...) ibdev_info_ratelimited((mr)->ibmr.device, \
/drivers/infiniband/sw/rdmavt/
A Dmr.h12 struct ib_mr ibmr; member
17 static inline struct rvt_mr *to_imr(struct ib_mr *ibmr) in to_imr() argument
19 return container_of(ibmr, struct rvt_mr, ibmr); in to_imr()
31 int rvt_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata);
34 int rvt_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg,
A Dmr.c259 mr->ibmr.lkey = mr->mr.lkey; in __rvt_alloc_mr()
260 mr->ibmr.rkey = mr->mr.lkey; in __rvt_alloc_mr()
314 ret = &mr->ibmr; in rvt_get_dma_mr()
392 return &mr->ibmr; in rvt_reg_user_mr()
516 struct rvt_mr *mr = to_imr(ibmr); in rvt_dereg_mr()
552 return &mr->ibmr; in rvt_alloc_mr()
602 mr->mr.user_base = ibmr->iova; in rvt_map_mr_sg()
603 mr->mr.iova = ibmr->iova; in rvt_map_mr_sg()
634 ibmr->lkey = key; in rvt_fast_reg_mr()
635 ibmr->rkey = key; in rvt_fast_reg_mr()
[all …]
A Dtrace_mr.h144 TP_PROTO(struct ib_mr *ibmr, int sg_nents, unsigned int *sg_offset),
145 TP_ARGS(ibmr, sg_nents, sg_offset),
147 RDI_DEV_ENTRY(ib_to_rvt(to_imr(ibmr)->mr.pd->device))
156 RDI_DEV_ASSIGN(ib_to_rvt(to_imr(ibmr)->mr.pd->device));
157 __entry->ibmr_iova = ibmr->iova;
158 __entry->iova = to_imr(ibmr)->mr.iova;
159 __entry->user_base = to_imr(ibmr)->mr.user_base;
160 __entry->ibmr_length = to_imr(ibmr)->mr.length;
/drivers/infiniband/hw/hns/
A Dhns_roce_mr.c221 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_get_dma_mr()
223 return &mr->ibmr; in hns_roce_get_dma_mr()
270 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_reg_user_mr()
272 return &mr->ibmr; in hns_roce_reg_user_mr()
293 struct hns_roce_mr *mr = to_hr_mr(ibmr); in hns_roce_rereg_user_mr()
370 struct hns_roce_mr *mr = to_hr_mr(ibmr); in hns_roce_dereg_mr()
419 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_alloc_mr()
420 mr->ibmr.length = mr->size; in hns_roce_alloc_mr()
422 return &mr->ibmr; in hns_roce_alloc_mr()
435 struct hns_roce_mr *mr = to_hr_mr(ibmr); in hns_roce_set_page()
[all …]
/drivers/infiniband/hw/mana/
A Dmr.c76 mr->ibmr.lkey = resp.lkey; in mana_ib_gd_create_mr()
77 mr->ibmr.rkey = resp.rkey; in mana_ib_gd_create_mr()
180 return &mr->ibmr; in mana_ib_reg_user_mr()
253 return &mr->ibmr; in mana_ib_reg_user_mr_dmabuf()
291 return &mr->ibmr; in mana_ib_get_dma_mr()
298 int mana_ib_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) in mana_ib_dereg_mr() argument
300 struct mana_ib_mr *mr = container_of(ibmr, struct mana_ib_mr, ibmr); in mana_ib_dereg_mr()
301 struct ib_device *ibdev = ibmr->device; in mana_ib_dereg_mr()
/drivers/infiniband/hw/usnic/
A Dusnic_ib.h69 struct ib_mr ibmr; member
125 struct usnic_ib_mr *to_umr(struct ib_mr *ibmr) in to_umr() argument
127 return container_of(ibmr, struct usnic_ib_mr, ibmr); in to_umr()
/drivers/infiniband/hw/cxgb4/
A Dmem.c384 mhp->ibmr.rkey = mhp->ibmr.lkey = stag; in finish_mem_reg()
385 mhp->ibmr.length = mhp->attr.len; in finish_mem_reg()
386 mhp->ibmr.page_size = 1U << (mhp->attr.page_size + 12); in finish_mem_reg()
478 return &mhp->ibmr; in c4iw_get_dma_mr()
586 return &mhp->ibmr; in c4iw_reg_user_mr()
657 mhp->ibmr.rkey = mhp->ibmr.lkey = stag; in c4iw_alloc_mr()
664 return &(mhp->ibmr); in c4iw_alloc_mr()
682 static int c4iw_set_page(struct ib_mr *ibmr, u64 addr) in c4iw_set_page() argument
684 struct c4iw_mr *mhp = to_c4iw_mr(ibmr); in c4iw_set_page()
697 struct c4iw_mr *mhp = to_c4iw_mr(ibmr); in c4iw_map_mr_sg()
[all …]
/drivers/infiniband/hw/mthca/
A Dmthca_provider.h74 struct ib_mr ibmr; member
286 static inline struct mthca_mr *to_mmr(struct ib_mr *ibmr) in to_mmr() argument
288 return container_of(ibmr, struct mthca_mr, ibmr); in to_mmr()
/drivers/infiniband/hw/erdma/
A Derdma_verbs.c1097 mr->ibmr.lkey = stag; in erdma_get_dma_mr()
1098 mr->ibmr.rkey = stag; in erdma_get_dma_mr()
1099 mr->ibmr.pd = ibpd; in erdma_get_dma_mr()
1105 return &mr->ibmr; in erdma_get_dma_mr()
1141 mr->ibmr.lkey = stag; in erdma_ib_alloc_mr()
1143 mr->ibmr.pd = ibpd; in erdma_ib_alloc_mr()
1160 return &mr->ibmr; in erdma_ib_alloc_mr()
1230 mr->ibmr.lkey = mr->ibmr.rkey = stag; in erdma_reg_user_mr()
1231 mr->ibmr.pd = ibpd; in erdma_reg_user_mr()
1242 return &mr->ibmr; in erdma_reg_user_mr()
[all …]
A Derdma_verbs.h125 struct ib_mr ibmr; member
394 static inline struct erdma_mr *to_emr(struct ib_mr *ibmr) in to_emr() argument
396 return container_of(ibmr, struct erdma_mr, ibmr); in to_emr()
458 int erdma_dereg_mr(struct ib_mr *ibmr, struct ib_udata *data);
472 int erdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents,
/drivers/infiniband/hw/irdma/
A Dverbs.c2701 iwmr->ibmr.rkey = stag; in irdma_alloc_mr()
2702 iwmr->ibmr.lkey = stag; in irdma_alloc_mr()
2703 iwmr->ibmr.pd = pd; in irdma_alloc_mr()
2723 return &iwmr->ibmr; in irdma_alloc_mr()
2912 iwmr->ibmr.pd = pd; in irdma_alloc_iwmr()
2914 iwmr->ibmr.iova = virt; in irdma_alloc_iwmr()
3082 return &iwmr->ibmr; in irdma_reg_user_mr()
3124 return &iwmr->ibmr; in irdma_reg_user_mr_dmabuf()
3203 iwmr->ibmr.pd = pd; in irdma_rereg_mr_trans()
3311 iwmr->ibmr.pd = pd; in irdma_reg_phys_mr()
[all …]
/drivers/infiniband/hw/ocrdma/
A Docrdma.h192 struct ib_mr ibmr; member
481 static inline struct ocrdma_mr *get_ocrdma_mr(struct ib_mr *ibmr) in get_ocrdma_mr() argument
483 return container_of(ibmr, struct ocrdma_mr, ibmr); in get_ocrdma_mr()
/drivers/i2c/busses/
A Di2c-pxa.c136 u32 ibmr; member
158 .ibmr = 0x00,
167 .ibmr = 0x00,
176 .ibmr = 0x14,
185 .ibmr = 0x00,
196 .ibmr = 0x00,
1297 u32 ibmr = readl(_IBMR(i2c)); in i2c_pxa_prepare_recovery() local
1303 gpiod_set_value(i2c->recovery.scl_gpiod, ibmr & IBMR_SCLS); in i2c_pxa_prepare_recovery()
1304 gpiod_set_value(i2c->recovery.sda_gpiod, ibmr & IBMR_SDAS); in i2c_pxa_prepare_recovery()
1469 i2c->reg_ibmr = i2c->reg_base + pxa_reg_layout[i2c_type].ibmr; in i2c_pxa_probe()

Completed in 101 milliseconds

123