Lines Matching refs:ibmr

29 	switch (mr->ibmr.type) {  in mr_check_range()
35 if (iova < mr->ibmr.iova || in mr_check_range()
36 iova + length > mr->ibmr.iova + mr->ibmr.length) { in mr_check_range()
62 mr->lkey = mr->ibmr.lkey = lkey; in rxe_mr_init()
63 mr->rkey = mr->ibmr.rkey = rkey; in rxe_mr_init()
66 mr->ibmr.page_size = PAGE_SIZE; in rxe_mr_init()
77 mr->ibmr.type = IB_MR_TYPE_DMA; in rxe_mr_init_dma()
82 return (iova >> mr->page_shift) - (mr->ibmr.iova >> mr->page_shift); in rxe_mr_iova_to_index()
158 mr->ibmr.type = IB_MR_TYPE_USER; in rxe_mr_init_user()
205 mr->ibmr.type = IB_MR_TYPE_MEM_REG; in rxe_mr_init_fast()
213 static int rxe_set_page(struct ib_mr *ibmr, u64 iova) in rxe_set_page() argument
215 struct rxe_mr *mr = to_rmr(ibmr); in rxe_set_page()
236 int rxe_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sgl, in rxe_map_mr_sg() argument
239 struct rxe_mr *mr = to_rmr(ibmr); in rxe_map_mr_sg()
245 mr->page_offset = mr->ibmr.iova & (page_size - 1); in rxe_map_mr_sg()
247 return ib_sg_to_pages(ibmr, sgl, sg_nents, sg_offset, rxe_set_page); in rxe_map_mr_sg()
320 if (mr->ibmr.type == IB_MR_TYPE_DMA) { in rxe_mr_copy()
445 if (mr->ibmr.type == IB_MR_TYPE_DMA) in rxe_flush_pmem_iova()
489 if (mr->ibmr.type == IB_MR_TYPE_DMA) { in rxe_mr_do_atomic_op()
546 if (mr->ibmr.type == IB_MR_TYPE_DMA) { in rxe_mr_do_atomic_write()
669 if (unlikely(mr->ibmr.type != IB_MR_TYPE_MEM_REG)) { in rxe_invalidate_mr()
670 rxe_dbg_mr(mr, "Type (%d) is wrong\n", mr->ibmr.type); in rxe_invalidate_mr()
704 if (unlikely(qp->ibqp.pd != mr->ibmr.pd)) { in rxe_reg_fast_mr()
719 mr->ibmr.iova = wqe->wr.wr.reg.mr->iova; in rxe_reg_fast_mr()
725 int rxe_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) in rxe_dereg_mr() argument
727 struct rxe_mr *mr = to_rmr(ibmr); in rxe_dereg_mr()
745 if (mr->ibmr.type != IB_MR_TYPE_DMA) in rxe_mr_cleanup()