| /drivers/infiniband/sw/rdmavt/ |
| A D | mr.c | 259 mr->ibmr.lkey = mr->mr.lkey; in __rvt_alloc_mr() 260 mr->ibmr.rkey = mr->mr.lkey; in __rvt_alloc_mr() 463 t, mr, mr->pd, mr->lkey, in rvt_check_refs() 566 u32 mapped_segs = mr->mr.length >> mr->mr.page_shift; in rvt_set_page() 599 mr->mr.length = 0; in rvt_map_mr_sg() 604 mr->mr.offset = ibmr->iova - (u64)mr->mr.map[0]->segs[0].vaddr; in rvt_map_mr_sg() 628 if (!mr->mr.lkey || mr->umem) in rvt_fast_reg_mr() 747 isge->mr = mr; in rvt_lkey_ok() 801 isge->mr = mr; in rvt_lkey_ok() 857 sge->mr = mr; in rvt_rkey_ok() [all …]
|
| A D | trace_mr.h | 22 TP_ARGS(mr, m, n, v, len), 40 __entry->iova = mr->iova; 42 __entry->lkey = mr->lkey; 46 __entry->length = mr->length; 47 __entry->offset = mr->offset; 68 TP_ARGS(mr, m, n, v, len)); 73 TP_ARGS(mr, m, n, v, len)); 78 TP_ARGS(mr, m, n, v, len)); 101 __entry->mr = sge->mr; 106 __entry->lkey = sge->mr->lkey; [all …]
|
| /drivers/infiniband/sw/rxe/ |
| A D | rxe_mr.c | 36 iova + length > mr->ibmr.iova + mr->ibmr.length) { in mr_check_range() 57 mr->lkey = mr->ibmr.lkey = key; in rxe_mr_init() 58 mr->rkey = mr->ibmr.rkey = key; in rxe_mr_init() 77 return (iova >> mr->page_shift) - (mr->ibmr.iova >> mr->page_shift); in rxe_mr_iova_to_index() 220 if (unlikely(mr->nbuf == mr->num_buf)) in rxe_set_page() 240 mr->page_offset = mr->ibmr.iova & (page_size - 1); in rxe_map_mr_sg() 415 if (mr) in copy_data() 421 if (mr) in copy_data() 654 key, (remote ? mr->rkey : mr->lkey)); in rxe_invalidate_mr() 689 struct rxe_mr *mr = to_rmr(wqe->wr.wr.reg.mr); in rxe_reg_fast_mr() local [all …]
|
| A D | rxe_odp.c | 69 ret = rxe_odp_do_pagefault_and_lock(mr, mr->umem->address, in rxe_odp_init_pages() 112 mr->ibmr.iova = iova; in rxe_odp_mr_init_user() 417 struct rxe_mr *mr; member 467 struct rxe_mr *mr; in rxe_ib_prefetch_sg_list() local 473 if (!mr) { in rxe_ib_prefetch_sg_list() 482 rxe_put(mr); in rxe_ib_prefetch_sg_list() 490 rxe_put(mr); in rxe_ib_prefetch_sg_list() 497 rxe_put(mr); in rxe_ib_prefetch_sg_list() 511 struct rxe_mr *mr; in rxe_ib_advise_mr_prefetch() local 538 if (!mr) { in rxe_ib_advise_mr_prefetch() [all …]
|
| A D | rxe_mw.c | 91 if (!mr) in rxe_check_bind_mw() 125 (mr->ibmr.iova + mr->ibmr.length)))) { in rxe_check_bind_mw() 146 if (mw->mr) { in rxe_do_bind_mw() 153 mw->mr = mr; in rxe_do_bind_mw() 155 rxe_get(mr); in rxe_do_bind_mw() 197 mr = NULL; in rxe_bind_mw() 216 if (mr) in rxe_bind_mw() 217 rxe_put(mr); in rxe_bind_mw() 247 mr = mw->mr; in rxe_do_invalidate_mw() 250 rxe_put(mr); in rxe_do_invalidate_mw() [all …]
|
| /drivers/infiniband/hw/mlx4/ |
| A D | mr.c | 63 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in mlx4_ib_get_dma_mr() 64 if (!mr) in mlx4_ib_get_dma_mr() 76 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_get_dma_mr() 85 kfree(mr); in mlx4_ib_get_dma_mr() 154 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in mlx4_ib_reg_user_mr() 155 if (!mr) in mlx4_ib_reg_user_mr() 183 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_reg_user_mr() 397 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in mlx4_ib_alloc_mr() 398 if (!mr) in mlx4_ib_alloc_mr() 415 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_alloc_mr() [all …]
|
| /drivers/vdpa/mlx5/core/ |
| A D | mr.c | 44 for_each_sg(mr->sg_head.sgl, sg, mr->nent, i) { in populate_mtts() 78 MLX5_SET64(mkc, mkc, len, mr->end - mr->start); in fill_create_direct_mr() 81 get_octo_len(mr->end - mr->start, mr->log_size)); in fill_create_direct_mr() 83 get_octo_len(mr->end - mr->start, mr->log_size)); in fill_create_direct_mr() 96 mr->mr = mlx5_idx_to_mkey(mkey_index); in create_direct_mr_end() 112 if (!mr->mr) in destroy_direct_mr() 115 mlx5_vdpa_destroy_mkey(mvdev, mr->mr); in destroy_direct_mr() 130 return map_end(map, mr) - map_start(map, mr); in maplen() 420 mr->nent = dma_map_sg_attrs(dma, mr->sg_head.sgl, mr->nsg, DMA_BIDIRECTIONAL, 0); in map_direct_mr() 754 mr, mr->mkey, refcount_read(&mr->refcount)); in mlx5_vdpa_show_mr_leaks() [all …]
|
| /drivers/infiniband/hw/vmw_pvrdma/ |
| A D | pvrdma_mr.c | 75 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in pvrdma_get_dma_mr() 76 if (!mr) in pvrdma_get_dma_mr() 89 kfree(mr); in pvrdma_get_dma_mr() 149 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in pvrdma_reg_user_mr() 150 if (!mr) { in pvrdma_reg_user_mr() 196 kfree(mr); in pvrdma_reg_user_mr() 225 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in pvrdma_alloc_mr() 226 if (!mr) in pvrdma_alloc_mr() 271 kfree(mr); in pvrdma_alloc_mr() 311 if (mr->npages == mr->max_pages) in pvrdma_set_page() [all …]
|
| /drivers/scsi/ |
| A D | mesh.c | 314 (mr->count_hi << 8) + mr->count_lo, mr->sequence, in mesh_dump_regs() 315 (mr->bus_status1 << 8) + mr->bus_status0, mr->fifo_count, in mesh_dump_regs() 316 mr->exception, mr->error, mr->intr_mask, mr->interrupt, in mesh_dump_regs() 437 MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count)); in mesh_start_cmd() 512 MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count)); in mesh_start_cmd() 517 MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count)); in mesh_start_cmd() 528 MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count)); in mesh_start_cmd() 806 MKWORD(mr->interrupt, mr->exception, mr->error, in start_phase() 908 MKWORD(0, mr->error, mr->exception, mr->fifo_count)); in reselected() 1367 MKWORD(mr->count_hi, mr->count_lo, mr->sequence, mr->fifo_count)); in phase_mismatch() [all …]
|
| /drivers/infiniband/hw/mlx5/ |
| A D | mr.c | 1086 mr->ibmr.lkey = mr->mmkey.key; in mlx5_ib_get_dma_mr() 1087 mr->ibmr.rkey = mr->mmkey.key; in mlx5_ib_get_dma_mr() 1122 mr->ibmr.lkey = mr->mmkey.key; in set_mr_fields() 1123 mr->ibmr.rkey = mr->mmkey.key; in set_mr_fields() 1917 if (!mr->umem || is_odp_mr(mr) || is_dmabuf_mr(mr)) in mlx5_ib_rereg_user_mr() 1939 if (!mr->umem || is_odp_mr(mr) || is_dmabuf_mr(mr)) in mlx5_ib_rereg_user_mr() 2325 mr->ibmr.lkey = mr->mmkey.key; in _mlx5_alloc_mkey_descs() 2326 mr->ibmr.rkey = mr->mmkey.key; in _mlx5_alloc_mkey_descs() 2739 mr->data_length = mr->ibmr.length; in mlx5_ib_sg_to_klms() 2760 mr->meta_length = mr->ibmr.length - mr->data_length; in mlx5_ib_sg_to_klms() [all …]
|
| A D | umr.c | 660 if (!mr->ibmr.length) in mlx5r_umr_final_update_xlt() 708 total_blocks = ib_umem_num_dma_blocks(mr->umem, 1UL << mr->page_shift); in _mlx5r_umr_update_mr_pas() 722 _mlx5r_umr_init_wqe(mr, &wqe, &sg, flags, mr->page_shift, dd); in _mlx5r_umr_update_mr_pas() 735 rdma_umem_for_each_dma_block(mr->umem, &biter, BIT(mr->page_shift)) { in _mlx5r_umr_update_mr_pas() 1090 mr->data_direct); in mlx5r_umr_dmabuf_update_pgsz() 1097 total_blocks = ib_umem_num_dma_blocks(mr->umem, 1UL << mr->page_shift); in mlx5r_umr_dmabuf_update_pgsz() 1103 mr, in mlx5r_umr_dmabuf_update_pgsz() 1107 mr->data_direct); in mlx5r_umr_dmabuf_update_pgsz() 1112 err = mlx5r_umr_update_mr_page_shift(mr, mr->page_shift, in mlx5r_umr_dmabuf_update_pgsz() 1113 mr->data_direct); in mlx5r_umr_dmabuf_update_pgsz() [all …]
|
| A D | odp.c | 255 mr) { in destroy_unused_implicit_child_mr() 298 if (!mr) in mlx5_ib_invalidate_range() 489 return mr; in implicit_get_child_mr() 496 mr->ibmr.lkey = mr->mmkey.key; in implicit_get_child_mr() 497 mr->ibmr.rkey = mr->mmkey.key; in implicit_get_child_mr() 546 return mr; in implicit_get_child_mr() 863 if (page_shift != mr->page_shift && mr->dmabuf_faulted) { in pagefault_dmabuf_mr() 938 ret = pagefault_real_mr(mr, to_ib_umem_odp(mr->umem), mr->umem->address, in mlx5_ib_init_odp_mr() 948 ret = pagefault_dmabuf_mr(mr, mr->umem->length, NULL, in mlx5_ib_init_dmabuf_mr() 1990 return mr; in get_prefetchable_mr() [all …]
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
| A D | gddr5.c | 75 ram->mr[0] &= ~0xf7f; in nvkm_gddr5_calc() 80 ram->mr[1] &= ~0x0bf; in nvkm_gddr5_calc() 91 ram->mr[1] &= ~0x030; in nvkm_gddr5_calc() 95 ram->mr[3] &= ~0x020; in nvkm_gddr5_calc() 98 ram->mr[5] &= ~0x004; in nvkm_gddr5_calc() 99 ram->mr[5] |= (l3 << 2); in nvkm_gddr5_calc() 103 if (ram->mr[6] & 0x001) in nvkm_gddr5_calc() 105 ram->mr[6] &= ~0xff1; in nvkm_gddr5_calc() 110 ram->mr[7] &= ~0x300; in nvkm_gddr5_calc() 113 ram->mr[7] &= ~0x088; in nvkm_gddr5_calc() [all …]
|
| A D | sddr3.c | 93 (ram->mr[1] & 0x040) >> 5 | in nvkm_sddr3_calc() 106 ram->mr[0] &= ~0xf74; in nvkm_sddr3_calc() 107 ram->mr[0] |= (WR & 0x07) << 9; in nvkm_sddr3_calc() 108 ram->mr[0] |= (CL & 0x0e) << 3; in nvkm_sddr3_calc() 109 ram->mr[0] |= (CL & 0x01) << 2; in nvkm_sddr3_calc() 111 ram->mr[1] &= ~0x245; in nvkm_sddr3_calc() 112 ram->mr[1] |= (ODT & 0x1) << 2; in nvkm_sddr3_calc() 113 ram->mr[1] |= (ODT & 0x2) << 5; in nvkm_sddr3_calc() 114 ram->mr[1] |= (ODT & 0x4) << 7; in nvkm_sddr3_calc() 115 ram->mr[1] |= !DLL; in nvkm_sddr3_calc() [all …]
|
| A D | gddr3.c | 89 DLL = !(ram->mr[1] & 0x1); in nvkm_gddr3_calc() 98 ODT = (ram->mr[1] & 0xc) >> 2; in nvkm_gddr3_calc() 101 hi = ram->mr[2] & 0x1; in nvkm_gddr3_calc() 107 ram->mr[0] &= ~0xf74; in nvkm_gddr3_calc() 108 ram->mr[0] |= (CWL & 0x07) << 9; in nvkm_gddr3_calc() 109 ram->mr[0] |= (CL & 0x07) << 4; in nvkm_gddr3_calc() 110 ram->mr[0] |= (CL & 0x08) >> 1; in nvkm_gddr3_calc() 112 ram->mr[1] &= ~0x3fc; in nvkm_gddr3_calc() 113 ram->mr[1] |= (ODT & 0x03) << 2; in nvkm_gddr3_calc() 114 ram->mr[1] |= (RON & 0x03) << 8; in nvkm_gddr3_calc() [all …]
|
| /drivers/infiniband/hw/hns/ |
| A D | hns_roce_mr.c | 121 mr->npages = mr->pbl_mtr.hem_cfg.buf_pg_count; in alloc_mr_pbl() 203 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in hns_roce_get_dma_mr() 204 if (!mr) in hns_roce_get_dma_mr() 221 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_get_dma_mr() 246 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in hns_roce_reg_user_mr() 270 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_reg_user_mr() 398 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in hns_roce_alloc_mr() 399 if (!mr) in hns_roce_alloc_mr() 419 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_alloc_mr() 420 mr->ibmr.length = mr->size; in hns_roce_alloc_mr() [all …]
|
| /drivers/infiniband/core/ |
| A D | uverbs_std_types_mr.c | 128 if (IS_ERR(mr)) in UVERBS_HANDLER() 152 &mr->rkey, sizeof(mr->rkey)); in UVERBS_HANDLER() 169 &mr->rkey, sizeof(mr->rkey)); in UVERBS_HANDLER() 175 &mr->length, sizeof(mr->length)); in UVERBS_HANDLER() 181 &mr->iova, sizeof(mr->iova)); in UVERBS_HANDLER() 247 mr->pd = pd; in UVERBS_HANDLER() 260 &mr->lkey, sizeof(mr->lkey)); in UVERBS_HANDLER() 265 &mr->rkey, sizeof(mr->rkey)); in UVERBS_HANDLER() 373 mr->pd = pd; in UVERBS_HANDLER() 389 &mr->lkey, sizeof(mr->lkey)); in UVERBS_HANDLER() [all …]
|
| A D | mr_pool.c | 10 struct ib_mr *mr; in ib_mr_pool_get() local 15 if (mr) { in ib_mr_pool_get() 16 list_del(&mr->qp_entry); in ib_mr_pool_get() 21 return mr; in ib_mr_pool_get() 30 list_add(&mr->qp_entry, list); in ib_mr_pool_put() 39 struct ib_mr *mr; in ib_mr_pool_init() local 49 if (IS_ERR(mr)) { in ib_mr_pool_init() 50 ret = PTR_ERR(mr); in ib_mr_pool_init() 68 struct ib_mr *mr; in ib_mr_pool_destroy() local 74 list_del(&mr->qp_entry); in ib_mr_pool_destroy() [all …]
|
| /drivers/infiniband/hw/mana/ |
| A D | mr.c | 133 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in mana_ib_reg_user_mr() 134 if (!mr) in mana_ib_reg_user_mr() 189 kfree(mr); in mana_ib_reg_user_mr() 216 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in mana_ib_reg_user_mr_dmabuf() 217 if (!mr) in mana_ib_reg_user_mr_dmabuf() 262 kfree(mr); in mana_ib_reg_user_mr_dmabuf() 280 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in mana_ib_get_dma_mr() 281 if (!mr) in mana_ib_get_dma_mr() 294 kfree(mr); in mana_ib_get_dma_mr() 311 if (mr->umem) in mana_ib_dereg_mr() [all …]
|
| /drivers/rtc/ |
| A D | rtc-at91sam9.c | 133 u32 offset, alarm, mr; in at91_rtc_settime() local 140 mr = rtt_readl(rtc, MR); in at91_rtc_settime() 164 mr &= ~AT91_RTT_ALMIEN; in at91_rtc_settime() 205 u32 mr; in at91_rtc_setalarm() local 214 mr = rtt_readl(rtc, MR); in at91_rtc_setalarm() 261 u32 sr, mr; in at91_rtc_cache_events() local 339 u32 mr; in at91_rtc_probe() local 395 mr = rtt_readl(rtc, MR); in at91_rtc_probe() 405 rtt_writel(rtc, MR, mr); in at91_rtc_probe() 502 u32 mr; in at91_rtc_resume() local [all …]
|
| /drivers/net/ethernet/mellanox/mlxsw/ |
| A D | spectrum_mr.c | 254 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_route_write() local 292 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_route_erase() local 485 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_route_ivif_resolve() local 519 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_route_ivif_unresolve() local 534 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_route_evif_resolve() local 593 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_route_evif_unresolve() local 751 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_rif_mtu_update() local 906 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_table_create() local 955 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_table_destroy() local 1045 mlxsw_sp->mr = mr; in mlxsw_sp_mr_init() [all …]
|
| /drivers/sh/intc/ |
| A D | handle.c | 55 if (mr->set_reg && mr->clr_reg) { in _intc_mask_data() 58 reg_e = mr->clr_reg; in _intc_mask_data() 59 reg_d = mr->set_reg; in _intc_mask_data() 62 if (mr->set_reg) { in _intc_mask_data() 64 reg_e = mr->set_reg; in _intc_mask_data() 65 reg_d = mr->set_reg; in _intc_mask_data() 68 reg_e = mr->clr_reg; in _intc_mask_data() 69 reg_d = mr->clr_reg; in _intc_mask_data() 183 mr = desc->hw.ack_regs + i; in intc_ack_data() 191 reg_e = mr->set_reg; in intc_ack_data() [all …]
|
| /drivers/infiniband/ulp/iser/ |
| A D | iser_memory.c | 270 ib_update_fast_reg_key(mr, ib_inc_rkey(mr->rkey)); in iser_reg_sig_mr() 286 wr->mr = mr; in iser_reg_sig_mr() 287 wr->key = mr->rkey; in iser_reg_sig_mr() 294 sig_reg->rkey = mr->rkey; in iser_reg_sig_mr() 312 struct ib_mr *mr = rsc->mr; in iser_fast_reg_mr() local 316 if (rsc->mr->need_inval) in iser_fast_reg_mr() 319 ib_update_fast_reg_key(mr, ib_inc_rkey(mr->rkey)); in iser_fast_reg_mr() 333 wr->mr = mr; in iser_fast_reg_mr() 334 wr->key = mr->rkey; in iser_fast_reg_mr() 341 reg->sge.lkey = mr->lkey; in iser_fast_reg_mr() [all …]
|
| /drivers/gpu/drm/i915/gem/ |
| A D | i915_gem_create.c | 51 struct drm_i915_private *i915 = mr->i915; in object_set_placements() 127 ret = mr->ops->init_object(mr, obj, I915_BO_INVALID_OFFSET, size, 0, flags); in __i915_gem_object_create_user_ext() 173 struct intel_memory_region *mr; in i915_gem_dumb_create() local 231 struct intel_memory_region *mr; in i915_gem_create_ioctl() local 264 mr->name, mr->type, mr->instance); in repr_placements() 306 struct intel_memory_region *mr; in set_placements() local 314 if (!mr || mr->private) { in set_placements() 321 if (mask & BIT(mr->id)) { in set_placements() 323 mr->name, region.memory_class, in set_placements() 329 placements[i] = mr; in set_placements() [all …]
|
| /drivers/watchdog/ |
| A D | sama5d4_wdt.c | 31 u32 mr; member 86 wdt->mr &= ~AT91_SAM9X60_WDDIS; in sama5d4_wdt_start() 88 wdt->mr &= ~AT91_WDT_WDDIS; in sama5d4_wdt_start() 101 wdt->mr |= AT91_SAM9X60_WDDIS; in sama5d4_wdt_stop() 103 wdt->mr |= AT91_WDT_WDDIS; in sama5d4_wdt_stop() 133 wdt->mr &= ~AT91_WDT_WDV; in sama5d4_wdt_set_timeout() 188 wdt->mr = AT91_SAM9X60_WDDIS; in of_sama5d4_wdt_init() 190 wdt->mr = AT91_WDT_WDDIS; in of_sama5d4_wdt_init() 200 wdt->mr |= AT91_WDT_WDDBGHLT; in of_sama5d4_wdt_init() 240 wdt->mr |= AT91_WDT_WDFIEN; in sama5d4_wdt_init() [all …]
|