| /drivers/gpu/drm/nouveau/nvkm/subdev/bios/ |
| A D | vmap.c | 32 u32 vmap = 0; in nvbios_vmap_table() local 37 if (vmap) { in nvbios_vmap_table() 38 *ver = nvbios_rd08(bios, vmap + 0); in nvbios_vmap_table() 45 return vmap; in nvbios_vmap_table() 62 switch (!!vmap * *ver) { in nvbios_vmap_parse() 77 return vmap; in nvbios_vmap_parse() 85 if (vmap && idx < cnt) { in nvbios_vmap_entry() 86 vmap = vmap + hdr + (idx * *len); in nvbios_vmap_entry() 87 return vmap; in nvbios_vmap_entry() 98 switch (!!vmap * *ver) { in nvbios_vmap_entry_parse() [all …]
|
| /drivers/gpu/drm/xe/tests/ |
| A D | xe_migrate.c | 113 retval = xe_map_rd(xe, &remote->vmap, 0, u64); in test_copy() 131 retval = xe_map_rd(xe, &bo->vmap, 0, u64); in test_copy() 148 retval = xe_map_rd(xe, &remote->vmap, 0, u64); in test_copy() 264 xe_map_wr(xe, &pt->vmap, 0, u32, 0xdeaddead); in xe_migrate_sanity_test() 272 retval = xe_map_rd(xe, &pt->vmap, 0, u32); in xe_migrate_sanity_test() 287 retval = xe_map_rd(xe, &tiny->vmap, 0, u32); in xe_migrate_sanity_test() 309 retval = xe_map_rd(xe, &big->vmap, 0, u32); in xe_migrate_sanity_test() 524 retval = xe_map_rd(xe, &vram_bo->vmap, 0, u64); in test_migrate() 532 retval = xe_map_rd(xe, &ccs_bo->vmap, 0, u64); in test_migrate() 563 retval = xe_map_rd(xe, &vram_bo->vmap, 0, u64); in test_migrate() [all …]
|
| /drivers/net/ |
| A D | vrf.c | 110 struct vrf_map vmap; member 136 return &nn_vrf->vmap; in netns_vrf_map() 211 static void vrf_map_lock(struct vrf_map *vmap) __acquires(&vmap->vmap_lock) in vrf_map_lock() argument 216 static void vrf_map_unlock(struct vrf_map *vmap) __releases(&vmap->vmap_lock) in vrf_map_unlock() argument 242 vrf_map_lock(vmap); in vrf_map_register_dev() 272 vrf_map_unlock(vmap); in vrf_map_register_dev() 290 vrf_map_lock(vmap); in vrf_map_unregister_dev() 309 vrf_map_unlock(vmap); in vrf_map_unregister_dev() 319 vrf_map_lock(vmap); in vrf_ifindex_lookup_by_table_id() 335 vrf_map_unlock(vmap); in vrf_ifindex_lookup_by_table_id() [all …]
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/volt/ |
| A D | base.c | 87 u32 vmap; in nvkm_volt_map_min() local 89 vmap = nvbios_vmap_entry_parse(bios, id, &ver, &len, &info); in nvkm_volt_map_min() 90 if (vmap) { in nvkm_volt_map_min() 109 u32 vmap; in nvkm_volt_map() local 111 vmap = nvbios_vmap_entry_parse(bios, id, &ver, &len, &info); in nvkm_volt_map() 112 if (vmap) { in nvkm_volt_map() 295 struct nvbios_vmap vmap; in nvkm_volt_ctor() local 301 if (nvbios_vmap_parse(bios, &ver, &hdr, &cnt, &len, &vmap)) { in nvkm_volt_ctor() 302 volt->max0_id = vmap.max0; in nvkm_volt_ctor() 303 volt->max1_id = vmap.max1; in nvkm_volt_ctor() [all …]
|
| /drivers/gpu/drm/xe/ |
| A D | xe_sa.c | 71 sa_manager->is_iomem = bo->vmap.is_iomem; in __xe_sa_bo_manager_init() 74 if (bo->vmap.is_iomem) { in __xe_sa_bo_manager_init() 79 sa_manager->cpu_ptr = bo->vmap.vaddr; in __xe_sa_bo_manager_init() 119 if (!sa_manager->bo->vmap.is_iomem) in xe_sa_bo_flush_write() 122 xe_map_memcpy_to(xe, &sa_manager->bo->vmap, drm_suballoc_soffset(sa_bo), in xe_sa_bo_flush_write()
|
| A D | xe_huc.c | 170 wr_offset = xe_gsc_emit_header(xe, &pkt->vmap, 0, HECI_MEADDRESS_PXP, 0, in huc_auth_via_gsccs() 172 wr_offset = huc_emit_pxp_auth_msg(xe, &pkt->vmap, wr_offset, in huc_auth_via_gsccs() 182 if (xe_gsc_check_and_update_pending(xe, &pkt->vmap, 0, &pkt->vmap, in huc_auth_via_gsccs() 194 err = xe_gsc_read_out_header(xe, &pkt->vmap, PXP43_HUC_AUTH_INOUT_SIZE, in huc_auth_via_gsccs() 206 out_status = huc_auth_msg_rd(xe, &pkt->vmap, rd_offset, header.status); in huc_auth_via_gsccs()
|
| A D | xe_gsc.c | 60 xe_map_memcpy_from(xe, storage, &gsc->fw.bo->vmap, 0, fw_size); in memcpy_fw() 61 xe_map_memcpy_to(xe, &gsc->private->vmap, 0, storage, fw_size); in memcpy_fw() 62 xe_map_memset(xe, &gsc->private->vmap, fw_size, 0, in memcpy_fw() 150 wr_offset = xe_gsc_emit_header(xe, &bo->vmap, 0, HECI_MEADDRESS_MKHI, 0, in query_compatibility_version() 152 wr_offset = emit_version_query_msg(xe, &bo->vmap, wr_offset); in query_compatibility_version() 164 err = xe_gsc_read_out_header(xe, &bo->vmap, GSC_VER_PKT_SZ, in query_compatibility_version() 172 compat->major = version_query_rd(xe, &bo->vmap, rd_offset, proj_major); in query_compatibility_version() 173 compat->minor = version_query_rd(xe, &bo->vmap, rd_offset, compat_major); in query_compatibility_version() 174 compat->patch = version_query_rd(xe, &bo->vmap, rd_offset, compat_minor); in query_compatibility_version()
|
| A D | xe_memirq.c | 194 iosys_map_memset(&bo->vmap, 0, 0, bo_size); in memirq_alloc_pages() 197 memirq->source = IOSYS_MAP_INIT_OFFSET(&bo->vmap, XE_MEMIRQ_SOURCE_OFFSET(0)); in memirq_alloc_pages() 198 memirq->status = IOSYS_MAP_INIT_OFFSET(&bo->vmap, XE_MEMIRQ_STATUS_OFFSET(0)); in memirq_alloc_pages() 199 memirq->mask = IOSYS_MAP_INIT_OFFSET(&bo->vmap, XE_MEMIRQ_ENABLE_OFFSET); in memirq_alloc_pages() 452 struct iosys_map src_offset = IOSYS_MAP_INIT_OFFSET(&memirq->bo->vmap, in xe_memirq_hwe_handler() 457 IOSYS_MAP_INIT_OFFSET(&memirq->bo->vmap, in xe_memirq_hwe_handler()
|
| A D | xe_pxp_submit.c | 143 gsc_res->batch = IOSYS_MAP_INIT_OFFSET(&bo->vmap, 0); in allocate_gsc_client_resources() 144 gsc_res->msg_in = IOSYS_MAP_INIT_OFFSET(&bo->vmap, PXP_BB_SIZE); in allocate_gsc_client_resources() 145 gsc_res->msg_out = IOSYS_MAP_INIT_OFFSET(&bo->vmap, PXP_BB_SIZE + inout_size); in allocate_gsc_client_resources() 296 offset = pxp_emit_session_termination(pxp->xe, &pxp->vcs_exec.bo->vmap, offset, id); in xe_pxp_submit_session_termination() 297 offset = pxp_emit_wait(pxp->xe, &pxp->vcs_exec.bo->vmap, offset); in xe_pxp_submit_session_termination() 298 emit_cmd(pxp->xe, &pxp->vcs_exec.bo->vmap, offset, MI_BATCH_BUFFER_END); in xe_pxp_submit_session_termination()
|
| A D | xe_lmtt.c | 84 xe_map_memset(lmtt_to_xe(lmtt), &bo->vmap, 0, 0, xe_bo_size(bo)); in lmtt_pt_alloc() 289 lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u32), u32)); in lmtt_write_pte() 291 xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u32), u32, pte); in lmtt_write_pte() 294 lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u64), u64)); in lmtt_write_pte() 296 xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u64), u64, pte); in lmtt_write_pte()
|
| A D | xe_bo.c | 1267 if (iosys_map_is_null(&bo->vmap)) { in xe_bo_evict_pinned() 1274 xe_map_memcpy_from(xe, backup->vmap.vaddr, &bo->vmap, 0, in xe_bo_evict_pinned() 1360 if (iosys_map_is_null(&bo->vmap)) { in xe_bo_restore_pinned() 1367 xe_map_memcpy_to(xe, &bo->vmap, 0, backup->vmap.vaddr, in xe_bo_restore_pinned() 1556 struct iosys_map vmap; in xe_ttm_access_memory() local 2209 xe_map_memcpy_to(xe, &bo->vmap, 0, data, size); in xe_managed_bo_create_from_data() 2236 xe_assert(xe, !(*src)->vmap.is_iomem); in xe_managed_bo_reinit_in_vram() 2519 if (!iosys_map_is_null(&bo->vmap)) in xe_bo_vmap() 2537 iosys_map_set_vaddr(&bo->vmap, virtual); in xe_bo_vmap() 2544 if (!iosys_map_is_null(&bo->vmap)) { in __xe_bo_vunmap() [all …]
|
| A D | xe_migrate.c | 165 xe_map_wr(xe, &bo->vmap, ofs, u64, entry); in xe_migrate_program_identity() 174 xe_map_wr(xe, &bo->vmap, ofs, u64, pos | flags); in xe_migrate_program_identity() 178 xe_map_wr(xe, &bo->vmap, ofs, u64, pos | flags); in xe_migrate_program_identity() 219 xe_pt_write(xe, &vm->pt_root[id]->bo->vmap, 0, entry); in xe_migrate_prepare_vm() 228 xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64, entry); in xe_migrate_prepare_vm() 245 xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64, in xe_migrate_prepare_vm() 262 xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64, in xe_migrate_prepare_vm() 287 xe_map_wr(xe, &bo->vmap, map_ofs + XE_PAGE_SIZE * level, u64, in xe_migrate_prepare_vm() 296 xe_map_wr(xe, &bo->vmap, map_ofs + XE_PAGE_SIZE + in xe_migrate_prepare_vm() 1294 &update->pt_bo->vmap, NULL, in xe_migrate_update_pgtables_cpu() [all …]
|
| A D | xe_bo_types.h | 48 struct iosys_map vmap; member
|
| A D | xe_guc_log.c | 163 xe_map_memcpy_from(xe, snapshot->copy[i], &log->bo->vmap, in xe_guc_log_snapshot_capture() 268 xe_map_memset(xe, &bo->vmap, 0, 0, guc_log_size()); in xe_guc_log_init()
|
| /drivers/gpu/drm/xe/display/ |
| A D | xe_hdcp_gsc.c | 88 xe_map_memset(xe, &bo->vmap, 0, 0, xe_bo_size(bo)); in intel_hdcp_gsc_initialize_message() 138 struct iosys_map *map = &gsc_context->hdcp_bo->vmap; in xe_gsc_send_sync() 180 addr_in_wr_off = xe_gsc_emit_header(xe, &gsc_context->hdcp_bo->vmap, in intel_hdcp_gsc_msg_send() 183 xe_map_memcpy_to(xe, &gsc_context->hdcp_bo->vmap, addr_in_wr_off, in intel_hdcp_gsc_msg_send() 206 xe_map_memcpy_from(xe, msg_out, &gsc_context->hdcp_bo->vmap, in intel_hdcp_gsc_msg_send()
|
| A D | xe_dsb_buffer.c | 20 iosys_map_wr(&dsb_buf->vma->bo->vmap, idx * 4, u32, val); in intel_dsb_buffer_write() 25 return iosys_map_rd(&dsb_buf->vma->bo->vmap, idx * 4, u32); in intel_dsb_buffer_read() 32 iosys_map_memset(&dsb_buf->vma->bo->vmap, idx * 4, val, size); in intel_dsb_buffer_memset()
|
| A D | intel_fbdev_fb.c | 107 XE_WARN_ON(iosys_map_is_null(&obj->vmap)); in intel_fbdev_fb_fill_info() 109 info->screen_base = obj->vmap.vaddr_iomem; in intel_fbdev_fb_fill_info()
|
| A D | xe_fb_pin.c | 137 iosys_map_wr(&dpt->vmap, x * 8, u64, pte | addr); in __xe_pin_fb_vma_dpt() 144 write_dpt_remapped(bo, &dpt->vmap, &dpt_ofs, in __xe_pin_fb_vma_dpt() 156 write_dpt_rotated(bo, &dpt->vmap, &dpt_ofs, in __xe_pin_fb_vma_dpt() 464 *map = vma->bo->vmap; in intel_fb_get_map()
|
| /drivers/gpu/drm/sitronix/ |
| A D | st7571-i2c.c | 104 const struct iosys_map *vmap, 271 const struct iosys_map *vmap, in st7571_prepare_buffer_monochrome() argument 285 drm_fb_xrgb8888_to_mono(&dst, &dst_pitch, vmap, fb, rect, fmtcnv_state); in st7571_prepare_buffer_monochrome() 290 memcpy(st7571->hwbuf, vmap->vaddr, size); in st7571_prepare_buffer_monochrome() 296 const struct iosys_map *vmap, in st7571_prepare_buffer_grayscale() argument 310 drm_fb_xrgb8888_to_mono(&dst, &dst_pitch, vmap, fb, rect, fmtcnv_state); in st7571_prepare_buffer_grayscale() 315 memcpy(st7571->hwbuf, vmap->vaddr, size); in st7571_prepare_buffer_grayscale() 320 memcpy(st7571->hwbuf, vmap->vaddr, size); in st7571_prepare_buffer_grayscale()
|
| A D | st7586.c | 74 struct iosys_map dst_map, vmap; in st7586_xrgb8888_to_gray332() local 81 iosys_map_set_vaddr(&vmap, vaddr); in st7586_xrgb8888_to_gray332() 82 drm_fb_xrgb8888_to_gray8(&dst_map, NULL, &vmap, fb, clip, fmtcnv_state); in st7586_xrgb8888_to_gray332()
|
| /drivers/gpu/drm/etnaviv/ |
| A D | etnaviv_gem.c | 336 etnaviv_obj->vaddr = etnaviv_obj->ops->vmap(etnaviv_obj); in etnaviv_gem_vmap() 365 return vmap(pages, obj->base.size >> PAGE_SHIFT, VM_MAP, prot); in etnaviv_gem_vmap_impl() 496 .vmap = etnaviv_gem_vmap_impl, 565 .vmap = etnaviv_gem_prime_vmap, 729 .vmap = etnaviv_gem_vmap_impl,
|
| /drivers/gpu/drm/xen/ |
| A D | xen_drm_front_gem.c | 108 .vmap = xen_drm_front_gem_prime_vmap, 292 vaddr = vmap(xen_obj->pages, xen_obj->num_pages, in xen_drm_front_gem_prime_vmap()
|
| /drivers/gpu/drm/rockchip/ |
| A D | rockchip_drm_gem.c | 137 rk_obj->kvaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP, in rockchip_gem_alloc_iommu() 276 .vmap = rockchip_gem_prime_vmap, 523 vaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP, in rockchip_gem_prime_vmap()
|
| /drivers/gpu/drm/mediatek/ |
| A D | mtk_gem.c | 28 .vmap = mtk_gem_prime_vmap, 260 mtk_gem->kvaddr = vmap(mtk_gem->pages, npages, VM_MAP, in mtk_gem_prime_vmap()
|
| /drivers/gpu/drm/v3d/ |
| A D | v3d_bo.c | 69 .vmap = drm_gem_shmem_object_vmap, 198 bo->vaddr = vmap(obj->pages, obj->base.size >> PAGE_SHIFT, VM_MAP, in v3d_get_bo_vaddr()
|