Home
last modified time | relevance | path

Searched refs:vmap (Results 1 – 25 of 122) sorted by relevance

12345

/drivers/gpu/drm/nouveau/nvkm/subdev/bios/
A Dvmap.c32 u32 vmap = 0; in nvbios_vmap_table() local
37 if (vmap) { in nvbios_vmap_table()
38 *ver = nvbios_rd08(bios, vmap + 0); in nvbios_vmap_table()
45 return vmap; in nvbios_vmap_table()
62 switch (!!vmap * *ver) { in nvbios_vmap_parse()
77 return vmap; in nvbios_vmap_parse()
85 if (vmap && idx < cnt) { in nvbios_vmap_entry()
86 vmap = vmap + hdr + (idx * *len); in nvbios_vmap_entry()
87 return vmap; in nvbios_vmap_entry()
98 switch (!!vmap * *ver) { in nvbios_vmap_entry_parse()
[all …]
/drivers/gpu/drm/xe/tests/
A Dxe_migrate.c113 retval = xe_map_rd(xe, &remote->vmap, 0, u64); in test_copy()
131 retval = xe_map_rd(xe, &bo->vmap, 0, u64); in test_copy()
148 retval = xe_map_rd(xe, &remote->vmap, 0, u64); in test_copy()
264 xe_map_wr(xe, &pt->vmap, 0, u32, 0xdeaddead); in xe_migrate_sanity_test()
272 retval = xe_map_rd(xe, &pt->vmap, 0, u32); in xe_migrate_sanity_test()
287 retval = xe_map_rd(xe, &tiny->vmap, 0, u32); in xe_migrate_sanity_test()
309 retval = xe_map_rd(xe, &big->vmap, 0, u32); in xe_migrate_sanity_test()
524 retval = xe_map_rd(xe, &vram_bo->vmap, 0, u64); in test_migrate()
532 retval = xe_map_rd(xe, &ccs_bo->vmap, 0, u64); in test_migrate()
563 retval = xe_map_rd(xe, &vram_bo->vmap, 0, u64); in test_migrate()
[all …]
/drivers/net/
A Dvrf.c110 struct vrf_map vmap; member
136 return &nn_vrf->vmap; in netns_vrf_map()
211 static void vrf_map_lock(struct vrf_map *vmap) __acquires(&vmap->vmap_lock) in vrf_map_lock() argument
216 static void vrf_map_unlock(struct vrf_map *vmap) __releases(&vmap->vmap_lock) in vrf_map_unlock() argument
242 vrf_map_lock(vmap); in vrf_map_register_dev()
272 vrf_map_unlock(vmap); in vrf_map_register_dev()
290 vrf_map_lock(vmap); in vrf_map_unregister_dev()
309 vrf_map_unlock(vmap); in vrf_map_unregister_dev()
319 vrf_map_lock(vmap); in vrf_ifindex_lookup_by_table_id()
335 vrf_map_unlock(vmap); in vrf_ifindex_lookup_by_table_id()
[all …]
/drivers/gpu/drm/nouveau/nvkm/subdev/volt/
A Dbase.c87 u32 vmap; in nvkm_volt_map_min() local
89 vmap = nvbios_vmap_entry_parse(bios, id, &ver, &len, &info); in nvkm_volt_map_min()
90 if (vmap) { in nvkm_volt_map_min()
109 u32 vmap; in nvkm_volt_map() local
111 vmap = nvbios_vmap_entry_parse(bios, id, &ver, &len, &info); in nvkm_volt_map()
112 if (vmap) { in nvkm_volt_map()
295 struct nvbios_vmap vmap; in nvkm_volt_ctor() local
301 if (nvbios_vmap_parse(bios, &ver, &hdr, &cnt, &len, &vmap)) { in nvkm_volt_ctor()
302 volt->max0_id = vmap.max0; in nvkm_volt_ctor()
303 volt->max1_id = vmap.max1; in nvkm_volt_ctor()
[all …]
/drivers/gpu/drm/xe/
A Dxe_sa.c71 sa_manager->is_iomem = bo->vmap.is_iomem; in __xe_sa_bo_manager_init()
74 if (bo->vmap.is_iomem) { in __xe_sa_bo_manager_init()
79 sa_manager->cpu_ptr = bo->vmap.vaddr; in __xe_sa_bo_manager_init()
119 if (!sa_manager->bo->vmap.is_iomem) in xe_sa_bo_flush_write()
122 xe_map_memcpy_to(xe, &sa_manager->bo->vmap, drm_suballoc_soffset(sa_bo), in xe_sa_bo_flush_write()
A Dxe_huc.c170 wr_offset = xe_gsc_emit_header(xe, &pkt->vmap, 0, HECI_MEADDRESS_PXP, 0, in huc_auth_via_gsccs()
172 wr_offset = huc_emit_pxp_auth_msg(xe, &pkt->vmap, wr_offset, in huc_auth_via_gsccs()
182 if (xe_gsc_check_and_update_pending(xe, &pkt->vmap, 0, &pkt->vmap, in huc_auth_via_gsccs()
194 err = xe_gsc_read_out_header(xe, &pkt->vmap, PXP43_HUC_AUTH_INOUT_SIZE, in huc_auth_via_gsccs()
206 out_status = huc_auth_msg_rd(xe, &pkt->vmap, rd_offset, header.status); in huc_auth_via_gsccs()
A Dxe_gsc.c60 xe_map_memcpy_from(xe, storage, &gsc->fw.bo->vmap, 0, fw_size); in memcpy_fw()
61 xe_map_memcpy_to(xe, &gsc->private->vmap, 0, storage, fw_size); in memcpy_fw()
62 xe_map_memset(xe, &gsc->private->vmap, fw_size, 0, in memcpy_fw()
150 wr_offset = xe_gsc_emit_header(xe, &bo->vmap, 0, HECI_MEADDRESS_MKHI, 0, in query_compatibility_version()
152 wr_offset = emit_version_query_msg(xe, &bo->vmap, wr_offset); in query_compatibility_version()
164 err = xe_gsc_read_out_header(xe, &bo->vmap, GSC_VER_PKT_SZ, in query_compatibility_version()
172 compat->major = version_query_rd(xe, &bo->vmap, rd_offset, proj_major); in query_compatibility_version()
173 compat->minor = version_query_rd(xe, &bo->vmap, rd_offset, compat_major); in query_compatibility_version()
174 compat->patch = version_query_rd(xe, &bo->vmap, rd_offset, compat_minor); in query_compatibility_version()
A Dxe_memirq.c194 iosys_map_memset(&bo->vmap, 0, 0, bo_size); in memirq_alloc_pages()
197 memirq->source = IOSYS_MAP_INIT_OFFSET(&bo->vmap, XE_MEMIRQ_SOURCE_OFFSET(0)); in memirq_alloc_pages()
198 memirq->status = IOSYS_MAP_INIT_OFFSET(&bo->vmap, XE_MEMIRQ_STATUS_OFFSET(0)); in memirq_alloc_pages()
199 memirq->mask = IOSYS_MAP_INIT_OFFSET(&bo->vmap, XE_MEMIRQ_ENABLE_OFFSET); in memirq_alloc_pages()
452 struct iosys_map src_offset = IOSYS_MAP_INIT_OFFSET(&memirq->bo->vmap, in xe_memirq_hwe_handler()
457 IOSYS_MAP_INIT_OFFSET(&memirq->bo->vmap, in xe_memirq_hwe_handler()
A Dxe_pxp_submit.c143 gsc_res->batch = IOSYS_MAP_INIT_OFFSET(&bo->vmap, 0); in allocate_gsc_client_resources()
144 gsc_res->msg_in = IOSYS_MAP_INIT_OFFSET(&bo->vmap, PXP_BB_SIZE); in allocate_gsc_client_resources()
145 gsc_res->msg_out = IOSYS_MAP_INIT_OFFSET(&bo->vmap, PXP_BB_SIZE + inout_size); in allocate_gsc_client_resources()
296 offset = pxp_emit_session_termination(pxp->xe, &pxp->vcs_exec.bo->vmap, offset, id); in xe_pxp_submit_session_termination()
297 offset = pxp_emit_wait(pxp->xe, &pxp->vcs_exec.bo->vmap, offset); in xe_pxp_submit_session_termination()
298 emit_cmd(pxp->xe, &pxp->vcs_exec.bo->vmap, offset, MI_BATCH_BUFFER_END); in xe_pxp_submit_session_termination()
A Dxe_lmtt.c84 xe_map_memset(lmtt_to_xe(lmtt), &bo->vmap, 0, 0, xe_bo_size(bo)); in lmtt_pt_alloc()
289 lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u32), u32)); in lmtt_write_pte()
291 xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u32), u32, pte); in lmtt_write_pte()
294 lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u64), u64)); in lmtt_write_pte()
296 xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u64), u64, pte); in lmtt_write_pte()
A Dxe_bo.c1267 if (iosys_map_is_null(&bo->vmap)) { in xe_bo_evict_pinned()
1274 xe_map_memcpy_from(xe, backup->vmap.vaddr, &bo->vmap, 0, in xe_bo_evict_pinned()
1360 if (iosys_map_is_null(&bo->vmap)) { in xe_bo_restore_pinned()
1367 xe_map_memcpy_to(xe, &bo->vmap, 0, backup->vmap.vaddr, in xe_bo_restore_pinned()
1556 struct iosys_map vmap; in xe_ttm_access_memory() local
2209 xe_map_memcpy_to(xe, &bo->vmap, 0, data, size); in xe_managed_bo_create_from_data()
2236 xe_assert(xe, !(*src)->vmap.is_iomem); in xe_managed_bo_reinit_in_vram()
2519 if (!iosys_map_is_null(&bo->vmap)) in xe_bo_vmap()
2537 iosys_map_set_vaddr(&bo->vmap, virtual); in xe_bo_vmap()
2544 if (!iosys_map_is_null(&bo->vmap)) { in __xe_bo_vunmap()
[all …]
A Dxe_migrate.c165 xe_map_wr(xe, &bo->vmap, ofs, u64, entry); in xe_migrate_program_identity()
174 xe_map_wr(xe, &bo->vmap, ofs, u64, pos | flags); in xe_migrate_program_identity()
178 xe_map_wr(xe, &bo->vmap, ofs, u64, pos | flags); in xe_migrate_program_identity()
219 xe_pt_write(xe, &vm->pt_root[id]->bo->vmap, 0, entry); in xe_migrate_prepare_vm()
228 xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64, entry); in xe_migrate_prepare_vm()
245 xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64, in xe_migrate_prepare_vm()
262 xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64, in xe_migrate_prepare_vm()
287 xe_map_wr(xe, &bo->vmap, map_ofs + XE_PAGE_SIZE * level, u64, in xe_migrate_prepare_vm()
296 xe_map_wr(xe, &bo->vmap, map_ofs + XE_PAGE_SIZE + in xe_migrate_prepare_vm()
1294 &update->pt_bo->vmap, NULL, in xe_migrate_update_pgtables_cpu()
[all …]
A Dxe_bo_types.h48 struct iosys_map vmap; member
A Dxe_guc_log.c163 xe_map_memcpy_from(xe, snapshot->copy[i], &log->bo->vmap, in xe_guc_log_snapshot_capture()
268 xe_map_memset(xe, &bo->vmap, 0, 0, guc_log_size()); in xe_guc_log_init()
/drivers/gpu/drm/xe/display/
A Dxe_hdcp_gsc.c88 xe_map_memset(xe, &bo->vmap, 0, 0, xe_bo_size(bo)); in intel_hdcp_gsc_initialize_message()
138 struct iosys_map *map = &gsc_context->hdcp_bo->vmap; in xe_gsc_send_sync()
180 addr_in_wr_off = xe_gsc_emit_header(xe, &gsc_context->hdcp_bo->vmap, in intel_hdcp_gsc_msg_send()
183 xe_map_memcpy_to(xe, &gsc_context->hdcp_bo->vmap, addr_in_wr_off, in intel_hdcp_gsc_msg_send()
206 xe_map_memcpy_from(xe, msg_out, &gsc_context->hdcp_bo->vmap, in intel_hdcp_gsc_msg_send()
A Dxe_dsb_buffer.c20 iosys_map_wr(&dsb_buf->vma->bo->vmap, idx * 4, u32, val); in intel_dsb_buffer_write()
25 return iosys_map_rd(&dsb_buf->vma->bo->vmap, idx * 4, u32); in intel_dsb_buffer_read()
32 iosys_map_memset(&dsb_buf->vma->bo->vmap, idx * 4, val, size); in intel_dsb_buffer_memset()
A Dintel_fbdev_fb.c107 XE_WARN_ON(iosys_map_is_null(&obj->vmap)); in intel_fbdev_fb_fill_info()
109 info->screen_base = obj->vmap.vaddr_iomem; in intel_fbdev_fb_fill_info()
A Dxe_fb_pin.c137 iosys_map_wr(&dpt->vmap, x * 8, u64, pte | addr); in __xe_pin_fb_vma_dpt()
144 write_dpt_remapped(bo, &dpt->vmap, &dpt_ofs, in __xe_pin_fb_vma_dpt()
156 write_dpt_rotated(bo, &dpt->vmap, &dpt_ofs, in __xe_pin_fb_vma_dpt()
464 *map = vma->bo->vmap; in intel_fb_get_map()
/drivers/gpu/drm/sitronix/
A Dst7571-i2c.c104 const struct iosys_map *vmap,
271 const struct iosys_map *vmap, in st7571_prepare_buffer_monochrome() argument
285 drm_fb_xrgb8888_to_mono(&dst, &dst_pitch, vmap, fb, rect, fmtcnv_state); in st7571_prepare_buffer_monochrome()
290 memcpy(st7571->hwbuf, vmap->vaddr, size); in st7571_prepare_buffer_monochrome()
296 const struct iosys_map *vmap, in st7571_prepare_buffer_grayscale() argument
310 drm_fb_xrgb8888_to_mono(&dst, &dst_pitch, vmap, fb, rect, fmtcnv_state); in st7571_prepare_buffer_grayscale()
315 memcpy(st7571->hwbuf, vmap->vaddr, size); in st7571_prepare_buffer_grayscale()
320 memcpy(st7571->hwbuf, vmap->vaddr, size); in st7571_prepare_buffer_grayscale()
A Dst7586.c74 struct iosys_map dst_map, vmap; in st7586_xrgb8888_to_gray332() local
81 iosys_map_set_vaddr(&vmap, vaddr); in st7586_xrgb8888_to_gray332()
82 drm_fb_xrgb8888_to_gray8(&dst_map, NULL, &vmap, fb, clip, fmtcnv_state); in st7586_xrgb8888_to_gray332()
/drivers/gpu/drm/etnaviv/
A Detnaviv_gem.c336 etnaviv_obj->vaddr = etnaviv_obj->ops->vmap(etnaviv_obj); in etnaviv_gem_vmap()
365 return vmap(pages, obj->base.size >> PAGE_SHIFT, VM_MAP, prot); in etnaviv_gem_vmap_impl()
496 .vmap = etnaviv_gem_vmap_impl,
565 .vmap = etnaviv_gem_prime_vmap,
729 .vmap = etnaviv_gem_vmap_impl,
/drivers/gpu/drm/xen/
A Dxen_drm_front_gem.c108 .vmap = xen_drm_front_gem_prime_vmap,
292 vaddr = vmap(xen_obj->pages, xen_obj->num_pages, in xen_drm_front_gem_prime_vmap()
/drivers/gpu/drm/rockchip/
A Drockchip_drm_gem.c137 rk_obj->kvaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP, in rockchip_gem_alloc_iommu()
276 .vmap = rockchip_gem_prime_vmap,
523 vaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP, in rockchip_gem_prime_vmap()
/drivers/gpu/drm/mediatek/
A Dmtk_gem.c28 .vmap = mtk_gem_prime_vmap,
260 mtk_gem->kvaddr = vmap(mtk_gem->pages, npages, VM_MAP, in mtk_gem_prime_vmap()
/drivers/gpu/drm/v3d/
A Dv3d_bo.c69 .vmap = drm_gem_shmem_object_vmap,
198 bo->vaddr = vmap(obj->pages, obj->base.size >> PAGE_SHIFT, VM_MAP, in v3d_get_bo_vaddr()

Completed in 57 milliseconds

12345