Home
last modified time | relevance | path

Searched refs:vmw (Results 1 – 15 of 15) sorted by relevance

/drivers/gpu/drm/vmwgfx/
A Dvmwgfx_devcaps.c86 int vmw_devcaps_create(struct vmw_private *vmw) in vmw_devcaps_create() argument
93 if (!vmw->devcaps) in vmw_devcaps_create()
96 vmw_write(vmw, SVGA_REG_DEV_CAP, i); in vmw_devcaps_create()
97 vmw->devcaps[i] = vmw_read(vmw, SVGA_REG_DEV_CAP); in vmw_devcaps_create()
105 vfree(vmw->devcaps); in vmw_devcaps_destroy()
106 vmw->devcaps = NULL; in vmw_devcaps_destroy()
119 else if (vmw->fifo_mem != NULL) in vmw_devcaps_size()
132 memcpy(dst, vmw->devcaps, dst_size); in vmw_devcaps_copy()
134 ret = vmw_fill_compat_cap(vmw, dst, dst_size); in vmw_devcaps_copy()
137 } else if (vmw->fifo_mem) { in vmw_devcaps_copy()
[all …]
A Dvmwgfx_vkms.c214 vmw->vkms_enabled = false; in vmw_vkms_init()
228 if (!vmw->crc_workq) { in vmw_vkms_init()
232 if (vmw->vkms_enabled) in vmw_vkms_init()
254 if (!vmw->vkms_enabled) in vmw_vkms_get_vblank_timestamp()
288 if (!vmw->vkms_enabled) in vmw_vkms_enable_vblank()
307 if (!vmw->vkms_enabled) in vmw_vkms_disable_vblank()
350 if (vmw->vkms_enabled) in vmw_vkms_crtc_atomic_begin()
361 if (!vmw->vkms_enabled) in vmw_vkms_crtc_atomic_flush()
386 if (vmw->vkms_enabled) in vmw_vkms_crtc_atomic_enable()
396 if (vmw->vkms_enabled) in vmw_vkms_crtc_atomic_disable()
[all …]
A Dvmwgfx_devcaps.h35 int vmw_devcaps_create(struct vmw_private *vmw);
36 void vmw_devcaps_destroy(struct vmw_private *vmw);
37 uint32_t vmw_devcaps_size(const struct vmw_private *vmw, bool gb_aware);
38 int vmw_devcaps_copy(struct vmw_private *vmw, bool gb_aware,
41 static inline uint32_t vmw_devcap_get(struct vmw_private *vmw, in vmw_devcap_get() argument
44 bool gb_objects = !!(vmw->capabilities & SVGA_CAP_GBOBJECTS); in vmw_devcap_get()
46 return vmw->devcaps[devcap]; in vmw_devcap_get()
A Dvmwgfx_drv.c459 atomic_set(&vmw->last_read_seqno, vmw_fence_read(vmw)); in vmw_device_fini()
464 vmw->enable_state); in vmw_device_fini()
466 vmw->traces_state); in vmw_device_fini()
468 vmw_fifo_destroy(vmw); in vmw_device_fini()
1313 drm_warn_once(&vmw->drm, in vmwgfx_supported()
1414 if (vmw->has_gmr) in vmw_debugfs_resource_managers_init()
1417 if (vmw->has_mob) { in vmw_debugfs_resource_managers_init()
1622 struct vmw_private *vmw; in vmw_probe() local
1635 if (IS_ERR(vmw)) { in vmw_probe()
1636 ret = PTR_ERR(vmw); in vmw_probe()
[all …]
A Dvmwgfx_drv.h728 bool vmwgfx_supported(struct vmw_private *vmw);
1427 BUG_ON(vmw_is_svga_v3(vmw)); in vmw_fifo_mem_read()
1428 return READ_ONCE(*(vmw->fifo_mem + fifo_reg)); in vmw_fifo_mem_read()
1442 BUG_ON(vmw_is_svga_v3(vmw)); in vmw_fifo_mem_write()
1443 WRITE_ONCE(*(vmw->fifo_mem + fifo_reg), value); in vmw_fifo_mem_write()
1466 if (vmw_is_svga_v3(vmw)) in vmw_irq_status_read()
1467 status = vmw_read(vmw, SVGA_REG_IRQ_STATUS); in vmw_irq_status_read()
1469 status = inl(vmw->io_start + SVGA_IRQSTATUS_PORT); in vmw_irq_status_read()
1476 if (vmw_is_svga_v3(vmw)) in vmw_irq_status_write()
1477 vmw_write(vmw, SVGA_REG_IRQ_STATUS, status); in vmw_irq_status_write()
[all …]
A Dvmwgfx_cmd.c386 struct vmw_private *vmw, in vmw_fifo_res_copy() argument
390 u32 *fifo_mem = vmw->fifo_mem; in vmw_fifo_res_copy()
399 vmw_fifo_mem_write(vmw, SVGA_FIFO_RESERVED, bytes); in vmw_fifo_res_copy()
408 struct vmw_private *vmw, in vmw_fifo_slow_copy() argument
416 vmw_fifo_mem_write(vmw, (next_cmd >> 2), *buffer++); in vmw_fifo_slow_copy()
421 vmw_fifo_mem_write(vmw, SVGA_FIFO_NEXT_CMD, next_cmd); in vmw_fifo_slow_copy()
677 bool vmw_cmd_supported(struct vmw_private *vmw) in vmw_cmd_supported() argument
680 (vmw->capabilities & (SVGA_CAP_COMMAND_BUFFERS | in vmw_cmd_supported()
682 if (vmw_is_svga_v3(vmw)) in vmw_cmd_supported()
684 (vmw->capabilities & SVGA_CAP_GBOBJECTS) != 0); in vmw_cmd_supported()
[all …]
A Dvmwgfx_cursor_plane.c74 vmw_cursor_plane_update_legacy(struct vmw_private *vmw, in vmw_cursor_plane_update_legacy() argument
85 vmw_send_define_cursor_cmd(vmw, surface->snooper.image, in vmw_cursor_plane_update_legacy()
100 if (vmw->has_mob) { in vmw_cursor_update_type()
101 if ((vmw->capabilities2 & SVGA_CAP2_CURSOR_MOB) != 0) in vmw_cursor_update_type()
108 static void vmw_cursor_update_mob(struct vmw_private *vmw, in vmw_cursor_update_mob() argument
599 struct vmw_private *vmw = vmw_priv(plane->dev); in vmw_cursor_plane_prepare_fb() local
619 vps->cursor.update_type = vmw_cursor_update_type(vmw, vps); in vmw_cursor_plane_prepare_fb()
693 struct vmw_private *vmw = vmw_priv(plane->dev); in vmw_cursor_plane_atomic_check() local
716 update_type = vmw_cursor_update_type(vmw, vps); in vmw_cursor_plane_atomic_check()
720 drm_warn(&vmw->drm, in vmw_cursor_plane_atomic_check()
[all …]
A Dvmwgfx_vkms.h40 void vmw_vkms_init(struct vmw_private *vmw);
41 void vmw_vkms_cleanup(struct vmw_private *vmw);
A Dvmwgfx_prime.c93 struct vmw_private *vmw = vmw_priv(dev); in vmw_prime_handle_to_fd() local
109 surf_handle = vmw_lookup_surface_handle_for_buffer(vmw, in vmw_prime_handle_to_fd()
A Dvmwgfx_bo.c452 int vmw_bo_create(struct vmw_private *vmw, in vmw_bo_create() argument
467 ret = vmw_bo_init(vmw, *p_bo, params, vmw_bo_free); in vmw_bo_create()
818 struct vmw_private *vmw = vmw_priv_from_ttm(bdev); in vmw_bo_placement_set() local
833 drm_warn(&vmw->drm, in vmw_bo_placement_set()
844 struct vmw_private *vmw = vmw_priv_from_ttm(bdev); in vmw_bo_placement_set_default_accelerated() local
847 if (vmw->has_mob) in vmw_bo_placement_set_default_accelerated()
A Dvmwgfx_blit.c427 struct vmw_private *vmw = in map_external() local
435 drm_dbg_driver(&vmw->drm, in map_external()
462 struct vmw_private *vmw = in vmw_external_bo_copy() local
475 drm_dbg_driver(&vmw->drm, "Wasn't able to map src\n"); in vmw_external_bo_copy()
482 drm_dbg_driver(&vmw->drm, "Wasn't able to map dst\n"); in vmw_external_bo_copy()
A Dvmwgfx_irq.c35 static u32 vmw_irqflag_fence_goal(struct vmw_private *vmw) in vmw_irqflag_fence_goal() argument
37 if ((vmw->capabilities2 & SVGA_CAP2_EXTRA_REGS) != 0) in vmw_irqflag_fence_goal()
A Dvmwgfx_surface.c873 vmw_lookup_user_surface_for_buffer(struct vmw_private *vmw, struct vmw_bo *bo, in vmw_lookup_user_surface_for_buffer() argument
886 drm_dbg_driver(&vmw->drm, in vmw_lookup_user_surface_for_buffer()
898 struct vmw_surface *vmw_lookup_surface_for_buffer(struct vmw_private *vmw, in vmw_lookup_surface_for_buffer() argument
903 vmw_lookup_user_surface_for_buffer(vmw, bo, handle); in vmw_lookup_surface_for_buffer()
915 u32 vmw_lookup_surface_handle_for_buffer(struct vmw_private *vmw, in vmw_lookup_surface_handle_for_buffer() argument
920 vmw_lookup_user_surface_for_buffer(vmw, bo, handle); in vmw_lookup_surface_handle_for_buffer()
2205 static SVGA3dSurfaceFormat vmw_format_bpp_to_svga(struct vmw_private *vmw, in vmw_format_bpp_to_svga() argument
2214 if (has_sm4_context(vmw)) in vmw_format_bpp_to_svga()
2218 drm_warn(&vmw->drm, "Unsupported format bpp: %d\n", bpp); in vmw_format_bpp_to_svga()
A Dvmwgfx_kms.c137 struct vmw_private *vmw = vmw_priv(crtc->dev); in vmw_du_crtc_atomic_check() local
150 drm_dbg_driver(&vmw->drm, in vmw_du_crtc_atomic_check()
1133 struct vmw_private *vmw = vmw_priv(old_state->dev); in vmw_atomic_commit_tail() local
1140 if (vmw->vkms_enabled) { in vmw_atomic_commit_tail()
A Dvmwgfx_stdu.c1465 struct vmw_private *vmw = vmw_priv(crtc->dev); in vmw_stdu_crtc_atomic_flush() local
1468 if (vmw->vkms_enabled) in vmw_stdu_crtc_atomic_flush()

Completed in 41 milliseconds