| /drivers/gpu/drm/virtio/ |
| A D | virtgpu_kms.c | 47 if (vgdev->has_edid) in virtio_gpu_config_changed_work_func() 77 if (!vgdev->capsets) { in virtio_gpu_get_capsets() 107 vgdev->capset_id_mask |= 1 << vgdev->capsets[i].id; in virtio_gpu_get_capsets() 133 if (!vgdev) in virtio_gpu_init() 136 vgdev->ddev = dev; in virtio_gpu_init() 138 vgdev->vdev = vdev; in virtio_gpu_init() 177 if (virtio_get_shm_region(vgdev->vdev, &vgdev->host_visible_region, in virtio_gpu_init() 256 wait_event_timeout(vgdev->resp_wq, !vgdev->display_info_pending, in virtio_gpu_init() 264 vgdev->vdev->config->del_vqs(vgdev->vdev); in virtio_gpu_init() 289 vgdev->vdev->config->del_vqs(vgdev->vdev); in virtio_gpu_deinit() [all …]
|
| A D | virtgpu_vq.c | 78 if (!vgdev->vbufs) in virtio_gpu_alloc_vbufs() 86 vgdev->vbufs = NULL; in virtio_gpu_free_vbufs() 271 free_vbuf(vgdev, entry); in virtio_gpu_dequeue_ctrl_func() 297 free_vbuf(vgdev, entry); in virtio_gpu_dequeue_cursor_func() 353 if (vgdev->has_indirect) in virtio_gpu_panic_queue_ctrl_sgs() 384 free_vbuf(vgdev, vbuf); in virtio_gpu_queue_ctrl_sgs() 388 if (vgdev->has_indirect) in virtio_gpu_queue_ctrl_sgs() 396 virtio_gpu_notify(vgdev); in virtio_gpu_queue_ctrl_sgs() 553 free_vbuf(vgdev, vbuf); in virtio_gpu_queue_cursor() 839 wake_up(&vgdev->resp_wq); in virtio_gpu_cmd_get_display_info_cb() [all …]
|
| A D | virtgpu_ioctl.c | 141 if (vgdev->has_virgl_3d) { in virtio_gpu_resource_create_ioctl() 172 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_resource_create_ioctl() 259 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_transfer_from_host_ioctl() 269 virtio_gpu_notify(vgdev); in virtio_gpu_transfer_from_host_ioctl() 303 (vgdev, offset, in virtio_gpu_transfer_to_host_ioctl() 319 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, in virtio_gpu_transfer_to_host_ioctl() 325 (vgdev, in virtio_gpu_transfer_to_host_ioctl() 331 virtio_gpu_notify(vgdev); in virtio_gpu_transfer_to_host_ioctl() 420 virtio_gpu_notify(vgdev); in virtio_gpu_get_caps_ioctl() 514 if (vgdev->has_virgl_3d) in virtio_gpu_resource_create_blob_ioctl() [all …]
|
| A D | virtgpu_debugfs.c | 49 virtio_gpu_add_bool(m, "virgl", vgdev->has_virgl_3d); in virtio_gpu_features() 50 virtio_gpu_add_bool(m, "edid", vgdev->has_edid); in virtio_gpu_features() 54 vgdev->has_resource_assign_uuid); in virtio_gpu_features() 58 virtio_gpu_add_int(m, "cap sets", vgdev->num_capsets); in virtio_gpu_features() 59 virtio_gpu_add_int(m, "scanouts", vgdev->num_scanouts); in virtio_gpu_features() 60 if (vgdev->host_visible_region.len) { in virtio_gpu_features() 62 (unsigned long)vgdev->host_visible_region.addr, in virtio_gpu_features() 63 (unsigned long)vgdev->host_visible_region.len); in virtio_gpu_features() 76 vgdev->fence_drv.current_fence_id); in virtio_gpu_debugfs_irq_info() 87 if (!vgdev->has_host_visible) { in virtio_gpu_debugfs_host_visible_mm() [all …]
|
| A D | virtgpu_object.c | 60 ida_free(&vgdev->resource_ida, id - 1); in virtio_gpu_resource_id_put() 74 spin_lock(&vgdev->host_visible_lock); in virtio_gpu_cleanup_object() 78 spin_unlock(&vgdev->host_visible_lock); in virtio_gpu_cleanup_object() 95 virtio_gpu_cmd_unref_resource(vgdev, bo); in virtio_gpu_free_object() 96 virtio_gpu_notify(vgdev); in virtio_gpu_free_object() 111 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_detach_object_fenced() 115 virtio_gpu_object_detach(vgdev, bo, fence); in virtio_gpu_detach_object_fenced() 116 virtio_gpu_notify(vgdev); in virtio_gpu_detach_object_fenced() 254 virtio_gpu_object_attach(vgdev, bo, ents, nents); in virtio_gpu_object_create() 256 virtio_gpu_cmd_create_resource(vgdev, bo, params, in virtio_gpu_object_create() [all …]
|
| A D | virtgpu_drv.h | 329 int virtio_gpu_alloc_vbufs(struct virtio_gpu_device *vgdev); 330 void virtio_gpu_free_vbufs(struct virtio_gpu_device *vgdev); 371 void virtio_gpu_cursor_ping(struct virtio_gpu_device *vgdev, 390 void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev, 422 void virtio_gpu_notify(struct virtio_gpu_device *vgdev); 428 int virtio_gpu_cmd_map(struct virtio_gpu_device *vgdev, 431 void virtio_gpu_cmd_unmap(struct virtio_gpu_device *vgdev, 449 int virtio_gpu_modeset_init(struct virtio_gpu_device *vgdev); 462 void virtio_gpu_fence_emit(struct virtio_gpu_device *vgdev, 472 int virtio_gpu_object_create(struct virtio_gpu_device *vgdev, [all …]
|
| A D | virtgpu_gem.c | 84 if (vgdev->has_resource_blob && !vgdev->has_virgl_3d) { in virtio_gpu_mode_dumb_create() 109 if (!vgdev->has_virgl_3d) in virtio_gpu_gem_object_open() 115 if (!vgdev->has_context_init) in virtio_gpu_gem_object_open() 128 virtio_gpu_notify(vgdev); in virtio_gpu_gem_object_open() 139 if (!vgdev->has_virgl_3d) in virtio_gpu_gem_object_close() 149 virtio_gpu_notify(vgdev); in virtio_gpu_gem_object_close() 276 spin_lock(&vgdev->obj_free_lock); in virtio_gpu_array_put_free_delayed() 278 spin_unlock(&vgdev->obj_free_lock); in virtio_gpu_array_put_free_delayed() 284 struct virtio_gpu_device *vgdev = in virtio_gpu_array_put_free_work() local 288 spin_lock(&vgdev->obj_free_lock); in virtio_gpu_array_put_free_work() [all …]
|
| A D | virtgpu_display.c | 96 virtio_gpu_notify(vgdev); in virtio_gpu_crtc_mode_set_nofb() 112 virtio_gpu_notify(vgdev); in virtio_gpu_crtc_atomic_disable() 255 struct drm_device *dev = vgdev->ddev; in vgdev_output_init() 282 if (vgdev->has_edid) in vgdev_output_init() 339 if (!vgdev->num_scanouts) in virtio_gpu_modeset_init() 342 ret = drmm_mode_config_init(vgdev->ddev); in virtio_gpu_modeset_init() 357 for (i = 0 ; i < vgdev->num_scanouts; ++i) in virtio_gpu_modeset_init() 358 vgdev_output_init(vgdev, i); in virtio_gpu_modeset_init() 360 drm_mode_config_reset(vgdev->ddev); in virtio_gpu_modeset_init() 368 if (!vgdev->num_scanouts) in virtio_gpu_modeset_fini() [all …]
|
| A D | virtgpu_vram.c | 14 spin_lock(&vgdev->host_visible_lock); in virtio_gpu_vram_free() 16 spin_unlock(&vgdev->host_visible_lock); in virtio_gpu_vram_free() 19 virtio_gpu_cmd_unmap(vgdev, bo); in virtio_gpu_vram_free() 21 virtio_gpu_cmd_unref_resource(vgdev, bo); in virtio_gpu_vram_free() 22 virtio_gpu_notify(vgdev); in virtio_gpu_vram_free() 152 if (!vgdev->has_host_visible) in virtio_gpu_vram_map() 155 spin_lock(&vgdev->host_visible_lock); in virtio_gpu_vram_map() 158 spin_unlock(&vgdev->host_visible_lock); in virtio_gpu_vram_map() 173 ret = virtio_gpu_cmd_map(vgdev, objs, offset); in virtio_gpu_vram_map() 182 spin_lock(&vgdev->host_visible_lock); in virtio_gpu_vram_map() [all …]
|
| A D | virtgpu_plane.c | 194 virtio_gpu_panic_notify(vgdev); in virtio_gpu_panic_resource_flush() 221 virtio_gpu_notify(vgdev); in virtio_gpu_resource_flush() 227 virtio_gpu_notify(vgdev); in virtio_gpu_resource_flush() 255 virtio_gpu_notify(vgdev); in virtio_gpu_primary_plane_update() 284 (vgdev, output->index, bo, in virtio_gpu_primary_plane_update() 291 virtio_gpu_cmd_set_scanout(vgdev, output->index, in virtio_gpu_primary_plane_update() 371 vgdev->fence_drv.context, in virtio_gpu_plane_prepare_fb() 463 (vgdev, 0, in virtio_gpu_cursor_plane_update() 467 virtio_gpu_notify(vgdev); in virtio_gpu_cursor_plane_update() 498 virtio_gpu_cursor_ping(vgdev, output); in virtio_gpu_cursor_plane_update() [all …]
|
| A D | virtgpu_prime.c | 37 struct virtio_gpu_device *vgdev = obj->dev->dev_private; in virtgpu_virtio_get_uuid() local 102 return virtio_gpu_cmd_resource_assign_uuid(vgdev, objs); in virtio_gpu_resource_assign_uuid() 110 struct virtio_gpu_device *vgdev = dev->dev_private; in virtgpu_gem_prime_export() local 117 if (vgdev->has_resource_assign_uuid) { in virtgpu_gem_prime_export() 118 ret = virtio_gpu_resource_assign_uuid(vgdev, bo); in virtgpu_gem_prime_export() 122 virtio_gpu_notify(vgdev); in virtgpu_gem_prime_export() 221 virtio_gpu_cmd_unref_resource(vgdev, bo); in virtgpu_dma_buf_free_obj() 222 virtio_gpu_notify(vgdev); in virtgpu_dma_buf_free_obj() 232 struct virtio_gpu_device *vgdev = dev->dev_private; in virtgpu_dma_buf_init_obj() local 297 struct virtio_gpu_device *vgdev = dev->dev_private; in virtgpu_gem_prime_import() local [all …]
|
| A D | virtgpu_fence.c | 58 struct virtio_gpu_fence *virtio_gpu_fence_alloc(struct virtio_gpu_device *vgdev, in virtio_gpu_fence_alloc() argument 63 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_alloc() 85 void virtio_gpu_fence_emit(struct virtio_gpu_device *vgdev, in virtio_gpu_fence_emit() argument 89 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_emit() 111 void virtio_gpu_fence_event_process(struct virtio_gpu_device *vgdev, in virtio_gpu_fence_event_process() argument 114 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_event_process() 119 atomic64_set(&vgdev->fence_drv.last_fence_id, fence_id); in virtio_gpu_fence_event_process() 140 drm_send_event(vgdev->ddev, &curr->e->base); in virtio_gpu_fence_event_process() 150 drm_send_event(vgdev->ddev, &signaled->e->base); in virtio_gpu_fence_event_process()
|
| A D | virtgpu_submit.c | 39 struct virtio_gpu_device *vgdev; member 353 virtio_gpu_cmd_submit(submit->vgdev, submit->buf, submit->exbuf->size, in virtio_gpu_submit() 356 virtio_gpu_notify(submit->vgdev); in virtio_gpu_submit() 374 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_init_submit() local 391 out_fence = virtio_gpu_fence_alloc(vgdev, fence_ctx, ring_idx); in virtio_gpu_init_submit() 408 submit->vgdev = vgdev; in virtio_gpu_init_submit() 476 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_execbuffer_ioctl() local 478 u64 fence_ctx = vgdev->fence_drv.context; in virtio_gpu_execbuffer_ioctl() 484 if (!vgdev->has_virgl_3d) in virtio_gpu_execbuffer_ioctl()
|
| A D | virtgpu_drv.c | 142 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_config_changed() local 144 schedule_work(&vgdev->config_changed_work); in virtio_gpu_config_changed()
|