| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_fence.c | 130 seq = ++ring->fence_drv.sync_seq; in amdgpu_fence_emit() 134 &ring->fence_drv.lock, in amdgpu_fence_emit() 140 &ring->fence_drv.lock, in amdgpu_fence_emit() 148 ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask]; in amdgpu_fence_emit() 322 ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask]; in amdgpu_fence_wait_empty() 490 ring->fence_drv.gpu_addr = 0; in amdgpu_fence_driver_init_ring() 491 ring->fence_drv.sync_seq = 0; in amdgpu_fence_driver_init_ring() 502 if (!ring->fence_drv.fences) in amdgpu_fence_driver_init_ring() 589 ring->fence_drv.irq_src && in amdgpu_fence_driver_hw_fini() 606 if (!ring || !ring->fence_drv.initialized || !ring->fence_drv.irq_src) in amdgpu_fence_driver_isr_toggle() [all …]
|
| A D | amdgpu_userq_fence.c | 77 fence_drv = kzalloc(sizeof(*fence_drv), GFP_KERNEL); in amdgpu_userq_fence_driver_alloc() 78 if (!fence_drv) in amdgpu_userq_fence_driver_alloc() 82 r = amdgpu_seq64_alloc(adev, &fence_drv->va, &fence_drv->gpu_addr, in amdgpu_userq_fence_driver_alloc() 104 userq->fence_drv = fence_drv; in amdgpu_userq_fence_driver_alloc() 111 kfree(fence_drv); in amdgpu_userq_fence_driver_alloc() 149 if (!fence_drv) in amdgpu_userq_fence_driver_process() 233 fence_drv = userq->fence_drv; in amdgpu_userq_fence_create() 234 if (!fence_drv) in amdgpu_userq_fence_create() 240 userq_fence->fence_drv = fence_drv; in amdgpu_userq_fence_create() 306 struct amdgpu_userq_fence_driver *fence_drv = fence->fence_drv; in amdgpu_userq_fence_signaled() local [all …]
|
| A D | amdgpu_userq_fence.h | 41 struct amdgpu_userq_fence_driver *fence_drv; member 64 void amdgpu_userq_fence_driver_get(struct amdgpu_userq_fence_driver *fence_drv); 65 void amdgpu_userq_fence_driver_put(struct amdgpu_userq_fence_driver *fence_drv); 69 void amdgpu_userq_fence_driver_process(struct amdgpu_userq_fence_driver *fence_drv);
|
| A D | umsch_mm_v4_0.c | 310 set_hw_resources.api_status.api_completion_fence_addr = umsch->ring.fence_drv.gpu_addr; in umsch_mm_v4_0_set_hw_resources() 311 set_hw_resources.api_status.api_completion_fence_value = ++umsch->ring.fence_drv.sync_seq; in umsch_mm_v4_0_set_hw_resources() 360 add_queue.api_status.api_completion_fence_addr = umsch->ring.fence_drv.gpu_addr; in umsch_mm_v4_0_add_queue() 361 add_queue.api_status.api_completion_fence_value = ++umsch->ring.fence_drv.sync_seq; in umsch_mm_v4_0_add_queue() 392 remove_queue.api_status.api_completion_fence_addr = umsch->ring.fence_drv.gpu_addr; in umsch_mm_v4_0_remove_queue() 393 remove_queue.api_status.api_completion_fence_value = ++umsch->ring.fence_drv.sync_seq; in umsch_mm_v4_0_remove_queue()
|
| A D | amdgpu_ring_mux.c | 99 last_seq = atomic_read(&e->ring->fence_drv.last_seq); in amdgpu_mux_resubmit_chunks() 109 le32_to_cpu(*(e->ring->fence_drv.cpu_addr + 2))) { in amdgpu_mux_resubmit_chunks() 474 last_seq = atomic_read(&ring->fence_drv.last_seq); in scan_and_remove_signaled_chunk() 537 chunk->sync_seq = READ_ONCE(ring->fence_drv.sync_seq); in amdgpu_ring_mux_end_ib() 570 mux->seqno_to_resubmit = ring->fence_drv.sync_seq; in amdgpu_mcbp_handle_trailing_fence_irq()
|
| A D | amdgpu_userq.h | 64 struct amdgpu_userq_fence_driver *fence_drv; member
|
| A D | amdgpu_debugfs.c | 1817 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_ib_preempt_fences_swap() 1820 last_seq = atomic_read(&ring->fence_drv.last_seq); in amdgpu_ib_preempt_fences_swap() 1821 sync_seq = ring->fence_drv.sync_seq; in amdgpu_ib_preempt_fences_swap() 1878 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_ib_preempt_mark_partial_job() 1932 length = ring->fence_drv.num_fences_mask + 1; in amdgpu_debugfs_ib_preempt() 1954 if (atomic_read(&ring->fence_drv.last_seq) != in amdgpu_debugfs_ib_preempt() 1955 ring->fence_drv.sync_seq) { in amdgpu_debugfs_ib_preempt()
|
| A D | amdgpu_job.c | 123 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout() 124 ring->fence_drv.sync_seq); in amdgpu_job_timedout()
|
| A D | sdma_v6_0.c | 1168 uint32_t seq = ring->fence_drv.sync_seq; in sdma_v6_0_ring_emit_pipeline_sync() 1169 uint64_t addr = ring->fence_drv.gpu_addr; in sdma_v6_0_ring_emit_pipeline_sync() 1654 struct amdgpu_userq_fence_driver *fence_drv = NULL; in sdma_v6_0_process_fence_irq() local 1661 fence_drv = xa_load(xa, doorbell_offset); in sdma_v6_0_process_fence_irq() 1662 if (fence_drv) in sdma_v6_0_process_fence_irq() 1663 amdgpu_userq_fence_driver_process(fence_drv); in sdma_v6_0_process_fence_irq()
|
| A D | sdma_v7_0.c | 1189 uint32_t seq = ring->fence_drv.sync_seq; in sdma_v7_0_ring_emit_pipeline_sync() 1190 uint64_t addr = ring->fence_drv.gpu_addr; in sdma_v7_0_ring_emit_pipeline_sync() 1578 struct amdgpu_userq_fence_driver *fence_drv = NULL; in sdma_v7_0_process_fence_irq() local 1585 fence_drv = xa_load(xa, doorbell_offset); in sdma_v7_0_process_fence_irq() 1586 if (fence_drv) in sdma_v7_0_process_fence_irq() 1587 amdgpu_userq_fence_driver_process(fence_drv); in sdma_v7_0_process_fence_irq()
|
| A D | amdgpu_umsch_mm.c | 56 r = amdgpu_fence_wait_polling(ring, ring->fence_drv.sync_seq, adev->usec_timeout); in amdgpu_umsch_mm_query_fence() 59 ring->fence_drv.sync_seq); in amdgpu_umsch_mm_query_fence()
|
| A D | uvd_v6_0.c | 1092 uint32_t seq = ring->fence_drv.sync_seq; in uvd_v6_0_ring_emit_pipeline_sync() 1093 uint64_t addr = ring->fence_drv.gpu_addr; in uvd_v6_0_ring_emit_pipeline_sync() 1121 uint32_t seq = ring->fence_drv.sync_seq; in uvd_v6_0_enc_ring_emit_pipeline_sync() 1122 uint64_t addr = ring->fence_drv.gpu_addr; in uvd_v6_0_enc_ring_emit_pipeline_sync()
|
| A D | si_dma.c | 431 uint32_t seq = ring->fence_drv.sync_seq; in si_dma_ring_emit_pipeline_sync() 432 uint64_t addr = ring->fence_drv.gpu_addr; in si_dma_ring_emit_pipeline_sync()
|
| A D | mes_userqueue.c | 239 userq_props->fence_address = queue->fence_drv->gpu_addr; in mes_userq_mqd_create()
|
| A D | vce_v3_0.c | 887 uint32_t seq = ring->fence_drv.sync_seq; in vce_v3_0_emit_pipeline_sync() 888 uint64_t addr = ring->fence_drv.gpu_addr; in vce_v3_0_emit_pipeline_sync()
|
| /drivers/gpu/drm/radeon/ |
| A D | radeon_fence.c | 119 &rdev->fence_drv[ring].lockup_work, in radeon_fence_schedule_check() 269 struct radeon_fence_driver *fence_drv; in radeon_fence_check_lockup() local 275 rdev = fence_drv->rdev; in radeon_fence_check_lockup() 276 ring = fence_drv - &rdev->fence_drv[0]; in radeon_fence_check_lockup() 287 fence_drv->delayed_irq = false; in radeon_fence_check_lockup() 301 fence_drv->sync_seq[ring], ring); in radeon_fence_check_lockup() 794 rdev->fence_drv[ring].scratch_reg - in radeon_fence_driver_start_ring() 822 rdev->fence_drv[ring].gpu_addr = 0; in radeon_fence_driver_init_ring() 829 rdev->fence_drv[ring].rdev = rdev; in radeon_fence_driver_init_ring() 911 if (!rdev->fence_drv[i].initialized) in radeon_debugfs_fence_info_show() [all …]
|
| A D | uvd_v2_2.c | 43 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v2_2_fence_emit()
|
| A D | evergreen_dma.c | 44 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in evergreen_dma_fence_ring_emit()
|
| A D | uvd_v1_0.c | 85 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v1_0_fence_emit()
|
| A D | r600_dma.c | 290 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in r600_dma_fence_ring_emit()
|
| /drivers/gpu/drm/virtio/ |
| A D | virtgpu_fence.c | 63 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_alloc() 89 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_emit() 114 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_event_process() 119 atomic64_set(&vgdev->fence_drv.last_fence_id, fence_id); in virtio_gpu_fence_event_process()
|
| A D | virtgpu_debugfs.c | 75 (u64)atomic64_read(&vgdev->fence_drv.last_fence_id), in virtio_gpu_debugfs_irq_info() 76 vgdev->fence_drv.current_fence_id); in virtio_gpu_debugfs_irq_info()
|
| A D | virtgpu_kms.c | 149 vgdev->fence_drv.context = dma_fence_context_alloc(1); in virtio_gpu_init() 150 spin_lock_init(&vgdev->fence_drv.lock); in virtio_gpu_init() 151 INIT_LIST_HEAD(&vgdev->fence_drv.fences); in virtio_gpu_init()
|
| A D | virtgpu_ioctl.c | 172 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_resource_create_ioctl() 259 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_transfer_from_host_ioctl() 319 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, in virtio_gpu_transfer_to_host_ioctl()
|
| A D | virtgpu_object.c | 111 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_detach_object_fenced()
|