| /drivers/dma-buf/ |
| A D | st-dma-resv.c | 40 struct dma_resv resv; in sanitycheck() local 53 dma_resv_init(&resv); in sanitycheck() 59 dma_resv_fini(&resv); in sanitycheck() 66 struct dma_resv resv; in test_signaling() local 76 dma_resv_init(&resv); in test_signaling() 104 dma_resv_fini(&resv); in test_signaling() 114 struct dma_resv resv; in test_for_each() local 123 dma_resv_init(&resv); in test_for_each() 165 dma_resv_fini(&resv); in test_for_each() 175 struct dma_resv resv; in test_for_each_unlocked() local [all …]
|
| A D | dma-buf.c | 163 dma_resv_fini(dmabuf->resv); in dma_buf_release() 315 struct dma_resv *resv; in dma_buf_poll() local 322 resv = dmabuf->resv; in dma_buf_poll() 330 dma_resv_lock(resv, NULL); in dma_buf_poll() 378 dma_resv_unlock(resv); in dma_buf_poll() 686 struct dma_resv *resv = exp_info->resv; in dma_buf_export() local 709 if (!exp_info->resv) in dma_buf_export() 731 if (!resv) { in dma_buf_export() 735 dmabuf->resv = resv; in dma_buf_export() 751 if (!resv) in dma_buf_export() [all …]
|
| /drivers/gpu/drm/ttm/tests/ |
| A D | ttm_bo_test.c | 70 dma_resv_unlock(bo->base.resv); in ttm_bo_reserve_optimistic_no_ticket() 86 dma_resv_unlock(bo->base.resv); in ttm_bo_reserve_locked_no_sleep() 126 dma_resv_unlock(bo->base.resv); in ttm_bo_reserve_double_resv() 156 bo2->base.resv->lock.ctx = &ctx2; in ttm_bo_reserve_deadlock() 166 dma_resv_unlock(bo1->base.resv); in ttm_bo_reserve_deadlock() 330 struct dma_resv *resv; in ttm_bo_unreserve_bulk() local 342 resv = kunit_kzalloc(test, sizeof(*resv), GFP_KERNEL); in ttm_bo_unreserve_bulk() 349 dma_resv_init(resv); in ttm_bo_unreserve_bulk() 379 dma_resv_fini(resv); in ttm_bo_unreserve_bulk() 410 dma_resv_unlock(bo->base.resv); in ttm_bo_put_basic() [all …]
|
| A D | ttm_bo_validate_test.c | 66 struct dma_resv *resv, in dma_resv_kunit_active_fence_init() argument 74 dma_resv_lock(resv, NULL); in dma_resv_kunit_active_fence_init() 75 dma_resv_reserve_fences(resv, 1); in dma_resv_kunit_active_fence_init() 77 dma_resv_unlock(resv); in dma_resv_kunit_active_fence_init() 127 dma_resv_unlock(bo->base.resv); in ttm_bo_init_reserved_sys_man() 176 dma_resv_unlock(bo->base.resv); in ttm_bo_init_reserved_mock_man() 202 struct dma_resv resv; in ttm_bo_init_reserved_resv() local 212 dma_resv_init(&resv); in ttm_bo_init_reserved_resv() 213 dma_resv_lock(&resv, NULL); in ttm_bo_init_reserved_resv() 221 KUNIT_EXPECT_PTR_EQ(test, bo->base.resv, &resv); in ttm_bo_init_reserved_resv() [all …]
|
| A D | ttm_tt_test.c | 161 dma_resv_lock(bo->base.resv, NULL); in ttm_tt_create_basic() 163 dma_resv_unlock(bo->base.resv); in ttm_tt_create_basic() 180 dma_resv_lock(bo->base.resv, NULL); in ttm_tt_create_invalid_bo_type() 182 dma_resv_unlock(bo->base.resv); in ttm_tt_create_invalid_bo_type() 204 dma_resv_lock(bo->base.resv, NULL); in ttm_tt_create_ttm_exists() 206 dma_resv_unlock(bo->base.resv); in ttm_tt_create_ttm_exists() 234 dma_resv_lock(bo->base.resv, NULL); in ttm_tt_create_failed() 236 dma_resv_unlock(bo->base.resv); in ttm_tt_create_failed() 249 dma_resv_lock(bo->base.resv, NULL); in ttm_tt_destroy_basic() 251 dma_resv_unlock(bo->base.resv); in ttm_tt_destroy_basic()
|
| /drivers/gpu/drm/ttm/ |
| A D | ttm_bo.c | 209 bo->base.resv = &bo->base._resv; in ttm_bo_individualize_resv() 244 dma_resv_unlock(bo->base.resv); in ttm_bo_delayed_delete() 314 dma_resv_unlock(bo->base.resv); in ttm_bo_release() 482 dma_resv_unlock(bo->base.resv); in ttm_bo_evict_first() 944 if (resv) in ttm_bo_init_reserved() 945 bo->base.resv = resv; in ttm_bo_init_reserved() 964 if (!resv) in ttm_bo_init_reserved() 967 dma_resv_assert_held(resv); in ttm_bo_init_reserved() 976 if (!resv) in ttm_bo_init_reserved() 1031 sg, resv, destroy); in ttm_bo_init_validate() [all …]
|
| A D | ttm_execbuf_util.c | 40 dma_resv_unlock(bo->base.resv); in ttm_eu_backoff_reservation_reverse() 56 dma_resv_unlock(bo->base.resv); in ttm_eu_backoff_reservation() 104 ret = dma_resv_reserve_fences(bo->base.resv, in ttm_eu_reserve_buffers() 121 ret = dma_resv_reserve_fences(bo->base.resv, in ttm_eu_reserve_buffers() 155 dma_resv_add_fence(bo->base.resv, fence, entry->num_shared ? in ttm_eu_fence_buffer_objects() 158 dma_resv_unlock(bo->base.resv); in ttm_eu_fence_buffer_objects()
|
| A D | ttm_bo_util.c | 254 fbo->base.base.resv = &fbo->base.base._resv; in ttm_buffer_object_transfer() 505 dma_resv_assert_held(bo->base.resv); in ttm_bo_vmap() 573 dma_resv_assert_held(bo->base.resv); in ttm_bo_vunmap() 810 if (dma_resv_trylock(bo->base.resv)) { in ttm_lru_walk_trylock() 815 if (bo->base.resv == ctx->resv && ctx->allow_res_evict) { in ttm_lru_walk_trylock() 816 dma_resv_assert_held(bo->base.resv); in ttm_lru_walk_trylock() 827 struct dma_resv *resv = bo->base.resv; in ttm_lru_walk_ticketlock() local 833 ret = dma_resv_lock(resv, arg->ticket); in ttm_lru_walk_ticketlock() 912 dma_resv_unlock(bo->base.resv); in ttm_bo_lru_cursor_cleanup_bo() 990 dma_resv_unlock(bo->base.resv); in __ttm_bo_lru_cursor_next() [all …]
|
| A D | ttm_bo_vm.c | 51 if (dma_resv_test_signaled(bo->base.resv, DMA_RESV_USAGE_KERNEL)) in ttm_bo_vm_fault_idle() 65 (void)dma_resv_wait_timeout(bo->base.resv, in ttm_bo_vm_fault_idle() 68 dma_resv_unlock(bo->base.resv); in ttm_bo_vm_fault_idle() 76 err = dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_KERNEL, true, in ttm_bo_vm_fault_idle() 127 if (unlikely(!dma_resv_trylock(bo->base.resv))) { in ttm_bo_vm_reserve() 137 if (!dma_resv_lock_interruptible(bo->base.resv, in ttm_bo_vm_reserve() 139 dma_resv_unlock(bo->base.resv); in ttm_bo_vm_reserve() 146 if (dma_resv_lock_interruptible(bo->base.resv, NULL)) in ttm_bo_vm_reserve() 156 dma_resv_unlock(bo->base.resv); in ttm_bo_vm_reserve() 345 dma_resv_unlock(bo->base.resv); in ttm_bo_vm_fault()
|
| /drivers/gpu/drm/ |
| A D | drm_gem_shmem_helper.c | 182 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_free() 302 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_pin() 326 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_unpin() 351 dma_resv_assert_held(obj->resv); in drm_gem_shmem_vmap_locked() 413 dma_resv_assert_held(obj->resv); in drm_gem_shmem_vunmap_locked() 563 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_fault() 585 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_vm_open() 597 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_vm_close() 647 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_mmap() 766 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_get_pages_sgt() [all …]
|
| A D | drm_exec.c | 62 dma_resv_unlock(obj->resv); in drm_exec_unlock_all() 175 ret = dma_resv_lock_slow_interruptible(obj->resv, in drm_exec_lock_contended() 180 dma_resv_lock_slow(obj->resv, &exec->ticket); in drm_exec_lock_contended() 191 dma_resv_unlock(obj->resv); in drm_exec_lock_contended() 224 ret = dma_resv_lock_interruptible(obj->resv, &exec->ticket); in drm_exec_lock_obj() 226 ret = dma_resv_lock(obj->resv, &exec->ticket); in drm_exec_lock_obj() 248 dma_resv_unlock(obj->resv); in drm_exec_lock_obj() 268 dma_resv_unlock(obj->resv); in drm_exec_unlock_obj() 300 ret = dma_resv_reserve_fences(obj->resv, num_fences); in drm_exec_prepare_obj()
|
| A D | drm_gem.c | 191 if (!obj->resv) in drm_gem_private_object_init() 192 obj->resv = &obj->_resv; in drm_gem_private_object_init() 1246 dma_resv_assert_held(obj->resv); in drm_gem_vmap_locked() 1263 dma_resv_assert_held(obj->resv); in drm_gem_vunmap_locked() 1278 dma_resv_lock(obj->resv, NULL); in drm_gem_lock() 1284 dma_resv_unlock(obj->resv); in drm_gem_unlock() 1292 dma_resv_lock(obj->resv, NULL); in drm_gem_vmap() 1294 dma_resv_unlock(obj->resv); in drm_gem_vmap() 1302 dma_resv_lock(obj->resv, NULL); in drm_gem_vunmap() 1304 dma_resv_unlock(obj->resv); in drm_gem_vunmap() [all …]
|
| /drivers/gpu/drm/virtio/ |
| A D | virtgpu_prime.c | 134 exp_info.resv = obj->resv; in virtgpu_gem_prime_export() 155 dma_resv_assert_held(attach->dmabuf->resv); in virtgpu_dma_buf_import_sgt() 157 ret = dma_resv_wait_timeout(attach->dmabuf->resv, in virtgpu_dma_buf_import_sgt() 190 dma_resv_assert_held(attach->dmabuf->resv); in virtgpu_dma_buf_unmap() 212 dma_resv_lock(dmabuf->resv, NULL); in virtgpu_dma_buf_free_obj() 214 dma_resv_unlock(dmabuf->resv); in virtgpu_dma_buf_free_obj() 234 struct dma_resv *resv = attach->dmabuf->resv; in virtgpu_dma_buf_init_obj() local 245 dma_resv_lock(resv, NULL); in virtgpu_dma_buf_init_obj() 265 dma_resv_unlock(resv); in virtgpu_dma_buf_init_obj() 272 dma_resv_unlock(resv); in virtgpu_dma_buf_init_obj() [all …]
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_dma_buf.c | 107 dma_resv_assert_held(dmabuf->resv); in amdgpu_dma_buf_pin() 343 struct dma_resv *resv = dma_buf->resv; in amdgpu_dma_buf_create_obj() local 350 dma_resv_lock(resv, NULL); in amdgpu_dma_buf_create_obj() 363 ttm_bo_type_sg, resv, &gobj, 0); in amdgpu_dma_buf_create_obj() 371 dma_resv_unlock(resv); in amdgpu_dma_buf_create_obj() 375 dma_resv_unlock(resv); in amdgpu_dma_buf_create_obj() 414 struct dma_resv *resv = vm->root.bo->tbo.base.resv; in amdgpu_dma_buf_move_notify() local 421 r = dma_resv_lock(resv, ticket); in amdgpu_dma_buf_move_notify() 430 if (!dma_resv_trylock(resv)) in amdgpu_dma_buf_move_notify() 435 r = dma_resv_reserve_fences(resv, 2); in amdgpu_dma_buf_move_notify() [all …]
|
| A D | amdgpu_eviction_fence.c | 201 struct dma_resv *resv = bo->tbo.base.resv; in amdgpu_eviction_fence_attach() local 204 if (!resv) in amdgpu_eviction_fence_attach() 207 ret = dma_resv_reserve_fences(resv, 1); in amdgpu_eviction_fence_attach() 216 dma_resv_add_fence(resv, &ev_fence->base, DMA_RESV_USAGE_BOOKKEEP); in amdgpu_eviction_fence_attach() 227 dma_resv_replace_fences(bo->tbo.base.resv, evf_mgr->ev_fence_ctx, in amdgpu_eviction_fence_detach()
|
| A D | amdgpu_object.c | 260 bp.resv = NULL; in amdgpu_bo_create_reserved() 635 .resv = bp->resv in amdgpu_bo_create() 705 bp->resv, bp->destroy); in amdgpu_bo_create() 728 if (!bp->resv) in amdgpu_bo_create() 741 if (!bp->resv) in amdgpu_bo_create() 742 dma_resv_unlock(bo->tbo.base.resv); in amdgpu_bo_create() 1143 dma_resv_assert_held(bo->tbo.base.resv); in amdgpu_bo_get_tiling_flags() 1390 struct dma_resv *resv = bo->tbo.base.resv; in amdgpu_bo_fence() local 1393 r = dma_resv_reserve_fences(resv, 1); in amdgpu_bo_fence() 1637 dma_resv_unlock(bo->tbo.base.resv); in amdgpu_bo_print_info() [all …]
|
| /drivers/gpu/drm/vgem/ |
| A D | vgem_fence.c | 116 struct dma_resv *resv; in vgem_fence_attach_ioctl() local 139 resv = obj->resv; in vgem_fence_attach_ioctl() 141 if (!dma_resv_test_signaled(resv, usage)) { in vgem_fence_attach_ioctl() 147 dma_resv_lock(resv, NULL); in vgem_fence_attach_ioctl() 148 ret = dma_resv_reserve_fences(resv, 1); in vgem_fence_attach_ioctl() 150 dma_resv_add_fence(resv, fence, arg->flags & VGEM_FENCE_WRITE ? in vgem_fence_attach_ioctl() 152 dma_resv_unlock(resv); in vgem_fence_attach_ioctl()
|
| /drivers/gpu/drm/msm/ |
| A D | msm_gem_shrinker.c | 59 struct dma_resv *resv = drm_gpuvm_resv(vm_bo->vm); in with_vm_locks() local 61 if (resv == obj->resv) in with_vm_locks() 64 ret = dma_resv_lock(resv, ticket); in with_vm_locks() 97 struct dma_resv *resv = drm_gpuvm_resv(vm_bo->vm); in with_vm_locks() local 99 if (resv == obj->resv) in with_vm_locks() 102 dma_resv_unlock(resv); in with_vm_locks() 143 return dma_resv_wait_timeout(obj->resv, usage, false, 10) > 0; in wait_for_idle()
|
| /drivers/gpu/drm/i915/gem/ |
| A D | i915_gem_wait.c | 35 i915_gem_object_boost(struct dma_resv *resv, unsigned int flags) in i915_gem_object_boost() argument 56 dma_resv_iter_begin(&cursor, resv, in i915_gem_object_boost() 66 i915_gem_object_wait_reservation(struct dma_resv *resv, in i915_gem_object_wait_reservation() argument 74 i915_gem_object_boost(resv, flags); in i915_gem_object_wait_reservation() 76 dma_resv_iter_begin(&cursor, resv, in i915_gem_object_wait_reservation() 149 dma_resv_iter_begin(&cursor, obj->base.resv, in i915_gem_object_wait_priority() 171 timeout = i915_gem_object_wait_reservation(obj->base.resv, in i915_gem_object_wait()
|
| /drivers/gpu/drm/loongson/ |
| A D | lsdc_gem.c | 22 dma_resv_assert_held(obj->resv); in lsdc_gem_prime_pin() 35 dma_resv_assert_held(obj->resv); in lsdc_gem_prime_unpin() 145 struct dma_resv *resv) in lsdc_gem_object_create() argument 152 lbo = lsdc_bo_create(ddev, domain, size, kerenl, sg, resv); in lsdc_gem_object_create() 179 struct dma_resv *resv = attach->dmabuf->resv; in lsdc_prime_import_sg_table() local 184 dma_resv_lock(resv, NULL); in lsdc_prime_import_sg_table() 186 sg, resv); in lsdc_prime_import_sg_table() 187 dma_resv_unlock(resv); in lsdc_prime_import_sg_table()
|
| /drivers/gpu/drm/radeon/ |
| A D | radeon_prime.c | 49 struct dma_resv *resv = attach->dmabuf->resv; in radeon_gem_prime_import_sg_table() local 54 dma_resv_lock(resv, NULL); in radeon_gem_prime_import_sg_table() 56 RADEON_GEM_DOMAIN_GTT, 0, sg, resv, &bo); in radeon_gem_prime_import_sg_table() 57 dma_resv_unlock(resv); in radeon_gem_prime_import_sg_table()
|
| A D | radeon_benchmark.c | 38 struct dma_resv *resv) in radeon_benchmark_do_move() argument 51 resv); in radeon_benchmark_do_move() 56 resv); in radeon_benchmark_do_move() 125 dobj->tbo.base.resv); in radeon_benchmark_move() 136 dobj->tbo.base.resv); in radeon_benchmark_move()
|
| /drivers/infiniband/core/ |
| A D | umem_dmabuf.c | 24 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_map_pages() 75 ret = dma_resv_wait_timeout(umem_dmabuf->attach->dmabuf->resv, in ib_umem_dmabuf_map_pages() 88 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_unmap_pages() 216 dma_resv_lock(umem_dmabuf->attach->dmabuf->resv, NULL); in ib_umem_dmabuf_get_pinned_with_dma_device() 225 dma_resv_unlock(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_get_pinned_with_dma_device() 232 dma_resv_unlock(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_get_pinned_with_dma_device() 252 dma_resv_lock(dmabuf->resv, NULL); in ib_umem_dmabuf_revoke() 262 dma_resv_unlock(dmabuf->resv); in ib_umem_dmabuf_revoke()
|
| /drivers/gpu/drm/xe/ |
| A D | xe_dma_buf.c | 209 struct dma_resv *resv = dma_buf->resv; in xe_dma_buf_init_obj() local 214 dma_resv_lock(resv, NULL); in xe_dma_buf_init_obj() 215 bo = ___xe_bo_create_locked(xe, storage, NULL, resv, NULL, dma_buf->size, in xe_dma_buf_init_obj() 222 dma_resv_unlock(resv); in xe_dma_buf_init_obj() 227 dma_resv_unlock(resv); in xe_dma_buf_init_obj()
|
| /drivers/gpu/drm/imagination/ |
| A D | pvr_gem.c | 203 dma_resv_lock(obj->resv, NULL); in pvr_gem_object_vmap() 219 dma_resv_unlock(obj->resv); in pvr_gem_object_vmap() 224 dma_resv_unlock(obj->resv); in pvr_gem_object_vmap() 247 dma_resv_lock(obj->resv, NULL); in pvr_gem_object_vunmap() 261 dma_resv_unlock(obj->resv); in pvr_gem_object_vunmap()
|