| /drivers/gpu/drm/ |
| A D | drm_gpuvm.c | 1019 gpuvm->ops = ops; in drm_gpuvm_init() 1036 __drm_gpuva_insert(gpuvm, &gpuvm->kernel_alloc_node); in drm_gpuvm_init() 1049 drm_WARN(gpuvm->drm, !RB_EMPTY_ROOT(&gpuvm->rb.tree.rb_root), in drm_gpuvm_fini() 1052 drm_WARN(gpuvm->drm, !list_empty(&gpuvm->extobj.list), in drm_gpuvm_fini() 1054 drm_WARN(gpuvm->drm, !list_empty(&gpuvm->evict.list), in drm_gpuvm_fini() 1067 if (drm_WARN_ON(gpuvm->drm, !gpuvm->ops->vm_free)) in drm_gpuvm_free() 1070 gpuvm->ops->vm_free(gpuvm); in drm_gpuvm_free() 1084 if (gpuvm) in drm_gpuvm_put() 1729 va->vm = gpuvm; in __drm_gpuva_insert() 2661 args.vm = gpuvm; in drm_gpuvm_sm_map_ops_create() [all …]
|
| A D | drm_debugfs.c | 210 struct drm_gpuvm *gpuvm) in drm_debugfs_gpuva_info() argument 212 struct drm_gpuva *va, *kva = &gpuvm->kernel_alloc_node; in drm_debugfs_gpuva_info() 214 if (!gpuvm->name) in drm_debugfs_gpuva_info() 218 gpuvm->name, gpuvm->mm_start, gpuvm->mm_start + gpuvm->mm_range); in drm_debugfs_gpuva_info() 224 drm_gpuvm_for_each_va(va, gpuvm) { in drm_debugfs_gpuva_info()
|
| /drivers/gpu/drm/xe/ |
| A D | xe_vm.h | 36 drm_gpuvm_get(&vm->gpuvm); in xe_vm_get() 42 drm_gpuvm_put(&vm->gpuvm); in xe_vm_put() 86 static inline struct xe_vm *gpuvm_to_vm(struct drm_gpuvm *gpuvm) in gpuvm_to_vm() argument 88 return container_of(gpuvm, struct xe_vm, gpuvm); in gpuvm_to_vm() 138 return container_of(vma->gpuva.vm, struct xe_vm, gpuvm); in xe_vma_vm() 284 return drm_gpuvm_resv(&vm->gpuvm); in xe_vm_resv()
|
| A D | xe_exec.c | 99 struct xe_vm *vm = container_of(vm_exec->vm, struct xe_vm, gpuvm); in xe_exec_fn() 240 vm_exec.vm = &vm->gpuvm; in xe_exec_ioctl() 312 drm_gpuvm_resv_add_fence(&vm->gpuvm, exec, &job->drm.s_fence->finished, in xe_exec_ioctl()
|
| A D | xe_vm.c | 47 return vm->gpuvm.r_obj; in xe_vm_obj() 225 .vm = &vm->gpuvm, in xe_vm_add_compute_exec_queue() 427 ret = drm_gpuvm_validate(&vm->gpuvm, exec); in xe_vm_validate_rebind() 434 } while (!list_empty(&vm->gpuvm.evict.list)); in xe_vm_validate_rebind() 1212 vma->gpuva.vm = &vm->gpuvm; in xe_vma_create() 2006 struct xe_vm *vm = container_of(gpuvm, struct xe_vm, gpuvm); in xe_vm_free() 2439 drm_gpuva_ops_free(&vm->gpuvm, ops); in vm_bind_ioctl_ops_create() 3717 drm_gpuva_ops_free(&vm->gpuvm, ops[i]); in xe_vm_bind_ioctl() 3801 drm_gpuva_ops_free(&vm->gpuvm, ops); in xe_vm_bind_kernel_bo() 3980 drm_gpuvm_for_each_va(gpuva, &vm->gpuvm) { in xe_vm_validate_protected() [all …]
|
| A D | xe_vm_types.h | 154 struct drm_gpuvm gpuvm; member
|
| /drivers/gpu/drm/msm/ |
| A D | msm_gem_vma.c | 159 msm_gem_vm_free(struct drm_gpuvm *gpuvm) in msm_gem_vm_free() argument 161 struct msm_gem_vm *vm = container_of(gpuvm, struct msm_gem_vm, base); in msm_gem_vm_free() 177 msm_gem_vm_unusable(struct drm_gpuvm *gpuvm) in msm_gem_vm_unusable() argument 179 struct msm_gem_vm *vm = to_msm_vm(gpuvm); in msm_gem_vm_unusable() 371 msm_gem_vma_new(struct drm_gpuvm *gpuvm, struct drm_gem_object *obj, in msm_gem_vma_new() argument 374 struct msm_gem_vm *vm = to_msm_vm(gpuvm); in msm_gem_vma_new() 840 msm_gem_vm_close(struct drm_gpuvm *gpuvm) in msm_gem_vm_close() argument 842 struct msm_gem_vm *vm = to_msm_vm(gpuvm); in msm_gem_vm_close() 863 drm_exec_lock_obj(&exec, drm_gpuvm_resv_obj(gpuvm)); in msm_gem_vm_close() 866 drm_gpuvm_for_each_va_safe (vma, tmp, gpuvm) { in msm_gem_vm_close() [all …]
|
| A D | msm_gem.h | 160 void msm_gem_vm_close(struct drm_gpuvm *gpuvm); 161 void msm_gem_vm_unusable(struct drm_gpuvm *gpuvm);
|
| /drivers/gpu/drm/imagination/ |
| A D | pvr_vm.c | 69 struct pvr_vm_context *to_pvr_vm_context(struct drm_gpuvm *gpuvm) in to_pvr_vm_context() argument 71 return container_of(gpuvm, struct pvr_vm_context, gpuvm_mgr); in to_pvr_vm_context() 501 static void pvr_gpuvm_free(struct drm_gpuvm *gpuvm) in pvr_gpuvm_free() argument 503 kfree(to_pvr_vm_context(gpuvm)); in pvr_gpuvm_free()
|
| /drivers/gpu/drm/amd/display/dc/dml/ |
| A D | display_mode_lib.c | 188 dml_print("DML PARAMS: gpuvm = %d\n", pipe_src->gpuvm); in dml_log_pipe_params()
|
| A D | display_mode_structs.h | 392 bool gpuvm; // gpuvm enabled member
|
| A D | display_mode_vba.c | 882 …mode_lib->vba.GPUVMEnable = mode_lib->vba.GPUVMEnable || !!pipes[k].pipe.src.gpuvm || !!pipes[k].p… in fetch_pipe_params()
|
| /drivers/gpu/drm/amd/display/dc/dml/dcn32/ |
| A D | display_rq_dlg_calc_32.c | 401 if (src->dynamic_metadata_enable && src->gpuvm) in dml32_rq_dlg_get_dlg_reg()
|
| /drivers/gpu/drm/nouveau/ |
| A D | nouveau_uvmm.c | 1794 nouveau_uvmm_free(struct drm_gpuvm *gpuvm) in nouveau_uvmm_free() argument 1796 struct nouveau_uvmm *uvmm = uvmm_from_gpuvm(gpuvm); in nouveau_uvmm_free()
|
| /drivers/gpu/drm/amd/display/dc/resource/dcn31/ |
| A D | dcn31_resource.c | 1629 pipes[i].pipe.src.gpuvm = 1; in dcn31x_populate_dml_pipes_from_context() 1674 pipes[pipe_cnt].pipe.src.gpuvm = true; in dcn31_populate_dml_pipes_from_context()
|
| /drivers/gpu/drm/panthor/ |
| A D | panthor_mmu.c | 1825 static void panthor_vm_free(struct drm_gpuvm *gpuvm) in panthor_vm_free() argument 1827 struct panthor_vm *vm = container_of(gpuvm, struct panthor_vm, base); in panthor_vm_free()
|
| /drivers/gpu/drm/amd/display/dc/dml/dcn30/ |
| A D | display_rq_dlg_calc_30.c | 1424 if (src->dynamic_metadata_enable && src->gpuvm) { in dml_rq_dlg_get_dlg_params()
|
| /drivers/gpu/drm/amd/display/dc/dml/dcn31/ |
| A D | display_rq_dlg_calc_31.c | 1272 if (src->dynamic_metadata_enable && src->gpuvm) in dml_rq_dlg_get_dlg_params()
|
| /drivers/gpu/drm/amd/display/dc/dml/dcn314/ |
| A D | display_rq_dlg_calc_314.c | 1360 if (src->dynamic_metadata_enable && src->gpuvm) in dml_rq_dlg_get_dlg_params()
|
| /drivers/gpu/drm/amd/display/dc/resource/dcn32/ |
| A D | dcn32_resource.c | 1921 pipes[pipe_cnt].pipe.src.gpuvm = true; in dcn32_populate_dml_pipes_from_context()
|
| /drivers/gpu/drm/amd/display/dc/dml/dcn20/ |
| A D | dcn20_fpu.c | 2170 pipes[i].pipe.src.gpuvm = 1; in dcn21_populate_dml_pipes_from_context()
|