Home
last modified time | relevance | path

Searched refs:gpuvm (Results 1 – 21 of 21) sorted by relevance

/drivers/gpu/drm/
A Ddrm_gpuvm.c1019 gpuvm->ops = ops; in drm_gpuvm_init()
1036 __drm_gpuva_insert(gpuvm, &gpuvm->kernel_alloc_node); in drm_gpuvm_init()
1049 drm_WARN(gpuvm->drm, !RB_EMPTY_ROOT(&gpuvm->rb.tree.rb_root), in drm_gpuvm_fini()
1052 drm_WARN(gpuvm->drm, !list_empty(&gpuvm->extobj.list), in drm_gpuvm_fini()
1054 drm_WARN(gpuvm->drm, !list_empty(&gpuvm->evict.list), in drm_gpuvm_fini()
1067 if (drm_WARN_ON(gpuvm->drm, !gpuvm->ops->vm_free)) in drm_gpuvm_free()
1070 gpuvm->ops->vm_free(gpuvm); in drm_gpuvm_free()
1084 if (gpuvm) in drm_gpuvm_put()
1729 va->vm = gpuvm; in __drm_gpuva_insert()
2661 args.vm = gpuvm; in drm_gpuvm_sm_map_ops_create()
[all …]
A Ddrm_debugfs.c210 struct drm_gpuvm *gpuvm) in drm_debugfs_gpuva_info() argument
212 struct drm_gpuva *va, *kva = &gpuvm->kernel_alloc_node; in drm_debugfs_gpuva_info()
214 if (!gpuvm->name) in drm_debugfs_gpuva_info()
218 gpuvm->name, gpuvm->mm_start, gpuvm->mm_start + gpuvm->mm_range); in drm_debugfs_gpuva_info()
224 drm_gpuvm_for_each_va(va, gpuvm) { in drm_debugfs_gpuva_info()
/drivers/gpu/drm/xe/
A Dxe_vm.h36 drm_gpuvm_get(&vm->gpuvm); in xe_vm_get()
42 drm_gpuvm_put(&vm->gpuvm); in xe_vm_put()
86 static inline struct xe_vm *gpuvm_to_vm(struct drm_gpuvm *gpuvm) in gpuvm_to_vm() argument
88 return container_of(gpuvm, struct xe_vm, gpuvm); in gpuvm_to_vm()
138 return container_of(vma->gpuva.vm, struct xe_vm, gpuvm); in xe_vma_vm()
284 return drm_gpuvm_resv(&vm->gpuvm); in xe_vm_resv()
A Dxe_exec.c99 struct xe_vm *vm = container_of(vm_exec->vm, struct xe_vm, gpuvm); in xe_exec_fn()
240 vm_exec.vm = &vm->gpuvm; in xe_exec_ioctl()
312 drm_gpuvm_resv_add_fence(&vm->gpuvm, exec, &job->drm.s_fence->finished, in xe_exec_ioctl()
A Dxe_vm.c47 return vm->gpuvm.r_obj; in xe_vm_obj()
225 .vm = &vm->gpuvm, in xe_vm_add_compute_exec_queue()
427 ret = drm_gpuvm_validate(&vm->gpuvm, exec); in xe_vm_validate_rebind()
434 } while (!list_empty(&vm->gpuvm.evict.list)); in xe_vm_validate_rebind()
1212 vma->gpuva.vm = &vm->gpuvm; in xe_vma_create()
2006 struct xe_vm *vm = container_of(gpuvm, struct xe_vm, gpuvm); in xe_vm_free()
2439 drm_gpuva_ops_free(&vm->gpuvm, ops); in vm_bind_ioctl_ops_create()
3717 drm_gpuva_ops_free(&vm->gpuvm, ops[i]); in xe_vm_bind_ioctl()
3801 drm_gpuva_ops_free(&vm->gpuvm, ops); in xe_vm_bind_kernel_bo()
3980 drm_gpuvm_for_each_va(gpuva, &vm->gpuvm) { in xe_vm_validate_protected()
[all …]
A Dxe_vm_types.h154 struct drm_gpuvm gpuvm; member
/drivers/gpu/drm/msm/
A Dmsm_gem_vma.c159 msm_gem_vm_free(struct drm_gpuvm *gpuvm) in msm_gem_vm_free() argument
161 struct msm_gem_vm *vm = container_of(gpuvm, struct msm_gem_vm, base); in msm_gem_vm_free()
177 msm_gem_vm_unusable(struct drm_gpuvm *gpuvm) in msm_gem_vm_unusable() argument
179 struct msm_gem_vm *vm = to_msm_vm(gpuvm); in msm_gem_vm_unusable()
371 msm_gem_vma_new(struct drm_gpuvm *gpuvm, struct drm_gem_object *obj, in msm_gem_vma_new() argument
374 struct msm_gem_vm *vm = to_msm_vm(gpuvm); in msm_gem_vma_new()
840 msm_gem_vm_close(struct drm_gpuvm *gpuvm) in msm_gem_vm_close() argument
842 struct msm_gem_vm *vm = to_msm_vm(gpuvm); in msm_gem_vm_close()
863 drm_exec_lock_obj(&exec, drm_gpuvm_resv_obj(gpuvm)); in msm_gem_vm_close()
866 drm_gpuvm_for_each_va_safe (vma, tmp, gpuvm) { in msm_gem_vm_close()
[all …]
A Dmsm_gem.h160 void msm_gem_vm_close(struct drm_gpuvm *gpuvm);
161 void msm_gem_vm_unusable(struct drm_gpuvm *gpuvm);
/drivers/gpu/drm/imagination/
A Dpvr_vm.c69 struct pvr_vm_context *to_pvr_vm_context(struct drm_gpuvm *gpuvm) in to_pvr_vm_context() argument
71 return container_of(gpuvm, struct pvr_vm_context, gpuvm_mgr); in to_pvr_vm_context()
501 static void pvr_gpuvm_free(struct drm_gpuvm *gpuvm) in pvr_gpuvm_free() argument
503 kfree(to_pvr_vm_context(gpuvm)); in pvr_gpuvm_free()
/drivers/gpu/drm/amd/display/dc/dml/
A Ddisplay_mode_lib.c188 dml_print("DML PARAMS: gpuvm = %d\n", pipe_src->gpuvm); in dml_log_pipe_params()
A Ddisplay_mode_structs.h392 bool gpuvm; // gpuvm enabled member
A Ddisplay_mode_vba.c882 …mode_lib->vba.GPUVMEnable = mode_lib->vba.GPUVMEnable || !!pipes[k].pipe.src.gpuvm || !!pipes[k].p… in fetch_pipe_params()
/drivers/gpu/drm/amd/display/dc/dml/dcn32/
A Ddisplay_rq_dlg_calc_32.c401 if (src->dynamic_metadata_enable && src->gpuvm) in dml32_rq_dlg_get_dlg_reg()
/drivers/gpu/drm/nouveau/
A Dnouveau_uvmm.c1794 nouveau_uvmm_free(struct drm_gpuvm *gpuvm) in nouveau_uvmm_free() argument
1796 struct nouveau_uvmm *uvmm = uvmm_from_gpuvm(gpuvm); in nouveau_uvmm_free()
/drivers/gpu/drm/amd/display/dc/resource/dcn31/
A Ddcn31_resource.c1629 pipes[i].pipe.src.gpuvm = 1; in dcn31x_populate_dml_pipes_from_context()
1674 pipes[pipe_cnt].pipe.src.gpuvm = true; in dcn31_populate_dml_pipes_from_context()
/drivers/gpu/drm/panthor/
A Dpanthor_mmu.c1825 static void panthor_vm_free(struct drm_gpuvm *gpuvm) in panthor_vm_free() argument
1827 struct panthor_vm *vm = container_of(gpuvm, struct panthor_vm, base); in panthor_vm_free()
/drivers/gpu/drm/amd/display/dc/dml/dcn30/
A Ddisplay_rq_dlg_calc_30.c1424 if (src->dynamic_metadata_enable && src->gpuvm) { in dml_rq_dlg_get_dlg_params()
/drivers/gpu/drm/amd/display/dc/dml/dcn31/
A Ddisplay_rq_dlg_calc_31.c1272 if (src->dynamic_metadata_enable && src->gpuvm) in dml_rq_dlg_get_dlg_params()
/drivers/gpu/drm/amd/display/dc/dml/dcn314/
A Ddisplay_rq_dlg_calc_314.c1360 if (src->dynamic_metadata_enable && src->gpuvm) in dml_rq_dlg_get_dlg_params()
/drivers/gpu/drm/amd/display/dc/resource/dcn32/
A Ddcn32_resource.c1921 pipes[pipe_cnt].pipe.src.gpuvm = true; in dcn32_populate_dml_pipes_from_context()
/drivers/gpu/drm/amd/display/dc/dml/dcn20/
A Ddcn20_fpu.c2170 pipes[i].pipe.src.gpuvm = 1; in dcn21_populate_dml_pipes_from_context()

Completed in 719 milliseconds