/linux-6.3-rc2/drivers/gpu/drm/radeon/ |
A D | radeon_vm.c | 301 return bo_va; in radeon_vm_bo_find() 330 bo_va->vm = vm; in radeon_vm_bo_add() 331 bo_va->bo = bo; in radeon_vm_bo_add() 343 return bo_va; in radeon_vm_bo_add() 497 if (bo_va->it.start || bo_va->it.last) { in radeon_vm_bo_set_addr() 966 nptes = bo_va->it.last - bo_va->it.start + 1; in radeon_vm_bo_update() 1126 if (bo_va->it.start || bo_va->it.last) in radeon_vm_bo_rmv() 1131 if (bo_va->it.start || bo_va->it.last) { in radeon_vm_bo_rmv() 1132 bo_va->bo = radeon_bo_ref(bo_va->bo); in radeon_vm_bo_rmv() 1159 (bo_va->it.start || bo_va->it.last)) in radeon_vm_bo_invalidate() [all …]
|
A D | radeon_gem.c | 204 struct radeon_bo_va *bo_va; in radeon_gem_object_open() local 218 if (!bo_va) { in radeon_gem_object_open() 221 ++bo_va->ref_count; in radeon_gem_object_open() 235 struct radeon_bo_va *bo_va; in radeon_gem_object_close() local 250 if (bo_va) { in radeon_gem_object_close() 637 tv.bo = &bo_va->bo->tbo; in radeon_gem_va_update_vm() 662 if (bo_va->it.start) in radeon_gem_va_update_vm() 663 r = radeon_vm_bo_update(rdev, bo_va, bo_va->bo->tbo.resource); in radeon_gem_va_update_vm() 686 struct radeon_bo_va *bo_va; in radeon_gem_va_ioctl() local 750 if (!bo_va) { in radeon_gem_va_ioctl() [all …]
|
A D | radeon_trace.h | 66 TP_PROTO(struct radeon_bo_va *bo_va), 67 TP_ARGS(bo_va), 75 __entry->soffset = bo_va->it.start; 76 __entry->eoffset = bo_va->it.last + 1; 77 __entry->flags = bo_va->flags;
|
A D | radeon_cs.c | 505 struct radeon_bo_va *bo_va; in radeon_bo_vm_update_pte() local 530 bo_va = radeon_vm_bo_find(vm, bo); in radeon_bo_vm_update_pte() 531 if (bo_va == NULL) { in radeon_bo_vm_update_pte() 536 r = radeon_vm_bo_update(rdev, bo_va, bo->tbo.resource); in radeon_bo_vm_update_pte() 540 radeon_sync_fence(&p->ib.sync, bo_va->last_pt_update); in radeon_bo_vm_update_pte()
|
A D | radeon.h | 2895 struct radeon_bo_va *bo_va, 2905 struct radeon_bo_va *bo_va, 2909 struct radeon_bo_va *bo_va);
|
/linux-6.3-rc2/drivers/gpu/drm/lima/ |
A D | lima_vm.c | 82 if (bo_va->vm == vm) { in lima_vm_bo_find() 83 ret = bo_va; in lima_vm_bo_find() 100 if (bo_va) { in lima_vm_bo_add() 101 bo_va->ref_count++; in lima_vm_bo_add() 112 bo_va = kzalloc(sizeof(*bo_va), GFP_KERNEL); in lima_vm_bo_add() 113 if (!bo_va) { in lima_vm_bo_add() 118 bo_va->vm = vm; in lima_vm_bo_add() 119 bo_va->ref_count = 1; in lima_vm_bo_add() 149 kfree(bo_va); in lima_vm_bo_add() 182 kfree(bo_va); in lima_vm_bo_del() [all …]
|
/linux-6.3-rc2/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_vm.c | 1036 list_splice_init(&bo_va->valids, &bo_va->invalids); in amdgpu_vm_bo_update() 1039 list_splice_init(&bo_va->valids, &bo_va->invalids); in amdgpu_vm_bo_update() 1083 list_splice_init(&bo_va->invalids, &bo_va->valids); in amdgpu_vm_bo_update() 1368 return bo_va; in amdgpu_vm_bo_add() 1377 return bo_va; in amdgpu_vm_bo_add() 1397 mapping->bo_va = bo_va; in amdgpu_vm_bo_insert_map() 1642 before->bo_va = tmp->bo_va; in amdgpu_vm_bo_clear_mappings() 1653 after->bo_va = tmp->bo_va; in amdgpu_vm_bo_clear_mappings() 1673 tmp->bo_va = NULL; in amdgpu_vm_bo_clear_mappings() 1734 if (mapping->bo_va && mapping->bo_va->base.bo) { in amdgpu_vm_bo_trace_cs() [all …]
|
A D | amdgpu_csa.c | 65 struct amdgpu_bo *bo, struct amdgpu_bo_va **bo_va, in amdgpu_map_static_csa() argument 88 *bo_va = amdgpu_vm_bo_add(adev, vm, bo); in amdgpu_map_static_csa() 89 if (!*bo_va) { in amdgpu_map_static_csa() 95 r = amdgpu_vm_bo_map(adev, *bo_va, csa_addr, 0, size, in amdgpu_map_static_csa() 101 amdgpu_vm_bo_del(adev, *bo_va); in amdgpu_map_static_csa()
|
A D | amdgpu_gem.c | 166 struct amdgpu_bo_va *bo_va; in amdgpu_gem_object_open() local 183 if (!bo_va) { in amdgpu_gem_object_open() 186 ++bo_va->ref_count; in amdgpu_gem_object_open() 205 struct amdgpu_bo_va *bo_va; in amdgpu_gem_object_close() local 223 bo_va = amdgpu_vm_bo_find(vm, bo); in amdgpu_gem_object_close() 224 if (!bo_va || --bo_va->ref_count) in amdgpu_gem_object_close() 227 amdgpu_vm_bo_del(adev, bo_va); in amdgpu_gem_object_close() 675 struct amdgpu_bo_va *bo_va; in amdgpu_gem_va_ioctl() local 756 if (!bo_va) { in amdgpu_gem_va_ioctl() 761 bo_va = fpriv->prt_va; in amdgpu_gem_va_ioctl() [all …]
|
A D | amdgpu_amdkfd_gpuvm.c | 79 if (entry->bo_va->base.vm == avm) in kfd_mem_is_attached() 851 if (unlikely(!attachment[i]->bo_va)) { in kfd_mem_attach() 871 if (attachment[i]->bo_va) { in kfd_mem_attach() 873 amdgpu_vm_bo_del(adev, attachment[i]->bo_va); in kfd_mem_attach() 1097 if ((vm && vm != entry->bo_va->base.vm) || in reserve_bo_and_cond_vms() 1119 if ((vm && vm != entry->bo_va->base.vm) || in reserve_bo_and_cond_vms() 1176 struct amdgpu_bo_va *bo_va = entry->bo_va; in unmap_bo_from_gpuvm() local 1178 struct amdgpu_vm *vm = bo_va->base.vm; in unmap_bo_from_gpuvm() 1180 amdgpu_vm_bo_unmap(adev, bo_va, entry->va); in unmap_bo_from_gpuvm() 1193 struct amdgpu_bo_va *bo_va = entry->bo_va; in update_gpuvm_pte() local [all …]
|
A D | amdgpu_trace.h | 246 TP_PROTO(struct amdgpu_bo_va *bo_va, 248 TP_ARGS(bo_va, mapping), 258 __entry->bo = bo_va ? bo_va->base.bo : NULL; 270 TP_PROTO(struct amdgpu_bo_va *bo_va, 272 TP_ARGS(bo_va, mapping), 282 __entry->bo = bo_va ? bo_va->base.bo : NULL;
|
A D | amdgpu_vm.h | 413 struct amdgpu_bo_va *bo_va, 425 struct amdgpu_bo_va *bo_va, 429 struct amdgpu_bo_va *bo_va, 433 struct amdgpu_bo_va *bo_va, 442 struct amdgpu_bo_va *bo_va);
|
A D | amdgpu_cs.c | 942 e->bo_va = amdgpu_vm_bo_find(vm, bo); in amdgpu_cs_parser_bos() 1095 struct amdgpu_bo_va *bo_va; in amdgpu_cs_vm_handling() local 1113 bo_va = fpriv->csa_va; in amdgpu_cs_vm_handling() 1114 BUG_ON(!bo_va); in amdgpu_cs_vm_handling() 1115 r = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_cs_vm_handling() 1119 r = amdgpu_sync_fence(&p->sync, bo_va->last_pt_update); in amdgpu_cs_vm_handling() 1130 bo_va = e->bo_va; in amdgpu_cs_vm_handling() 1131 if (bo_va == NULL) in amdgpu_cs_vm_handling() 1134 r = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_cs_vm_handling() 1782 if (!mapping || !mapping->bo_va || !mapping->bo_va->base.bo) in amdgpu_cs_find_mapping() [all …]
|
A D | amdgpu_mes.c | 1128 struct amdgpu_bo_va *bo_va; in amdgpu_mes_ctx_map_meta_data() local 1152 bo_va = amdgpu_vm_bo_add(adev, vm, ctx_data->meta_data_obj); in amdgpu_mes_ctx_map_meta_data() 1153 if (!bo_va) { in amdgpu_mes_ctx_map_meta_data() 1159 r = amdgpu_vm_bo_map(adev, bo_va, ctx_data->meta_data_gpu_addr, 0, in amdgpu_mes_ctx_map_meta_data() 1169 r = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_mes_ctx_map_meta_data() 1174 amdgpu_sync_fence(&sync, bo_va->last_pt_update); in amdgpu_mes_ctx_map_meta_data() 1187 ctx_data->meta_data_va = bo_va; in amdgpu_mes_ctx_map_meta_data() 1191 amdgpu_vm_bo_del(adev, bo_va); in amdgpu_mes_ctx_map_meta_data() 1200 struct amdgpu_bo_va *bo_va = ctx_data->meta_data_va; in amdgpu_mes_ctx_unmap_meta_data() local 1202 struct amdgpu_vm *vm = bo_va->base.vm; in amdgpu_mes_ctx_unmap_meta_data() [all …]
|
A D | amdgpu_csa.h | 35 struct amdgpu_bo *bo, struct amdgpu_bo_va **bo_va,
|
A D | amdgpu_bo_list.h | 40 struct amdgpu_bo_va *bo_va; member
|
A D | amdgpu_object.h | 63 struct amdgpu_bo_va *bo_va; member
|
A D | amdgpu_amdkfd.h | 59 struct amdgpu_bo_va *bo_va; member
|
A D | gmc_v9_0.c | 1167 if (mapping->bo_va->is_xgmi) in gmc_v9_0_get_coherence_flags() 1204 struct amdgpu_bo *bo = mapping->bo_va->base.bo; in gmc_v9_0_get_vm_pte() 1218 gmc_v9_0_get_coherence_flags(adev, mapping->bo_va->base.bo, in gmc_v9_0_get_vm_pte()
|
A D | vcn_v1_0.c | 1928 if (!mapping || !mapping->bo_va || !mapping->bo_va->base.bo) in vcn_v1_0_validate_bo() 1931 bo = mapping->bo_va->base.bo; in vcn_v1_0_validate_bo()
|
A D | gmc_v11_0.c | 518 struct amdgpu_bo *bo = mapping->bo_va->base.bo; in gmc_v11_0_get_vm_pte()
|
A D | gmc_v10_0.c | 623 struct amdgpu_bo *bo = mapping->bo_va->base.bo; in gmc_v10_0_get_vm_pte()
|
/linux-6.3-rc2/drivers/gpu/drm/amd/amdkfd/ |
A D | kfd_chardev.c | 359 wptr_bo = wptr_mapping->bo_va->base.bo; in kfd_ioctl_create_queue()
|
A D | kfd_svm.c | 2629 bo = mapping->bo_va->base.bo; in svm_range_check_vm_userptr()
|