Home
last modified time | relevance | path

Searched refs:vm (Results 1 – 25 of 348) sorted by relevance

12345678910>>...14

/drivers/virtio/
A Dvirtio_mem.c854 return start < vm->addr + vm->region_size && vm->addr < start + size; in virtio_mem_overlaps_range()
864 return start >= vm->addr && start + size <= vm->addr + vm->region_size; in virtio_mem_contains_range()
1400 sg_init_one(&sg_req, &vm->req, sizeof(vm->req)); in virtio_mem_send_request()
1416 return virtio16_to_cpu(vm->vdev, vm->resp.type); in virtio_mem_send_request()
2450 if (vm->requested_size > vm->plugged_size) { in virtio_mem_run_wq()
2451 diff = vm->requested_size - vm->plugged_size; in virtio_mem_run_wq()
2575 vm->sbm.next_mb_id = vm->sbm.first_mb_id; in virtio_mem_init_hotplug()
2597 vm->bbm.next_bb_id = vm->bbm.first_bb_id; in virtio_mem_init_hotplug()
2868 dev_info(&vm->vdev->dev, "nid: %d", vm->nid); in virtio_mem_init()
2891 vm->parent_resource = __request_mem_region(vm->addr, vm->region_size, in virtio_mem_create_resource()
[all …]
/drivers/gpu/drm/amd/amdgpu/
A Damdgpu_vm.c180 struct amdgpu_vm *vm = vm_bo->vm; in amdgpu_vm_bo_evicted() local
324 struct amdgpu_vm *vm = base->vm; in amdgpu_vm_update_shared() local
373 struct amdgpu_vm *vm = base->vm; in amdgpu_vm_update_stats_locked() local
412 struct amdgpu_vm *vm = base->vm; in amdgpu_vm_update_stats() local
432 base->vm = vm; in amdgpu_vm_bo_base_init()
909 if (base->vm != vm) in amdgpu_vm_bo_find()
977 params.vm = vm; in amdgpu_vm_update_pdes()
1042 struct amdgpu_vm *vm = params->vm; in amdgpu_vm_tlb_flush() local
1044 tlb_cb->vm = vm; in amdgpu_vm_tlb_flush()
1127 params.vm = vm; in amdgpu_vm_update_range()
[all …]
/drivers/gpu/drm/xe/
A Dxe_vm.h36 drm_gpuvm_get(&vm->gpuvm); in xe_vm_get()
37 return vm; in xe_vm_get()
42 drm_gpuvm_put(&vm->gpuvm); in xe_vm_put()
52 return !vm->size; in xe_vm_is_closed()
63 return xe_vm_is_closed(vm) || xe_vm_is_banned(vm); in xe_vm_is_closed_or_banned()
209 return xe_vm_in_lr_mode(vm) && !xe_vm_in_fault_mode(vm); in xe_vm_in_preempt_fence_mode()
240 xe_assert(vm->xe, xe_vm_in_preempt_fence_mode(vm)); in xe_vm_queue_rebind_worker()
241 queue_work(vm->xe->ordered_wq, &vm->preempt.rebind_work); in xe_vm_queue_rebind_worker()
255 if (xe_vm_in_preempt_fence_mode(vm) && vm->preempt.rebind_deactivated) { in xe_vm_reactivate_rebind()
293 #define xe_vm_assert_held(vm) dma_resv_assert_held(xe_vm_resv(vm)) argument
[all …]
A Dxe_vm.c225 .vm = &vm->gpuvm, in xe_vm_add_compute_exec_queue()
387 struct xe_vm *vm = gpuvm_to_vm(vm_bo->vm); in xe_gpuvm_validate() local
1183 xe_assert(vm->xe, end < vm->size); in xe_vma_create()
1212 vma->gpuva.vm = &vm->gpuvm; in xe_vma_create()
1421 xe_assert(vm->xe, xe_vma_vm(vma) == vm); in xe_vm_insert_vma()
1434 xe_assert(vm->xe, xe_vma_vm(vma) == vm); in xe_vm_remove_vma()
1657 vm = kzalloc(sizeof(*vm), GFP_KERNEL); in xe_vm_create()
2017 vm = xa_load(&xef->vm.xa, id); in xe_vm_lookup()
2157 vm = xa_load(&xef->vm.xa, args->vm_id); in xe_vm_destroy_ioctl()
3436 to_wait_exec_queue(vm, q), vm); in vm_bind_ioctl_signal_fences()
[all …]
A Dxe_svm.c245 xe_vm_unlock(vm); in __xe_svm_garbage_collector()
299 up_write(&vm->lock); in xe_svm_garbage_collector_work_func()
575 err = drm_gpusvm_init(&vm->svm.gpusvm, "Xe SVM", &vm->xe->drm, in xe_svm_init()
583 drm_gpusvm_driver_set_lock(&vm->svm.gpusvm, &vm->lock); in xe_svm_init()
596 xe_assert(vm->xe, xe_vm_is_closed(vm)); in xe_svm_close()
608 xe_assert(vm->xe, xe_vm_is_closed(vm)); in xe_svm_fini()
778 xe_assert(vm->xe, IS_DGFX(vm->xe)); in xe_svm_range_needs_migrate_to_vram()
880 vm->usm.asid, &vm->svm.gpusvm, ERR_PTR(err)); in xe_svm_handle_pagefault()
886 vm->usm.asid, &vm->svm.gpusvm, ERR_PTR(err)); in xe_svm_handle_pagefault()
900 xe_vm_unlock(vm); in xe_svm_handle_pagefault()
[all …]
A Dxe_exec.c99 struct xe_vm *vm = container_of(vm_exec->vm, struct xe_vm, gpuvm); in xe_exec_fn() local
119 struct xe_vm *vm; in xe_exec_ioctl() local
159 vm = q->vm; in xe_exec_ioctl()
197 if (!xe_vm_in_lr_mode(vm) && xe_vm_userptr_check_repin(vm)) { in xe_exec_ioctl()
227 xe_vm_unlock(vm); in xe_exec_ioctl()
236 xe_vm_unlock(vm); in xe_exec_ioctl()
240 vm_exec.vm = &vm->gpuvm; in xe_exec_ioctl()
281 xe_vm_resv(vm), in xe_exec_ioctl()
311 if (!xe_vm_in_lr_mode(vm)) in xe_exec_ioctl()
323 if (!xe_vm_in_lr_mode(vm)) in xe_exec_ioctl()
[all …]
A Dxe_trace_bo.h20 #define __dev_name_vm(vm) dev_name((vm)->xe->drm.dev) argument
38 __entry->vm = bo->vm;
201 __entry->vm = vm;
213 TP_ARGS(vm)
218 TP_ARGS(vm)
223 TP_ARGS(vm)
228 TP_ARGS(vm)
233 TP_ARGS(vm)
238 TP_ARGS(vm)
243 TP_ARGS(vm)
[all …]
/drivers/gpu/drm/panthor/
A Dpanthor_mmu.c102 } vm; member
445 drm_WARN_ON(&vm->ptdev->base, vm->op_ctx); in alloc_pt()
681 ptdev->mmu->as.slots[vm->as.id].vm = NULL; in panthor_vm_release_as_locked()
753 ptdev->mmu->as.slots[as].vm = vm; in panthor_vm_active()
777 ret = panthor_mmu_as_enable(vm->ptdev, vm->as.id, transtab, transcfg, vm->memattr); in panthor_vm_active()
1059 struct drm_gpuvm *vm = vm_bo->vm; in panthor_vm_bo_put() local
1524 panthor_vm_unmap_range(vm, vm->base.mm_start, vm->base.mm_range)); in panthor_vm_destroy()
1877 drm_gpuvm_put(vm ? &vm->base : NULL); in panthor_vm_put()
2307 vm = kzalloc(sizeof(*vm), GFP_KERNEL); in panthor_vm_create()
2486 if (vm->destroyed || vm->unusable) in panthor_vm_bind_job_create()
[all …]
/drivers/gpu/drm/i915/gt/
A Dintel_ggtt.c73 ggtt->vm.cleanup(&ggtt->vm); in ggtt_init_hw()
120 drm_WARN_ON(&vm->i915->drm, !vm->is_ggtt && !vm->is_dpt); in i915_ggtt_suspend_vm()
171 vm->clear_range(vm, 0, vm->total); in i915_ggtt_suspend_vm()
927 ggtt->vm.scratch_range(&ggtt->vm, start, size); in init_ggtt()
947 ggtt->vm.clear_range(&ggtt->vm, hole_start, in init_ggtt()
952 ggtt->vm.clear_range(&ggtt->vm, ggtt->vm.total - PAGE_SIZE, PAGE_SIZE); in init_ggtt()
975 ppgtt_bind_vma(&i915_vm_to_ggtt(vm)->alias->vm, in aliasing_gtt_bind_vma()
1025 ppgtt->vm.allocate_va_range(&ppgtt->vm, &stash, 0, ggtt->vm.total); in init_aliasing_ppgtt()
1107 ggtt->vm.cleanup(&ggtt->vm); in ggtt_cleanup_hw()
1620 drm_WARN_ON(&vm->i915->drm, !vm->is_ggtt && !vm->is_dpt); in i915_ggtt_resume_vm()
[all …]
A Dintel_gtt.c68 if (vm->fpriv) in alloc_pt_lmem()
92 if (vm->fpriv) in alloc_pt_dma()
200 if (vm->scratch[0]->base.resv == &vm->_resv) { in i915_vm_lock_objects()
231 kfree(vm); in i915_vm_resv_release()
239 __i915_vm_close(vm); in __i915_vm_release()
244 vm->cleanup(vm); in __i915_vm_release()
258 queue_work(vm->i915->wq, &vm->release_work); in i915_vm_release()
263 kref_init(&vm->ref); in i915_address_space_init()
284 i915_gem_shrinker_taints_mutex(vm->i915, &vm->mutex); in i915_address_space_init()
301 drm_mm_init(&vm->mm, 0, vm->total); in i915_address_space_init()
[all …]
A Dgen8_ppgtt.c234 gen8_pd_top_count(vm), vm->top); in gen8_ppgtt_cleanup()
318 __gen8_ppgtt_clear(vm, i915_vm_to_ppgtt(vm)->pd, in gen8_ppgtt_clear()
402 __gen8_ppgtt_alloc(vm, stash, i915_vm_to_ppgtt(vm)->pd, in gen8_ppgtt_alloc()
449 __gen8_ppgtt_foreach(vm, i915_vm_to_ppgtt(vm)->pd, in gen8_ppgtt_foreach()
836 if (vm->has_read_only && vm->gt->vm && !i915_is_ggtt(vm->gt->vm)) { in gen8_init_scratch()
837 struct i915_address_space *clone = vm->gt->vm; in gen8_init_scratch()
857 vm->pte_encode(px_dma(vm->scratch[0]), in gen8_init_scratch()
865 obj = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K); in gen8_init_scratch()
894 struct i915_address_space *vm = &ppgtt->vm; in gen8_preallocate_top_level_pdp() local
937 pd->pt.base = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K); in gen8_alloc_top_pd()
[all …]
A Dintel_gtt.h418 #define i915_is_ggtt(vm) ((vm)->is_ggtt) argument
419 #define i915_is_dpt(vm) ((vm)->is_dpt) argument
420 #define i915_is_ggtt_or_dpt(vm) (i915_is_ggtt(vm) || i915_is_dpt(vm)) argument
461 return i915_is_ggtt(vm) && vm->mm.color_adjust; in i915_vm_has_cache_coloring()
469 return container_of(vm, struct i915_ggtt, vm); in i915_vm_to_ggtt()
477 return container_of(vm, struct i915_ppgtt, vm); in i915_vm_to_ppgtt()
484 return vm; in i915_vm_get()
490 return kref_get_unless_zero(&vm->ref) ? vm : NULL; in i915_vm_tryget()
507 return &vm->_resv; in i915_vm_resv_get()
645 #define free_pt(vm, px) free_px(vm, px, 0) argument
[all …]
A Dgen6_ppgtt.c221 struct i915_address_space * const vm = &ppgtt->base.vm; in gen6_ppgtt_init_scratch() local
224 ret = setup_scratch_page(vm); in gen6_ppgtt_init_scratch()
228 vm->scratch[0]->encode = in gen6_ppgtt_init_scratch()
229 vm->pte_encode(px_dma(vm->scratch[0]), in gen6_ppgtt_init_scratch()
234 vm->scratch[1] = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K); in gen6_ppgtt_init_scratch()
240 ret = map_pt_dma(vm, vm->scratch[1]); in gen6_ppgtt_init_scratch()
244 fill32_px(vm->scratch[1], vm->scratch[0]->encode); in gen6_ppgtt_init_scratch()
252 vm->scratch[0] = NULL; in gen6_ppgtt_init_scratch()
272 free_scratch(vm); in gen6_ppgtt_cleanup()
442 ppgtt->base.vm.top = 1; in gen6_ppgtt_create()
[all …]
A Dintel_ppgtt.c24 pt->base = vm->alloc_pt_dma(vm, sz); in alloc_pt()
61 pd->pt.base = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K); in alloc_pd()
191 vm->allocate_va_range(vm, stash, vma_res->start, in ppgtt_bind_vma()
203 vm->insert_entries(vm, vma_res, pat_index, pte_flags); in ppgtt_bind_vma()
213 vm->clear_range(vm, vma_res->start, vma_res->vma_size); in ppgtt_unbind_vma()
230 shift = vm->pd_shift; in i915_vm_alloc_pt_stash()
246 pt = alloc_pt(vm, pt_sz); in i915_vm_alloc_pt_stash()
262 pd = alloc_pd(vm); in i915_vm_alloc_pt_stash()
302 free_px(vm, pt, n); in i915_vm_free_pt_stash()
312 ppgtt->vm.gt = gt; in ppgtt_init()
[all …]
/drivers/gpu/drm/lima/
A Dlima_vm.c18 struct lima_vm *vm; member
82 if (bo_va->vm == vm) { in lima_vm_bo_find()
118 bo_va->vm = vm; in lima_vm_bo_add()
204 vm = kzalloc(sizeof(*vm), GFP_KERNEL); in lima_vm_create()
205 if (!vm) in lima_vm_create()
208 vm->dev = dev; in lima_vm_create()
226 return vm; in lima_vm_create()
231 kfree(vm); in lima_vm_create()
245 vm->bts[i].cpu, vm->bts[i].dma); in lima_vm_release()
249 dma_free_wc(vm->dev->dev, LIMA_PAGE_SIZE, vm->pd.cpu, vm->pd.dma); in lima_vm_release()
[all …]
/drivers/virt/acrn/
A Dvm.c40 vm->vmid = vm_param->vmid; in acrn_vm_create()
53 acrn_ioeventfd_init(vm); in acrn_vm_create()
54 acrn_irqfd_init(vm); in acrn_vm_create()
56 return vm; in acrn_vm_create()
77 list_del_init(&vm->list); in acrn_vm_destroy()
80 acrn_ioeventfd_deinit(vm); in acrn_vm_destroy()
81 acrn_irqfd_deinit(vm); in acrn_vm_destroy()
82 acrn_ioreq_deinit(vm); in acrn_vm_destroy()
84 if (vm->monitor_page) { in acrn_vm_destroy()
86 vm->monitor_page = NULL; in acrn_vm_destroy()
[all …]
A Dirqfd.c30 struct acrn_vm *vm; member
41 struct acrn_vm *vm = irqfd->vm; in acrn_irqfd_inject() local
63 struct acrn_vm *vm; in hsm_irqfd_shutdown_work() local
66 vm = irqfd->vm; in hsm_irqfd_shutdown_work()
67 mutex_lock(&vm->irqfds_lock); in hsm_irqfd_shutdown_work()
79 struct acrn_vm *vm; in hsm_irqfd_wakeup() local
82 vm = irqfd->vm; in hsm_irqfd_wakeup()
119 irqfd->vm = vm; in acrn_irqfd_assign()
207 INIT_LIST_HEAD(&vm->irqfds); in acrn_irqfd_init()
209 vm->irqfd_wq = alloc_workqueue("acrn_irqfd-%u", 0, 0, vm->vmid); in acrn_irqfd_init()
[all …]
A Dioreq.c98 if (vm->default_client) in acrn_ioreq_request_default_complete()
432 client->vm = vm; in acrn_ioreq_client_create()
468 struct acrn_vm *vm = client->vm; in acrn_ioreq_client_destroy() local
546 struct acrn_vm *vm; in ioreq_dispatcher() local
550 if (!vm->ioreq_buf) in ioreq_dispatcher()
602 if (vm->ioreq_buf) in acrn_ioreq_init()
618 vm->ioreq_page = page; in acrn_ioreq_init()
624 vm->ioreq_buf = NULL; in acrn_ioreq_init()
645 if (vm->default_client) in acrn_ioreq_deinit()
648 if (vm->ioreq_buf && vm->ioreq_page) { in acrn_ioreq_deinit()
[all …]
A Dioeventfd.c121 mutex_lock(&vm->ioeventfds_lock); in acrn_ioeventfd_assign()
135 mutex_unlock(&vm->ioeventfds_lock); in acrn_ioeventfd_assign()
140 mutex_unlock(&vm->ioeventfds_lock); in acrn_ioeventfd_assign()
157 mutex_lock(&vm->ioeventfds_lock); in acrn_ioeventfd_deassign()
164 acrn_ioeventfd_shutdown(vm, p); in acrn_ioeventfd_deassign()
248 mutex_init(&vm->ioeventfds_lock); in acrn_ioeventfd_init()
249 INIT_LIST_HEAD(&vm->ioeventfds); in acrn_ioeventfd_init()
251 vm->ioeventfd_client = acrn_ioreq_client_create(vm, in acrn_ioeventfd_init()
254 if (!vm->ioeventfd_client) { in acrn_ioeventfd_init()
269 mutex_lock(&vm->ioeventfds_lock); in acrn_ioeventfd_deinit()
[all …]
/drivers/gpu/drm/i915/selftests/
A Dmock_gtt.c73 ppgtt->vm.gt = to_gt(i915); in mock_ppgtt()
74 ppgtt->vm.i915 = i915; in mock_ppgtt()
76 ppgtt->vm.dma = i915->drm.dev; in mock_ppgtt()
86 ppgtt->vm.cleanup = mock_cleanup; in mock_ppgtt()
111 ggtt->vm.gt = gt; in mock_init_ggtt()
112 ggtt->vm.i915 = gt->i915; in mock_init_ggtt()
113 ggtt->vm.is_ggtt = true; in mock_init_ggtt()
117 ggtt->vm.total = 4096 * PAGE_SIZE; in mock_init_ggtt()
119 ggtt->vm.alloc_pt_dma = alloc_pt_dma; in mock_init_ggtt()
125 ggtt->vm.cleanup = mock_cleanup; in mock_init_ggtt()
[all …]
A Di915_gem_gtt.c206 ppgtt->vm.clear_range(&ppgtt->vm, 0, size); in igt_ppgtt_alloc()
225 ppgtt->vm.allocate_va_range(&ppgtt->vm, &stash, in igt_ppgtt_alloc()
342 vm->allocate_va_range(vm, &stash, in lowlevel_hole()
362 vm->insert_entries(vm, mock_vma_res, in lowlevel_hole()
376 vm->clear_range(vm, addr, BIT_ULL(size)); in lowlevel_hole()
999 memset(&vm->fault_attr, 0, sizeof(vm->fault_attr)); in shrink_hole()
1078 memset(&vm->fault_attr, 0, sizeof(vm->fault_attr)); in shrink_boom()
1383 ggtt->vm.insert_page(&ggtt->vm, in igt_ggtt_page()
1472 err = func(vm, 0, min(vm->total, limit), end_time); in exercise_mock()
1509 struct i915_address_space *vm = vma->vm; in reserve_gtt_with_resource() local
[all …]
/drivers/gpu/drm/i915/display/
A Dintel_dpt.c26 #define i915_is_dpt(vm) ((vm)->is_dpt) argument
32 drm_WARN_ON(&vm->i915->drm, !i915_is_dpt(vm)); in i915_vm_to_dpt()
33 return container_of(vm, struct i915_dpt, vm); in i915_vm_to_dpt()
101 vm->insert_entries(vm, vma_res, pat_index, pte_flags); in dpt_bind_vma()
116 vm->clear_range(vm, vma_res->start, vma_res->vma_size); in dpt_unbind_vma()
289 vm = &dpt->vm; in intel_dpt_create()
291 vm->gt = to_gt(i915); in intel_dpt_create()
292 vm->i915 = i915; in intel_dpt_create()
295 vm->is_dpt = true; in intel_dpt_create()
307 vm->pte_encode = vm->gt->ggtt->vm.pte_encode; in intel_dpt_create()
[all …]
/drivers/gpu/drm/radeon/
A Dradeon_vm.c298 if (bo_va->vm == vm) in radeon_vm_bo_find()
328 bo_va->vm = vm; in radeon_vm_bo_add()
337 mutex_lock(&vm->mutex); in radeon_vm_bo_add()
451 struct radeon_vm *vm = bo_va->vm; in radeon_vm_bo_set_addr() local
476 mutex_lock(&vm->mutex); in radeon_vm_bo_set_addr()
506 tmp->vm = vm; in radeon_vm_bo_set_addr()
913 struct radeon_vm *vm = bo_va->vm; in radeon_vm_bo_update() local
922 bo_va->bo, vm); in radeon_vm_bo_update()
1119 struct radeon_vm *vm = bo_va->vm; in radeon_vm_bo_rmv() local
1178 vm->ib_bo_va = NULL; in radeon_vm_init()
[all …]
/drivers/gpu/drm/msm/
A Dmsm_gem_vma.c165 vm->mmu->funcs->destroy(vm->mmu); in msm_gem_vm_free()
188 if (!vm->log || !vm->log[0].op) in msm_gem_vm_unusable()
242 vm->log_idx = (vm->log_idx + 1) & ((1 << vm->log_shift) - 1); in vm_log()
270 struct msm_gem_vm *vm = to_msm_vm(vma->vm); in msm_gem_vma_unmap() local
301 struct msm_gem_vm *vm = to_msm_vm(vma->vm); in msm_gem_vma_map() local
350 struct msm_gem_vm *vm = to_msm_vm(vma->vm); in msm_gem_vma_close() local
525 struct drm_gpuvm *vm = job->vm; in msm_gem_vm_sm_step_remap() local
647 struct msm_gem_vm *vm = to_msm_vm(job->vm); in msm_vma_job_run() local
703 struct msm_gem_vm *vm = to_msm_vm(job->vm); in msm_vma_job_free() local
767 vm = kzalloc(sizeof(*vm), GFP_KERNEL); in msm_gem_vm_create()
[all …]
/drivers/gpu/drm/imx/dcss/
A Ddcss-ss.c129 lrc_x = vm->hfront_porch + vm->hback_porch + vm->hsync_len + in dcss_ss_sync_set()
130 vm->hactive - 1; in dcss_ss_sync_set()
131 lrc_y = vm->vfront_porch + vm->vback_porch + vm->vsync_len + in dcss_ss_sync_set()
132 vm->vactive - 1; in dcss_ss_sync_set()
136 hsync_start = vm->hfront_porch + vm->hback_porch + vm->hsync_len + in dcss_ss_sync_set()
145 vsync_end = vm->vfront_porch + vm->vsync_len - 1; in dcss_ss_sync_set()
151 de_ulc_x = vm->hsync_len + vm->hback_porch - 1; in dcss_ss_sync_set()
152 de_ulc_y = vm->vsync_len + vm->vfront_porch + vm->vback_porch; in dcss_ss_sync_set()
157 de_lrc_x = vm->hsync_len + vm->hback_porch + vm->hactive - 1; in dcss_ss_sync_set()
158 de_lrc_y = vm->vsync_len + vm->vfront_porch + vm->vback_porch + in dcss_ss_sync_set()
[all …]

Completed in 873 milliseconds

12345678910>>...14