Lines Matching refs:vm_bo
178 static void amdgpu_vm_bo_evicted(struct amdgpu_vm_bo_base *vm_bo) in amdgpu_vm_bo_evicted() argument
180 struct amdgpu_vm *vm = vm_bo->vm; in amdgpu_vm_bo_evicted()
181 struct amdgpu_bo *bo = vm_bo->bo; in amdgpu_vm_bo_evicted()
183 vm_bo->moved = true; in amdgpu_vm_bo_evicted()
184 spin_lock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_evicted()
186 list_move(&vm_bo->vm_status, &vm->evicted); in amdgpu_vm_bo_evicted()
188 list_move_tail(&vm_bo->vm_status, &vm->evicted); in amdgpu_vm_bo_evicted()
189 spin_unlock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_evicted()
199 static void amdgpu_vm_bo_moved(struct amdgpu_vm_bo_base *vm_bo) in amdgpu_vm_bo_moved() argument
201 spin_lock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_moved()
202 list_move(&vm_bo->vm_status, &vm_bo->vm->moved); in amdgpu_vm_bo_moved()
203 spin_unlock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_moved()
214 static void amdgpu_vm_bo_idle(struct amdgpu_vm_bo_base *vm_bo) in amdgpu_vm_bo_idle() argument
216 spin_lock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_idle()
217 list_move(&vm_bo->vm_status, &vm_bo->vm->idle); in amdgpu_vm_bo_idle()
218 spin_unlock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_idle()
219 vm_bo->moved = false; in amdgpu_vm_bo_idle()
230 static void amdgpu_vm_bo_invalidated(struct amdgpu_vm_bo_base *vm_bo) in amdgpu_vm_bo_invalidated() argument
232 spin_lock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_invalidated()
233 list_move(&vm_bo->vm_status, &vm_bo->vm->invalidated); in amdgpu_vm_bo_invalidated()
234 spin_unlock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_invalidated()
245 static void amdgpu_vm_bo_evicted_user(struct amdgpu_vm_bo_base *vm_bo) in amdgpu_vm_bo_evicted_user() argument
247 vm_bo->moved = true; in amdgpu_vm_bo_evicted_user()
248 spin_lock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_evicted_user()
249 list_move(&vm_bo->vm_status, &vm_bo->vm->evicted_user); in amdgpu_vm_bo_evicted_user()
250 spin_unlock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_evicted_user()
261 static void amdgpu_vm_bo_relocated(struct amdgpu_vm_bo_base *vm_bo) in amdgpu_vm_bo_relocated() argument
263 if (vm_bo->bo->parent) { in amdgpu_vm_bo_relocated()
264 spin_lock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_relocated()
265 list_move(&vm_bo->vm_status, &vm_bo->vm->relocated); in amdgpu_vm_bo_relocated()
266 spin_unlock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_relocated()
268 amdgpu_vm_bo_idle(vm_bo); in amdgpu_vm_bo_relocated()
280 static void amdgpu_vm_bo_done(struct amdgpu_vm_bo_base *vm_bo) in amdgpu_vm_bo_done() argument
282 spin_lock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_done()
283 list_move(&vm_bo->vm_status, &vm_bo->vm->done); in amdgpu_vm_bo_done()
284 spin_unlock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_done()
296 struct amdgpu_vm_bo_base *vm_bo, *tmp; in amdgpu_vm_bo_reset_state_machine() local
300 list_for_each_entry(vm_bo, &vm->invalidated, vm_status) in amdgpu_vm_bo_reset_state_machine()
301 vm_bo->moved = true; in amdgpu_vm_bo_reset_state_machine()
302 list_for_each_entry_safe(vm_bo, tmp, &vm->idle, vm_status) { in amdgpu_vm_bo_reset_state_machine()
303 struct amdgpu_bo *bo = vm_bo->bo; in amdgpu_vm_bo_reset_state_machine()
305 vm_bo->moved = true; in amdgpu_vm_bo_reset_state_machine()
307 list_move(&vm_bo->vm_status, &vm_bo->vm->moved); in amdgpu_vm_bo_reset_state_machine()
309 list_move(&vm_bo->vm_status, &vm_bo->vm->relocated); in amdgpu_vm_bo_reset_state_machine()
356 for (base = bo->vm_bo; base; base = base->next) in amdgpu_vm_bo_update_shared()
439 base->next = bo->vm_bo; in amdgpu_vm_bo_base_init()
440 bo->vm_bo = base; in amdgpu_vm_bo_base_init()
908 for (base = bo->vm_bo; base; base = base->next) { in amdgpu_vm_bo_find()
2170 for (base = &bo_va->base.bo->vm_bo; *base; in amdgpu_vm_bo_del()
2216 struct amdgpu_vm_bo_base *bo_base = bo->vm_bo; in amdgpu_vm_evictable()
2253 for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next) { in amdgpu_vm_bo_invalidate()
2288 for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next) { in amdgpu_vm_bo_move()