Lines Matching refs:preempt
88 list_for_each_entry(q, &vm->preempt.exec_queues, lr.link) { in preempt_fences_waiting()
113 if (*count >= vm->preempt.num_exec_queues) in alloc_preempt_fences()
116 for (; *count < vm->preempt.num_exec_queues; ++(*count)) { in alloc_preempt_fences()
134 list_for_each_entry(q, &vm->preempt.exec_queues, lr.link) { in wait_for_existing_preempt_fences()
155 list_for_each_entry(q, &vm->preempt.exec_queues, lr.link) { in xe_vm_is_idle()
168 list_for_each_entry(q, &vm->preempt.exec_queues, lr.link) { in arm_preempt_fences()
189 if (!vm->preempt.num_exec_queues) in add_preempt_fences()
192 err = dma_resv_reserve_fences(bo->ttm.base.resv, vm->preempt.num_exec_queues); in add_preempt_fences()
196 list_for_each_entry(q, &vm->preempt.exec_queues, lr.link) in add_preempt_fences()
214 list_for_each_entry(q, &vm->preempt.exec_queues, lr.link) { in resume_and_reinstall_preempt_fences()
248 list_add(&q->lr.link, &vm->preempt.exec_queues); in xe_vm_add_compute_exec_queue()
249 ++vm->preempt.num_exec_queues; in xe_vm_add_compute_exec_queue()
292 --vm->preempt.num_exec_queues; in xe_vm_remove_compute_exec_queue()
343 list_for_each_entry(q, &vm->preempt.exec_queues, lr.link) in xe_vm_kill()
455 vm->preempt.rebind_deactivated = true; in xe_preempt_work_begin()
479 return xe_vm_validate_rebind(vm, exec, vm->preempt.num_exec_queues); in xe_preempt_work_begin()
484 struct xe_vm *vm = container_of(w, struct xe_vm, preempt.rebind_work); in preempt_rebind_work_func()
1695 INIT_LIST_HEAD(&vm->preempt.exec_queues); in xe_vm_create()
1696 vm->preempt.min_run_period_ms = 10; /* FIXME: Wire up to uAPI */ in xe_vm_create()
1710 INIT_WORK(&vm->preempt.rebind_work, preempt_rebind_work_func); in xe_vm_create()
1874 xe_assert(xe, !vm->preempt.num_exec_queues); in xe_vm_close_and_put()
1878 flush_work(&vm->preempt.rebind_work); in xe_vm_close_and_put()
1984 flush_work(&vm->preempt.rebind_work); in vm_destroy_work_func()
2160 else if (XE_IOCTL_DBG(xe, vm->preempt.num_exec_queues)) in xe_vm_destroy_ioctl()