| /drivers/gpu/drm/msm/adreno/ |
| A D | a5xx_preempt.c | 149 a5xx_gpu->preempt[ring->id]->wptr = get_wptr(ring); in a5xx_preempt_trigger() 231 a5xx_gpu->preempt[i]->data = 0; in a5xx_preempt_hw_init() 232 a5xx_gpu->preempt[i]->info = 0; in a5xx_preempt_hw_init() 233 a5xx_gpu->preempt[i]->wptr = 0; in a5xx_preempt_hw_init() 234 a5xx_gpu->preempt[i]->rptr = 0; in a5xx_preempt_hw_init() 235 a5xx_gpu->preempt[i]->rbase = gpu->rb[i]->iova; in a5xx_preempt_hw_init() 236 a5xx_gpu->preempt[i]->rptr_addr = shadowptr(a5xx_gpu, gpu->rb[i]); in a5xx_preempt_hw_init() 278 a5xx_gpu->preempt[ring->id] = ptr; in preempt_init_ring()
|
| A D | a6xx_preempt.c | 199 struct a6xx_preempt_record *record_ptr = a6xx_gpu->preempt[i]; in a6xx_preempt_hw_init() 283 struct a6xx_preempt_record *record_ptr = a6xx_gpu->preempt[ring->id]; in a6xx_preempt_trigger() 358 a6xx_gpu->preempt[ring->id] = ptr; in preempt_init_ring()
|
| A D | a5xx_gpu.h | 35 struct a5xx_preempt_record *preempt[MSM_GPU_MAX_RINGS]; member
|
| A D | a6xx_gpu.h | 64 void *preempt[MSM_GPU_MAX_RINGS]; member
|
| /drivers/gpu/drm/xe/ |
| A D | Kconfig.profile | 35 int "Default max preempt timeout (us)" 38 Configures the default max preempt timeout after which context 42 int "Default min preempt timeout (us)" 45 Configures the default min preempt timeout after which context
|
| A D | xe_vm.h | 241 queue_work(vm->xe->ordered_wq, &vm->preempt.rebind_work); in xe_vm_queue_rebind_worker() 255 if (xe_vm_in_preempt_fence_mode(vm) && vm->preempt.rebind_deactivated) { in xe_vm_reactivate_rebind() 256 vm->preempt.rebind_deactivated = false; in xe_vm_reactivate_rebind()
|
| A D | xe_vm.c | 113 if (*count >= vm->preempt.num_exec_queues) in alloc_preempt_fences() 116 for (; *count < vm->preempt.num_exec_queues; ++(*count)) { in alloc_preempt_fences() 189 if (!vm->preempt.num_exec_queues) in add_preempt_fences() 248 list_add(&q->lr.link, &vm->preempt.exec_queues); in xe_vm_add_compute_exec_queue() 249 ++vm->preempt.num_exec_queues; in xe_vm_add_compute_exec_queue() 292 --vm->preempt.num_exec_queues; in xe_vm_remove_compute_exec_queue() 455 vm->preempt.rebind_deactivated = true; in xe_preempt_work_begin() 1695 INIT_LIST_HEAD(&vm->preempt.exec_queues); in xe_vm_create() 1874 xe_assert(xe, !vm->preempt.num_exec_queues); in xe_vm_close_and_put() 1878 flush_work(&vm->preempt.rebind_work); in xe_vm_close_and_put() [all …]
|
| A D | xe_vm_types.h | 296 } preempt; member
|
| /drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
| A D | chan.c | 141 chan->func->preempt(chan); in nvkm_chan_preempt_locked() 153 if (!chan->func->preempt) in nvkm_chan_preempt() 182 nvkm_chan_remove(struct nvkm_chan *chan, bool preempt) in nvkm_chan_remove() argument 187 if (preempt && chan->func->preempt) in nvkm_chan_remove() 227 nvkm_chan_error(struct nvkm_chan *chan, bool preempt) in nvkm_chan_error() argument 235 if (preempt) in nvkm_chan_error() 236 chan->func->preempt(chan); in nvkm_chan_error()
|
| A D | gk110.c | 41 cgrp->func->preempt(cgrp); in gk110_chan_preempt() 57 .preempt = gk110_chan_preempt, 68 .preempt = gk110_cgrp_preempt,
|
| A D | chan.h | 52 void (*preempt)(struct nvkm_chan *); member 63 void nvkm_chan_error(struct nvkm_chan *, bool preempt); 65 void nvkm_chan_remove(struct nvkm_chan *, bool preempt);
|
| A D | tu102.c | 58 .preempt = gk110_chan_preempt, 94 .preempt = gv100_runl_preempt,
|
| A D | runl.c | 87 if (runl->func->preempt) { in nvkm_runl_rc() 137 if (runl->func->preempt) in nvkm_runl_rc_runl() 138 runl->func->preempt(runl); in nvkm_runl_rc_runl()
|
| A D | ga100.c | 105 .preempt = gk110_chan_preempt, 119 .preempt = ga100_cgrp_preempt, 402 runl->func->preempt(runl); in ga100_runl_init() 435 .preempt = ga100_runl_preempt,
|
| A D | gv100.c | 86 .preempt = gk110_chan_preempt, 214 .preempt = gv100_runl_preempt,
|
| A D | cgrp.h | 31 void (*preempt)(struct nvkm_cgrp *); member
|
| A D | gm107.c | 42 .preempt = gk110_chan_preempt,
|
| A D | runl.h | 59 void (*preempt)(struct nvkm_runl *); member
|
| A D | gk104.c | 130 .preempt = gf100_chan_preempt,
|
| /drivers/parport/ |
| A D | share.c | 686 if (!par_dev_cb->preempt || !par_dev_cb->wakeup) { in parport_register_dev_model() 727 par_dev->preempt = par_dev_cb->preempt; in parport_register_dev_model() 971 if (oldcad->preempt) { in parport_claim() 972 if (oldcad->preempt(oldcad->private)) in parport_claim()
|
| /drivers/gpu/drm/i915/display/ |
| A D | intel_flipq.c | 158 static void intel_flipq_preempt(struct intel_crtc *crtc, bool preempt) in intel_flipq_preempt() argument 163 PIPEDMC_FQ_CTRL_PREEMPT, preempt ? PIPEDMC_FQ_CTRL_PREEMPT : 0); in intel_flipq_preempt() 165 if (preempt && in intel_flipq_preempt()
|
| /drivers/gpu/drm/i915/gt/ |
| A D | selftest_lrc.c | 617 bool preempt) in __live_lrc_gpr() argument 647 if (preempt) { in __live_lrc_gpr() 798 static int __lrc_timestamp(const struct lrc_timestamp *arg, bool preempt) in __lrc_timestamp() argument 814 if (preempt) { in __lrc_timestamp() 831 arg->engine->name, preempt ? "preempt" : "simple", in __lrc_timestamp() 839 arg->engine->name, preempt ? "preempt" : "simple", in __lrc_timestamp()
|
| A D | intel_execlists_submission.c | 1252 set_timer_ms(&engine->execlists.preempt, in set_preempt_timeout() 1652 cancel_timer(&execlists->preempt); in cancel_port_requests() 2406 const struct timer_list *t = &engine->execlists.preempt; in preempt_timeout() 2446 cancel_timer(&engine->execlists.preempt); in execlists_submission_tasklet() 2450 set_timer_ms(&engine->execlists.preempt, in execlists_submission_tasklet() 2538 execlists_kick(timer, preempt); in execlists_preempt() 3265 cancel_timer(&engine->execlists.preempt); in execlists_park() 3374 timer_delete_sync(&engine->execlists.preempt); in execlists_shutdown() 3536 timer_setup(&engine->execlists.preempt, execlists_preempt, 0); in intel_execlists_submission_setup()
|
| A D | intel_engine_types.h | 170 struct timer_list preempt; member
|
| A D | sysfs_engines.c | 333 set_timer_ms(&engine->execlists.preempt, timeout); in preempt_timeout_store()
|