| /drivers/vhost/ |
| A D | vhost.c | 253 worker->ops->wakeup(worker); in vhost_worker_queue() 263 worker = rcu_dereference(vq->worker); in vhost_vq_work_queue() 284 if (!worker->attachment_cnt || worker->killed) in __vhost_worker_flush() 324 worker = rcu_dereference(vq->worker); in vhost_vq_has_work() 325 if (worker && !llist_empty(&worker->work_list)) in vhost_vq_has_work() 722 worker->ops->stop(worker); in vhost_worker_destroy() 875 rcu_assign_pointer(vq->worker, worker); in __vhost_vq_attach_worker() 940 if (!worker || worker->id != info->worker_id) in vhost_vq_attach_worker() 969 if (!worker || worker->id != info->worker_id) in vhost_free_worker() 973 if (worker->attachment_cnt || worker->killed) { in vhost_free_worker() [all …]
|
| A D | vhost.h | 33 int (*create)(struct vhost_worker *worker, struct vhost_dev *dev, 35 void (*stop)(struct vhost_worker *worker); 36 void (*wakeup)(struct vhost_worker *worker); 95 struct vhost_worker __rcu *worker; member
|
| /drivers/net/wireguard/ |
| A D | queueing.c | 13 struct multicore_worker __percpu *worker = alloc_percpu(struct multicore_worker); in wg_packet_percpu_multicore_worker_alloc() local 15 if (!worker) in wg_packet_percpu_multicore_worker_alloc() 19 per_cpu_ptr(worker, cpu)->ptr = ptr; in wg_packet_percpu_multicore_worker_alloc() 20 INIT_WORK(&per_cpu_ptr(worker, cpu)->work, function); in wg_packet_percpu_multicore_worker_alloc() 22 return worker; in wg_packet_percpu_multicore_worker_alloc() 35 queue->worker = wg_packet_percpu_multicore_worker_alloc(function, queue); in wg_packet_queue_init() 36 if (!queue->worker) { in wg_packet_queue_init() 45 free_percpu(queue->worker); in wg_packet_queue_free()
|
| /drivers/gpu/drm/ |
| A D | drm_vblank_work.c | 62 kthread_queue_work(vblank->worker, &work->base); in drm_handle_vblank_works() 152 ret = kthread_queue_work(vblank->worker, &work->base); in drm_vblank_work_schedule() 255 kthread_flush_worker(vblank->worker); in drm_vblank_work_flush_all() 278 struct kthread_worker *worker; in drm_vblank_worker_init() local 282 worker = kthread_run_worker(0, "card%d-crtc%d", in drm_vblank_worker_init() 285 if (IS_ERR(worker)) in drm_vblank_worker_init() 286 return PTR_ERR(worker); in drm_vblank_worker_init() 288 vblank->worker = worker; in drm_vblank_worker_init() 290 sched_set_fifo(worker->task); in drm_vblank_worker_init()
|
| A D | drm_flip_work.c | 98 queue_work(wq, &work->worker); in drm_flip_work_commit() 104 struct drm_flip_work *work = container_of(w, struct drm_flip_work, worker); in flip_worker() 144 INIT_WORK(&work->worker, flip_worker); in drm_flip_work_init()
|
| A D | drm_internal.h | 109 kthread_flush_worker(vblank->worker); in drm_vblank_flush_worker() 114 if (vblank->worker) in drm_vblank_destroy_worker() 115 kthread_destroy_worker(vblank->worker); in drm_vblank_destroy_worker()
|
| /drivers/gpu/drm/msm/ |
| A D | msm_atomic.c | 118 timer->worker = kthread_run_worker(0, "atomic-worker-%d", crtc_idx); in msm_atomic_init_pending_timer() 119 if (IS_ERR(timer->worker)) { in msm_atomic_init_pending_timer() 120 int ret = PTR_ERR(timer->worker); in msm_atomic_init_pending_timer() 121 timer->worker = NULL; in msm_atomic_init_pending_timer() 124 sched_set_fifo(timer->worker->task); in msm_atomic_init_pending_timer() 126 msm_hrtimer_work_init(&timer->work, timer->worker, in msm_atomic_init_pending_timer() 135 if (timer->worker) in msm_atomic_destroy_pending_timer() 136 kthread_destroy_worker(timer->worker); in msm_atomic_destroy_pending_timer()
|
| A D | msm_kms.c | 249 if (kms->event_thread[i].worker) in msm_drm_kms_uninit() 250 kthread_destroy_worker(kms->event_thread[i].worker); in msm_drm_kms_uninit() 305 ev_thread->worker = kthread_run_worker(0, "crtc_event:%d", crtc->base.id); in msm_drm_kms_init() 306 if (IS_ERR(ev_thread->worker)) { in msm_drm_kms_init() 307 ret = PTR_ERR(ev_thread->worker); in msm_drm_kms_init() 309 ev_thread->worker = NULL; in msm_drm_kms_init() 313 sched_set_fifo(ev_thread->worker->task); in msm_drm_kms_init()
|
| A D | msm_io_utils.c | 120 kthread_queue_work(work->worker, &work->work); in msm_hrtimer_worktimer() 133 struct kthread_worker *worker, in msm_hrtimer_work_init() argument 139 work->worker = worker; in msm_hrtimer_work_init()
|
| A D | msm_gpu.c | 649 kthread_queue_work(gpu->worker, &gpu->recover_work); in hangcheck_handler() 859 kthread_queue_work(gpu->worker, &gpu->retire_work); in msm_gpu_retire() 977 gpu->worker = kthread_run_worker(0, "gpu-worker"); in msm_gpu_init() 978 if (IS_ERR(gpu->worker)) { in msm_gpu_init() 979 ret = PTR_ERR(gpu->worker); in msm_gpu_init() 980 gpu->worker = NULL; in msm_gpu_init() 984 sched_set_fifo_low(gpu->worker->task); in msm_gpu_init() 1129 if (gpu->worker) { in msm_gpu_cleanup() 1130 kthread_destroy_worker(gpu->worker); in msm_gpu_cleanup()
|
| /drivers/block/ |
| A D | loop.c | 804 worker = cur_worker; in loop_queue_work() 812 if (worker) in loop_queue_work() 820 if (!worker) { in loop_queue_work() 829 css_get(worker->blkcg_css); in loop_queue_work() 833 worker->lo = lo; in loop_queue_work() 837 if (worker) { in loop_queue_work() 845 work = &worker->work; in loop_queue_work() 874 css_put(worker->blkcg_css); in loop_free_idle_workers() 875 kfree(worker); in loop_free_idle_workers() 1956 if (worker && !work_pending(&worker->work)) { in loop_process_work() [all …]
|
| /drivers/md/ |
| A D | dm-delay.c | 39 struct task_struct *worker; member 71 return !!dc->worker; in delay_is_fast() 174 if (dc->worker) in delay_dtr() 175 kthread_stop(dc->worker); in delay_dtr() 284 dc->worker = kthread_run(&flush_worker_fn, dc, "dm-delay-flush-worker"); in delay_ctr() 285 if (IS_ERR(dc->worker)) { in delay_ctr() 286 ret = PTR_ERR(dc->worker); in delay_ctr() 287 dc->worker = NULL; in delay_ctr() 335 wake_up_process(dc->worker); in delay_bio()
|
| /drivers/gpu/drm/i915/selftests/ |
| A D | i915_request.c | 496 if (IS_ERR(worker)) { in mock_breadcrumbs_smoketest() 497 ret = PTR_ERR(worker); in mock_breadcrumbs_smoketest() 502 threads[n].worker = worker; in mock_breadcrumbs_smoketest() 1650 if (IS_ERR(worker)) { in live_parallel_engines() 1651 err = PTR_ERR(worker); in live_parallel_engines() 1655 threads[idx].worker = worker; in live_parallel_engines() 1810 if (IS_ERR(worker)) { in live_breadcrumbs_smoketest() 1811 ret = PTR_ERR(worker); in live_breadcrumbs_smoketest() 1815 threads[i].worker = worker; in live_breadcrumbs_smoketest() 3224 if (IS_ERR(worker)) { in perf_parallel_engines() [all …]
|
| /drivers/i2c/ |
| A D | i2c-slave-testunit.c | 50 struct delayed_work worker; member 127 queue_delayed_work(system_long_wq, &tu->worker, in i2c_slave_testunit_slave_cb() 170 struct testunit_data *tu = container_of(work, struct testunit_data, worker.work); in i2c_slave_testunit_work() 247 INIT_DELAYED_WORK(&tu->worker, i2c_slave_testunit_work); in i2c_slave_testunit_probe() 268 cancel_delayed_work_sync(&tu->worker); in i2c_slave_testunit_remove()
|
| /drivers/crypto/caam/ |
| A D | caamrng.c | 43 struct work_struct worker; member 143 worker); in caam_rng_worker() 161 schedule_work(&ctx->worker); in caam_read() 170 flush_work(&ctx->worker); in caam_cleanup() 239 INIT_WORK(&ctx->worker, caam_rng_worker); in caam_init()
|
| /drivers/gpu/drm/i915/gt/ |
| A D | selftest_slpc.c | 20 struct kthread_worker *worker; member 507 threads[i].worker = kthread_run_worker(0, "igt/slpc_parallel:%d", gt->info.id); in live_slpc_tile_interaction() 509 if (IS_ERR(threads[i].worker)) { in live_slpc_tile_interaction() 510 ret = PTR_ERR(threads[i].worker); in live_slpc_tile_interaction() 516 kthread_queue_work(threads[i].worker, &threads[i].work); in live_slpc_tile_interaction() 522 if (IS_ERR_OR_NULL(threads[i].worker)) in live_slpc_tile_interaction() 531 kthread_destroy_worker(threads[i].worker); in live_slpc_tile_interaction()
|
| /drivers/hid/ |
| A D | hid-bigbenff.c | 185 struct work_struct worker; member 194 schedule_work(&bigben->worker); in bigben_schedule_work() 201 struct bigben_device, worker); in bigben_worker() 365 cancel_work_sync(&bigben->worker); in bigben_remove() 414 INIT_WORK(&bigben->worker, bigben_worker); in bigben_probe()
|
| /drivers/gpu/drm/xe/ |
| A D | xe_sriov_vf.c | 144 INIT_WORK(&xe->sriov.vf.migration.worker, migration_worker_func); in xe_sriov_vf_init_early() 266 sriov.vf.migration.worker); in migration_worker_func() 305 started = queue_work(xe->sriov.wq, &xe->sriov.vf.migration.worker); in xe_sriov_vf_start_migration_recovery()
|
| A D | xe_gt_types.h | 185 struct work_struct worker; member 275 struct work_struct worker; member 302 struct work_struct worker; member
|
| A D | xe_sync.c | 26 struct work_struct worker; member 78 struct xe_user_fence *ufence = container_of(w, struct xe_user_fence, worker); in user_fence_worker() 101 INIT_WORK(&ufence->worker, user_fence_worker); in kick_ufence() 102 queue_work(ufence->xe->ordered_wq, &ufence->worker); in kick_ufence()
|
| /drivers/platform/olpc/ |
| A D | olpc-ec.c | 36 struct work_struct worker; member 79 struct olpc_ec_priv *ec = container_of(w, struct olpc_ec_priv, worker); in olpc_ec_worker() 105 schedule_work(&ec->worker); in olpc_ec_worker() 123 schedule_work(&ec->worker); in queue_ec_descriptor() 416 INIT_WORK(&ec->worker, olpc_ec_worker); in olpc_ec_probe()
|
| /drivers/media/platform/chips-media/wave5/ |
| A D | wave5-vpu.c | 118 kthread_queue_work(dev->worker, &dev->work); in wave5_vpu_timer_callback() 274 dev->worker = kthread_run_worker(0, "vpu_irq_thread"); in wave5_vpu_probe() 275 if (IS_ERR(dev->worker)) { in wave5_vpu_probe() 277 ret = PTR_ERR(dev->worker); in wave5_vpu_probe() 355 kthread_destroy_worker(dev->worker); in wave5_vpu_remove()
|
| /drivers/macintosh/ams/ |
| A D | ams-core.c | 65 schedule_work(&ams_info.worker); in ams_handle_irq() 185 INIT_WORK(&ams_info.worker, ams_worker); in ams_init() 216 flush_work(&ams_info.worker); in ams_sensor_detach()
|
| /drivers/net/wwan/t7xx/ |
| A D | t7xx_hif_cldma.c | 578 queue_work(md_ctrl->txq[i].worker, in t7xx_cldma_irq_work_cb() 1274 if (md_ctrl->txq[i].worker) { in t7xx_cldma_destroy_wqs() 1275 destroy_workqueue(md_ctrl->txq[i].worker); in t7xx_cldma_destroy_wqs() 1276 md_ctrl->txq[i].worker = NULL; in t7xx_cldma_destroy_wqs() 1281 if (md_ctrl->rxq[i].worker) { in t7xx_cldma_destroy_wqs() 1282 destroy_workqueue(md_ctrl->rxq[i].worker); in t7xx_cldma_destroy_wqs() 1283 md_ctrl->rxq[i].worker = NULL; in t7xx_cldma_destroy_wqs() 1317 md_ctrl->txq[i].worker = in t7xx_cldma_init() 1321 if (!md_ctrl->txq[i].worker) in t7xx_cldma_init() 1331 md_ctrl->rxq[i].worker = in t7xx_cldma_init() [all …]
|
| A D | t7xx_hif_dpmaif_tx.c | 177 queue_work(dpmaif_ctrl->txq[txq->index].worker, in t7xx_dpmaif_tx_done() 533 queue_work(dpmaif_ctrl->txq[i].worker, &dpmaif_ctrl->txq[i].dpmaif_tx_work); in t7xx_dpmaif_irq_tx_done() 621 txq->worker = alloc_ordered_workqueue("md_dpmaif_tx%d_worker", in t7xx_dpmaif_txq_init() 624 if (!txq->worker) in t7xx_dpmaif_txq_init() 635 if (txq->worker) in t7xx_dpmaif_txq_free() 636 destroy_workqueue(txq->worker); in t7xx_dpmaif_txq_free()
|