| /drivers/dma/bestcomm/ |
| A D | fec.c | 88 if (!tsk) in bcom_fec_rx_init() 102 return tsk; in bcom_fec_rx_init() 127 var->bd_last = tsk->bd_pa + ((tsk->num_bd-1) * tsk->bd_size); in bcom_fec_rx_reset() 136 tsk->index = 0; in bcom_fec_rx_reset() 139 memset_io(tsk->bd, 0x00, tsk->num_bd * tsk->bd_size); in bcom_fec_rx_reset() 143 bcom_set_task_auto_start(tsk->tasknum, tsk->tasknum); in bcom_fec_rx_reset() 190 if (!tsk) in bcom_fec_tx_init() 203 return tsk; in bcom_fec_tx_init() 229 var->bd_last = tsk->bd_pa + ((tsk->num_bd-1) * tsk->bd_size); in bcom_fec_tx_reset() 240 memset_io(tsk->bd, 0x00, tsk->num_bd * tsk->bd_size); in bcom_fec_tx_reset() [all …]
|
| A D | gen_bd.c | 93 if (!tsk) in bcom_gen_bd_rx_init() 109 return tsk; in bcom_gen_bd_rx_init() 134 var->bd_last = tsk->bd_pa + ((tsk->num_bd-1) * tsk->bd_size); in bcom_gen_bd_rx_reset() 142 tsk->index = 0; in bcom_gen_bd_rx_reset() 145 memset_io(tsk->bd, 0x00, tsk->num_bd * tsk->bd_size); in bcom_gen_bd_rx_reset() 149 bcom_set_task_auto_start(tsk->tasknum, tsk->tasknum); in bcom_gen_bd_rx_reset() 178 if (!tsk) in bcom_gen_bd_tx_init() 193 return tsk; in bcom_gen_bd_tx_init() 218 var->bd_last = tsk->bd_pa + ((tsk->num_bd-1) * tsk->bd_size); in bcom_gen_bd_tx_reset() 229 memset_io(tsk->bd, 0x00, tsk->num_bd * tsk->bd_size); in bcom_gen_bd_tx_reset() [all …]
|
| A D | ata.c | 56 struct bcom_task *tsk; in bcom_ata_init() local 64 if (!tsk) in bcom_ata_init() 69 bcom_ata_reset_bd(tsk); in bcom_ata_init() 75 bcom_task_free(tsk); in bcom_ata_init() 82 var->bd_last = tsk->bd_pa + ((tsk->num_bd-1) * tsk->bd_size); in bcom_ata_init() 88 bcom_set_task_auto_start(tsk->tasknum, tsk->tasknum); in bcom_ata_init() 95 return tsk; in bcom_ata_init() 132 memset_io(tsk->bd, 0x00, tsk->num_bd * tsk->bd_size); in bcom_ata_reset_bd() 134 tsk->index = 0; in bcom_ata_reset_bd() 135 tsk->outdex = 0; in bcom_ata_reset_bd() [all …]
|
| A D | bestcomm.c | 73 if (!tsk) in bcom_task_alloc() 78 tsk->priv = (void*)tsk + sizeof(struct bcom_task); in bcom_task_alloc() 81 tsk->irq = irq_of_parse_and_map(bcom_eng->ofnode, tsk->tasknum); in bcom_task_alloc() 82 if (!tsk->irq) in bcom_task_alloc() 92 tsk->bd = bcom_sram_alloc(bd_count * bd_size, 4, &tsk->bd_pa); in bcom_task_alloc() 93 if (!tsk->bd) in bcom_task_alloc() 101 return tsk; in bcom_task_alloc() 104 if (tsk) { in bcom_task_alloc() 105 if (tsk->irq) in bcom_task_alloc() 109 kfree(tsk); in bcom_task_alloc() [all …]
|
| /drivers/dma-buf/ |
| A D | st-dma-fence-chain.c | 583 if (IS_ERR(tsk)) { in wait_forward() 584 err = PTR_ERR(tsk); in wait_forward() 587 get_task_struct(tsk); in wait_forward() 588 yield_to(tsk, true); in wait_forward() 612 if (IS_ERR(tsk)) { in wait_backward() 613 err = PTR_ERR(tsk); in wait_backward() 616 get_task_struct(tsk); in wait_backward() 617 yield_to(tsk, true); in wait_backward() 659 if (IS_ERR(tsk)) { in wait_random() 660 err = PTR_ERR(tsk); in wait_random() [all …]
|
| /drivers/powercap/ |
| A D | idle_inject.c | 56 struct task_struct *tsk; member 110 wake_up_process(iit->tsk); in idle_inject_wakeup() 283 wait_task_inactive(iit->tsk, TASK_ANY); in idle_inject_stop() 410 .store = &idle_inject_thread.tsk,
|
| /drivers/gpu/drm/i915/gt/ |
| A D | selftest_migrate.c | 678 struct task_struct *tsk; member 697 struct task_struct *tsk; in threaded_migrate() local 703 tsk = kthread_run(fn, &thread[i], "igt-%d", i); in threaded_migrate() 704 if (IS_ERR(tsk)) { in threaded_migrate() 705 err = PTR_ERR(tsk); in threaded_migrate() 709 get_task_struct(tsk); in threaded_migrate() 710 thread[i].tsk = tsk; in threaded_migrate() 716 struct task_struct *tsk = thread[i].tsk; in threaded_migrate() local 719 if (IS_ERR_OR_NULL(tsk)) in threaded_migrate() 722 status = kthread_stop_put(tsk); in threaded_migrate()
|
| A D | selftest_hangcheck.c | 1433 struct task_struct *tsk = NULL; in __igt_reset_evict_vma() local 1533 tsk = kthread_run(fn, &arg, "igt/evict_vma"); in __igt_reset_evict_vma() 1534 if (IS_ERR(tsk)) { in __igt_reset_evict_vma() 1535 err = PTR_ERR(tsk); in __igt_reset_evict_vma() 1537 tsk = NULL; in __igt_reset_evict_vma() 1540 get_task_struct(tsk); in __igt_reset_evict_vma() 1559 if (tsk) { in __igt_reset_evict_vma() 1564 err = kthread_stop(tsk); in __igt_reset_evict_vma() 1566 put_task_struct(tsk); in __igt_reset_evict_vma()
|
| /drivers/net/ppp/ |
| A D | ppp_synctty.c | 70 struct tasklet_struct tsk; member 180 tasklet_setup(&ap->tsk, ppp_sync_process); in ppp_sync_open() 234 tasklet_kill(&ap->tsk); in ppp_sync_close() 337 tasklet_schedule(&ap->tsk); in ppp_sync_receive() 351 tasklet_schedule(&ap->tsk); in ppp_sync_wakeup() 479 struct syncppp *ap = from_tasklet(ap, t, tsk); in ppp_sync_process()
|
| A D | ppp_async.c | 66 struct tasklet_struct tsk; member 183 tasklet_setup(&ap->tsk, ppp_async_process); in ppp_asynctty_open() 236 tasklet_kill(&ap->tsk); in ppp_asynctty_close() 345 tasklet_schedule(&ap->tsk); in ppp_asynctty_receive() 359 tasklet_schedule(&ap->tsk); in ppp_asynctty_wakeup() 488 struct asyncppp *ap = from_tasklet(ap, t, tsk); in ppp_async_process()
|
| /drivers/tty/ |
| A D | tty_ldsem.c | 77 struct task_struct *tsk; in __ldsem_wake_readers() local 95 tsk = waiter->task; in __ldsem_wake_readers() 97 wake_up_process(tsk); in __ldsem_wake_readers() 98 put_task_struct(tsk); in __ldsem_wake_readers()
|
| A D | tty_jobctrl.c | 345 struct task_struct *tsk = current; in no_tty() local 348 proc_clear_tty(tsk); in no_tty()
|
| /drivers/pcmcia/ |
| A D | cs.c | 105 struct task_struct *tsk; in pcmcia_register_socket() local 169 tsk = kthread_run(pccardd, socket, "pccardd"); in pcmcia_register_socket() 170 if (IS_ERR(tsk)) { in pcmcia_register_socket() 171 ret = PTR_ERR(tsk); in pcmcia_register_socket()
|
| /drivers/scsi/qla2xxx/ |
| A D | qla_mbx.c | 3372 tsk->p.tsk.entry_type = TSK_MGMT_IOCB_TYPE; in __qla24xx_issue_tmf() 3373 tsk->p.tsk.entry_count = 1; in __qla24xx_issue_tmf() 3374 tsk->p.tsk.handle = make_handle(req->id, tsk->p.tsk.handle); in __qla24xx_issue_tmf() 3377 tsk->p.tsk.control_flags = cpu_to_le32(type); in __qla24xx_issue_tmf() 3378 tsk->p.tsk.port_id[0] = fcport->d_id.b.al_pa; in __qla24xx_issue_tmf() 3379 tsk->p.tsk.port_id[1] = fcport->d_id.b.area; in __qla24xx_issue_tmf() 3380 tsk->p.tsk.port_id[2] = fcport->d_id.b.domain; in __qla24xx_issue_tmf() 3381 tsk->p.tsk.vp_index = fcport->vha->vp_idx; in __qla24xx_issue_tmf() 3383 int_to_scsilun(l, &tsk->p.tsk.lun); in __qla24xx_issue_tmf() 3384 host_to_fcp_swap((uint8_t *)&tsk->p.tsk.lun, in __qla24xx_issue_tmf() [all …]
|
| A D | qla_iocb.c | 2553 tsk->entry_type = TSK_MGMT_IOCB_TYPE; in qla24xx_tm_iocb() 2554 tsk->entry_count = 1; in qla24xx_tm_iocb() 2555 tsk->handle = make_handle(req->id, tsk->handle); in qla24xx_tm_iocb() 2558 tsk->control_flags = cpu_to_le32(flags); in qla24xx_tm_iocb() 2559 tsk->port_id[0] = fcport->d_id.b.al_pa; in qla24xx_tm_iocb() 2560 tsk->port_id[1] = fcport->d_id.b.area; in qla24xx_tm_iocb() 2561 tsk->port_id[2] = fcport->d_id.b.domain; in qla24xx_tm_iocb() 2562 tsk->vp_index = fcport->vha->vp_idx; in qla24xx_tm_iocb() 2566 int_to_scsilun(lun, &tsk->lun); in qla24xx_tm_iocb() 2567 host_to_fcp_swap((uint8_t *)&tsk->lun, in qla24xx_tm_iocb() [all …]
|
| A D | qla_isr.c | 2642 qla24xx_tm_iocb_entry(scsi_qla_host_t *vha, struct req_que *req, void *tsk) in qla24xx_tm_iocb_entry() argument 2649 struct sts_entry_24xx *sts = (struct sts_entry_24xx *)tsk; in qla24xx_tm_iocb_entry() 2652 sp = qla2x00_get_sp_from_handle(vha, func, req, tsk); in qla24xx_tm_iocb_entry() 2718 void *tsk, srb_t *sp) in qla24xx_nvme_iocb_entry() argument 2722 struct sts_entry_24xx *sts = (struct sts_entry_24xx *)tsk; in qla24xx_nvme_iocb_entry()
|
| /drivers/android/ |
| A D | binder_internal.h | 425 struct task_struct *tsk; member
|
| A D | binder.c | 2794 if (t1->to_proc->tsk == t2->to_proc->tsk && t1->code == t2->code && in binder_can_update_transaction() 3284 t->sender_euid = task_euid(proc->tsk); in binder_transaction() 4981 struct task_struct *sender = t_from->proc->tsk; in binder_thread_read() 5248 put_task_struct(proc->tsk); in binder_free_proc() 5955 if (proc->tsk != current->group_leader) in binder_mmap() 5996 proc->tsk = current->group_leader; in binder_open()
|
| /drivers/gpu/drm/i915/ |
| A D | i915_request.c | 1951 struct task_struct *tsk; member 1958 wake_up_process(fetch_and_zero(&wait->tsk)); in request_wait_wake() 2049 wait.tsk = current; in i915_request_wait_timeout() 2091 if (READ_ONCE(wait.tsk)) in i915_request_wait_timeout()
|
| /drivers/base/power/ |
| A D | main.c | 512 struct task_struct *tsk; member 536 show_stack(wd->tsk, NULL, KERN_EMERG); in dpm_watchdog_handler() 544 show_stack(wd->tsk, NULL, KERN_WARNING); in dpm_watchdog_handler() 560 wd->tsk = current; in dpm_watchdog_set()
|
| /drivers/scsi/ibmvscsi_tgt/ |
| A D | ibmvscsi_tgt.c | 2376 struct srp_tsk_mgmt *tsk; in ibmvscsis_srp_cmd() local 2410 tsk = &vio_iu(iue)->srp.tsk_mgmt; in ibmvscsis_srp_cmd() 2412 tsk->tag, tsk->tag); in ibmvscsis_srp_cmd() 2413 cmd->rsp.tag = tsk->tag; in ibmvscsis_srp_cmd()
|
| /drivers/md/ |
| A D | md.h | 865 struct task_struct *tsk; member
|
| A D | raid5-cache.c | 1583 kthread_park(thread->tsk); in r5l_quiesce() 1587 kthread_unpark(thread->tsk); in r5l_quiesce()
|
| A D | md.c | 8135 wake_up_process(t->tsk); in md_wakeup_thread_directly() 8146 pr_debug("md: waking up MD thread %s.\n", t->tsk->comm); in md_wakeup_thread() 8169 thread->tsk = kthread_run(md_thread, thread, in md_register_thread() 8173 if (IS_ERR(thread->tsk)) { in md_register_thread() 8192 pr_debug("interrupting MD-thread pid %d\n", task_pid_nr(thread->tsk)); in md_unregister_thread() 8193 kthread_stop(thread->tsk); in md_unregister_thread()
|
| A D | raid10.c | 978 if (thread->tsk == current) { in stop_waiting_barrier()
|