| /drivers/dma/ |
| A D | virt-dma.h | 62 vd->tx.flags = tx_flags; in vchan_tx_prep() 63 vd->tx.tx_submit = vchan_tx_submit; in vchan_tx_prep() 67 vd->tx_result.residue = 0; in vchan_tx_prep() 73 return &vd->tx; in vchan_tx_prep() 103 cookie = vd->tx.cookie; in vchan_cookie_complete() 104 dma_cookie_complete(&vd->tx); in vchan_cookie_complete() 106 vd, cookie); in vchan_cookie_complete() 127 vc->desc_free(vd); in vchan_vdesc_fini() 139 vc->cyclic = vd; in vchan_cyclic_callback() 157 if (vc->cyclic == vd) in vchan_terminate_vdesc() [all …]
|
| A D | virt-dma.c | 33 vc, vd, cookie); in vchan_tx_submit() 56 list_del(&vd->node); in vchan_tx_desc_free() 60 vc, vd, vd->tx.cookie); in vchan_tx_desc_free() 61 vc->desc_free(vd); in vchan_tx_desc_free() 73 return vd; in vchan_find_desc() 92 vd = vc->cyclic; in vchan_complete() 93 if (vd) { in vchan_complete() 106 list_del(&vd->node); in vchan_complete() 108 vchan_vdesc_fini(vd); in vchan_complete() 117 list_del(&vd->node); in vchan_dma_desc_free_list() [all …]
|
| A D | loongson1-apb-dma.c | 60 struct virt_dma_desc vd; member 400 struct virt_dma_desc *vd; in ls1x_dma_terminate_all() local 407 if (vd) in ls1x_dma_terminate_all() 408 vchan_terminate_vdesc(vd); in ls1x_dma_terminate_all() 428 struct virt_dma_desc *vd; in ls1x_dma_tx_status() local 438 if (vd) { in ls1x_dma_tx_status() 477 if (vd) { in ls1x_dma_issue_pending() 493 struct virt_dma_desc *vd; in ls1x_dma_irq_handler() local 497 if (!vd) { in ls1x_dma_irq_handler() 505 vchan_cyclic_callback(vd); in ls1x_dma_irq_handler() [all …]
|
| A D | uniphier-mdmac.c | 52 struct virt_dma_desc vd; member 83 return container_of(vd, struct uniphier_mdmac_desc, vd); in to_uniphier_mdmac_desc() 90 struct virt_dma_desc *vd; in uniphier_mdmac_next_desc() local 92 vd = vchan_next_desc(&mc->vc); in uniphier_mdmac_next_desc() 93 if (!vd) { in uniphier_mdmac_next_desc() 98 list_del(&vd->node); in uniphier_mdmac_next_desc() 100 mc->md = to_uniphier_mdmac_desc(vd); in uniphier_mdmac_next_desc() 210 vchan_cookie_complete(&md->vd); in uniphier_mdmac_interrupt() 286 struct virt_dma_desc *vd; in uniphier_mdmac_tx_status() local 312 vd = vchan_find_desc(vc, cookie); in uniphier_mdmac_tx_status() [all …]
|
| A D | pxa_dma.c | 623 __func__, vd, vd->tx.cookie, vd_completed, in pxad_chan_handler() 627 vchan_cyclic_callback(vd); in pxad_chan_handler() 631 list_del(&vd->node); in pxad_chan_handler() 632 vchan_cookie_complete(vd); in pxad_chan_handler() 771 pxad_free_desc(&sw_desc->vd); in pxad_alloc_desc() 793 __func__, vd, cookie); in pxad_tx_submit() 854 INIT_LIST_HEAD(&vd->node); in pxad_tx_prep() 859 vc, vd, vd->tx.cookie, in pxad_tx_prep() 1095 vd, vd->tx.cookie, is_desc_completed(vd)); in pxad_terminate_all() 1134 if (!vd) in pxad_residue() [all …]
|
| A D | milbeaut-xdmac.c | 62 struct virt_dma_desc vd; member 89 return container_of(vd, struct milbeaut_xdmac_desc, vd); in to_milbeaut_xdmac_desc() 96 struct virt_dma_desc *vd; in milbeaut_xdmac_next_desc() local 98 vd = vchan_next_desc(&mc->vc); in milbeaut_xdmac_next_desc() 99 if (!vd) { in milbeaut_xdmac_next_desc() 104 list_del(&vd->node); in milbeaut_xdmac_next_desc() 106 mc->md = to_milbeaut_xdmac_desc(vd); in milbeaut_xdmac_next_desc() 175 vchan_cookie_complete(&md->vd); in milbeaut_xdmac_interrupt() 203 return vchan_tx_prep(vc, &md->vd, flags); in milbeaut_xdmac_prep_memcpy() 224 vchan_terminate_vdesc(&mc->md->vd); in milbeaut_xdmac_terminate_all() [all …]
|
| A D | milbeaut-hdmac.c | 60 struct virt_dma_desc vd; member 92 return container_of(vd, struct milbeaut_hdmac_desc, vd); in to_milbeaut_hdmac_desc() 99 struct virt_dma_desc *vd; in milbeaut_hdmac_next_desc() local 101 vd = vchan_next_desc(&mc->vc); in milbeaut_hdmac_next_desc() 102 if (!vd) { in milbeaut_hdmac_next_desc() 107 list_del(&vd->node); in milbeaut_hdmac_next_desc() 109 mc->md = to_milbeaut_hdmac_desc(vd); in milbeaut_hdmac_next_desc() 194 vchan_cookie_complete(&md->vd); in milbeaut_hdmac_interrupt() 326 struct virt_dma_desc *vd; in milbeaut_hdmac_tx_status() local 364 vd = vchan_find_desc(vc, cookie); in milbeaut_hdmac_tx_status() [all …]
|
| A D | amba-pl08x.c | 394 if (pl08x->vd->pl080s) in pl08x_write_lli() 514 if (pl08x->vd->pl080s) in pl08x_write_lli() 1570 if (vd) { in pl08x_dma_tx_status() 2720 pl08x->vd = vd; in pl08x_probe() 2728 if (vd->ftdmac020) { in pl08x_probe() 2766 if (vd->ftdmac020) in pl08x_probe() 2774 if (vd->signals) { in pl08x_probe() 2824 if (vd->pl080s) in pl08x_probe() 2842 if (vd->ftdmac020) in pl08x_probe() 2886 if (vd->pl080s) in pl08x_probe() [all …]
|
| A D | moxart-dma.c | 125 struct virt_dma_desc vd; member 175 kfree(container_of(vd, struct moxart_desc, vd)); in moxart_dma_desc_free() 190 moxart_dma_desc_free(&ch->desc->vd); in moxart_terminate_all() 413 struct virt_dma_desc *vd; in moxart_dma_start_desc() local 415 vd = vchan_next_desc(&ch->vc); in moxart_dma_start_desc() 417 if (!vd) { in moxart_dma_start_desc() 422 list_del(&vd->node); in moxart_dma_start_desc() 473 struct virt_dma_desc *vd; in moxart_tx_status() local 484 vd = vchan_find_desc(&ch->vc, cookie); in moxart_tx_status() 485 if (vd) { in moxart_tx_status() [all …]
|
| A D | arm-dma350.c | 172 struct virt_dma_desc vd; member 207 return container_of(vd, struct d350_desc, vd); in to_d350_desc() 212 kfree(to_d350_desc(vd)); in d350_desc_free() 347 vchan_terminate_vdesc(&dch->desc->vd); in d350_terminate_all() 374 struct virt_dma_desc *vd; in d350_tx_status() local 407 list_del(&dch->desc->vd.node); in d350_start_next() 409 dch->cookie = dch->desc->vd.tx.cookie; in d350_start_next() 468 struct virt_dma_desc *vd = &dch->desc->vd; in d350_irq() local 483 vd->tx_result.result = DMA_TRANS_ABORTED; in d350_irq() 492 vchan_cookie_complete(vd); in d350_irq() [all …]
|
| A D | uniphier-xdmac.c | 78 struct virt_dma_desc vd; member 112 return container_of(vd, struct uniphier_xdmac_desc, vd); in to_uniphier_xdmac_desc() 119 struct virt_dma_desc *vd; in uniphier_xdmac_next_desc() local 121 vd = vchan_next_desc(&xc->vc); in uniphier_xdmac_next_desc() 122 if (!vd) in uniphier_xdmac_next_desc() 125 list_del(&vd->node); in uniphier_xdmac_next_desc() 127 return to_uniphier_xdmac_desc(vd); in uniphier_xdmac_next_desc() 251 vchan_cookie_complete(&xc->xd->vd); in uniphier_xdmac_chan_irq() 315 return vchan_tx_prep(vc, &xd->vd, flags); in uniphier_xdmac_prep_dma_memcpy() 416 vchan_terminate_vdesc(&xc->xd->vd); in uniphier_xdmac_terminate_all() [all …]
|
| A D | img-mdc-dma.c | 105 struct virt_dma_desc vd; member 334 mdc_desc_free(&mdesc->vd); in mdc_prep_dma_memcpy() 441 mdc_desc_free(&mdesc->vd); in mdc_prep_dma_cyclic() 518 mdc_desc_free(&mdesc->vd); in mdc_prep_slave_sg() 526 struct virt_dma_desc *vd; in mdc_issue_desc() local 531 if (!vd) in mdc_issue_desc() 534 list_del(&vd->node); in mdc_issue_desc() 536 mdesc = to_mdc_desc(&vd->tx); in mdc_issue_desc() 575 struct virt_dma_desc *vd; in mdc_tx_status() local 589 if (vd) { in mdc_tx_status() [all …]
|
| A D | bcm2835-dma.c | 87 struct virt_dma_desc vd; member 218 container_of(vd, struct bcm2835_desc, vd)); in bcm2835_dma_desc_free() 443 if (!vd) { in bcm2835_dma_start_desc() 448 list_del(&vd->node); in bcm2835_dma_start_desc() 450 c->desc = d = to_bcm2835_dma_desc(&vd->tx); in bcm2835_dma_start_desc() 488 vchan_cyclic_callback(&d->vd); in bcm2835_dma_callback() 490 vchan_cookie_complete(&c->desc->vd); in bcm2835_dma_callback() 566 struct virt_dma_desc *vd; in bcm2835_dma_tx_status() local 575 vd = vchan_find_desc(&c->vc, cookie); in bcm2835_dma_tx_status() 576 if (vd) { in bcm2835_dma_tx_status() [all …]
|
| A D | sa11x0-dma.c | 73 struct virt_dma_desc vd; member 140 return vd ? container_of(vd, struct sa11x0_dma_desc, vd) : NULL; in sa11x0_dma_next_desc() 145 kfree(container_of(vd, struct sa11x0_dma_desc, vd)); in sa11x0_dma_free_desc() 150 list_del(&txd->vd.node); in sa11x0_dma_start_desc() 155 p->num, &txd->vd, txd->vd.tx.cookie, txd->ddar); in sa11x0_dma_start_desc() 229 vchan_cookie_complete(&txd->vd); in sa11x0_dma_complete() 238 vchan_cyclic_callback(&txd->vd); in sa11x0_dma_complete() 423 struct virt_dma_desc *vd; in sa11x0_dma_tx_status() local 441 vd = vchan_find_desc(&c->vc, cookie); in sa11x0_dma_tx_status() 442 if (vd) { in sa11x0_dma_tx_status() [all …]
|
| /drivers/dma/mediatek/ |
| A D | mtk-hsdma.c | 131 struct virt_dma_desc vd; member 147 struct virt_dma_desc *vd; member 260 return container_of(vd, struct mtk_hsdma_vdesc, vd); in to_hsdma_vdesc() 301 kfree(container_of(vd, struct mtk_hsdma_vdesc, vd)); in mtk_hsdma_vdesc_free() 459 ring->cb[ring->cur_tptr].vd = &hvd->vd; in mtk_hsdma_issue_pending_vdesc() 572 if (unlikely(!cb->vd)) { in mtk_hsdma_free_rooms_in_ring() 603 cb->vd = NULL; in mtk_hsdma_free_rooms_in_ring() 663 struct virt_dma_desc *vd; in mtk_hsdma_find_active_desc() local 667 return vd; in mtk_hsdma_find_active_desc() 671 return vd; in mtk_hsdma_find_active_desc() [all …]
|
| A D | mtk-cqdma.c | 79 struct virt_dma_desc vd; member 160 return container_of(vd, struct mtk_cqdma_vdesc, vd); in to_cqdma_vdesc() 201 kfree(to_cqdma_vdesc(vd)); in mtk_cqdma_vdesc_free() 280 cvd = to_cqdma_vdesc(vd); in mtk_cqdma_issue_vchan_pending() 290 list_del(&vd->node); in mtk_cqdma_issue_vchan_pending() 422 struct virt_dma_desc *vd; in mtk_cqdma_find_active_desc() local 426 return vd; in mtk_cqdma_find_active_desc() 431 return vd; in mtk_cqdma_find_active_desc() 442 struct virt_dma_desc *vd; in mtk_cqdma_tx_status() local 457 if (vd) { in mtk_cqdma_tx_status() [all …]
|
| /drivers/dma/amd/ptdma/ |
| A D | ptdma-dmaengine.c | 72 return container_of(vd, struct pt_dma_desc, vd); in to_pt_desc() 193 return vd ? to_pt_desc(vd) : NULL; in pt_next_dma_desc() 200 struct virt_dma_desc *vd; in pt_handle_active_desc() local 214 tx_desc = &desc->vd.tx; in pt_handle_active_desc() 215 vd = &desc->vd; in pt_handle_active_desc() 243 vchan_vdesc_fini(vd); in pt_handle_active_desc() 354 tx_desc = &desc->vd.tx; in pt_cmd_callback_work() 355 vd = &desc->vd; in pt_cmd_callback_work() 376 vchan_vdesc_fini(vd); in pt_cmd_callback_work() 432 return &desc->vd.tx; in pt_prep_dma_memcpy() [all …]
|
| /drivers/clk/versatile/ |
| A D | icst.c | 76 unsigned int vd; in icst_hz_to_vco() local 81 vd = (f + fref_div / 2) / fref_div; in icst_hz_to_vco() 82 if (vd < p->vd_min || vd > p->vd_max) in icst_hz_to_vco() 85 f_pll = fref_div * vd; in icst_hz_to_vco() 91 vco.v = vd - 8; in icst_hz_to_vco()
|
| /drivers/dma/sh/ |
| A D | usb-dmac.c | 52 struct virt_dma_desc vd; member 63 #define to_usb_dmac_desc(vd) container_of(vd, struct usb_dmac_desc, vd) argument 224 struct virt_dma_desc *vd; in usb_dmac_chan_start_desc() local 226 vd = vchan_next_desc(&chan->vc); in usb_dmac_chan_start_desc() 227 if (!vd) { in usb_dmac_chan_start_desc() 237 list_del(&vd->node); in usb_dmac_chan_start_desc() 239 chan->desc = to_usb_dmac_desc(vd); in usb_dmac_chan_start_desc() 507 struct virt_dma_desc *vd; in usb_dmac_chan_get_residue() local 513 if (!vd) in usb_dmac_chan_get_residue() 515 desc = to_usb_dmac_desc(vd); in usb_dmac_chan_get_residue() [all …]
|
| /drivers/media/radio/ |
| A D | tea575x.c | 522 tea->vd = tea575x_radio; in snd_tea575x_init() 523 video_set_drvdata(&tea->vd, tea); in snd_tea575x_init() 525 strscpy(tea->vd.name, tea->v4l2_dev->name, sizeof(tea->vd.name)); in snd_tea575x_init() 526 tea->vd.lock = &tea->mutex; in snd_tea575x_init() 527 tea->vd.v4l2_dev = tea->v4l2_dev; in snd_tea575x_init() 530 tea->vd.device_caps |= V4L2_CAP_HW_FREQ_SEEK; in snd_tea575x_init() 533 tea->vd.fops = &tea->fops; in snd_tea575x_init() 539 tea->vd.ctrl_handler = &tea->ctrl_handler; in snd_tea575x_init() 564 v4l2_ctrl_handler_free(tea->vd.ctrl_handler); in snd_tea575x_init() 574 video_unregister_device(&tea->vd); in snd_tea575x_exit() [all …]
|
| A D | radio-tea5777.c | 547 tea->vd = tea575x_radio; in radio_tea5777_init() 548 video_set_drvdata(&tea->vd, tea); in radio_tea5777_init() 550 strscpy(tea->vd.name, tea->v4l2_dev->name, sizeof(tea->vd.name)); in radio_tea5777_init() 551 tea->vd.lock = &tea->mutex; in radio_tea5777_init() 552 tea->vd.v4l2_dev = tea->v4l2_dev; in radio_tea5777_init() 553 tea->vd.device_caps = V4L2_CAP_TUNER | V4L2_CAP_RADIO | in radio_tea5777_init() 557 tea->vd.fops = &tea->fops; in radio_tea5777_init() 559 tea->vd.ctrl_handler = &tea->ctrl_handler; in radio_tea5777_init() 574 v4l2_ctrl_handler_free(tea->vd.ctrl_handler); in radio_tea5777_init() 584 video_unregister_device(&tea->vd); in radio_tea5777_exit() [all …]
|
| /drivers/dma/dw-edma/ |
| A D | dw-edma-core.c | 29 return container_of(vd, struct dw_edma_desc, vd); in vd2dw_edma_desc() 179 struct virt_dma_desc *vd; in dw_edma_start_transfer() local 182 if (!vd) in dw_edma_start_transfer() 185 desc = vd2dw_edma_desc(vd); in dw_edma_start_transfer() 320 struct virt_dma_desc *vd; in dw_edma_device_tx_status() local 337 if (vd) { in dw_edma_device_tx_status() 595 if (vd) { in dw_edma_done_interrupt() 600 list_del(&vd->node); in dw_edma_done_interrupt() 610 list_del(&vd->node); in dw_edma_done_interrupt() 635 if (vd) { in dw_edma_abort_interrupt() [all …]
|
| /drivers/dma/xilinx/ |
| A D | xdma.c | 338 if (!vd || xchan->busy) in xdma_xfer_start() 347 desc = to_xdma_desc(vd); in xdma_xfer_start() 510 struct virt_dma_desc *vd; in xdma_terminate_all() local 521 if (vd) { in xdma_terminate_all() 522 list_del(&vd->node); in xdma_terminate_all() 851 if (!vd) in xdma_tx_status() 854 desc = to_xdma_desc(vd); in xdma_tx_status() 891 if (!vd) in xdma_channel_isr() 899 desc = to_xdma_desc(vd); in xdma_channel_isr() 937 list_del(&vd->node); in xdma_channel_isr() [all …]
|
| /drivers/dma/amd/qdma/ |
| A D | qdma.c | 433 kfree(vd); in qdma_free_vdesc() 711 struct virt_dma_desc *vd; in qdma_fill_pending_vdesc() local 717 vdesc = to_qdma_vdesc(vd); in qdma_fill_pending_vdesc() 733 vd = list_first_entry(&vc->desc_submitted, typeof(*vd), node); in qdma_fill_pending_vdesc() 736 vdesc = to_qdma_vdesc(vd); in qdma_fill_pending_vdesc() 748 struct virt_dma_desc *vd; in qdma_tx_submit() local 845 struct virt_dma_desc *vd; in qdma_queue_isr() local 868 if (!vd) in qdma_queue_isr() 871 vdesc = to_qdma_vdesc(vd); in qdma_queue_isr() 873 list_del(&vd->node); in qdma_queue_isr() [all …]
|
| /drivers/dma/qcom/ |
| A D | bam_dma.c | 63 struct virt_dma_desc vd; member 883 list_add(&async_desc->vd.node, in process_channel_irqs() 949 struct virt_dma_desc *vd; in bam_tx_status() local 964 if (vd) { in bam_tx_status() 965 residue = container_of(vd, struct bam_async_desc, vd)->length; in bam_tx_status() 1029 if (!vd) in bam_start_dma() 1036 while (vd && !IS_BUSY(bchan)) { in bam_start_dma() 1037 list_del(&vd->node); in bam_start_dma() 1039 async_desc = container_of(vd, struct bam_async_desc, vd); in bam_start_dma() 1063 vd = vchan_next_desc(&bchan->vc); in bam_start_dma() [all …]
|