Home
last modified time | relevance | path

Searched refs:vd (Results 1 – 25 of 67) sorted by relevance

123

/drivers/dma/
A Dvirt-dma.h62 vd->tx.flags = tx_flags; in vchan_tx_prep()
63 vd->tx.tx_submit = vchan_tx_submit; in vchan_tx_prep()
67 vd->tx_result.residue = 0; in vchan_tx_prep()
73 return &vd->tx; in vchan_tx_prep()
103 cookie = vd->tx.cookie; in vchan_cookie_complete()
104 dma_cookie_complete(&vd->tx); in vchan_cookie_complete()
106 vd, cookie); in vchan_cookie_complete()
127 vc->desc_free(vd); in vchan_vdesc_fini()
139 vc->cyclic = vd; in vchan_cyclic_callback()
157 if (vc->cyclic == vd) in vchan_terminate_vdesc()
[all …]
A Dvirt-dma.c33 vc, vd, cookie); in vchan_tx_submit()
56 list_del(&vd->node); in vchan_tx_desc_free()
60 vc, vd, vd->tx.cookie); in vchan_tx_desc_free()
61 vc->desc_free(vd); in vchan_tx_desc_free()
73 return vd; in vchan_find_desc()
92 vd = vc->cyclic; in vchan_complete()
93 if (vd) { in vchan_complete()
106 list_del(&vd->node); in vchan_complete()
108 vchan_vdesc_fini(vd); in vchan_complete()
117 list_del(&vd->node); in vchan_dma_desc_free_list()
[all …]
A Dloongson1-apb-dma.c60 struct virt_dma_desc vd; member
400 struct virt_dma_desc *vd; in ls1x_dma_terminate_all() local
407 if (vd) in ls1x_dma_terminate_all()
408 vchan_terminate_vdesc(vd); in ls1x_dma_terminate_all()
428 struct virt_dma_desc *vd; in ls1x_dma_tx_status() local
438 if (vd) { in ls1x_dma_tx_status()
477 if (vd) { in ls1x_dma_issue_pending()
493 struct virt_dma_desc *vd; in ls1x_dma_irq_handler() local
497 if (!vd) { in ls1x_dma_irq_handler()
505 vchan_cyclic_callback(vd); in ls1x_dma_irq_handler()
[all …]
A Duniphier-mdmac.c52 struct virt_dma_desc vd; member
83 return container_of(vd, struct uniphier_mdmac_desc, vd); in to_uniphier_mdmac_desc()
90 struct virt_dma_desc *vd; in uniphier_mdmac_next_desc() local
92 vd = vchan_next_desc(&mc->vc); in uniphier_mdmac_next_desc()
93 if (!vd) { in uniphier_mdmac_next_desc()
98 list_del(&vd->node); in uniphier_mdmac_next_desc()
100 mc->md = to_uniphier_mdmac_desc(vd); in uniphier_mdmac_next_desc()
210 vchan_cookie_complete(&md->vd); in uniphier_mdmac_interrupt()
286 struct virt_dma_desc *vd; in uniphier_mdmac_tx_status() local
312 vd = vchan_find_desc(vc, cookie); in uniphier_mdmac_tx_status()
[all …]
A Dpxa_dma.c623 __func__, vd, vd->tx.cookie, vd_completed, in pxad_chan_handler()
627 vchan_cyclic_callback(vd); in pxad_chan_handler()
631 list_del(&vd->node); in pxad_chan_handler()
632 vchan_cookie_complete(vd); in pxad_chan_handler()
771 pxad_free_desc(&sw_desc->vd); in pxad_alloc_desc()
793 __func__, vd, cookie); in pxad_tx_submit()
854 INIT_LIST_HEAD(&vd->node); in pxad_tx_prep()
859 vc, vd, vd->tx.cookie, in pxad_tx_prep()
1095 vd, vd->tx.cookie, is_desc_completed(vd)); in pxad_terminate_all()
1134 if (!vd) in pxad_residue()
[all …]
A Dmilbeaut-xdmac.c62 struct virt_dma_desc vd; member
89 return container_of(vd, struct milbeaut_xdmac_desc, vd); in to_milbeaut_xdmac_desc()
96 struct virt_dma_desc *vd; in milbeaut_xdmac_next_desc() local
98 vd = vchan_next_desc(&mc->vc); in milbeaut_xdmac_next_desc()
99 if (!vd) { in milbeaut_xdmac_next_desc()
104 list_del(&vd->node); in milbeaut_xdmac_next_desc()
106 mc->md = to_milbeaut_xdmac_desc(vd); in milbeaut_xdmac_next_desc()
175 vchan_cookie_complete(&md->vd); in milbeaut_xdmac_interrupt()
203 return vchan_tx_prep(vc, &md->vd, flags); in milbeaut_xdmac_prep_memcpy()
224 vchan_terminate_vdesc(&mc->md->vd); in milbeaut_xdmac_terminate_all()
[all …]
A Dmilbeaut-hdmac.c60 struct virt_dma_desc vd; member
92 return container_of(vd, struct milbeaut_hdmac_desc, vd); in to_milbeaut_hdmac_desc()
99 struct virt_dma_desc *vd; in milbeaut_hdmac_next_desc() local
101 vd = vchan_next_desc(&mc->vc); in milbeaut_hdmac_next_desc()
102 if (!vd) { in milbeaut_hdmac_next_desc()
107 list_del(&vd->node); in milbeaut_hdmac_next_desc()
109 mc->md = to_milbeaut_hdmac_desc(vd); in milbeaut_hdmac_next_desc()
194 vchan_cookie_complete(&md->vd); in milbeaut_hdmac_interrupt()
326 struct virt_dma_desc *vd; in milbeaut_hdmac_tx_status() local
364 vd = vchan_find_desc(vc, cookie); in milbeaut_hdmac_tx_status()
[all …]
A Damba-pl08x.c394 if (pl08x->vd->pl080s) in pl08x_write_lli()
514 if (pl08x->vd->pl080s) in pl08x_write_lli()
1570 if (vd) { in pl08x_dma_tx_status()
2720 pl08x->vd = vd; in pl08x_probe()
2728 if (vd->ftdmac020) { in pl08x_probe()
2766 if (vd->ftdmac020) in pl08x_probe()
2774 if (vd->signals) { in pl08x_probe()
2824 if (vd->pl080s) in pl08x_probe()
2842 if (vd->ftdmac020) in pl08x_probe()
2886 if (vd->pl080s) in pl08x_probe()
[all …]
A Dmoxart-dma.c125 struct virt_dma_desc vd; member
175 kfree(container_of(vd, struct moxart_desc, vd)); in moxart_dma_desc_free()
190 moxart_dma_desc_free(&ch->desc->vd); in moxart_terminate_all()
413 struct virt_dma_desc *vd; in moxart_dma_start_desc() local
415 vd = vchan_next_desc(&ch->vc); in moxart_dma_start_desc()
417 if (!vd) { in moxart_dma_start_desc()
422 list_del(&vd->node); in moxart_dma_start_desc()
473 struct virt_dma_desc *vd; in moxart_tx_status() local
484 vd = vchan_find_desc(&ch->vc, cookie); in moxart_tx_status()
485 if (vd) { in moxart_tx_status()
[all …]
A Darm-dma350.c172 struct virt_dma_desc vd; member
207 return container_of(vd, struct d350_desc, vd); in to_d350_desc()
212 kfree(to_d350_desc(vd)); in d350_desc_free()
347 vchan_terminate_vdesc(&dch->desc->vd); in d350_terminate_all()
374 struct virt_dma_desc *vd; in d350_tx_status() local
407 list_del(&dch->desc->vd.node); in d350_start_next()
409 dch->cookie = dch->desc->vd.tx.cookie; in d350_start_next()
468 struct virt_dma_desc *vd = &dch->desc->vd; in d350_irq() local
483 vd->tx_result.result = DMA_TRANS_ABORTED; in d350_irq()
492 vchan_cookie_complete(vd); in d350_irq()
[all …]
A Duniphier-xdmac.c78 struct virt_dma_desc vd; member
112 return container_of(vd, struct uniphier_xdmac_desc, vd); in to_uniphier_xdmac_desc()
119 struct virt_dma_desc *vd; in uniphier_xdmac_next_desc() local
121 vd = vchan_next_desc(&xc->vc); in uniphier_xdmac_next_desc()
122 if (!vd) in uniphier_xdmac_next_desc()
125 list_del(&vd->node); in uniphier_xdmac_next_desc()
127 return to_uniphier_xdmac_desc(vd); in uniphier_xdmac_next_desc()
251 vchan_cookie_complete(&xc->xd->vd); in uniphier_xdmac_chan_irq()
315 return vchan_tx_prep(vc, &xd->vd, flags); in uniphier_xdmac_prep_dma_memcpy()
416 vchan_terminate_vdesc(&xc->xd->vd); in uniphier_xdmac_terminate_all()
[all …]
A Dimg-mdc-dma.c105 struct virt_dma_desc vd; member
334 mdc_desc_free(&mdesc->vd); in mdc_prep_dma_memcpy()
441 mdc_desc_free(&mdesc->vd); in mdc_prep_dma_cyclic()
518 mdc_desc_free(&mdesc->vd); in mdc_prep_slave_sg()
526 struct virt_dma_desc *vd; in mdc_issue_desc() local
531 if (!vd) in mdc_issue_desc()
534 list_del(&vd->node); in mdc_issue_desc()
536 mdesc = to_mdc_desc(&vd->tx); in mdc_issue_desc()
575 struct virt_dma_desc *vd; in mdc_tx_status() local
589 if (vd) { in mdc_tx_status()
[all …]
A Dbcm2835-dma.c87 struct virt_dma_desc vd; member
218 container_of(vd, struct bcm2835_desc, vd)); in bcm2835_dma_desc_free()
443 if (!vd) { in bcm2835_dma_start_desc()
448 list_del(&vd->node); in bcm2835_dma_start_desc()
450 c->desc = d = to_bcm2835_dma_desc(&vd->tx); in bcm2835_dma_start_desc()
488 vchan_cyclic_callback(&d->vd); in bcm2835_dma_callback()
490 vchan_cookie_complete(&c->desc->vd); in bcm2835_dma_callback()
566 struct virt_dma_desc *vd; in bcm2835_dma_tx_status() local
575 vd = vchan_find_desc(&c->vc, cookie); in bcm2835_dma_tx_status()
576 if (vd) { in bcm2835_dma_tx_status()
[all …]
A Dsa11x0-dma.c73 struct virt_dma_desc vd; member
140 return vd ? container_of(vd, struct sa11x0_dma_desc, vd) : NULL; in sa11x0_dma_next_desc()
145 kfree(container_of(vd, struct sa11x0_dma_desc, vd)); in sa11x0_dma_free_desc()
150 list_del(&txd->vd.node); in sa11x0_dma_start_desc()
155 p->num, &txd->vd, txd->vd.tx.cookie, txd->ddar); in sa11x0_dma_start_desc()
229 vchan_cookie_complete(&txd->vd); in sa11x0_dma_complete()
238 vchan_cyclic_callback(&txd->vd); in sa11x0_dma_complete()
423 struct virt_dma_desc *vd; in sa11x0_dma_tx_status() local
441 vd = vchan_find_desc(&c->vc, cookie); in sa11x0_dma_tx_status()
442 if (vd) { in sa11x0_dma_tx_status()
[all …]
/drivers/dma/mediatek/
A Dmtk-hsdma.c131 struct virt_dma_desc vd; member
147 struct virt_dma_desc *vd; member
260 return container_of(vd, struct mtk_hsdma_vdesc, vd); in to_hsdma_vdesc()
301 kfree(container_of(vd, struct mtk_hsdma_vdesc, vd)); in mtk_hsdma_vdesc_free()
459 ring->cb[ring->cur_tptr].vd = &hvd->vd; in mtk_hsdma_issue_pending_vdesc()
572 if (unlikely(!cb->vd)) { in mtk_hsdma_free_rooms_in_ring()
603 cb->vd = NULL; in mtk_hsdma_free_rooms_in_ring()
663 struct virt_dma_desc *vd; in mtk_hsdma_find_active_desc() local
667 return vd; in mtk_hsdma_find_active_desc()
671 return vd; in mtk_hsdma_find_active_desc()
[all …]
A Dmtk-cqdma.c79 struct virt_dma_desc vd; member
160 return container_of(vd, struct mtk_cqdma_vdesc, vd); in to_cqdma_vdesc()
201 kfree(to_cqdma_vdesc(vd)); in mtk_cqdma_vdesc_free()
280 cvd = to_cqdma_vdesc(vd); in mtk_cqdma_issue_vchan_pending()
290 list_del(&vd->node); in mtk_cqdma_issue_vchan_pending()
422 struct virt_dma_desc *vd; in mtk_cqdma_find_active_desc() local
426 return vd; in mtk_cqdma_find_active_desc()
431 return vd; in mtk_cqdma_find_active_desc()
442 struct virt_dma_desc *vd; in mtk_cqdma_tx_status() local
457 if (vd) { in mtk_cqdma_tx_status()
[all …]
/drivers/dma/amd/ptdma/
A Dptdma-dmaengine.c72 return container_of(vd, struct pt_dma_desc, vd); in to_pt_desc()
193 return vd ? to_pt_desc(vd) : NULL; in pt_next_dma_desc()
200 struct virt_dma_desc *vd; in pt_handle_active_desc() local
214 tx_desc = &desc->vd.tx; in pt_handle_active_desc()
215 vd = &desc->vd; in pt_handle_active_desc()
243 vchan_vdesc_fini(vd); in pt_handle_active_desc()
354 tx_desc = &desc->vd.tx; in pt_cmd_callback_work()
355 vd = &desc->vd; in pt_cmd_callback_work()
376 vchan_vdesc_fini(vd); in pt_cmd_callback_work()
432 return &desc->vd.tx; in pt_prep_dma_memcpy()
[all …]
/drivers/clk/versatile/
A Dicst.c76 unsigned int vd; in icst_hz_to_vco() local
81 vd = (f + fref_div / 2) / fref_div; in icst_hz_to_vco()
82 if (vd < p->vd_min || vd > p->vd_max) in icst_hz_to_vco()
85 f_pll = fref_div * vd; in icst_hz_to_vco()
91 vco.v = vd - 8; in icst_hz_to_vco()
/drivers/dma/sh/
A Dusb-dmac.c52 struct virt_dma_desc vd; member
63 #define to_usb_dmac_desc(vd) container_of(vd, struct usb_dmac_desc, vd) argument
224 struct virt_dma_desc *vd; in usb_dmac_chan_start_desc() local
226 vd = vchan_next_desc(&chan->vc); in usb_dmac_chan_start_desc()
227 if (!vd) { in usb_dmac_chan_start_desc()
237 list_del(&vd->node); in usb_dmac_chan_start_desc()
239 chan->desc = to_usb_dmac_desc(vd); in usb_dmac_chan_start_desc()
507 struct virt_dma_desc *vd; in usb_dmac_chan_get_residue() local
513 if (!vd) in usb_dmac_chan_get_residue()
515 desc = to_usb_dmac_desc(vd); in usb_dmac_chan_get_residue()
[all …]
/drivers/media/radio/
A Dtea575x.c522 tea->vd = tea575x_radio; in snd_tea575x_init()
523 video_set_drvdata(&tea->vd, tea); in snd_tea575x_init()
525 strscpy(tea->vd.name, tea->v4l2_dev->name, sizeof(tea->vd.name)); in snd_tea575x_init()
526 tea->vd.lock = &tea->mutex; in snd_tea575x_init()
527 tea->vd.v4l2_dev = tea->v4l2_dev; in snd_tea575x_init()
530 tea->vd.device_caps |= V4L2_CAP_HW_FREQ_SEEK; in snd_tea575x_init()
533 tea->vd.fops = &tea->fops; in snd_tea575x_init()
539 tea->vd.ctrl_handler = &tea->ctrl_handler; in snd_tea575x_init()
564 v4l2_ctrl_handler_free(tea->vd.ctrl_handler); in snd_tea575x_init()
574 video_unregister_device(&tea->vd); in snd_tea575x_exit()
[all …]
A Dradio-tea5777.c547 tea->vd = tea575x_radio; in radio_tea5777_init()
548 video_set_drvdata(&tea->vd, tea); in radio_tea5777_init()
550 strscpy(tea->vd.name, tea->v4l2_dev->name, sizeof(tea->vd.name)); in radio_tea5777_init()
551 tea->vd.lock = &tea->mutex; in radio_tea5777_init()
552 tea->vd.v4l2_dev = tea->v4l2_dev; in radio_tea5777_init()
553 tea->vd.device_caps = V4L2_CAP_TUNER | V4L2_CAP_RADIO | in radio_tea5777_init()
557 tea->vd.fops = &tea->fops; in radio_tea5777_init()
559 tea->vd.ctrl_handler = &tea->ctrl_handler; in radio_tea5777_init()
574 v4l2_ctrl_handler_free(tea->vd.ctrl_handler); in radio_tea5777_init()
584 video_unregister_device(&tea->vd); in radio_tea5777_exit()
[all …]
/drivers/dma/dw-edma/
A Ddw-edma-core.c29 return container_of(vd, struct dw_edma_desc, vd); in vd2dw_edma_desc()
179 struct virt_dma_desc *vd; in dw_edma_start_transfer() local
182 if (!vd) in dw_edma_start_transfer()
185 desc = vd2dw_edma_desc(vd); in dw_edma_start_transfer()
320 struct virt_dma_desc *vd; in dw_edma_device_tx_status() local
337 if (vd) { in dw_edma_device_tx_status()
595 if (vd) { in dw_edma_done_interrupt()
600 list_del(&vd->node); in dw_edma_done_interrupt()
610 list_del(&vd->node); in dw_edma_done_interrupt()
635 if (vd) { in dw_edma_abort_interrupt()
[all …]
/drivers/dma/xilinx/
A Dxdma.c338 if (!vd || xchan->busy) in xdma_xfer_start()
347 desc = to_xdma_desc(vd); in xdma_xfer_start()
510 struct virt_dma_desc *vd; in xdma_terminate_all() local
521 if (vd) { in xdma_terminate_all()
522 list_del(&vd->node); in xdma_terminate_all()
851 if (!vd) in xdma_tx_status()
854 desc = to_xdma_desc(vd); in xdma_tx_status()
891 if (!vd) in xdma_channel_isr()
899 desc = to_xdma_desc(vd); in xdma_channel_isr()
937 list_del(&vd->node); in xdma_channel_isr()
[all …]
/drivers/dma/amd/qdma/
A Dqdma.c433 kfree(vd); in qdma_free_vdesc()
711 struct virt_dma_desc *vd; in qdma_fill_pending_vdesc() local
717 vdesc = to_qdma_vdesc(vd); in qdma_fill_pending_vdesc()
733 vd = list_first_entry(&vc->desc_submitted, typeof(*vd), node); in qdma_fill_pending_vdesc()
736 vdesc = to_qdma_vdesc(vd); in qdma_fill_pending_vdesc()
748 struct virt_dma_desc *vd; in qdma_tx_submit() local
845 struct virt_dma_desc *vd; in qdma_queue_isr() local
868 if (!vd) in qdma_queue_isr()
871 vdesc = to_qdma_vdesc(vd); in qdma_queue_isr()
873 list_del(&vd->node); in qdma_queue_isr()
[all …]
/drivers/dma/qcom/
A Dbam_dma.c63 struct virt_dma_desc vd; member
883 list_add(&async_desc->vd.node, in process_channel_irqs()
949 struct virt_dma_desc *vd; in bam_tx_status() local
964 if (vd) { in bam_tx_status()
965 residue = container_of(vd, struct bam_async_desc, vd)->length; in bam_tx_status()
1029 if (!vd) in bam_start_dma()
1036 while (vd && !IS_BUSY(bchan)) { in bam_start_dma()
1037 list_del(&vd->node); in bam_start_dma()
1039 async_desc = container_of(vd, struct bam_async_desc, vd); in bam_start_dma()
1063 vd = vchan_next_desc(&bchan->vc); in bam_start_dma()
[all …]

Completed in 83 milliseconds

123