Home
last modified time | relevance | path

Searched refs:dchan (Results 1 – 22 of 22) sorted by relevance

/drivers/dma/
A Dloongson1-apb-dma.c86 #define to_ls1x_dma_chan(dchan) \ argument
100 struct dma_chan *dchan = &chan->vc.chan; in ls1x_dma_query() local
105 val |= dchan->chan_id; in ls1x_dma_query()
119 struct device *dev = chan2dev(dchan); in ls1x_dma_start()
124 val |= dchan->chan_id; in ls1x_dma_start()
146 struct device *dev = chan2dev(dchan); in ls1x_dma_free_chan_resources()
159 struct device *dev = chan2dev(dchan); in ls1x_dma_alloc_chan_resources()
222 struct device *dev = chan2dev(dchan); in ls1x_dma_prep_lli()
328 dev_dbg(chan2dev(dchan), in ls1x_dma_prep_dma_cyclic()
492 struct device *dev = chan2dev(dchan); in ls1x_dma_irq_handler()
[all …]
A Dnbpfaxi.c603 struct nbpf_channel *chan = nbpf_to_chan(dchan); in nbpf_issue_pending()
691 struct dma_chan *dchan = &chan->dma_chan; in nbpf_desc_page_alloc() local
699 struct device *dev = dchan->device->dev; in nbpf_desc_page_alloc()
714 if (dma_mapping_error(dchan->device->dev, in nbpf_desc_page_alloc()
876 static int nbpf_pause(struct dma_chan *dchan) in nbpf_pause() argument
895 dev_dbg(dchan->device->dev, "Terminating\n"); in nbpf_terminate_all()
903 static int nbpf_config(struct dma_chan *dchan, in nbpf_config() argument
1104 struct dma_chan *dchan; in nbpf_of_xlate() local
1111 if (!dchan) in nbpf_of_xlate()
1117 chan = nbpf_to_chan(dchan); in nbpf_of_xlate()
[all …]
A Dmmp_pdma.c134 #define to_mmp_pdma_chan(dchan) \ argument
135 container_of(dchan, struct mmp_pdma_chan, chan)
443 mmp_pdma_prep_memcpy(struct dma_chan *dchan, in mmp_pdma_prep_memcpy() argument
451 if (!dchan) in mmp_pdma_prep_memcpy()
457 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_memcpy()
604 mmp_pdma_prep_dma_cyclic(struct dma_chan *dchan, in mmp_pdma_prep_dma_cyclic() argument
613 if (!dchan || !len || !period_len) in mmp_pdma_prep_dma_cyclic()
623 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_dma_cyclic()
698 if (!dchan) in mmp_pdma_config_write()
747 if (!dchan) in mmp_pdma_terminate_all()
[all …]
A Dfsldma.c373 if (!dchan) in fsl_dma_external_start()
376 chan = to_fsl_chan(dchan); in fsl_dma_external_start()
692 struct fsldma_chan *chan = to_fsl_chan(dchan); in fsl_dma_alloc_chan_resources()
745 struct fsldma_chan *chan = to_fsl_chan(dchan); in fsl_dma_free_chan_resources()
760 fsl_dma_prep_memcpy(struct dma_chan *dchan, in fsl_dma_prep_memcpy() argument
768 if (!dchan) in fsl_dma_prep_memcpy()
774 chan = to_fsl_chan(dchan); in fsl_dma_prep_memcpy()
828 if (!dchan) in fsl_dma_device_terminate_all()
831 chan = to_fsl_chan(dchan); in fsl_dma_device_terminate_all()
854 if (!dchan) in fsl_dma_device_config()
[all …]
A Dpxa_dma.c139 #define to_pxad_chan(dchan) \ argument
140 container_of(dchan, struct pxad_chan, vc.chan)
686 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_alloc_chan_resources()
709 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_free_chan_resources()
827 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_issue_pending()
916 pxad_prep_memcpy(struct dma_chan *dchan, in pxad_prep_memcpy() argument
920 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_prep_memcpy()
927 if (!dchan || !len) in pxad_prep_memcpy()
1013 pxad_prep_dma_cyclic(struct dma_chan *dchan, in pxad_prep_dma_cyclic() argument
1024 if (!dchan || !len || !period_len) in pxad_prep_dma_cyclic()
[all …]
A Daltera-msgdma.c332 msgdma_prep_memcpy(struct dma_chan *dchan, dma_addr_t dma_dst, in msgdma_prep_memcpy() argument
335 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_memcpy()
388 msgdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, in msgdma_prep_slave_sg() argument
393 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_slave_sg()
461 static int msgdma_dma_config(struct dma_chan *dchan, in msgdma_dma_config() argument
464 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_dma_config()
639 static void msgdma_free_chan_resources(struct dma_chan *dchan) in msgdma_free_chan_resources() argument
641 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_free_chan_resources()
656 static int msgdma_alloc_chan_resources(struct dma_chan *dchan) in msgdma_alloc_chan_resources() argument
658 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_alloc_chan_resources()
A Dxgene-dma.c191 #define to_dma_chan(dchan) \ argument
192 container_of(dchan, struct xgene_dma_chan, dma_chan)
777 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_alloc_chan_resources()
814 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_free_chan_resources()
839 struct dma_chan *dchan, dma_addr_t dst, dma_addr_t *src, in xgene_dma_prep_xor() argument
847 if (unlikely(!dchan || !len)) in xgene_dma_prep_xor()
850 chan = to_dma_chan(dchan); in xgene_dma_prep_xor()
896 if (unlikely(!dchan || !len)) in xgene_dma_prep_pq()
899 chan = to_dma_chan(dchan); in xgene_dma_prep_pq()
965 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_issue_pending()
[all …]
A Dbcm-sba-raid.c87 #define to_sba_device(dchan) \ argument
88 container_of(dchan, struct sba_device, dma_chan)
499 sba_cleanup_pending_requests(to_sba_device(dchan)); in sba_device_terminate_all()
507 struct sba_device *sba = to_sba_device(dchan); in sba_issue_pending()
544 struct sba_device *sba = to_sba_device(dchan); in sba_tx_status()
546 ret = dma_cookie_status(dchan, cookie, txstate); in sba_tx_status()
552 return dma_cookie_status(dchan, cookie, txstate); in sba_tx_status()
618 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_interrupt()
733 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_memcpy()
872 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_xor()
[all …]
A Dmmp_tdma.c137 #define to_mmp_tdma_chan(dchan) container_of(dchan, struct mmp_tdma_chan, chan) argument
A Dste_dma40.c2231 d40_prep_sg(struct dma_chan *dchan, struct scatterlist *sg_src, in d40_prep_sg() argument
2235 struct d40_chan *chan = container_of(dchan, struct d40_chan, chan); in d40_prep_sg()
2247 d40_set_runtime_config_write(dchan, &chan->slave_config, direction); in d40_prep_sg()
A Dat_xdmac.c322 static inline struct at_xdmac_chan *to_at_xdmac_chan(struct dma_chan *dchan) in to_at_xdmac_chan() argument
324 return container_of(dchan, struct at_xdmac_chan, chan); in to_at_xdmac_chan()
/drivers/rapidio/devices/
A Dtsi721_dma.c308 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_start_dma()
315 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_start_dma()
413 struct dma_chan *dchan = desc->txd.chan; in tsi721_submit_sg() local
588 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_dma_tasklet()
652 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_dma_tasklet()
701 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_tx_submit()
758 dma_cookie_init(dchan); in tsi721_alloc_chan_resources()
862 tsi_err(&dchan->dev->device, in tsi721_prep_rio_sg()
887 tsi_debug(DMA, &dchan->dev->device, in tsi721_prep_rio_sg()
982 bdma_chan->dchan.cookie = 1; in tsi721_register_dma()
[all …]
A Dtsi721.h701 struct dma_chan dchan; member
/drivers/dma/dw-edma/
A Ddw-edma-core.c204 static void dw_edma_device_caps(struct dma_chan *dchan, in dw_edma_device_caps() argument
207 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_caps()
225 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_config()
233 static int dw_edma_device_pause(struct dma_chan *dchan) in dw_edma_device_pause() argument
235 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_pause()
252 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_resume()
325 ret = dma_cookie_status(dchan, cookie, txstate); in dw_edma_device_tx_status()
542 xfer.dchan = dchan; in dw_edma_device_prep_slave_sg()
560 xfer.dchan = dchan; in dw_edma_device_prep_dma_cyclic()
578 xfer.dchan = dchan; in dw_edma_device_prep_interleaved_dma()
[all …]
A Ddw-edma-core.h143 struct dma_chan *dchan; member
161 struct dw_edma_chan *dchan2dw_edma_chan(struct dma_chan *dchan) in dchan2dw_edma_chan() argument
163 return vc2dw_edma_chan(to_virt_chan(dchan)); in dchan2dw_edma_chan()
/drivers/dma/sf-pdma/
A Dsf-pdma.c88 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_prep_dma_memcpy()
112 static int sf_pdma_slave_config(struct dma_chan *dchan, in sf_pdma_slave_config() argument
115 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_slave_config()
124 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_alloc_chan_resources()
127 dma_cookie_init(dchan); in sf_pdma_alloc_chan_resources()
142 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_free_chan_resources()
195 sf_pdma_tx_status(struct dma_chan *dchan, in sf_pdma_tx_status() argument
199 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_tx_status()
202 status = dma_cookie_status(dchan, cookie, txstate); in sf_pdma_tx_status()
212 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_terminate_all()
[all …]
/drivers/dma/dw-axi-dmac/
A Ddw-axi-dmac-platform.c368 status = dma_cookie_status(dchan, cookie, txstate); in dma_chan_tx_status()
549 dev_err(dchan2dev(dchan), "%s is non-idle!\n", in dma_chan_free_chan_resources()
559 dev_vdbg(dchan2dev(dchan), in dma_chan_free_chan_resources()
1195 dev_warn(dchan2dev(dchan), in dma_chan_terminate_all()
1217 static int dma_chan_pause(struct dma_chan *dchan) in dma_chan_pause() argument
1302 static int dma_chan_resume(struct dma_chan *dchan) in dma_chan_resume() argument
1365 struct dma_chan *dchan; in dw_axi_dma_of_xlate() local
1367 dchan = dma_get_any_slave_channel(&dw->dma); in dw_axi_dma_of_xlate()
1368 if (!dchan) in dw_axi_dma_of_xlate()
1371 chan = dchan_to_axi_dma_chan(dchan); in dw_axi_dma_of_xlate()
[all …]
A Ddw-axi-dmac.h121 static inline struct device *dchan2dev(struct dma_chan *dchan) in dchan2dev() argument
123 return &dchan->dev->device; in dchan2dev()
141 static inline struct axi_dma_chan *dchan_to_axi_dma_chan(struct dma_chan *dchan) in dchan_to_axi_dma_chan() argument
143 return vc_to_axi_dma_chan(to_virt_chan(dchan)); in dchan_to_axi_dma_chan()
/drivers/dma/xilinx/
A Dxilinx_dpdma.c1276 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_prep_dma_cyclic()
1293 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_prep_interleaved_dma()
1324 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_alloc_chan_resources()
1350 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_free_chan_resources()
1360 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_issue_pending()
1371 static int xilinx_dpdma_config(struct dma_chan *dchan, in xilinx_dpdma_config() argument
1374 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_config()
1402 static int xilinx_dpdma_pause(struct dma_chan *dchan) in xilinx_dpdma_pause() argument
1404 xilinx_dpdma_chan_pause(to_xilinx_chan(dchan)); in xilinx_dpdma_pause()
1409 static int xilinx_dpdma_resume(struct dma_chan *dchan) in xilinx_dpdma_resume() argument
[all …]
A Dzynqmp_dma.c477 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_alloc_chan_resources()
574 static int zynqmp_dma_device_config(struct dma_chan *dchan, in zynqmp_dma_device_config() argument
577 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_device_config()
660 static void zynqmp_dma_issue_pending(struct dma_chan *dchan) in zynqmp_dma_issue_pending() argument
662 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_issue_pending()
691 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_free_chan_resources()
804 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_device_terminate_all()
816 static void zynqmp_dma_synchronize(struct dma_chan *dchan) in zynqmp_dma_synchronize() argument
818 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_synchronize()
834 struct dma_chan *dchan, dma_addr_t dma_dst, in zynqmp_dma_prep_memcpy() argument
[all …]
A Dxilinx_dma.c933 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_free_chan_resources()
1116 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_alloc_chan_resources()
1217 dma_cookie_init(dchan); in xilinx_dma_alloc_chan_resources()
1274 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_tx_status()
1280 ret = dma_cookie_status(dchan, cookie, txstate); in xilinx_dma_tx_status()
1688 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_issue_pending()
2047 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_vdma_dma_prep_interleaved()
2133 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_cdma_prep_memcpy()
2191 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_prep_slave_sg()
2293 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_prep_dma_cyclic()
[all …]
/drivers/rapidio/
A Drio.c1781 void rio_release_dma(struct dma_chan *dchan) in rio_release_dma() argument
1783 dma_release_channel(dchan); in rio_release_dma()
1803 struct dma_async_tx_descriptor *rio_dma_prep_xfer(struct dma_chan *dchan, in rio_dma_prep_xfer() argument
1809 if (!dchan->device->device_prep_slave_sg) { in rio_dma_prep_xfer()
1819 return dmaengine_prep_rio_sg(dchan, data->sg, data->sg_len, in rio_dma_prep_xfer()

Completed in 157 milliseconds