| /drivers/dma/ |
| A D | loongson1-apb-dma.c | 86 #define to_ls1x_dma_chan(dchan) \ argument 100 struct dma_chan *dchan = &chan->vc.chan; in ls1x_dma_query() local 105 val |= dchan->chan_id; in ls1x_dma_query() 119 struct device *dev = chan2dev(dchan); in ls1x_dma_start() 124 val |= dchan->chan_id; in ls1x_dma_start() 146 struct device *dev = chan2dev(dchan); in ls1x_dma_free_chan_resources() 159 struct device *dev = chan2dev(dchan); in ls1x_dma_alloc_chan_resources() 222 struct device *dev = chan2dev(dchan); in ls1x_dma_prep_lli() 328 dev_dbg(chan2dev(dchan), in ls1x_dma_prep_dma_cyclic() 492 struct device *dev = chan2dev(dchan); in ls1x_dma_irq_handler() [all …]
|
| A D | nbpfaxi.c | 603 struct nbpf_channel *chan = nbpf_to_chan(dchan); in nbpf_issue_pending() 691 struct dma_chan *dchan = &chan->dma_chan; in nbpf_desc_page_alloc() local 699 struct device *dev = dchan->device->dev; in nbpf_desc_page_alloc() 714 if (dma_mapping_error(dchan->device->dev, in nbpf_desc_page_alloc() 876 static int nbpf_pause(struct dma_chan *dchan) in nbpf_pause() argument 895 dev_dbg(dchan->device->dev, "Terminating\n"); in nbpf_terminate_all() 903 static int nbpf_config(struct dma_chan *dchan, in nbpf_config() argument 1104 struct dma_chan *dchan; in nbpf_of_xlate() local 1111 if (!dchan) in nbpf_of_xlate() 1117 chan = nbpf_to_chan(dchan); in nbpf_of_xlate() [all …]
|
| A D | mmp_pdma.c | 134 #define to_mmp_pdma_chan(dchan) \ argument 135 container_of(dchan, struct mmp_pdma_chan, chan) 443 mmp_pdma_prep_memcpy(struct dma_chan *dchan, in mmp_pdma_prep_memcpy() argument 451 if (!dchan) in mmp_pdma_prep_memcpy() 457 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_memcpy() 604 mmp_pdma_prep_dma_cyclic(struct dma_chan *dchan, in mmp_pdma_prep_dma_cyclic() argument 613 if (!dchan || !len || !period_len) in mmp_pdma_prep_dma_cyclic() 623 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_dma_cyclic() 698 if (!dchan) in mmp_pdma_config_write() 747 if (!dchan) in mmp_pdma_terminate_all() [all …]
|
| A D | fsldma.c | 373 if (!dchan) in fsl_dma_external_start() 376 chan = to_fsl_chan(dchan); in fsl_dma_external_start() 692 struct fsldma_chan *chan = to_fsl_chan(dchan); in fsl_dma_alloc_chan_resources() 745 struct fsldma_chan *chan = to_fsl_chan(dchan); in fsl_dma_free_chan_resources() 760 fsl_dma_prep_memcpy(struct dma_chan *dchan, in fsl_dma_prep_memcpy() argument 768 if (!dchan) in fsl_dma_prep_memcpy() 774 chan = to_fsl_chan(dchan); in fsl_dma_prep_memcpy() 828 if (!dchan) in fsl_dma_device_terminate_all() 831 chan = to_fsl_chan(dchan); in fsl_dma_device_terminate_all() 854 if (!dchan) in fsl_dma_device_config() [all …]
|
| A D | pxa_dma.c | 139 #define to_pxad_chan(dchan) \ argument 140 container_of(dchan, struct pxad_chan, vc.chan) 686 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_alloc_chan_resources() 709 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_free_chan_resources() 827 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_issue_pending() 916 pxad_prep_memcpy(struct dma_chan *dchan, in pxad_prep_memcpy() argument 920 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_prep_memcpy() 927 if (!dchan || !len) in pxad_prep_memcpy() 1013 pxad_prep_dma_cyclic(struct dma_chan *dchan, in pxad_prep_dma_cyclic() argument 1024 if (!dchan || !len || !period_len) in pxad_prep_dma_cyclic() [all …]
|
| A D | altera-msgdma.c | 332 msgdma_prep_memcpy(struct dma_chan *dchan, dma_addr_t dma_dst, in msgdma_prep_memcpy() argument 335 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_memcpy() 388 msgdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, in msgdma_prep_slave_sg() argument 393 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_slave_sg() 461 static int msgdma_dma_config(struct dma_chan *dchan, in msgdma_dma_config() argument 464 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_dma_config() 639 static void msgdma_free_chan_resources(struct dma_chan *dchan) in msgdma_free_chan_resources() argument 641 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_free_chan_resources() 656 static int msgdma_alloc_chan_resources(struct dma_chan *dchan) in msgdma_alloc_chan_resources() argument 658 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_alloc_chan_resources()
|
| A D | xgene-dma.c | 191 #define to_dma_chan(dchan) \ argument 192 container_of(dchan, struct xgene_dma_chan, dma_chan) 777 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_alloc_chan_resources() 814 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_free_chan_resources() 839 struct dma_chan *dchan, dma_addr_t dst, dma_addr_t *src, in xgene_dma_prep_xor() argument 847 if (unlikely(!dchan || !len)) in xgene_dma_prep_xor() 850 chan = to_dma_chan(dchan); in xgene_dma_prep_xor() 896 if (unlikely(!dchan || !len)) in xgene_dma_prep_pq() 899 chan = to_dma_chan(dchan); in xgene_dma_prep_pq() 965 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_issue_pending() [all …]
|
| A D | bcm-sba-raid.c | 87 #define to_sba_device(dchan) \ argument 88 container_of(dchan, struct sba_device, dma_chan) 499 sba_cleanup_pending_requests(to_sba_device(dchan)); in sba_device_terminate_all() 507 struct sba_device *sba = to_sba_device(dchan); in sba_issue_pending() 544 struct sba_device *sba = to_sba_device(dchan); in sba_tx_status() 546 ret = dma_cookie_status(dchan, cookie, txstate); in sba_tx_status() 552 return dma_cookie_status(dchan, cookie, txstate); in sba_tx_status() 618 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_interrupt() 733 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_memcpy() 872 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_xor() [all …]
|
| A D | mmp_tdma.c | 137 #define to_mmp_tdma_chan(dchan) container_of(dchan, struct mmp_tdma_chan, chan) argument
|
| A D | ste_dma40.c | 2231 d40_prep_sg(struct dma_chan *dchan, struct scatterlist *sg_src, in d40_prep_sg() argument 2235 struct d40_chan *chan = container_of(dchan, struct d40_chan, chan); in d40_prep_sg() 2247 d40_set_runtime_config_write(dchan, &chan->slave_config, direction); in d40_prep_sg()
|
| A D | at_xdmac.c | 322 static inline struct at_xdmac_chan *to_at_xdmac_chan(struct dma_chan *dchan) in to_at_xdmac_chan() argument 324 return container_of(dchan, struct at_xdmac_chan, chan); in to_at_xdmac_chan()
|
| /drivers/rapidio/devices/ |
| A D | tsi721_dma.c | 308 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_start_dma() 315 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_start_dma() 413 struct dma_chan *dchan = desc->txd.chan; in tsi721_submit_sg() local 588 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_dma_tasklet() 652 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_dma_tasklet() 701 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_tx_submit() 758 dma_cookie_init(dchan); in tsi721_alloc_chan_resources() 862 tsi_err(&dchan->dev->device, in tsi721_prep_rio_sg() 887 tsi_debug(DMA, &dchan->dev->device, in tsi721_prep_rio_sg() 982 bdma_chan->dchan.cookie = 1; in tsi721_register_dma() [all …]
|
| A D | tsi721.h | 701 struct dma_chan dchan; member
|
| /drivers/dma/dw-edma/ |
| A D | dw-edma-core.c | 204 static void dw_edma_device_caps(struct dma_chan *dchan, in dw_edma_device_caps() argument 207 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_caps() 225 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_config() 233 static int dw_edma_device_pause(struct dma_chan *dchan) in dw_edma_device_pause() argument 235 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_pause() 252 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_resume() 325 ret = dma_cookie_status(dchan, cookie, txstate); in dw_edma_device_tx_status() 542 xfer.dchan = dchan; in dw_edma_device_prep_slave_sg() 560 xfer.dchan = dchan; in dw_edma_device_prep_dma_cyclic() 578 xfer.dchan = dchan; in dw_edma_device_prep_interleaved_dma() [all …]
|
| A D | dw-edma-core.h | 143 struct dma_chan *dchan; member 161 struct dw_edma_chan *dchan2dw_edma_chan(struct dma_chan *dchan) in dchan2dw_edma_chan() argument 163 return vc2dw_edma_chan(to_virt_chan(dchan)); in dchan2dw_edma_chan()
|
| /drivers/dma/sf-pdma/ |
| A D | sf-pdma.c | 88 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_prep_dma_memcpy() 112 static int sf_pdma_slave_config(struct dma_chan *dchan, in sf_pdma_slave_config() argument 115 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_slave_config() 124 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_alloc_chan_resources() 127 dma_cookie_init(dchan); in sf_pdma_alloc_chan_resources() 142 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_free_chan_resources() 195 sf_pdma_tx_status(struct dma_chan *dchan, in sf_pdma_tx_status() argument 199 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_tx_status() 202 status = dma_cookie_status(dchan, cookie, txstate); in sf_pdma_tx_status() 212 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_terminate_all() [all …]
|
| /drivers/dma/dw-axi-dmac/ |
| A D | dw-axi-dmac-platform.c | 368 status = dma_cookie_status(dchan, cookie, txstate); in dma_chan_tx_status() 549 dev_err(dchan2dev(dchan), "%s is non-idle!\n", in dma_chan_free_chan_resources() 559 dev_vdbg(dchan2dev(dchan), in dma_chan_free_chan_resources() 1195 dev_warn(dchan2dev(dchan), in dma_chan_terminate_all() 1217 static int dma_chan_pause(struct dma_chan *dchan) in dma_chan_pause() argument 1302 static int dma_chan_resume(struct dma_chan *dchan) in dma_chan_resume() argument 1365 struct dma_chan *dchan; in dw_axi_dma_of_xlate() local 1367 dchan = dma_get_any_slave_channel(&dw->dma); in dw_axi_dma_of_xlate() 1368 if (!dchan) in dw_axi_dma_of_xlate() 1371 chan = dchan_to_axi_dma_chan(dchan); in dw_axi_dma_of_xlate() [all …]
|
| A D | dw-axi-dmac.h | 121 static inline struct device *dchan2dev(struct dma_chan *dchan) in dchan2dev() argument 123 return &dchan->dev->device; in dchan2dev() 141 static inline struct axi_dma_chan *dchan_to_axi_dma_chan(struct dma_chan *dchan) in dchan_to_axi_dma_chan() argument 143 return vc_to_axi_dma_chan(to_virt_chan(dchan)); in dchan_to_axi_dma_chan()
|
| /drivers/dma/xilinx/ |
| A D | xilinx_dpdma.c | 1276 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_prep_dma_cyclic() 1293 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_prep_interleaved_dma() 1324 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_alloc_chan_resources() 1350 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_free_chan_resources() 1360 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_issue_pending() 1371 static int xilinx_dpdma_config(struct dma_chan *dchan, in xilinx_dpdma_config() argument 1374 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_config() 1402 static int xilinx_dpdma_pause(struct dma_chan *dchan) in xilinx_dpdma_pause() argument 1404 xilinx_dpdma_chan_pause(to_xilinx_chan(dchan)); in xilinx_dpdma_pause() 1409 static int xilinx_dpdma_resume(struct dma_chan *dchan) in xilinx_dpdma_resume() argument [all …]
|
| A D | zynqmp_dma.c | 477 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_alloc_chan_resources() 574 static int zynqmp_dma_device_config(struct dma_chan *dchan, in zynqmp_dma_device_config() argument 577 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_device_config() 660 static void zynqmp_dma_issue_pending(struct dma_chan *dchan) in zynqmp_dma_issue_pending() argument 662 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_issue_pending() 691 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_free_chan_resources() 804 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_device_terminate_all() 816 static void zynqmp_dma_synchronize(struct dma_chan *dchan) in zynqmp_dma_synchronize() argument 818 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_synchronize() 834 struct dma_chan *dchan, dma_addr_t dma_dst, in zynqmp_dma_prep_memcpy() argument [all …]
|
| A D | xilinx_dma.c | 933 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_free_chan_resources() 1116 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_alloc_chan_resources() 1217 dma_cookie_init(dchan); in xilinx_dma_alloc_chan_resources() 1274 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_tx_status() 1280 ret = dma_cookie_status(dchan, cookie, txstate); in xilinx_dma_tx_status() 1688 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_issue_pending() 2047 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_vdma_dma_prep_interleaved() 2133 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_cdma_prep_memcpy() 2191 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_prep_slave_sg() 2293 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_prep_dma_cyclic() [all …]
|
| /drivers/rapidio/ |
| A D | rio.c | 1781 void rio_release_dma(struct dma_chan *dchan) in rio_release_dma() argument 1783 dma_release_channel(dchan); in rio_release_dma() 1803 struct dma_async_tx_descriptor *rio_dma_prep_xfer(struct dma_chan *dchan, in rio_dma_prep_xfer() argument 1809 if (!dchan->device->device_prep_slave_sg) { in rio_dma_prep_xfer() 1819 return dmaengine_prep_rio_sg(dchan, data->sg, data->sg_len, in rio_dma_prep_xfer()
|