| /linux/drivers/rapidio/devices/ |
| A D | tsi721_dma.c | 306 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_start_dma() 313 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_start_dma() 411 struct dma_chan *dchan = desc->txd.chan; in tsi721_submit_sg() local 586 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_dma_tasklet() 650 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_dma_tasklet() 699 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_tx_submit() 756 dma_cookie_init(dchan); in tsi721_alloc_chan_resources() 860 tsi_err(&dchan->dev->device, in tsi721_prep_rio_sg() 885 tsi_debug(DMA, &dchan->dev->device, in tsi721_prep_rio_sg() 980 bdma_chan->dchan.cookie = 1; in tsi721_register_dma() [all …]
|
| A D | tsi721.h | 701 struct dma_chan dchan; member
|
| /linux/drivers/dma/ |
| A D | nbpfaxi.c | 604 struct nbpf_channel *chan = nbpf_to_chan(dchan); in nbpf_issue_pending() 629 struct nbpf_channel *chan = nbpf_to_chan(dchan); in nbpf_tx_status() 692 struct dma_chan *dchan = &chan->dma_chan; in nbpf_desc_page_alloc() local 700 struct device *dev = dchan->device->dev; in nbpf_desc_page_alloc() 864 static int nbpf_pause(struct dma_chan *dchan) in nbpf_pause() argument 883 dev_dbg(dchan->device->dev, "Terminating\n"); in nbpf_terminate_all() 891 static int nbpf_config(struct dma_chan *dchan, in nbpf_config() argument 1092 struct dma_chan *dchan; in nbpf_of_xlate() local 1099 if (!dchan) in nbpf_of_xlate() 1105 chan = nbpf_to_chan(dchan); in nbpf_of_xlate() [all …]
|
| A D | mmp_pdma.c | 135 #define to_mmp_pdma_chan(dchan) \ argument 136 container_of(dchan, struct mmp_pdma_chan, chan) 444 mmp_pdma_prep_memcpy(struct dma_chan *dchan, in mmp_pdma_prep_memcpy() argument 452 if (!dchan) in mmp_pdma_prep_memcpy() 458 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_memcpy() 605 mmp_pdma_prep_dma_cyclic(struct dma_chan *dchan, in mmp_pdma_prep_dma_cyclic() argument 614 if (!dchan || !len || !period_len) in mmp_pdma_prep_dma_cyclic() 624 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_dma_cyclic() 699 if (!dchan) in mmp_pdma_config_write() 754 if (!dchan) in mmp_pdma_terminate_all() [all …]
|
| A D | fsldma.c | 372 if (!dchan) in fsl_dma_external_start() 375 chan = to_fsl_chan(dchan); in fsl_dma_external_start() 691 struct fsldma_chan *chan = to_fsl_chan(dchan); in fsl_dma_alloc_chan_resources() 744 struct fsldma_chan *chan = to_fsl_chan(dchan); in fsl_dma_free_chan_resources() 759 fsl_dma_prep_memcpy(struct dma_chan *dchan, in fsl_dma_prep_memcpy() argument 767 if (!dchan) in fsl_dma_prep_memcpy() 773 chan = to_fsl_chan(dchan); in fsl_dma_prep_memcpy() 827 if (!dchan) in fsl_dma_device_terminate_all() 830 chan = to_fsl_chan(dchan); in fsl_dma_device_terminate_all() 853 if (!dchan) in fsl_dma_device_config() [all …]
|
| A D | pxa_dma.c | 138 #define to_pxad_chan(dchan) \ argument 139 container_of(dchan, struct pxad_chan, vc.chan) 686 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_alloc_chan_resources() 709 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_free_chan_resources() 826 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_issue_pending() 922 pxad_prep_memcpy(struct dma_chan *dchan, in pxad_prep_memcpy() argument 926 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_prep_memcpy() 933 if (!dchan || !len) in pxad_prep_memcpy() 1019 pxad_prep_dma_cyclic(struct dma_chan *dchan, in pxad_prep_dma_cyclic() argument 1030 if (!dchan || !len || !period_len) in pxad_prep_dma_cyclic() [all …]
|
| A D | altera-msgdma.c | 332 msgdma_prep_memcpy(struct dma_chan *dchan, dma_addr_t dma_dst, in msgdma_prep_memcpy() argument 335 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_memcpy() 388 msgdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, in msgdma_prep_slave_sg() argument 393 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_slave_sg() 461 static int msgdma_dma_config(struct dma_chan *dchan, in msgdma_dma_config() argument 464 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_dma_config() 636 static void msgdma_free_chan_resources(struct dma_chan *dchan) in msgdma_free_chan_resources() argument 638 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_free_chan_resources() 653 static int msgdma_alloc_chan_resources(struct dma_chan *dchan) in msgdma_alloc_chan_resources() argument 655 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_alloc_chan_resources()
|
| A D | xgene-dma.c | 190 #define to_dma_chan(dchan) \ argument 191 container_of(dchan, struct xgene_dma_chan, dma_chan) 776 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_alloc_chan_resources() 813 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_free_chan_resources() 838 struct dma_chan *dchan, dma_addr_t dst, dma_addr_t *src, in xgene_dma_prep_xor() argument 846 if (unlikely(!dchan || !len)) in xgene_dma_prep_xor() 849 chan = to_dma_chan(dchan); in xgene_dma_prep_xor() 895 if (unlikely(!dchan || !len)) in xgene_dma_prep_pq() 898 chan = to_dma_chan(dchan); in xgene_dma_prep_pq() 964 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_issue_pending() [all …]
|
| A D | bcm-sba-raid.c | 95 #define to_sba_device(dchan) \ argument 96 container_of(dchan, struct sba_device, dma_chan) 507 sba_cleanup_pending_requests(to_sba_device(dchan)); in sba_device_terminate_all() 515 struct sba_device *sba = to_sba_device(dchan); in sba_issue_pending() 552 struct sba_device *sba = to_sba_device(dchan); in sba_tx_status() 554 ret = dma_cookie_status(dchan, cookie, txstate); in sba_tx_status() 560 return dma_cookie_status(dchan, cookie, txstate); in sba_tx_status() 626 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_interrupt() 741 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_memcpy() 880 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_xor() [all …]
|
| A D | at_hdmac_regs.h | 296 static inline struct at_dma_chan *to_at_dma_chan(struct dma_chan *dchan) in to_at_dma_chan() argument 298 return container_of(dchan, struct at_dma_chan, chan_common); in to_at_dma_chan()
|
| A D | mmp_tdma.c | 137 #define to_mmp_tdma_chan(dchan) container_of(dchan, struct mmp_tdma_chan, chan) argument
|
| /linux/drivers/dma/dw-edma/ |
| A D | dw-edma-core.c | 26 struct device *dchan2dev(struct dma_chan *dchan) in dchan2dev() argument 28 return &dchan->dev->device; in dchan2dev() 204 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_config() 212 static int dw_edma_device_pause(struct dma_chan *dchan) in dw_edma_device_pause() argument 214 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_pause() 231 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_resume() 301 ret = dma_cookie_status(dchan, cookie, txstate); in dw_edma_device_tx_status() 494 xfer.dchan = dchan; in dw_edma_device_prep_slave_sg() 512 xfer.dchan = dchan; in dw_edma_device_prep_dma_cyclic() 530 xfer.dchan = dchan; in dw_edma_device_prep_interleaved_dma() [all …]
|
| A D | dw-edma-core.h | 157 struct dma_chan *dchan; member 175 struct dw_edma_chan *dchan2dw_edma_chan(struct dma_chan *dchan) in dchan2dw_edma_chan() argument 177 return vc2dw_edma_chan(to_virt_chan(dchan)); in dchan2dw_edma_chan()
|
| /linux/drivers/dma/sf-pdma/ |
| A D | sf-pdma.c | 95 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_prep_dma_memcpy() 121 static int sf_pdma_slave_config(struct dma_chan *dchan, in sf_pdma_slave_config() argument 124 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_slave_config() 133 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_alloc_chan_resources() 136 dma_cookie_init(dchan); in sf_pdma_alloc_chan_resources() 151 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_free_chan_resources() 198 sf_pdma_tx_status(struct dma_chan *dchan, in sf_pdma_tx_status() argument 202 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_tx_status() 205 status = dma_cookie_status(dchan, cookie, txstate); in sf_pdma_tx_status() 215 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_terminate_all() [all …]
|
| /linux/drivers/dma/dw-axi-dmac/ |
| A D | dw-axi-dmac-platform.c | 315 status = dma_cookie_status(dchan, cookie, txstate); in dma_chan_tx_status() 498 dev_err(dchan2dev(dchan), "%s is non-idle!\n", in dma_chan_free_chan_resources() 508 dev_vdbg(dchan2dev(dchan), in dma_chan_free_chan_resources() 1130 dev_warn(dchan2dev(dchan), in dma_chan_terminate_all() 1152 static int dma_chan_pause(struct dma_chan *dchan) in dma_chan_pause() argument 1207 static int dma_chan_resume(struct dma_chan *dchan) in dma_chan_resume() argument 1270 struct dma_chan *dchan; in dw_axi_dma_of_xlate() local 1272 dchan = dma_get_any_slave_channel(&dw->dma); in dw_axi_dma_of_xlate() 1273 if (!dchan) in dw_axi_dma_of_xlate() 1276 chan = dchan_to_axi_dma_chan(dchan); in dw_axi_dma_of_xlate() [all …]
|
| A D | dw-axi-dmac.h | 119 static inline struct device *dchan2dev(struct dma_chan *dchan) in dchan2dev() argument 121 return &dchan->dev->device; in dchan2dev() 139 static inline struct axi_dma_chan *dchan_to_axi_dma_chan(struct dma_chan *dchan) in dchan_to_axi_dma_chan() argument 141 return vc_to_axi_dma_chan(to_virt_chan(dchan)); in dchan_to_axi_dma_chan()
|
| /linux/drivers/dma/xilinx/ |
| A D | xilinx_dpdma.c | 1196 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_prep_interleaved_dma() 1227 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_alloc_chan_resources() 1253 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_free_chan_resources() 1263 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_issue_pending() 1272 static int xilinx_dpdma_config(struct dma_chan *dchan, in xilinx_dpdma_config() argument 1275 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_config() 1299 static int xilinx_dpdma_pause(struct dma_chan *dchan) in xilinx_dpdma_pause() argument 1301 xilinx_dpdma_chan_pause(to_xilinx_chan(dchan)); in xilinx_dpdma_pause() 1306 static int xilinx_dpdma_resume(struct dma_chan *dchan) in xilinx_dpdma_resume() argument 1308 xilinx_dpdma_chan_unpause(to_xilinx_chan(dchan)); in xilinx_dpdma_resume() [all …]
|
| A D | zynqmp_dma.c | 461 static int zynqmp_dma_alloc_chan_resources(struct dma_chan *dchan) in zynqmp_dma_alloc_chan_resources() argument 463 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_alloc_chan_resources() 559 static int zynqmp_dma_device_config(struct dma_chan *dchan, in zynqmp_dma_device_config() argument 562 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_device_config() 645 static void zynqmp_dma_issue_pending(struct dma_chan *dchan) in zynqmp_dma_issue_pending() argument 647 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_issue_pending() 674 static void zynqmp_dma_free_chan_resources(struct dma_chan *dchan) in zynqmp_dma_free_chan_resources() argument 676 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_free_chan_resources() 789 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_device_terminate_all() 808 struct dma_chan *dchan, dma_addr_t dma_dst, in zynqmp_dma_prep_memcpy() argument [all …]
|
| A D | xilinx_dma.c | 900 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_free_chan_resources() 1083 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_alloc_chan_resources() 1184 dma_cookie_init(dchan); in xilinx_dma_alloc_chan_resources() 1241 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_tx_status() 1247 ret = dma_cookie_status(dchan, cookie, txstate); in xilinx_dma_tx_status() 1650 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_issue_pending() 2006 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_vdma_dma_prep_interleaved() 2088 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_cdma_prep_memcpy() 2146 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_prep_slave_sg() 2245 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_prep_dma_cyclic() [all …]
|
| /linux/include/linux/ |
| A D | rio_drv.h | 396 extern void rio_release_dma(struct dma_chan *dchan); 398 struct rio_dev *rdev, struct dma_chan *dchan, 402 struct dma_chan *dchan, u16 destid,
|
| A D | fsldma.h | 8 int fsl_dma_external_start(struct dma_chan *dchan, int enable);
|
| /linux/include/linux/dma/ |
| A D | xilinx_dma.h | 42 int xilinx_vdma_channel_set_config(struct dma_chan *dchan,
|
| /linux/arch/mips/alchemy/common/ |
| A D | irq.c | 581 void au1300_set_dbdma_gpio(int dchan, unsigned int gpio) in au1300_set_dbdma_gpio() argument 585 if ((dchan >= 0) && (dchan <= 1)) { in au1300_set_dbdma_gpio() 587 r &= ~(0xff << (8 * dchan)); in au1300_set_dbdma_gpio() 588 r |= (gpio & 0x7f) << (8 * dchan); in au1300_set_dbdma_gpio()
|
| /linux/drivers/rapidio/ |
| A D | rio.c | 1794 void rio_release_dma(struct dma_chan *dchan) in rio_release_dma() argument 1796 dma_release_channel(dchan); in rio_release_dma() 1816 struct dma_async_tx_descriptor *rio_dma_prep_xfer(struct dma_chan *dchan, in rio_dma_prep_xfer() argument 1822 if (!dchan->device->device_prep_slave_sg) { in rio_dma_prep_xfer() 1832 return dmaengine_prep_rio_sg(dchan, data->sg, data->sg_len, in rio_dma_prep_xfer() 1854 struct dma_chan *dchan, struct rio_dma_data *data, in rio_dma_prep_slave_sg() argument 1857 return rio_dma_prep_xfer(dchan, rdev->destid, data, direction, flags); in rio_dma_prep_slave_sg()
|
| /linux/arch/mips/include/asm/mach-au1x00/ |
| A D | au1000.h | 879 extern void au1300_set_dbdma_gpio(int dchan, unsigned int gpio);
|