| /drivers/tty/serial/ |
| A D | sprd_serial.c | 135 struct sprd_uart_dma rx_dma; member 202 if (sp->rx_dma.enable) in sprd_stop_rx() 378 if (!sp->rx_dma.virt) in sprd_rx_alloc_buf() 386 if (sp->rx_dma.virt) in sprd_rx_free_buf() 388 sp->rx_dma.virt, sp->rx_dma.phys_addr); in sprd_rx_free_buf() 389 sp->rx_dma.virt = NULL; in sprd_rx_free_buf() 483 if (!sp->rx_dma.enable) in sprd_start_dma_rx() 504 if (sp->rx_dma.enable) in sprd_release_dma() 520 sp->rx_dma.enable = true; in sprd_request_dma() 622 if (sp->rx_dma.enable) { in sprd_rx() [all …]
|
| A D | msm_serial.c | 187 struct msm_dma rx_dma; member 298 dma = &msm_port->rx_dma; in msm_release_dma() 365 dma = &msm_port->rx_dma; in msm_request_rx_dma() 564 struct msm_dma *dma = &msm_port->rx_dma; in msm_complete_rx_dma() 622 struct msm_dma *dma = &msm_port->rx_dma; in msm_start_rx_dma() 709 struct msm_dma *dma = &msm_port->rx_dma; in msm_stop_rx() 956 struct msm_dma *dma = &msm_port->rx_dma; in msm_uart_irq() 1274 struct msm_dma *dma = &msm_port->rx_dma; in msm_set_termios()
|
| /drivers/spi/ |
| A D | spi-fsl-cpm.c | 82 if (mspi->rx_dma == mspi->dma_dummy_rx) in fsl_spi_cpm_bufs_start() 83 iowrite32be(mspi->rx_dma, &rx_bd->cbd_bufaddr); in fsl_spi_cpm_bufs_start() 85 iowrite32be(mspi->rx_dma + xfer_ofs, &rx_bd->cbd_bufaddr); in fsl_spi_cpm_bufs_start() 115 mspi->rx_dma = mspi->dma_dummy_rx; in fsl_spi_cpm_bufs() 148 mspi->rx_dma = dma_map_single(dev, mspi->rx, t->len, in fsl_spi_cpm_bufs() 150 if (dma_mapping_error(dev, mspi->rx_dma)) { in fsl_spi_cpm_bufs() 155 mspi->rx_dma = t->rx_dma; in fsl_spi_cpm_bufs() 184 dma_unmap_single(dev, mspi->rx_dma, t->len, DMA_FROM_DEVICE); in fsl_spi_cpm_bufs_complete()
|
| A D | spi-s3c64xx.c | 220 struct s3c64xx_spi_dma_data rx_dma; member 389 if (IS_ERR(sdd->rx_dma.ch)) { in s3c64xx_spi_prepare_transfer() 391 sdd->rx_dma.ch = NULL; in s3c64xx_spi_prepare_transfer() 398 dma_release_channel(sdd->rx_dma.ch); in s3c64xx_spi_prepare_transfer() 400 sdd->rx_dma.ch = NULL; in s3c64xx_spi_prepare_transfer() 404 spi->dma_rx = sdd->rx_dma.ch; in s3c64xx_spi_prepare_transfer() 419 dma_release_channel(sdd->rx_dma.ch); in s3c64xx_spi_unprepare_transfer() 421 sdd->rx_dma.ch = NULL; in s3c64xx_spi_unprepare_transfer() 434 if (sdd->rx_dma.ch && sdd->tx_dma.ch) in s3c64xx_spi_can_dma() 916 dmaengine_pause(sdd->rx_dma.ch); in s3c64xx_spi_transfer_one() [all …]
|
| A D | spi-atmel.c | 833 xfer->rx_buf, (unsigned long long)xfer->rx_dma); in atmel_spi_next_xfer_dma_submit() 864 dma_addr_t *rx_dma, in atmel_spi_next_xfer_data() argument 867 *rx_dma = xfer->rx_dma + xfer->len - *plen; in atmel_spi_next_xfer_data() 930 dma_addr_t tx_dma, rx_dma; in atmel_spi_pdc_next_xfer() local 938 spi_writel(as, RPR, rx_dma); in atmel_spi_pdc_next_xfer() 950 (unsigned long long)xfer->rx_dma); in atmel_spi_pdc_next_xfer() 957 spi_writel(as, RNPR, rx_dma); in atmel_spi_pdc_next_xfer() 969 (unsigned long long)xfer->rx_dma); in atmel_spi_pdc_next_xfer() 1009 xfer->rx_dma = dma_map_single(dev, in atmel_spi_dma_map_xfer() 1012 if (dma_mapping_error(dev, xfer->rx_dma)) { in atmel_spi_dma_map_xfer() [all …]
|
| A D | spi-meson-spicc.c | 200 dma_addr_t rx_dma; member 245 t->rx_dma = dma_map_single(dev, t->rx_buf, t->len, DMA_FROM_DEVICE); in meson_spicc_dma_map() 246 if (dma_mapping_error(dev, t->rx_dma)) in meson_spicc_dma_map() 250 spicc->rx_dma = t->rx_dma; in meson_spicc_dma_map() 262 if (t->rx_dma) in meson_spicc_dma_unmap() 263 dma_unmap_single(dev, t->rx_dma, t->len, DMA_FROM_DEVICE); in meson_spicc_dma_unmap() 315 writel_relaxed(spicc->rx_dma, spicc->base + SPICC_DWADDR); in meson_spicc_setup_dma() 338 if (spicc->rx_dma) { in meson_spicc_setup_dma() 339 spicc->rx_dma += len; in meson_spicc_setup_dma()
|
| A D | spi-mt65xx.c | 177 dma_addr_t rx_dma; member 638 writel((u32)(xfer->rx_dma & MTK_SPI_32BITS_MASK), in mtk_spi_setup_dma_addr() 642 writel((u32)(xfer->rx_dma >> 32), in mtk_spi_setup_dma_addr() 711 xfer->rx_dma = sg_dma_address(mdata->rx_sgl); in mtk_spi_dma_transfer() 824 xfer->rx_dma += mdata->xfer_len; in mtk_spi_interrupt_thread() 836 xfer->rx_dma = sg_dma_address(mdata->rx_sgl); in mtk_spi_interrupt_thread() 935 writel((u32)(mdata->rx_dma & MTK_SPI_32BITS_MASK), in mtk_spi_mem_setup_dma_xfer() 939 writel((u32)(mdata->rx_dma >> 32), in mtk_spi_mem_setup_dma_xfer() 1090 mdata->rx_dma = dma_map_single(mdata->dev, in mtk_spi_mem_exec_op() 1094 if (dma_mapping_error(mdata->dev, mdata->rx_dma)) { in mtk_spi_mem_exec_op() [all …]
|
| A D | spi-slave-mt27xx.c | 224 xfer->rx_dma = dma_map_single(dev, xfer->rx_buf, in mtk_spi_slave_dma_transfer() 226 if (dma_mapping_error(dev, xfer->rx_dma)) { in mtk_spi_slave_dma_transfer() 233 writel(xfer->rx_dma, mdata->base + SPIS_RX_DST_REG); in mtk_spi_slave_dma_transfer() 266 dma_unmap_single(dev, xfer->rx_dma, in mtk_spi_slave_dma_transfer() 349 dma_unmap_single(mdata->dev, trans->rx_dma, in mtk_spi_slave_interrupt()
|
| A D | spi-sunplus-sp7021.c | 149 writel(xfer->rx_dma, pspim->s_base + SP7021_SLAVE_DMA_ADDR_REG); in sp7021_spi_target_rx() 377 xfer->rx_dma = dma_map_single(dev, xfer->rx_buf, xfer->len, in sp7021_spi_target_transfer_one() 379 if (dma_mapping_error(dev, xfer->rx_dma)) in sp7021_spi_target_transfer_one() 382 dma_unmap_single(dev, xfer->rx_dma, xfer->len, DMA_FROM_DEVICE); in sp7021_spi_target_transfer_one()
|
| A D | spi-fsl-lib.h | 39 dma_addr_t rx_dma; member
|
| A D | spi-amlogic-spisg.c | 280 if (xfer->rx_buf || xfer->rx_dma) { in aml_spisg_setup_transfer() 357 } else if (xfer->rx_buf || xfer->rx_dma) { in aml_spisg_setup_transfer() 358 paddr = xfer->rx_dma; in aml_spisg_setup_transfer() 403 } else if (!xfer->rx_dma) { in aml_spisg_cleanup_transfer()
|
| A D | spi-tegra210-quad.c | 516 rx_dma_phys = t->rx_dma; in tegra_qspi_start_rx_dma() 583 t->rx_dma = dma_map_single(tqspi->dev, (void *)rx_buf, len, DMA_FROM_DEVICE); in tegra_qspi_dma_map_xfer() 584 if (dma_mapping_error(tqspi->dev, t->rx_dma)) { in tegra_qspi_dma_map_xfer() 602 dma_unmap_single(tqspi->dev, t->rx_dma, len, DMA_FROM_DEVICE); in tegra_qspi_dma_unmap_xfer() 701 rx_dma_phys = t->rx_dma; in tegra_qspi_start_dma_based_transfer()
|
| /drivers/iio/adc/ |
| A D | ad7606_spi.c | 279 struct dma_chan *rx_dma; in ad7606_spi_offload_probe() local 313 rx_dma = devm_spi_offload_rx_stream_request_dma_chan(dev, in ad7606_spi_offload_probe() 315 if (IS_ERR(rx_dma)) in ad7606_spi_offload_probe() 316 return dev_err_probe(dev, PTR_ERR(rx_dma), in ad7606_spi_offload_probe() 320 rx_dma, IIO_BUFFER_DIRECTION_IN); in ad7606_spi_offload_probe()
|
| A D | ad7944.c | 810 struct dma_chan *rx_dma; in ad7944_probe() local 831 rx_dma = devm_spi_offload_rx_stream_request_dma_chan(dev, in ad7944_probe() 833 if (IS_ERR(rx_dma)) in ad7944_probe() 834 return dev_err_probe(dev, PTR_ERR(rx_dma), in ad7944_probe() 846 indio_dev, rx_dma, IIO_BUFFER_DIRECTION_IN); in ad7944_probe()
|
| A D | ad_sigma_delta.c | 812 struct dma_chan *rx_dma; in devm_ad_sd_setup_buffer_and_trigger() local 814 rx_dma = devm_spi_offload_rx_stream_request_dma_chan(dev, in devm_ad_sd_setup_buffer_and_trigger() 816 if (IS_ERR(rx_dma)) in devm_ad_sd_setup_buffer_and_trigger() 817 return dev_err_probe(dev, PTR_ERR(rx_dma), in devm_ad_sd_setup_buffer_and_trigger() 821 rx_dma, IIO_BUFFER_DIRECTION_IN); in devm_ad_sd_setup_buffer_and_trigger()
|
| A D | ad4000.c | 854 struct dma_chan *rx_dma; in ad4000_spi_offload_setup() local 868 rx_dma = devm_spi_offload_rx_stream_request_dma_chan(dev, st->offload); in ad4000_spi_offload_setup() 869 if (IS_ERR(rx_dma)) in ad4000_spi_offload_setup() 870 return dev_err_probe(dev, PTR_ERR(rx_dma), in ad4000_spi_offload_setup() 873 ret = devm_iio_dmaengine_buffer_setup_with_handle(dev, indio_dev, rx_dma, in ad4000_spi_offload_setup()
|
| A D | ad4695.c | 1696 struct dma_chan *rx_dma; in ad4695_probe_spi_offload() local 1743 rx_dma = devm_spi_offload_rx_stream_request_dma_chan(dev, st->offload); in ad4695_probe_spi_offload() 1744 if (IS_ERR(rx_dma)) in ad4695_probe_spi_offload() 1745 return dev_err_probe(dev, PTR_ERR(rx_dma), in ad4695_probe_spi_offload() 1789 rx_dma, IIO_BUFFER_DIRECTION_IN); in ad4695_probe_spi_offload()
|
| A D | ad7380.c | 1821 struct dma_chan *rx_dma; in ad7380_probe_spi_offload() local 1850 rx_dma = devm_spi_offload_rx_stream_request_dma_chan(dev, st->offload); in ad7380_probe_spi_offload() 1851 if (IS_ERR(rx_dma)) in ad7380_probe_spi_offload() 1852 return dev_err_probe(dev, PTR_ERR(rx_dma), in ad7380_probe_spi_offload() 1856 rx_dma, IIO_BUFFER_DIRECTION_IN); in ad7380_probe_spi_offload()
|
| /drivers/net/ethernet/amd/ |
| A D | au1000_eth.h | 46 struct rx_dma { struct 73 struct rx_dma *rx_dma_ring[NUM_RX_DMA];
|
| /drivers/net/ethernet/apple/ |
| A D | bmac.c | 62 volatile struct dbdma_regs __iomem *rx_dma; member 224 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_enable_and_reset_chip() 401 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_start_chip() 472 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_suspend() 603 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_init_rx_ring() 673 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_rxdma_intr() 1252 if (!bp->rx_dma) in bmac_probe() 1304 iounmap(bp->rx_dma); in bmac_probe() 1332 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_close() 1416 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_tx_timeout() [all …]
|
| A D | mace.c | 48 volatile struct dbdma_regs __iomem *rx_dma; member 188 mp->rx_dma = ioremap(macio_resource_start(mdev, 2), 0x1000); in mace_probe() 189 if (mp->rx_dma == NULL) { in mace_probe() 262 iounmap(mp->rx_dma); in mace_probe() 292 iounmap(mp->rx_dma); in mace_remove() 435 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_open() 505 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_close() 812 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_tx_timeout() 881 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_rxdma_intr()
|
| /drivers/net/ethernet/agere/ |
| A D | et131x.c | 1558 writel(0, &rx_dma->psr_full_offset); in et131x_config_rx_dma_regs() 1562 &rx_dma->psr_min_des); in et131x_config_rx_dma_regs() 1578 num_des = &rx_dma->fbr0_num_des; in et131x_config_rx_dma_regs() 1580 min_des = &rx_dma->fbr0_min_des; in et131x_config_rx_dma_regs() 1581 base_hi = &rx_dma->fbr0_base_hi; in et131x_config_rx_dma_regs() 1582 base_lo = &rx_dma->fbr0_base_lo; in et131x_config_rx_dma_regs() 1584 num_des = &rx_dma->fbr1_num_des; in et131x_config_rx_dma_regs() 1586 min_des = &rx_dma->fbr1_min_des; in et131x_config_rx_dma_regs() 1587 base_hi = &rx_dma->fbr1_base_hi; in et131x_config_rx_dma_regs() 1588 base_lo = &rx_dma->fbr1_base_lo; in et131x_config_rx_dma_regs() [all …]
|
| /drivers/net/fddi/ |
| A D | defza.c | 425 writel_o((fp->rx_dma[i] + 0x1000) >> 9, in fza_rx_init() 427 writel_o(fp->rx_dma[i] >> 9 | FZA_RING_OWN_FZA, in fza_rx_init() 693 dma = fp->rx_dma[i]; in fza_rx() 767 fp->rx_dma[i] = dma; in fza_rx() 1173 dma_unmap_single(fp->bdev, fp->rx_dma[i], in fza_open() 1177 fp->rx_dma[i] = 0; in fza_open() 1183 fp->rx_dma[i] = dma; in fza_open() 1256 dma_unmap_single(fp->bdev, fp->rx_dma[i], in fza_close() 1259 fp->rx_dma[i] = 0; in fza_close()
|
| /drivers/net/can/peak_canfd/ |
| A D | peak_pciefd_main.c | 315 struct pciefd_rx_dma *rx_dma = priv->rx_dma_vaddr; in pciefd_irq_handler() local 322 priv->irq_status = le32_to_cpu(rx_dma->irq_status); in pciefd_irq_handler() 330 rx_dma->msg, in pciefd_irq_handler()
|
| /drivers/atm/ |
| A D | eni.h | 81 void __iomem *rx_dma; /* RX DMA queue */ member
|