Searched refs:tx_sgl (Results 1 – 9 of 9) sorted by relevance
| /drivers/spi/ |
| A D | spi-mt65xx.c | 168 struct scatterlist *tx_sgl, *rx_sgl; member 627 if (mdata->tx_sgl) { in mtk_spi_setup_dma_addr() 685 mdata->tx_sgl = NULL; in mtk_spi_dma_transfer() 702 mdata->tx_sgl = xfer->tx_sg.sgl; in mtk_spi_dma_transfer() 706 if (mdata->tx_sgl) { in mtk_spi_dma_transfer() 707 xfer->tx_dma = sg_dma_address(mdata->tx_sgl); in mtk_spi_dma_transfer() 708 mdata->tx_sgl_len = sg_dma_len(mdata->tx_sgl); in mtk_spi_dma_transfer() 821 if (mdata->tx_sgl) in mtk_spi_interrupt_thread() 827 mdata->tx_sgl = sg_next(mdata->tx_sgl); in mtk_spi_interrupt_thread() 828 if (mdata->tx_sgl) { in mtk_spi_interrupt_thread() [all …]
|
| A D | spi-pci1xxxx.c | 149 struct scatterlist *tx_sgl, *rx_sgl; member 549 p->tx_sgl = xfer->tx_sg.sgl; in pci1xxxx_spi_transfer_with_dma() 556 if (!xfer->tx_buf || !p->tx_sgl) { in pci1xxxx_spi_transfer_with_dma() 568 tx_dma_addr = sg_dma_address(p->tx_sgl); in pci1xxxx_spi_transfer_with_dma() 569 p->tx_sgl_len = sg_dma_len(p->tx_sgl); in pci1xxxx_spi_transfer_with_dma() 666 p->tx_sgl = sg_next(p->tx_sgl); in pci1xxxx_spi_setup_next_dma_to_io_transfer() 667 if (p->tx_sgl) { in pci1xxxx_spi_setup_next_dma_to_io_transfer() 668 tx_dma_addr = sg_dma_address(p->tx_sgl); in pci1xxxx_spi_setup_next_dma_to_io_transfer() 670 p->tx_sgl_len = sg_dma_len(p->tx_sgl); in pci1xxxx_spi_setup_next_dma_to_io_transfer()
|
| A D | spi-qup.c | 475 struct scatterlist *tx_sgl, *rx_sgl; in spi_qup_do_dma() local 490 tx_sgl = xfer->tx_sg.sgl; in spi_qup_do_dma() 498 if (tx_sgl) in spi_qup_do_dma() 499 qup->n_words = spi_qup_sgl_get_nents_len(tx_sgl, in spi_qup_do_dma() 522 if (tx_sgl) { in spi_qup_do_dma() 523 ret = spi_qup_prep_sg(host, tx_sgl, tx_nents, in spi_qup_do_dma() 536 for (; tx_sgl && tx_nents--; tx_sgl = sg_next(tx_sgl)) in spi_qup_do_dma() 539 } while (rx_sgl || tx_sgl); in spi_qup_do_dma()
|
| /drivers/tty/serial/ |
| A D | imx.c | 221 struct scatterlist rx_sgl, tx_sgl[2]; member 588 struct scatterlist *sgl = &sport->tx_sgl[0]; in imx_uart_dma_tx_callback() 625 struct scatterlist *sgl = sport->tx_sgl; in imx_uart_dma_tx() 639 sg_init_table(sgl, ARRAY_SIZE(sport->tx_sgl)); in imx_uart_dma_tx() 642 ARRAY_SIZE(sport->tx_sgl), sport->tx_bytes); in imx_uart_dma_tx() 1599 dma_unmap_sg(sport->port.dev, &sport->tx_sgl[0], in imx_uart_shutdown() 1717 struct scatterlist *sgl = &sport->tx_sgl[0]; in imx_uart_flush_buffer()
|
| A D | mxs-auart.c | 431 struct scatterlist tx_sgl; member 522 dma_unmap_sg(s->dev, &s->tx_sgl, 1, DMA_TO_DEVICE); in dma_tx_callback() 538 struct scatterlist *sgl = &s->tx_sgl; in mxs_auart_dma_tx()
|
| A D | fsl_lpuart.c | 285 struct scatterlist rx_sgl, tx_sgl[2]; member 481 struct scatterlist *sgl = sport->tx_sgl; in lpuart_dma_tx() 489 sg_init_table(sgl, ARRAY_SIZE(sport->tx_sgl)); in lpuart_dma_tx() 492 ARRAY_SIZE(sport->tx_sgl), sport->dma_tx_bytes); in lpuart_dma_tx() 527 struct scatterlist *sgl = &sport->tx_sgl[0]; in lpuart_dma_tx_complete() 609 dma_unmap_sg(chan->device->dev, &sport->tx_sgl[0], in lpuart_flush_buffer()
|
| /drivers/net/ethernet/hisilicon/hns3/ |
| A D | hns3_enet.h | 434 u64 tx_sgl; member
|
| A D | hns3_enet.c | 59 static unsigned int tx_sgl = 1; variable 60 module_param(tx_sgl, uint, 0600); 61 MODULE_PARM_DESC(tx_sgl, "Minimum number of frags when using dma_map_sg() to optimize the IOMMU map… 1027 if (skb->len <= ring->tx_copybreak || !tx_sgl || in hns3_can_use_tx_sgl() 1029 skb_shinfo(skb)->nr_frags < tx_sgl)) in hns3_can_use_tx_sgl() 2255 hns3_ring_stats_update(ring, tx_sgl); in hns3_handle_tx_sgl()
|
| A D | hns3_ethtool.c | 41 HNS3_TQP_STAT("sgl", tx_sgl),
|
Completed in 58 milliseconds