| /linux/drivers/dma/ |
| A D | sa11x0-dma.c | 155 p->num, &txd->vd, txd->vd.tx.cookie, txd->ddar); in sa11x0_dma_start_desc() 167 if (!txd) in sa11x0_dma_start_sg() 304 if (txd) { in sa11x0_dma_start_txd() 458 if (txd) { in sa11x0_dma_tx_status() 466 i, txd->sg[i].addr, txd->sg[i].len); in sa11x0_dma_tx_status() 468 addr < txd->sg[i].addr + txd->sg[i].len) { in sa11x0_dma_tx_status() 482 i, txd->sg[i].addr, txd->sg[i].len); in sa11x0_dma_tx_status() 556 txd = kzalloc(struct_size(txd, sg, j), GFP_ATOMIC); in sa11x0_dma_prep_slave_sg() 557 if (!txd) { in sa11x0_dma_prep_slave_sg() 599 &c->vc, &txd->vd, txd->size, txd->sglen); in sa11x0_dma_prep_slave_sg() [all …]
|
| A D | owl-dma.c | 547 vchan->txd = txd; in owl_dma_start_next_txd() 654 txd = vchan->txd; in owl_dma_interrupt() 655 if (txd) { in owl_dma_interrupt() 686 kfree(txd); in owl_dma_free_txd() 780 txd = vchan->txd; in owl_dma_getbytes_chan() 881 txd = kzalloc(sizeof(*txd), GFP_NOWAIT); in owl_dma_prep_memcpy() 882 if (!txd) in owl_dma_prep_memcpy() 932 txd = kzalloc(sizeof(*txd), GFP_NOWAIT); in owl_dma_prep_slave_sg() 933 if (!txd) in owl_dma_prep_slave_sg() 996 txd = kzalloc(sizeof(*txd), GFP_NOWAIT); in owl_prep_dma_cyclic() [all …]
|
| A D | timb_dma.c | 64 struct dma_async_tx_descriptor txd; member 231 txd = &td_desc->txd; in __td_finish() 234 txd->cookie); in __td_finish() 243 dma_cookie_complete(txd); in __td_finish() 250 dma_descriptor_unmap(txd); in __td_finish() 288 __func__, td_desc->txd.cookie); in __td_start_next() 297 txd); in td_tx_submit() 307 txd->cookie); in td_tx_submit() 312 txd->cookie); in td_tx_submit() 363 dma_unmap_single(chan2dmadev(td_desc->txd.chan), td_desc->txd.phys, in td_free_desc() [all …]
|
| A D | amba-pl08x.c | 1274 bd.txd = txd; in pl08x_fill_llis_for_desc() 1498 dma_pool_free(pl08x->pool, txd->llis_va, txd->llis_bus); in pl08x_free_txd() 1505 kfree(txd); in pl08x_free_txd() 1746 struct pl08x_txd *txd = kzalloc(sizeof(*txd), GFP_NOWAIT); in pl08x_get_txd() local 1748 if (txd) in pl08x_get_txd() 1750 return txd; in pl08x_get_txd() 1892 if (!txd) { in pl08x_prep_dma_memcpy() 1942 if (!txd) { in pl08x_init_txd() 2012 return txd; in pl08x_init_txd() 2057 if (!txd) in pl08x_prep_slave_sg() [all …]
|
| A D | txx9dmac.c | 141 return container_of(txd, struct txx9dmac_desc, txd); in txd_to_txx9dmac_desc() 403 struct dma_async_tx_descriptor *txd = &desc->txd; in txx9dmac_descriptor_complete() local 406 txd->cookie, desc); in txx9dmac_descriptor_complete() 408 dma_cookie_complete(txd); in txx9dmac_descriptor_complete() 415 dma_descriptor_unmap(txd); in txx9dmac_descriptor_complete() 421 dma_run_dependencies(txd); in txx9dmac_descriptor_complete() 528 bad_desc->txd.cookie); in txx9dmac_handle_error() 796 first->txd.flags = flags; in txx9dmac_prep_dma_memcpy() 799 return &first->txd; in txx9dmac_prep_dma_memcpy() 874 prev->txd.phys, in txx9dmac_prep_slave_sg() [all …]
|
| A D | sun6i-dma.c | 431 kfree(txd); in sun6i_dma_free_desc() 666 txd = kzalloc(sizeof(*txd), GFP_NOWAIT); in sun6i_dma_prep_dma_memcpy() 667 if (!txd) in sun6i_dma_prep_dma_memcpy() 696 kfree(txd); in sun6i_dma_prep_dma_memcpy() 724 txd = kzalloc(sizeof(*txd), GFP_NOWAIT); in sun6i_dma_prep_slave_sg() 725 if (!txd) in sun6i_dma_prep_slave_sg() 769 for (p_lli = txd->p_lli, v_lli = txd->v_lli; v_lli; in sun6i_dma_prep_slave_sg() 776 for (p_lli = txd->p_lli, v_lli = txd->v_lli; v_lli; in sun6i_dma_prep_slave_sg() 807 txd = kzalloc(sizeof(*txd), GFP_NOWAIT); in sun6i_dma_prep_dma_cyclic() 808 if (!txd) in sun6i_dma_prep_dma_cyclic() [all …]
|
| A D | plx_dma.c | 98 struct dma_async_tx_descriptor txd; member 126 return container_of(txd, struct plx_dma_desc, txd); in to_plx_desc() 159 dma_cookie_complete(&desc->txd); in plx_dma_process_desc() 160 dma_descriptor_unmap(&desc->txd); in plx_dma_process_desc() 162 desc->txd.callback = NULL; in plx_dma_process_desc() 163 desc->txd.callback_result = NULL; in plx_dma_process_desc() 186 dma_cookie_complete(&desc->txd); in plx_dma_abort_desc() 187 dma_descriptor_unmap(&desc->txd); in plx_dma_abort_desc() 189 desc->txd.callback = NULL; in plx_dma_abort_desc() 283 plxdesc->txd.flags = flags; in plx_dma_prep_memcpy() [all …]
|
| A D | ep93xx_dma.c | 318 d->txd.callback = desc->txd.callback; in ep93xx_dma_set_active() 319 d->txd.callback_param = desc->txd.callback_param; in ep93xx_dma_set_active() 362 return !desc->txd.cookie; in ep93xx_dma_advance_active() 750 desc->txd.cookie = 0; in ep93xx_dma_desc_get() 1065 first->txd.cookie = -EBUSY; in ep93xx_dma_prep_dma_memcpy() 1066 first->txd.flags = flags; in ep93xx_dma_prep_dma_memcpy() 1068 return &first->txd; in ep93xx_dma_prep_dma_memcpy() 1140 first->txd.cookie = -EBUSY; in ep93xx_dma_prep_slave_sg() 1141 first->txd.flags = flags; in ep93xx_dma_prep_slave_sg() 1143 return &first->txd; in ep93xx_dma_prep_slave_sg() [all …]
|
| A D | pch_dma.c | 85 struct dma_async_tx_descriptor txd; member 140 return container_of(txd, struct pch_dma_desc, txd); in to_pd_desc() 346 struct dma_async_tx_descriptor *txd = &desc->txd; in pdc_chain_complete() local 387 bad_desc->txd.cookie); in pdc_handle_error() 431 desc->txd.tx_submit = pd_tx_submit; in pdc_alloc_desc() 432 desc->txd.flags = DMA_CTRL_ACK; in pdc_alloc_desc() 433 desc->txd.phys = addr; in pdc_alloc_desc() 624 prev->regs.next |= desc->txd.phys; in pd_prep_slave_sg() 636 first->txd.cookie = -EBUSY; in pd_prep_slave_sg() 637 desc->txd.flags = flags; in pd_prep_slave_sg() [all …]
|
| /linux/drivers/net/ethernet/sfc/ |
| A D | ef100_tx.c | 104 if (likely(tx_queue->txd.addr)) in ef100_tx_desc() 153 EFX_OR_OWORD(*txd, *txd, csum); in ef100_set_tx_csum_partial() 164 EFX_OR_OWORD(*txd, *txd, vlan); in ef100_set_tx_hw_vlan() 173 EFX_POPULATE_OWORD_3(*txd, in ef100_make_send_desc() 182 ef100_set_tx_hw_vlan(skb, txd); in ef100_make_send_desc() 231 EFX_POPULATE_OWORD_19(*txd, in ef100_make_tso_desc() 265 efx_oword_t *txd; in ef100_tx_make_descriptors() local 283 EFX_POPULATE_OWORD_3(*txd, in ef100_tx_make_descriptors() 306 buffer, txd, nr_descs); in ef100_tx_make_descriptors() 312 buffer, txd, nr_descs); in ef100_tx_make_descriptors() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/nfd3/ |
| A D | dp.c | 103 txd->mss = cpu_to_le16(mss); in nfp_nfd3_tx_tso() 262 struct nfp_nfd3_tx_desc *txd; in nfp_nfd3_tx() local 324 txd->flags = 0; in nfp_nfd3_tx() 325 txd->mss = 0; in nfp_nfd3_tx() 326 txd->lso_hdrlen = 0; in nfp_nfd3_tx() 886 txd->flags = 0; in nfp_nfd3_tx_xdp_buf() 887 txd->mss = 0; in nfp_nfd3_tx_xdp_buf() 888 txd->lso_hdrlen = 0; in nfp_nfd3_tx_xdp_buf() 1270 txd->flags = 0; in nfp_nfd3_ctrl_tx_one() 1271 txd->mss = 0; in nfp_nfd3_ctrl_tx_one() [all …]
|
| A D | ipsec.c | 10 void nfp_nfd3_ipsec_tx(struct nfp_nfd3_tx_desc *txd, struct sk_buff *skb) in nfp_nfd3_ipsec_tx() argument 18 txd->flags |= NFD3_DESC_TX_CSUM; in nfp_nfd3_ipsec_tx() 21 txd->flags |= NFD3_DESC_TX_IP4_CSUM; in nfp_nfd3_ipsec_tx() 32 txd->flags |= NFD3_DESC_TX_UDP_CSUM; in nfp_nfd3_ipsec_tx() 35 txd->flags |= NFD3_DESC_TX_TCP_CSUM; in nfp_nfd3_ipsec_tx()
|
| A D | xsk.c | 23 struct nfp_nfd3_tx_desc *txd; in nfp_nfd3_xsk_tx_xdp() local 40 txd = &tx_ring->txds[wr_idx]; in nfp_nfd3_xsk_tx_xdp() 41 txd->offset_eop = NFD3_DESC_TX_EOP; in nfp_nfd3_xsk_tx_xdp() 42 txd->dma_len = cpu_to_le16(pkt_len); in nfp_nfd3_xsk_tx_xdp() 44 txd->data_len = cpu_to_le16(pkt_len); in nfp_nfd3_xsk_tx_xdp() 46 txd->flags = 0; in nfp_nfd3_xsk_tx_xdp() 47 txd->mss = 0; in nfp_nfd3_xsk_tx_xdp() 48 txd->lso_hdrlen = 0; in nfp_nfd3_xsk_tx_xdp() 338 struct nfp_nfd3_tx_desc *txd; in nfp_nfd3_xsk_tx() local 366 txd = &tx_ring->txds[wr_idx]; in nfp_nfd3_xsk_tx() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/nfdk/ |
| A D | dp.c | 77 return txd.raw; in nfp_nfdk_tx_tso() 349 txd++; in nfp_nfdk_tx() 367 txd++; in nfp_nfdk_tx() 396 txd++; in nfp_nfdk_tx() 405 txd += 2; in nfp_nfdk_tx() 985 txd++; in nfp_nfdk_tx_xdp_buf() 996 txd++; in nfp_nfdk_tx_xdp_buf() 1003 txd++; in nfp_nfdk_tx_xdp_buf() 1399 txd++; in nfp_nfdk_ctrl_tx_one() 1410 txd++; in nfp_nfdk_ctrl_tx_one() [all …]
|
| /linux/Documentation/devicetree/bindings/pinctrl/ |
| A D | marvell,armada-39x-pinctrl.txt | 19 mpp1 1 gpio, ua0(txd) 22 mpp4 4 gpio, ua1(txd), ua0(rts), smi(mdc) 32 mpp14 14 gpio, dram(vttctrl), dev(we1), ua1(txd) 36 mpp18 18 gpio, ua1(txd), spi0(cs0), i2c2(sck) 38 mpp20 20 gpio, sata0(prsnt) [1], ua0(rts), ua1(txd), smi(mdc) 44 mpp25 25 gpio, spi0(cs0), ua0(rts), ua1(txd), sd0(d5), dev(cs0) 61 mpp42 42 gpio, ua1(txd), ua0(rts), dev(ad7) 66 mpp46 46 gpio, ref(clk), pcie0(rstout), ua1(txd), led(stb) 73 mpp51 51 gpio, tdm(dtx) [2], audio(sdo) [2], dram(deccerr), ua2(txd) 78 mpp56 56 gpio, ua1(rts), dram(deccerr), spi1(mosi), ua1(txd) [all …]
|
| A D | marvell,kirkwood-pinctrl.txt | 36 mpp10 10 gpo, spi(sck), uart0(txd), ptp(trig) 40 mpp13 13 gpio, sdio(cmd), uart1(txd) 78 mpp13 13 gpio, sdio(cmd), uart1(txd) 122 mpp13 13 gpio, sdio(cmd), uart1(txd) 171 mpp13 13 gpio, sdio(cmd), uart1(txd) 293 mpp5 5 gpo, nand(io7), uart0(txd) 298 mpp10 10 gpo, spi(sck), uart0(txd) 300 mpp13 13 gpio, uart1(txd) 330 mpp5 5 gpo, nand(io7), uart0(txd) 335 mpp10 10 gpo, uart0(txd) [all …]
|
| /linux/drivers/net/wireless/mediatek/mt76/mt7603/ |
| A D | mcu.c | 43 struct mt7603_mcu_txd *txd; in mt7603_mcu_skb_send_msg() local 54 txd->len = cpu_to_le16(skb->len); in mt7603_mcu_skb_send_msg() 58 txd->pq_id = cpu_to_le16(MCU_PORT_QUEUE); in mt7603_mcu_skb_send_msg() 59 txd->pkt_type = MCU_PKT_ID; in mt7603_mcu_skb_send_msg() 60 txd->seq = seq; in mt7603_mcu_skb_send_msg() 63 txd->cid = -cmd; in mt7603_mcu_skb_send_msg() 64 txd->set_query = MCU_Q_NA; in mt7603_mcu_skb_send_msg() 66 txd->cid = MCU_CMD_EXT_CID; in mt7603_mcu_skb_send_msg() 67 txd->ext_cid = cmd; in mt7603_mcu_skb_send_msg() 68 txd->set_query = MCU_Q_SET; in mt7603_mcu_skb_send_msg() [all …]
|
| /linux/drivers/rapidio/devices/ |
| A D | tsi721_dma.c | 60 return container_of(txd, struct tsi721_tx_desc, txd); in to_tsi721_desc() 377 struct dma_async_tx_descriptor *txd = &desc->txd; in tsi721_dma_tx_err() local 379 void *param = txd->callback_param; in tsi721_dma_tx_err() 643 dma_cookie_complete(&desc->txd); in tsi721_dma_tasklet() 671 callback = desc->txd.callback; in tsi721_dma_tasklet() 703 bdma_chan->id, txd); in tsi721_tx_submit() 714 cookie = dma_cookie_assign(txd); in tsi721_tx_submit() 880 txd = &desc->txd; in tsi721_prep_rio_sg() 881 txd->flags = flags; in tsi721_prep_rio_sg() 886 if (!txd) { in tsi721_prep_rio_sg() [all …]
|
| /linux/drivers/net/ethernet/intel/fm10k/ |
| A D | fm10k_debugfs.c | 60 struct fm10k_tx_desc *txd = FM10K_TX_DESC(ring, i); in fm10k_dbg_tx_desc_seq_show() local 63 i, txd->buffer_addr, txd->buflen, txd->vlan, in fm10k_dbg_tx_desc_seq_show() 64 txd->mss, txd->hdrlen, txd->flags); in fm10k_dbg_tx_desc_seq_show()
|
| /linux/include/linux/ |
| A D | dmaengine.h | 666 static inline void txd_lock(struct dma_async_tx_descriptor *txd) in txd_lock() argument 669 static inline void txd_unlock(struct dma_async_tx_descriptor *txd) in txd_unlock() argument 692 static inline void txd_lock(struct dma_async_tx_descriptor *txd) in txd_lock() argument 694 spin_lock_bh(&txd->lock); in txd_lock() 698 spin_unlock_bh(&txd->lock); in txd_unlock() 702 txd->next = next; in txd_chain() 703 next->parent = txd; in txd_chain() 707 txd->parent = NULL; in txd_clear_parent() 711 txd->next = NULL; in txd_clear_next() 715 return txd->parent; in txd_parent() [all …]
|
| /linux/drivers/media/pci/mantis/ |
| A D | mantis_i2c.c | 75 u32 txd = 0, stat, trials; in mantis_i2c_write() local 82 txd = (msg->addr << 25) | (msg->buf[i] << 8) in mantis_i2c_write() 88 txd &= ~MANTIS_I2C_STOP; in mantis_i2c_write() 91 mmwrite(txd, MANTIS_I2CDATA_CTL); in mantis_i2c_write() 119 u32 stat, data, txd; in mantis_i2c_xfer() local 142 txd = msgs[i].addr << 25 | (0x1 << 24) in mantis_i2c_xfer() 146 mmwrite(txd, MANTIS_I2CDATA_CTL); in mantis_i2c_xfer()
|
| /linux/Documentation/devicetree/bindings/arm/marvell/ |
| A D | cp110-system-controller.txt | 92 mpp1 1 gpio, dev(ale0), au(i2sdo_spdifo), ge0(rxd2), tdm(drx), ptp(clk), mss_i2c(sck), uart0(txd), … 94 …ev(ad14), au(i2slrclk), ge0(rxd0), tdm(fsync), mss_uart(txd), pcie(rstoutn), i2c1(sda), uart1(txd)… 96 … dev(ad12), au(i2sdi), ge0(rxclk), tdm(intn), mss_uart(txd), uart1(rts), pcie1(clkreq), uart3(txd)… 98 mpp7 7 gpio, dev(ad10), ge0(txd2), spi0(csn1), spi1(csn1), sata0(present_act), led(data), uart0(txd… 102 mpp11 11 gpio, dev(wen1), ge0(txclkout), spi0(clk), spi1(clk), uart0(rts), led(clk), uart2(txd), sa… 136 mpp45 45 gpio, ge1(txd3), uart0(txd), pcie(rstoutn) 138 mpp47 47 gpio, ge1(txd0), spi1(clk), uart1(txd), ge(mdc) 141 mpp50 50 gpio, ge1(rxclk), mss_i2c(sda), spi1(csn0), uart2(txd), uart0(rxd), xg(mdio), sdio(pwr11) 148 mpp57 57 gpio, mss_i2c(sda), ptp(pclk_out), tdm(intn), au(i2sbclk), spi0(mosi), uart1(txd), sata0(p… 150 …pio7, synce2(clk), tdm(fsync), au(i2slrclk), spi0(csn0), uart0(cts), led(stb), uart1(txd), sdio(d1) [all …]
|
| /linux/drivers/net/ethernet/aquantia/atlantic/hw_atl/ |
| A D | hw_atl_a0.c | 442 struct hw_atl_txd_s *txd = NULL; in hw_atl_a0_hw_ring_tx_xmit() local 454 txd->ctl = 0; in hw_atl_a0_hw_ring_tx_xmit() 455 txd->ctl2 = 0; in hw_atl_a0_hw_ring_tx_xmit() 456 txd->buf_addr = 0; in hw_atl_a0_hw_ring_tx_xmit() 461 txd->ctl |= (buff->len_l3 << 31) | in hw_atl_a0_hw_ring_tx_xmit() 465 txd->ctl2 |= (buff->mss << 16) | in hw_atl_a0_hw_ring_tx_xmit() 479 txd->buf_addr = buff->pa; in hw_atl_a0_hw_ring_tx_xmit() 480 txd->ctl |= (HW_ATL_A0_TXD_CTL_BLEN & in hw_atl_a0_hw_ring_tx_xmit() 487 txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_LSO; in hw_atl_a0_hw_ring_tx_xmit() 499 txd->ctl |= HW_ATL_A0_TXD_CTL_EOP; in hw_atl_a0_hw_ring_tx_xmit() [all …]
|
| /linux/drivers/dma/dw/ |
| A D | core.c | 73 __func__, desc->txd.cookie); in dwc_tx_submit() 92 desc->txd.flags = DMA_CTRL_ACK; in dwc_desc_get() 93 desc->txd.phys = phys; in dwc_desc_get() 236 struct dma_async_tx_descriptor *txd = &desc->txd; in dwc_descriptor_complete() local 244 dma_cookie_complete(txd); in dwc_descriptor_complete() 252 async_tx_ack(&child->txd); in dwc_descriptor_complete() 253 async_tx_ack(&desc->txd); in dwc_descriptor_complete() 604 first->txd.flags = flags; in dwc_prep_dma_memcpy() 607 return &first->txd; in dwc_prep_dma_memcpy() 754 return &first->txd; in dwc_prep_slave_sg() [all …]
|
| /linux/drivers/dma/idxd/ |
| A D | dma.c | 46 tx = &desc->txd; in idxd_dma_complete_txd() 102 desc->txd.flags = flags; in idxd_dma_prep_interrupt() 103 return &desc->txd; in idxd_dma_prep_interrupt() 130 desc->txd.flags = flags; in idxd_dma_submit_memcpy() 132 return &desc->txd; in idxd_dma_submit_memcpy() 177 struct idxd_desc *desc = container_of(tx, struct idxd_desc, txd); in idxd_dma_tx_submit() 268 dma_async_tx_descriptor_init(&desc->txd, chan); in idxd_register_dma_channel() 269 desc->txd.tx_submit = idxd_dma_tx_submit; in idxd_register_dma_channel()
|