Lines Matching refs:tx_skb

828 	struct nv_skb_map *tx_skb;  member
1038 kfree(np->tx_skb); in free_rings()
1957 np->get_tx_ctx = np->tx_skb; in nv_init_tx()
1958 np->put_tx_ctx = np->tx_skb; in nv_init_tx()
1959 np->last_tx_ctx = &np->tx_skb[np->tx_ring_size-1]; in nv_init_tx()
1976 np->tx_skb[i].skb = NULL; in nv_init_tx()
1977 np->tx_skb[i].dma = 0; in nv_init_tx()
1978 np->tx_skb[i].dma_len = 0; in nv_init_tx()
1979 np->tx_skb[i].dma_single = 0; in nv_init_tx()
1980 np->tx_skb[i].first_tx_desc = NULL; in nv_init_tx()
1981 np->tx_skb[i].next_tx_ctx = NULL; in nv_init_tx()
1998 static void nv_unmap_txskb(struct fe_priv *np, struct nv_skb_map *tx_skb) in nv_unmap_txskb() argument
2000 if (tx_skb->dma) { in nv_unmap_txskb()
2001 if (tx_skb->dma_single) in nv_unmap_txskb()
2002 dma_unmap_single(&np->pci_dev->dev, tx_skb->dma, in nv_unmap_txskb()
2003 tx_skb->dma_len, in nv_unmap_txskb()
2006 dma_unmap_page(&np->pci_dev->dev, tx_skb->dma, in nv_unmap_txskb()
2007 tx_skb->dma_len, in nv_unmap_txskb()
2009 tx_skb->dma = 0; in nv_unmap_txskb()
2013 static int nv_release_txskb(struct fe_priv *np, struct nv_skb_map *tx_skb) in nv_release_txskb() argument
2015 nv_unmap_txskb(np, tx_skb); in nv_release_txskb()
2016 if (tx_skb->skb) { in nv_release_txskb()
2017 dev_kfree_skb_any(tx_skb->skb); in nv_release_txskb()
2018 tx_skb->skb = NULL; in nv_release_txskb()
2039 if (nv_release_txskb(np, &np->tx_skb[i])) { in nv_drain_tx()
2044 np->tx_skb[i].dma = 0; in nv_drain_tx()
2045 np->tx_skb[i].dma_len = 0; in nv_drain_tx()
2046 np->tx_skb[i].dma_single = 0; in nv_drain_tx()
2047 np->tx_skb[i].first_tx_desc = NULL; in nv_drain_tx()
2048 np->tx_skb[i].next_tx_ctx = NULL; in nv_drain_tx()
2279 np->put_tx_ctx = np->tx_skb; in nv_start_xmit()
2305 tmp_tx_ctx = np->tx_skb; in nv_start_xmit()
2328 np->put_tx_ctx = np->tx_skb; in nv_start_xmit()
2337 if (unlikely(np->put_tx_ctx == np->tx_skb)) in nv_start_xmit()
2457 np->put_tx_ctx = np->tx_skb; in nv_start_xmit_optimized()
2483 tmp_tx_ctx = np->tx_skb; in nv_start_xmit_optimized()
2506 np->put_tx_ctx = np->tx_skb; in nv_start_xmit_optimized()
2515 if (unlikely(np->put_tx_ctx == np->tx_skb)) in nv_start_xmit_optimized()
2664 np->get_tx_ctx = np->tx_skb; in nv_tx_done()
2721 np->get_tx_ctx = np->tx_skb; in nv_tx_done_optimized()
4759 np->tx_skb = (struct nv_skb_map *)tx_skbuff; in nv_set_ringparam()
4763 memset(np->tx_skb, 0, sizeof(struct nv_skb_map) * np->tx_ring_size); in nv_set_ringparam()
5141 struct sk_buff *tx_skb, *rx_skb; in nv_loopback_test() local
5179 tx_skb = netdev_alloc_skb(dev, pkt_len); in nv_loopback_test()
5180 if (!tx_skb) { in nv_loopback_test()
5184 test_dma_addr = dma_map_single(&np->pci_dev->dev, tx_skb->data, in nv_loopback_test()
5185 skb_tailroom(tx_skb), in nv_loopback_test()
5189 dev_kfree_skb_any(tx_skb); in nv_loopback_test()
5192 pkt_data = skb_put(tx_skb, pkt_len); in nv_loopback_test()
5244 (skb_end_pointer(tx_skb) - tx_skb->data), in nv_loopback_test()
5246 dev_kfree_skb_any(tx_skb); in nv_loopback_test()
5870 np->tx_skb = kcalloc(np->tx_ring_size, sizeof(struct nv_skb_map), GFP_KERNEL); in nv_probe()
5871 if (!np->rx_skb || !np->tx_skb) in nv_probe()