Lines Matching refs:xsk_pool

501 		nb_buffs_extra = ice_fill_rx_descs(rx_ring->xsk_pool, xdp,  in __ice_alloc_rx_bufs_zc()
515 nb_buffs = ice_fill_rx_descs(rx_ring->xsk_pool, xdp, rx_desc, count); in __ice_alloc_rx_bufs_zc()
655 xsk_tx_completed(xdp_ring->xsk_pool, xsk_frames); in ice_clean_xdp_irq_zc()
688 xsk_buff_raw_dma_sync_for_device(xdp_ring->xsk_pool, dma, size); in ice_xmit_xdp_tx_zc()
728 if (xsk_uses_need_wakeup(rx_ring->xsk_pool) && err == -ENOBUFS) in ice_run_xdp_zc()
818 xsk_buff_dma_sync_for_cpu(xdp, rx_ring->xsk_pool); in ice_clean_rx_irq_zc()
872 if (xsk_uses_need_wakeup(rx_ring->xsk_pool)) { in ice_clean_rx_irq_zc()
874 xsk_set_rx_need_wakeup(rx_ring->xsk_pool); in ice_clean_rx_irq_zc()
876 xsk_clear_rx_need_wakeup(rx_ring->xsk_pool); in ice_clean_rx_irq_zc()
896 dma = xsk_buff_raw_get_dma(xdp_ring->xsk_pool, desc->addr); in ice_xmit_pkt()
897 xsk_buff_raw_dma_sync_for_device(xdp_ring->xsk_pool, dma, desc->len); in ice_xmit_pkt()
923 dma = xsk_buff_raw_get_dma(xdp_ring->xsk_pool, descs[i].addr); in ice_xmit_pkt_batch()
924 xsk_buff_raw_dma_sync_for_device(xdp_ring->xsk_pool, dma, descs[i].len); in ice_xmit_pkt_batch()
965 struct xdp_desc *descs = xdp_ring->xsk_pool->tx_descs; in ice_xmit_zc()
975 nb_pkts = xsk_tx_peek_release_desc_batch(xdp_ring->xsk_pool, budget); in ice_xmit_zc()
992 if (xsk_uses_need_wakeup(xdp_ring->xsk_pool)) in ice_xmit_zc()
993 xsk_set_tx_need_wakeup(xdp_ring->xsk_pool); in ice_xmit_zc()
1026 if (!ring->xsk_pool) in ice_xsk_wakeup()
1104 xsk_tx_completed(xdp_ring->xsk_pool, xsk_frames); in ice_xsk_clean_xdp_ring()