Home
last modified time | relevance | path

Searched refs:xsk_pool (Results 1 – 25 of 35) sorted by relevance

12

/linux-6.3-rc2/drivers/net/ethernet/intel/i40e/
A Di40e_xsk.c433 xsk_buff_dma_sync_for_cpu(bi, rx_ring->xsk_pool); in i40e_clean_rx_irq_zc()
455 if (xsk_uses_need_wakeup(rx_ring->xsk_pool)) { in i40e_clean_rx_irq_zc()
457 xsk_set_rx_need_wakeup(rx_ring->xsk_pool); in i40e_clean_rx_irq_zc()
459 xsk_clear_rx_need_wakeup(rx_ring->xsk_pool); in i40e_clean_rx_irq_zc()
472 dma = xsk_buff_raw_get_dma(xdp_ring->xsk_pool, desc->addr); in i40e_xmit_pkt()
538 struct xdp_desc *descs = xdp_ring->xsk_pool->tx_descs; in i40e_xmit_zc()
589 struct xsk_buff_pool *bp = tx_ring->xsk_pool; in i40e_clean_xdp_tx_irq()
634 if (xsk_uses_need_wakeup(tx_ring->xsk_pool)) in i40e_clean_xdp_tx_irq()
635 xsk_set_tx_need_wakeup(tx_ring->xsk_pool); in i40e_clean_xdp_tx_irq()
667 if (!vsi->xdp_rings[queue_id]->xsk_pool) in i40e_xsk_wakeup()
[all …]
A Di40e_txrx.h395 struct xsk_buff_pool *xsk_pool; member
A Di40e_txrx.c793 if (ring_is_xdp(tx_ring) && tx_ring->xsk_pool) { in i40e_clean_tx_ring()
1483 if (rx_ring->xsk_pool) { in i40e_clean_rx_ring()
1517 if (rx_ring->xsk_pool) in i40e_clean_rx_ring()
2717 bool wd = ring->xsk_pool ? in i40e_napi_poll()
2745 int cleaned = ring->xsk_pool ? in i40e_napi_poll()
/linux-6.3-rc2/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
A Drx.c29 if (unlikely(!xsk_buff_can_alloc(rq->xsk_pool, rq->mpwqe.pages_per_wqe))) in mlx5e_xsk_alloc_rx_mpwqe()
34 batch = xsk_buff_alloc_batch(rq->xsk_pool, (struct xdp_buff **)wi->alloc_units, in mlx5e_xsk_alloc_rx_mpwqe()
44 wi->alloc_units[batch].xsk = xsk_buff_alloc(rq->xsk_pool); in mlx5e_xsk_alloc_rx_mpwqe()
101 rq->xsk_pool->chunk_size); in mlx5e_xsk_alloc_rx_mpwqe()
102 __be32 frame_size = cpu_to_be32(rq->xsk_pool->chunk_size); in mlx5e_xsk_alloc_rx_mpwqe()
175 alloc = xsk_buff_alloc_batch(rq->xsk_pool, buffs + ix, wqe_bulk); in mlx5e_xsk_alloc_rx_wqes_batched()
177 alloc = xsk_buff_alloc_batch(rq->xsk_pool, buffs + ix, contig); in mlx5e_xsk_alloc_rx_wqes_batched()
179 alloc += xsk_buff_alloc_batch(rq->xsk_pool, buffs, wqe_bulk - contig); in mlx5e_xsk_alloc_rx_wqes_batched()
214 frag->au->xsk = xsk_buff_alloc(rq->xsk_pool); in mlx5e_xsk_alloc_rx_wqes()
273 xsk_buff_dma_sync_for_cpu(&mxbuf->xdp, rq->xsk_pool); in mlx5e_xsk_skb_from_cqe_mpwrq_linear()
[all …]
A Dtx.c63 struct xsk_buff_pool *pool = sq->xsk_pool; in mlx5e_xsk_tx()
A Dsetup.c81 rq->xsk_pool = pool; in mlx5e_init_xsk_rq()
/linux-6.3-rc2/drivers/net/ethernet/intel/ice/
A Dice_xsk.c501 nb_buffs_extra = ice_fill_rx_descs(rx_ring->xsk_pool, xdp, in __ice_alloc_rx_bufs_zc()
655 xsk_tx_completed(xdp_ring->xsk_pool, xsk_frames); in ice_clean_xdp_irq_zc()
818 xsk_buff_dma_sync_for_cpu(xdp, rx_ring->xsk_pool); in ice_clean_rx_irq_zc()
872 if (xsk_uses_need_wakeup(rx_ring->xsk_pool)) { in ice_clean_rx_irq_zc()
874 xsk_set_rx_need_wakeup(rx_ring->xsk_pool); in ice_clean_rx_irq_zc()
876 xsk_clear_rx_need_wakeup(rx_ring->xsk_pool); in ice_clean_rx_irq_zc()
965 struct xdp_desc *descs = xdp_ring->xsk_pool->tx_descs; in ice_xmit_zc()
992 if (xsk_uses_need_wakeup(xdp_ring->xsk_pool)) in ice_xmit_zc()
993 xsk_set_tx_need_wakeup(xdp_ring->xsk_pool); in ice_xmit_zc()
1026 if (!ring->xsk_pool) in ice_xsk_wakeup()
[all …]
A Dice_base.c437 if (ring->xsk_pool) in ice_setup_rx_ctx()
508 ring->xsk_pool = ice_xsk_pool(ring); in ice_vsi_cfg_rxq()
509 if (ring->xsk_pool) { in ice_vsi_cfg_rxq()
513 xsk_pool_get_rx_frame_size(ring->xsk_pool); in ice_vsi_cfg_rxq()
519 xsk_pool_set_rxq_info(ring->xsk_pool, &ring->xdp_rxq); in ice_vsi_cfg_rxq()
549 if (ring->xsk_pool) { in ice_vsi_cfg_rxq()
552 if (!xsk_buff_can_alloc(ring->xsk_pool, num_bufs)) { in ice_vsi_cfg_rxq()
A Dice_txrx.h338 struct xsk_buff_pool *xsk_pool; member
363 struct xsk_buff_pool *xsk_pool; member
A Dice_txrx.c156 if (ice_ring_is_xdp(tx_ring) && tx_ring->xsk_pool) { in ice_clean_tx_ring()
394 if (rx_ring->xsk_pool) { in ice_clean_rx_ring()
429 if (rx_ring->xsk_pool) in ice_clean_rx_ring()
460 if (rx_ring->xsk_pool) { in ice_free_rx_ring()
1528 if (tx_ring->xsk_pool) in ice_napi_poll()
1561 cleaned = rx_ring->xsk_pool ? in ice_napi_poll()
A Dice.h750 ring->xsk_pool = NULL; in ice_tx_xsk_pool()
754 ring->xsk_pool = xsk_get_pool_from_qid(vsi->netdev, qid); in ice_tx_xsk_pool()
/linux-6.3-rc2/drivers/net/ethernet/netronome/nfp/nfd3/
A Dxsk.c21 struct xsk_buff_pool *pool = r_vec->xsk_pool; in nfp_nfd3_xsk_tx_xdp()
187 xsk_buff_dma_sync_for_cpu(xrxbuf->xdp, r_vec->xsk_pool); in nfp_nfd3_xsk_rx()
324 xsk_tx_completed(r_vec->xsk_pool, done_pkts - reused); in nfp_nfd3_xsk_complete()
337 struct xsk_buff_pool *xsk_pool; in nfp_nfd3_xsk_tx() local
342 xsk_pool = r_vec->xsk_pool; in nfp_nfd3_xsk_tx()
346 if (!xsk_tx_peek_desc(xsk_pool, &desc[i])) in nfp_nfd3_xsk_tx()
356 xsk_buff_raw_dma_sync_for_device(xsk_pool, desc[i].addr, in nfp_nfd3_xsk_tx()
368 xsk_buff_raw_get_dma(xsk_pool, desc[i].addr)); in nfp_nfd3_xsk_tx()
381 xsk_tx_release(xsk_pool); in nfp_nfd3_xsk_tx()
A Drings.c25 if (tx_ring->r_vec->xsk_pool) { in nfp_nfd3_xsk_tx_bufs_free()
29 xsk_tx_completed(tx_ring->r_vec->xsk_pool, 1); in nfp_nfd3_xsk_tx_bufs_free()
/linux-6.3-rc2/drivers/net/ethernet/freescale/dpaa2/
A Ddpaa2-xsk.c49 ch->xsk_pool->umem->headroom); in dpaa2_xsk_run_xdp()
58 xsk_buff_dma_sync_for_cpu(xdp_buff, ch->xsk_pool); in dpaa2_xsk_run_xdp()
183 if (!ch->xsk_pool) in dpaa2_xsk_disable_pool()
200 ch->xsk_pool = NULL; in dpaa2_xsk_disable_pool()
265 ch->xsk_pool = pool; in dpaa2_xsk_enable_pool()
354 addr = xsk_buff_raw_get_dma(ch->xsk_pool, xdp_desc->addr); in dpaa2_xsk_tx_build_fd()
355 xsk_buff_raw_dma_sync_for_device(ch->xsk_pool, addr, xdp_desc->len); in dpaa2_xsk_tx_build_fd()
393 struct xdp_desc *xdp_descs = ch->xsk_pool->tx_descs; in dpaa2_xsk_tx()
410 batch = xsk_tx_peek_release_desc_batch(ch->xsk_pool, budget); in dpaa2_xsk_tx()
451 xsk_tx_release(ch->xsk_pool); in dpaa2_xsk_tx()
/linux-6.3-rc2/drivers/net/ethernet/mellanox/mlx5/core/
A Den_txrx.c90 bool need_wakeup = xsk_uses_need_wakeup(xskrq->xsk_pool); in mlx5e_napi_xsk_post()
98 xsk_set_tx_need_wakeup(xsksq->xsk_pool); in mlx5e_napi_xsk_post()
102 xsk_clear_tx_need_wakeup(xsksq->xsk_pool); in mlx5e_napi_xsk_post()
108 xsk_set_rx_need_wakeup(xskrq->xsk_pool); in mlx5e_napi_xsk_post()
117 xsk_set_rx_need_wakeup(xskrq->xsk_pool); in mlx5e_napi_xsk_post()
119 xsk_clear_rx_need_wakeup(xskrq->xsk_pool); in mlx5e_napi_xsk_post()
A Den_main.c459 u32 xsk_chunk_size = rq->xsk_pool ? rq->xsk_pool->chunk_size : 0; in mlx5e_create_rq_umr_mkey()
502 if (rq->xsk_pool) { in mlx5e_init_frags_partition()
1291 struct xsk_buff_pool *xsk_pool, in mlx5e_alloc_xdpsq() argument
1307 sq->xsk_pool = xsk_pool; in mlx5e_alloc_xdpsq()
1309 sq->stats = sq->xsk_pool ? in mlx5e_alloc_xdpsq()
2404 struct xsk_buff_pool *xsk_pool, in mlx5e_open_channel() argument
2446 if (xsk_pool) { in mlx5e_open_channel()
2447 mlx5e_build_xsk_param(xsk_pool, &xsk); in mlx5e_open_channel()
2533 struct xsk_buff_pool *xsk_pool = NULL; in mlx5e_open_channels() local
4147 struct xsk_buff_pool *xsk_pool = in mlx5e_xsk_validate_mtu() local
[all …]
A Den.h546 struct xsk_buff_pool *xsk_pool; member
758 struct xsk_buff_pool *xsk_pool; member
1060 struct mlx5e_sq_param *param, struct xsk_buff_pool *xsk_pool,
A Den_rx.c432 if (rq->xsk_pool) { in mlx5e_free_rx_wqe()
511 if (rq->xsk_pool) { in mlx5e_free_rx_mpwqe()
841 if (!rq->xsk_pool) in mlx5e_post_rx_wqes()
843 else if (likely(!rq->xsk_pool->dma_need_sync)) in mlx5e_post_rx_wqes()
1032 alloc_err = rq->xsk_pool ? mlx5e_xsk_alloc_rx_mpwqe(rq, head) : in mlx5e_post_rx_mpwqes()
1055 if (unlikely(alloc_err == -ENOMEM && rq->xsk_pool)) in mlx5e_post_rx_mpwqes()
/linux-6.3-rc2/drivers/net/ethernet/intel/ixgbe/
A Dixgbe_xsk.c114 if (xsk_uses_need_wakeup(rx_ring->xsk_pool) && err == -ENOBUFS) in ixgbe_run_xdp_zc()
168 bi->xdp = xsk_buff_alloc(rx_ring->xsk_pool); in ixgbe_alloc_rx_buffers_zc()
307 xsk_buff_dma_sync_for_cpu(bi->xdp, rx_ring->xsk_pool); in ixgbe_clean_rx_irq_zc()
369 if (xsk_uses_need_wakeup(rx_ring->xsk_pool)) { in ixgbe_clean_rx_irq_zc()
371 xsk_set_rx_need_wakeup(rx_ring->xsk_pool); in ixgbe_clean_rx_irq_zc()
373 xsk_clear_rx_need_wakeup(rx_ring->xsk_pool); in ixgbe_clean_rx_irq_zc()
398 struct xsk_buff_pool *pool = xdp_ring->xsk_pool; in ixgbe_xmit_zc()
466 struct xsk_buff_pool *pool = tx_ring->xsk_pool; in ixgbe_clean_xdp_tx_irq()
538 if (!ring->xsk_pool) in ixgbe_xsk_wakeup()
553 struct xsk_buff_pool *pool = tx_ring->xsk_pool; in ixgbe_xsk_clean_tx_ring()
/linux-6.3-rc2/drivers/net/ethernet/netronome/nfp/
A Dnfp_net_xsk.c22 headroom = xsk_pool_get_headroom(rx_ring->r_vec->xsk_pool); in nfp_net_xsk_rx_bufs_stash()
60 struct xsk_buff_pool *pool = r_vec->xsk_pool; in nfp_net_xsk_rx_ring_fill_freelist()
A Dnfp_net_debugfs.c46 if (!r_vec->xsk_pool) { in nfp_rx_q_show()
/linux-6.3-rc2/drivers/net/ethernet/stmicro/stmmac/
A Dstmmac_main.c238 if (rx_q->xsk_pool) { in stmmac_disable_all_queues()
1667 if (rx_q->xsk_pool) { in __init_dma_rx_desc_rings()
1684 if (rx_q->xsk_pool) { in __init_dma_rx_desc_rings()
1735 if (rx_q->xsk_pool) in init_dma_rx_desc_rings()
1741 rx_q->xsk_pool = NULL; in init_dma_rx_desc_rings()
1871 tx_q->xsk_pool = NULL; in dma_free_tx_skbufs()
1901 if (rx_q->xsk_pool) in __free_dma_rx_desc_resources()
1907 rx_q->xsk_pool = NULL; in __free_dma_rx_desc_resources()
2388 if (rx_q->xsk_pool) { in stmmac_dma_operation_mode()
2658 if (tx_q->xsk_pool) { in stmmac_tx_clean()
[all …]
A Dstmmac.h73 struct xsk_buff_pool *xsk_pool; member
99 struct xsk_buff_pool *xsk_pool; member
/linux-6.3-rc2/drivers/net/ethernet/intel/igc/
A Digc_main.c252 if (tx_ring->xsk_pool && xsk_frames) in igc_clean_tx_ring()
431 if (ring->xsk_pool) in igc_clean_rx_ring()
604 if (ring->xsk_pool) { in igc_configure_rx_ring()
637 if (ring->xsk_pool) in igc_configure_rx_ring()
2209 bi->xdp = xsk_buff_alloc(ring->xsk_pool); in igc_alloc_rx_buffers_zc()
2741 xsk_set_rx_need_wakeup(ring->xsk_pool); in igc_clean_rx_irq_zc()
2927 if (tx_ring->xsk_pool) { in igc_clean_tx_irq()
3831 if (ring->xsk_pool) in igc_configure()
4457 int cleaned = rx_ring->xsk_pool ? in igc_poll()
6353 if (!ring->xsk_pool) in igc_xsk_wakeup()
[all …]
/linux-6.3-rc2/drivers/net/ethernet/mellanox/mlx5/core/en/
A Dxdp.c598 xsk_tx_completed(sq->xsk_pool, xsk_frames); in mlx5e_poll_xdpsq_cq()
636 xsk_tx_completed(sq->xsk_pool, xsk_frames); in mlx5e_free_xdpsq_descs()

Completed in 92 milliseconds

12