Home
last modified time | relevance | path

Searched refs:page_pool (Results 1 – 25 of 86) sorted by relevance

1234

/drivers/net/ethernet/apm/xgene/
A Dxgene_enet_main.c690 page_pool = rx_ring->page_pool; in xgene_enet_rx_frame()
1081 page_pool = ring->page_pool; in xgene_enet_delete_desc_rings()
1082 if (page_pool) { in xgene_enet_delete_desc_rings()
1179 page_pool = ring->page_pool; in xgene_enet_free_desc_rings()
1180 if (page_pool) { in xgene_enet_free_desc_rings()
1373 if (!page_pool) { in xgene_enet_create_desc_rings()
1395 page_pool->dst_ring_num = xgene_enet_dst_ring_num(page_pool); in xgene_enet_create_desc_rings()
1396 rx_ring->page_pool = page_pool; in xgene_enet_create_desc_rings()
1835 page_pool = pdata->rx_ring[i]->page_pool; in xgene_enet_init_hw()
1869 page_pool = pdata->rx_ring[0]->page_pool; in xgene_enet_init_hw()
[all …]
/drivers/net/vmxnet3/
A Dvmxnet3_xdp.c282 page_pool_recycle_direct(rq->page_pool, page); in vmxnet3_run_xdp()
290 page_pool_recycle_direct(rq->page_pool, page); in vmxnet3_run_xdp()
307 page_pool_recycle_direct(rq->page_pool, page); in vmxnet3_run_xdp()
320 page_pool_recycle_direct(rq->page_pool, page); in vmxnet3_build_skb()
345 page = page_pool_alloc_pages(rq->page_pool, GFP_ATOMIC); in vmxnet3_process_xdp_small()
352 xdp_prepare_buff(&xdp, page_address(page), rq->page_pool->p.offset, in vmxnet3_process_xdp_small()
395 rq->page_pool->p.offset, rbi->len, in vmxnet3_process_xdp()
396 page_pool_get_dma_dir(rq->page_pool)); in vmxnet3_process_xdp()
399 xdp_prepare_buff(&xdp, page_address(page), rq->page_pool->p.offset, in vmxnet3_process_xdp()
417 new_data = vmxnet3_pp_get_buff(rq->page_pool, &new_dma_addr, in vmxnet3_process_xdp()
A Dvmxnet3_xdp.h39 void *vmxnet3_pp_get_buff(struct page_pool *pp, dma_addr_t *dma_addr,
/drivers/net/ethernet/microchip/lan966x/
A Dlan966x_fdma.c16 page = page_pool_dev_alloc_pages(rx->page_pool); in lan966x_fdma_rx_dataptr_cb()
58 page_pool_put_full_page(rx->page_pool, in lan966x_fdma_rx_free_pages()
93 rx->page_pool = page_pool_create(&pp_params); in lan966x_fdma_rx_alloc_page_pool()
104 rx->page_pool); in lan966x_fdma_rx_alloc_page_pool()
107 return PTR_ERR_OR_ZERO(rx->page_pool); in lan966x_fdma_rx_alloc_page_pool()
117 return PTR_ERR(rx->page_pool); in lan966x_fdma_rx_alloc()
376 page_pool_recycle_direct(rx->page_pool, in lan966x_fdma_tx_clear_buf()
812 struct page_pool *page_pool; in lan966x_fdma_reload() local
818 page_pool = lan966x->rx.page_pool; in lan966x_fdma_reload()
835 page_pool_destroy(page_pool); in lan966x_fdma_reload()
[all …]
/drivers/net/ethernet/google/gve/
A Dgve_buffer_mgmt_dqo.c229 netmem = page_pool_alloc_netmem(rx->dqo.page_pool, in gve_alloc_from_page_pool()
240 buf_state->page_info.pad = rx->dqo.page_pool->p.offset; in gve_alloc_from_page_pool()
245 struct page_pool *gve_rx_create_page_pool(struct gve_priv *priv, in gve_rx_create_page_pool()
269 if (rx->dqo.page_pool) { in gve_free_buffer()
281 if (rx->dqo.page_pool) { in gve_reuse_buffer()
311 } else if (rx->dqo.page_pool) { in gve_alloc_buffer()
A Dgve_rx_dqo.c100 if (rx->dqo.page_pool) in gve_rx_reset_ring_dqo()
119 if (rx->dqo.page_pool) in gve_rx_stop_ring_dqo()
149 if (rx->dqo.page_pool) in gve_rx_free_ring_dqo()
183 if (rx->dqo.page_pool) { in gve_rx_free_ring_dqo()
185 rx->dqo.page_pool = NULL; in gve_rx_free_ring_dqo()
220 struct page_pool *pool; in gve_rx_alloc_ring_dqo()
522 if (rx->dqo.page_pool) { in gve_skb_add_rx_frag()
554 if (rx->dqo.page_pool) in gve_rx_append_frags()
766 if (rx->dqo.page_pool) { in gve_rx_dqo()
786 if (rx->dqo.page_pool) in gve_rx_dqo()
[all …]
/drivers/net/ethernet/microchip/sparx5/lan969x/
A Dlan969x_fdma.c31 page = page_pool_dev_alloc_pages(rx->page_pool); in lan969x_fdma_rx_dataptr_cb()
95 page_pool_put_full_page(rx->page_pool, in lan969x_fdma_free_pages()
145 page_pool_recycle_direct(rx->page_pool, page); in lan969x_fdma_rx_get_frame()
168 rx->page_pool = page_pool_create(&pp_params); in lan969x_fdma_rx_alloc()
169 if (IS_ERR(rx->page_pool)) in lan969x_fdma_rx_alloc()
170 return PTR_ERR(rx->page_pool); in lan969x_fdma_rx_alloc()
403 page_pool_destroy(rx->page_pool); in lan969x_fdma_deinit()
/drivers/net/ethernet/huawei/hinic3/
A Dhinic3_rx.c47 static int rx_alloc_mapped_page(struct page_pool *page_pool, in rx_alloc_mapped_page() argument
53 page = page_pool_dev_alloc_frag(page_pool, &page_offset, buf_len); in rx_alloc_mapped_page()
83 err = rx_alloc_mapped_page(rxq->page_pool, rx_info, in hinic3_rx_fill_buffers()
116 page_pool_dma_sync_for_cpu(rxq->page_pool, page, rx_info->page_offset, in hinic3_add_rx_frag()
122 page_pool_put_full_page(rxq->page_pool, page, false); in hinic3_add_rx_frag()
A Dhinic3_rx.h73 struct page_pool *page_pool; member
/drivers/net/netdevsim/
A Dnetdev.c455 struct page_pool *pool; in nsim_create_page_pool()
480 err = nsim_create_page_pool(&rq->page_pool, &rq->napi); in nsim_init_napi()
489 page_pool_destroy(ns->rq[i]->page_pool); in nsim_init_napi()
490 ns->rq[i]->page_pool = NULL; in nsim_init_napi()
558 page_pool_destroy(ns->rq[i]->page_pool); in nsim_del_napi()
559 ns->rq[i]->page_pool = NULL; in nsim_del_napi()
732 struct page_pool *pp;
779 page_pool_destroy(qmem->rq->page_pool); in nsim_queue_mem_free()
793 ns->rq[idx]->page_pool = qmem->pp; in nsim_queue_start()
827 qmem->pp = ns->rq[idx]->page_pool; in nsim_queue_stop()
[all …]
/drivers/net/ethernet/mellanox/mlxsw/
A Dpci.c89 struct page_pool *page_pool; member
399 struct page_pool *page_pool; in mlxsw_pci_rdq_build_skb() local
410 page_pool = cq->u.cq.page_pool; in mlxsw_pci_rdq_build_skb()
411 page_pool_dma_sync_for_cpu(page_pool, pages[page_index], in mlxsw_pci_rdq_build_skb()
454 page = page_pool_dev_alloc_pages(cq->u.cq.page_pool); in mlxsw_pci_rdq_page_alloc()
997 struct page_pool *page_pool; in mlxsw_pci_cq_page_pool_init() local
1010 page_pool = page_pool_create(&pp_params); in mlxsw_pci_cq_page_pool_init()
1011 if (IS_ERR(page_pool)) in mlxsw_pci_cq_page_pool_init()
1012 return PTR_ERR(page_pool); in mlxsw_pci_cq_page_pool_init()
1014 q->u.cq.page_pool = page_pool; in mlxsw_pci_cq_page_pool_init()
[all …]
/drivers/net/ethernet/socionext/
A Dnetsec.c286 struct page_pool *page_pool; member
742 page = page_pool_dev_alloc_pages(dring->page_pool); in netsec_alloc_rx_data()
863 page_pool_get_dma_dir(rx_ring->page_pool); in netsec_xdp_queue_one()
966 dma_dir = page_pool_get_dma_dir(dring->page_pool); in netsec_process_rx()
1041 page_pool_put_page(dring->page_pool, page, pkt_len, in netsec_process_rx()
1230 page_pool_destroy(dring->page_pool); in netsec_uninit_pkt_dring()
1313 dring->page_pool = page_pool_create(&pp_params); in netsec_setup_rx_dring()
1314 if (IS_ERR(dring->page_pool)) { in netsec_setup_rx_dring()
1315 err = PTR_ERR(dring->page_pool); in netsec_setup_rx_dring()
1316 dring->page_pool = NULL; in netsec_setup_rx_dring()
[all …]
/drivers/net/
A Dveth.c71 struct page_pool *page_pool; member
189 if (!priv->rq[i].page_pool) in veth_get_page_pool_stats()
191 page_pool_get_stats(priv->rq[i].page_pool, &pp_stats); in veth_get_page_pool_stats()
1011 rq->page_pool = page_pool_create(&pp_params); in veth_create_page_pool()
1012 if (IS_ERR(rq->page_pool)) { in veth_create_page_pool()
1013 int err = PTR_ERR(rq->page_pool); in veth_create_page_pool()
1015 rq->page_pool = NULL; in veth_create_page_pool()
1056 page_pool_destroy(priv->rq[i].page_pool); in __veth_napi_enable_range()
1057 priv->rq[i].page_pool = NULL; in __veth_napi_enable_range()
1090 page_pool_destroy(priv->rq[i].page_pool); in veth_napi_del_range()
[all …]
/drivers/net/ethernet/ti/
A Dcpsw_priv.c1086 struct page_pool *pool; in cpsw_fill_rx_channels()
1093 pool = cpsw->page_pool[ch]; in cpsw_fill_rx_channels()
1127 static struct page_pool *cpsw_create_page_pool(struct cpsw_common *cpsw, in cpsw_create_page_pool()
1131 struct page_pool *pool; in cpsw_create_page_pool()
1149 struct page_pool *pool; in cpsw_create_rx_pool()
1157 cpsw->page_pool[ch] = pool; in cpsw_create_rx_pool()
1166 struct page_pool *pool; in cpsw_ndev_create_xdp_rxq()
1169 pool = cpsw->page_pool[ch]; in cpsw_ndev_create_xdp_rxq()
1207 page_pool_destroy(cpsw->page_pool[ch]); in cpsw_destroy_xdp_rxqs()
1208 cpsw->page_pool[ch] = NULL; in cpsw_destroy_xdp_rxqs()
[all …]
A Dam65-cpsw-nuss.h103 struct page_pool *page_pool; member
/drivers/net/ethernet/meta/fbnic/
A Dfbnic_txrx.h132 struct page_pool *page_pool; member
/drivers/net/ethernet/marvell/octeontx2/nic/
A Dotx2_txrx.h133 struct page_pool *page_pool; member
/drivers/net/ethernet/hisilicon/hns3/
A Dhns3_debugfs.c778 READ_ONCE(ring->page_pool->pages_state_hold_cnt), in hns3_dump_page_pool_info()
779 atomic_read(&ring->page_pool->pages_state_release_cnt), in hns3_dump_page_pool_info()
780 ring->page_pool->p.pool_size, in hns3_dump_page_pool_info()
781 ring->page_pool->p.order, in hns3_dump_page_pool_info()
782 ring->page_pool->p.nid, in hns3_dump_page_pool_info()
783 ring->page_pool->p.max_len / 1024); in hns3_dump_page_pool_info()
798 if (!priv->ring[h->kinfo.num_tqps].page_pool) { in hns3_dbg_page_pool_info()
/drivers/net/ethernet/broadcom/bnxt/
A Dbnxt_xdp.c157 page_pool_recycle_direct(rxr->page_pool, tx_buf->page); in bnxt_tx_int_xdp()
215 page_pool_recycle_direct(rxr->page_pool, page); in bnxt_xdp_buff_frags_free()
311 page_pool_recycle_direct(rxr->page_pool, page); in bnxt_rx_xdp()
464 struct page_pool *pool, struct xdp_buff *xdp) in bnxt_xdp_build_skb()
/drivers/net/ethernet/engleder/
A Dtsnep.h129 struct page_pool *page_pool; member
A Dtsnep_main.c968 if (rx->page_pool) in tsnep_rx_ring_cleanup()
969 page_pool_destroy(rx->page_pool); in tsnep_rx_ring_cleanup()
1018 rx->page_pool = page_pool_create(&pp_params); in tsnep_rx_ring_create()
1019 if (IS_ERR(rx->page_pool)) { in tsnep_rx_ring_create()
1020 retval = PTR_ERR(rx->page_pool); in tsnep_rx_ring_create()
1021 rx->page_pool = NULL; in tsnep_rx_ring_create()
1125 page = page_pool_dev_alloc_pages(rx->page_pool); in tsnep_rx_alloc_buffer()
1428 page_pool_recycle_direct(rx->page_pool, page); in tsnep_rx_page()
1450 dma_dir = page_pool_get_dma_dir(rx->page_pool); in tsnep_rx_poll()
1627 page = page_pool_dev_alloc_pages(rx->page_pool); in tsnep_rx_poll_zc()
[all …]
/drivers/net/ethernet/tehuti/
A Dtn40.h146 struct page_pool *page_pool; member
/drivers/net/ethernet/realtek/rtase/
A Drtase_main.c378 page = page_pool_dev_alloc_pages(tp->page_pool); in rtase_alloc_rx_data_buf()
425 static void rtase_rx_ring_clear(struct page_pool *page_pool, in rtase_rx_ring_clear() argument
437 page_pool_put_full_page(page_pool, page, true); in rtase_rx_ring_clear()
615 rtase_rx_ring_clear(tp->page_pool, &tp->rx_ring[i]); in rtase_rx_clear()
617 page_pool_destroy(tp->page_pool); in rtase_rx_clear()
618 tp->page_pool = NULL; in rtase_rx_clear()
625 struct page_pool *page_pool; in rtase_init_ring() local
638 page_pool = page_pool_create(&pp_params); in rtase_init_ring()
639 if (IS_ERR(page_pool)) { in rtase_init_ring()
644 tp->page_pool = page_pool; in rtase_init_ring()
A Drtase.h329 struct page_pool *page_pool; member
/drivers/net/ethernet/marvell/mvpp2/
A Dmvpp2_main.c104 static struct page_pool *
361 struct page_pool *page_pool) in mvpp2_frag_alloc() argument
363 if (page_pool) in mvpp2_frag_alloc()
373 struct page_pool *page_pool, void *data) in mvpp2_frag_free() argument
375 if (page_pool) in mvpp2_frag_free()
495 struct page_pool *pp = NULL; in mvpp2_bm_bufs_free()
662 priv->page_pool[i] = in mvpp2_bm_init()
763 struct page_pool *page_pool, in mvpp2_buf_alloc() argument
776 if (page_pool) { in mvpp2_buf_alloc()
3591 struct page_pool *page_pool, int pool) in mvpp2_rx_refill() argument
[all …]

Completed in 81 milliseconds

1234