| /drivers/net/ethernet/apm/xgene/ |
| A D | xgene_enet_main.c | 690 page_pool = rx_ring->page_pool; in xgene_enet_rx_frame() 1081 page_pool = ring->page_pool; in xgene_enet_delete_desc_rings() 1082 if (page_pool) { in xgene_enet_delete_desc_rings() 1179 page_pool = ring->page_pool; in xgene_enet_free_desc_rings() 1180 if (page_pool) { in xgene_enet_free_desc_rings() 1373 if (!page_pool) { in xgene_enet_create_desc_rings() 1395 page_pool->dst_ring_num = xgene_enet_dst_ring_num(page_pool); in xgene_enet_create_desc_rings() 1396 rx_ring->page_pool = page_pool; in xgene_enet_create_desc_rings() 1835 page_pool = pdata->rx_ring[i]->page_pool; in xgene_enet_init_hw() 1869 page_pool = pdata->rx_ring[0]->page_pool; in xgene_enet_init_hw() [all …]
|
| /drivers/net/vmxnet3/ |
| A D | vmxnet3_xdp.c | 282 page_pool_recycle_direct(rq->page_pool, page); in vmxnet3_run_xdp() 290 page_pool_recycle_direct(rq->page_pool, page); in vmxnet3_run_xdp() 307 page_pool_recycle_direct(rq->page_pool, page); in vmxnet3_run_xdp() 320 page_pool_recycle_direct(rq->page_pool, page); in vmxnet3_build_skb() 345 page = page_pool_alloc_pages(rq->page_pool, GFP_ATOMIC); in vmxnet3_process_xdp_small() 352 xdp_prepare_buff(&xdp, page_address(page), rq->page_pool->p.offset, in vmxnet3_process_xdp_small() 395 rq->page_pool->p.offset, rbi->len, in vmxnet3_process_xdp() 396 page_pool_get_dma_dir(rq->page_pool)); in vmxnet3_process_xdp() 399 xdp_prepare_buff(&xdp, page_address(page), rq->page_pool->p.offset, in vmxnet3_process_xdp() 417 new_data = vmxnet3_pp_get_buff(rq->page_pool, &new_dma_addr, in vmxnet3_process_xdp()
|
| A D | vmxnet3_xdp.h | 39 void *vmxnet3_pp_get_buff(struct page_pool *pp, dma_addr_t *dma_addr,
|
| /drivers/net/ethernet/microchip/lan966x/ |
| A D | lan966x_fdma.c | 16 page = page_pool_dev_alloc_pages(rx->page_pool); in lan966x_fdma_rx_dataptr_cb() 58 page_pool_put_full_page(rx->page_pool, in lan966x_fdma_rx_free_pages() 93 rx->page_pool = page_pool_create(&pp_params); in lan966x_fdma_rx_alloc_page_pool() 104 rx->page_pool); in lan966x_fdma_rx_alloc_page_pool() 107 return PTR_ERR_OR_ZERO(rx->page_pool); in lan966x_fdma_rx_alloc_page_pool() 117 return PTR_ERR(rx->page_pool); in lan966x_fdma_rx_alloc() 376 page_pool_recycle_direct(rx->page_pool, in lan966x_fdma_tx_clear_buf() 812 struct page_pool *page_pool; in lan966x_fdma_reload() local 818 page_pool = lan966x->rx.page_pool; in lan966x_fdma_reload() 835 page_pool_destroy(page_pool); in lan966x_fdma_reload() [all …]
|
| /drivers/net/ethernet/google/gve/ |
| A D | gve_buffer_mgmt_dqo.c | 229 netmem = page_pool_alloc_netmem(rx->dqo.page_pool, in gve_alloc_from_page_pool() 240 buf_state->page_info.pad = rx->dqo.page_pool->p.offset; in gve_alloc_from_page_pool() 245 struct page_pool *gve_rx_create_page_pool(struct gve_priv *priv, in gve_rx_create_page_pool() 269 if (rx->dqo.page_pool) { in gve_free_buffer() 281 if (rx->dqo.page_pool) { in gve_reuse_buffer() 311 } else if (rx->dqo.page_pool) { in gve_alloc_buffer()
|
| A D | gve_rx_dqo.c | 100 if (rx->dqo.page_pool) in gve_rx_reset_ring_dqo() 119 if (rx->dqo.page_pool) in gve_rx_stop_ring_dqo() 149 if (rx->dqo.page_pool) in gve_rx_free_ring_dqo() 183 if (rx->dqo.page_pool) { in gve_rx_free_ring_dqo() 185 rx->dqo.page_pool = NULL; in gve_rx_free_ring_dqo() 220 struct page_pool *pool; in gve_rx_alloc_ring_dqo() 522 if (rx->dqo.page_pool) { in gve_skb_add_rx_frag() 554 if (rx->dqo.page_pool) in gve_rx_append_frags() 766 if (rx->dqo.page_pool) { in gve_rx_dqo() 786 if (rx->dqo.page_pool) in gve_rx_dqo() [all …]
|
| /drivers/net/ethernet/microchip/sparx5/lan969x/ |
| A D | lan969x_fdma.c | 31 page = page_pool_dev_alloc_pages(rx->page_pool); in lan969x_fdma_rx_dataptr_cb() 95 page_pool_put_full_page(rx->page_pool, in lan969x_fdma_free_pages() 145 page_pool_recycle_direct(rx->page_pool, page); in lan969x_fdma_rx_get_frame() 168 rx->page_pool = page_pool_create(&pp_params); in lan969x_fdma_rx_alloc() 169 if (IS_ERR(rx->page_pool)) in lan969x_fdma_rx_alloc() 170 return PTR_ERR(rx->page_pool); in lan969x_fdma_rx_alloc() 403 page_pool_destroy(rx->page_pool); in lan969x_fdma_deinit()
|
| /drivers/net/ethernet/huawei/hinic3/ |
| A D | hinic3_rx.c | 47 static int rx_alloc_mapped_page(struct page_pool *page_pool, in rx_alloc_mapped_page() argument 53 page = page_pool_dev_alloc_frag(page_pool, &page_offset, buf_len); in rx_alloc_mapped_page() 83 err = rx_alloc_mapped_page(rxq->page_pool, rx_info, in hinic3_rx_fill_buffers() 116 page_pool_dma_sync_for_cpu(rxq->page_pool, page, rx_info->page_offset, in hinic3_add_rx_frag() 122 page_pool_put_full_page(rxq->page_pool, page, false); in hinic3_add_rx_frag()
|
| A D | hinic3_rx.h | 73 struct page_pool *page_pool; member
|
| /drivers/net/netdevsim/ |
| A D | netdev.c | 455 struct page_pool *pool; in nsim_create_page_pool() 480 err = nsim_create_page_pool(&rq->page_pool, &rq->napi); in nsim_init_napi() 489 page_pool_destroy(ns->rq[i]->page_pool); in nsim_init_napi() 490 ns->rq[i]->page_pool = NULL; in nsim_init_napi() 558 page_pool_destroy(ns->rq[i]->page_pool); in nsim_del_napi() 559 ns->rq[i]->page_pool = NULL; in nsim_del_napi() 732 struct page_pool *pp; 779 page_pool_destroy(qmem->rq->page_pool); in nsim_queue_mem_free() 793 ns->rq[idx]->page_pool = qmem->pp; in nsim_queue_start() 827 qmem->pp = ns->rq[idx]->page_pool; in nsim_queue_stop() [all …]
|
| /drivers/net/ethernet/mellanox/mlxsw/ |
| A D | pci.c | 89 struct page_pool *page_pool; member 399 struct page_pool *page_pool; in mlxsw_pci_rdq_build_skb() local 410 page_pool = cq->u.cq.page_pool; in mlxsw_pci_rdq_build_skb() 411 page_pool_dma_sync_for_cpu(page_pool, pages[page_index], in mlxsw_pci_rdq_build_skb() 454 page = page_pool_dev_alloc_pages(cq->u.cq.page_pool); in mlxsw_pci_rdq_page_alloc() 997 struct page_pool *page_pool; in mlxsw_pci_cq_page_pool_init() local 1010 page_pool = page_pool_create(&pp_params); in mlxsw_pci_cq_page_pool_init() 1011 if (IS_ERR(page_pool)) in mlxsw_pci_cq_page_pool_init() 1012 return PTR_ERR(page_pool); in mlxsw_pci_cq_page_pool_init() 1014 q->u.cq.page_pool = page_pool; in mlxsw_pci_cq_page_pool_init() [all …]
|
| /drivers/net/ethernet/socionext/ |
| A D | netsec.c | 286 struct page_pool *page_pool; member 742 page = page_pool_dev_alloc_pages(dring->page_pool); in netsec_alloc_rx_data() 863 page_pool_get_dma_dir(rx_ring->page_pool); in netsec_xdp_queue_one() 966 dma_dir = page_pool_get_dma_dir(dring->page_pool); in netsec_process_rx() 1041 page_pool_put_page(dring->page_pool, page, pkt_len, in netsec_process_rx() 1230 page_pool_destroy(dring->page_pool); in netsec_uninit_pkt_dring() 1313 dring->page_pool = page_pool_create(&pp_params); in netsec_setup_rx_dring() 1314 if (IS_ERR(dring->page_pool)) { in netsec_setup_rx_dring() 1315 err = PTR_ERR(dring->page_pool); in netsec_setup_rx_dring() 1316 dring->page_pool = NULL; in netsec_setup_rx_dring() [all …]
|
| /drivers/net/ |
| A D | veth.c | 71 struct page_pool *page_pool; member 189 if (!priv->rq[i].page_pool) in veth_get_page_pool_stats() 191 page_pool_get_stats(priv->rq[i].page_pool, &pp_stats); in veth_get_page_pool_stats() 1011 rq->page_pool = page_pool_create(&pp_params); in veth_create_page_pool() 1012 if (IS_ERR(rq->page_pool)) { in veth_create_page_pool() 1013 int err = PTR_ERR(rq->page_pool); in veth_create_page_pool() 1015 rq->page_pool = NULL; in veth_create_page_pool() 1056 page_pool_destroy(priv->rq[i].page_pool); in __veth_napi_enable_range() 1057 priv->rq[i].page_pool = NULL; in __veth_napi_enable_range() 1090 page_pool_destroy(priv->rq[i].page_pool); in veth_napi_del_range() [all …]
|
| /drivers/net/ethernet/ti/ |
| A D | cpsw_priv.c | 1086 struct page_pool *pool; in cpsw_fill_rx_channels() 1093 pool = cpsw->page_pool[ch]; in cpsw_fill_rx_channels() 1127 static struct page_pool *cpsw_create_page_pool(struct cpsw_common *cpsw, in cpsw_create_page_pool() 1131 struct page_pool *pool; in cpsw_create_page_pool() 1149 struct page_pool *pool; in cpsw_create_rx_pool() 1157 cpsw->page_pool[ch] = pool; in cpsw_create_rx_pool() 1166 struct page_pool *pool; in cpsw_ndev_create_xdp_rxq() 1169 pool = cpsw->page_pool[ch]; in cpsw_ndev_create_xdp_rxq() 1207 page_pool_destroy(cpsw->page_pool[ch]); in cpsw_destroy_xdp_rxqs() 1208 cpsw->page_pool[ch] = NULL; in cpsw_destroy_xdp_rxqs() [all …]
|
| A D | am65-cpsw-nuss.h | 103 struct page_pool *page_pool; member
|
| /drivers/net/ethernet/meta/fbnic/ |
| A D | fbnic_txrx.h | 132 struct page_pool *page_pool; member
|
| /drivers/net/ethernet/marvell/octeontx2/nic/ |
| A D | otx2_txrx.h | 133 struct page_pool *page_pool; member
|
| /drivers/net/ethernet/hisilicon/hns3/ |
| A D | hns3_debugfs.c | 778 READ_ONCE(ring->page_pool->pages_state_hold_cnt), in hns3_dump_page_pool_info() 779 atomic_read(&ring->page_pool->pages_state_release_cnt), in hns3_dump_page_pool_info() 780 ring->page_pool->p.pool_size, in hns3_dump_page_pool_info() 781 ring->page_pool->p.order, in hns3_dump_page_pool_info() 782 ring->page_pool->p.nid, in hns3_dump_page_pool_info() 783 ring->page_pool->p.max_len / 1024); in hns3_dump_page_pool_info() 798 if (!priv->ring[h->kinfo.num_tqps].page_pool) { in hns3_dbg_page_pool_info()
|
| /drivers/net/ethernet/broadcom/bnxt/ |
| A D | bnxt_xdp.c | 157 page_pool_recycle_direct(rxr->page_pool, tx_buf->page); in bnxt_tx_int_xdp() 215 page_pool_recycle_direct(rxr->page_pool, page); in bnxt_xdp_buff_frags_free() 311 page_pool_recycle_direct(rxr->page_pool, page); in bnxt_rx_xdp() 464 struct page_pool *pool, struct xdp_buff *xdp) in bnxt_xdp_build_skb()
|
| /drivers/net/ethernet/engleder/ |
| A D | tsnep.h | 129 struct page_pool *page_pool; member
|
| A D | tsnep_main.c | 968 if (rx->page_pool) in tsnep_rx_ring_cleanup() 969 page_pool_destroy(rx->page_pool); in tsnep_rx_ring_cleanup() 1018 rx->page_pool = page_pool_create(&pp_params); in tsnep_rx_ring_create() 1019 if (IS_ERR(rx->page_pool)) { in tsnep_rx_ring_create() 1020 retval = PTR_ERR(rx->page_pool); in tsnep_rx_ring_create() 1021 rx->page_pool = NULL; in tsnep_rx_ring_create() 1125 page = page_pool_dev_alloc_pages(rx->page_pool); in tsnep_rx_alloc_buffer() 1428 page_pool_recycle_direct(rx->page_pool, page); in tsnep_rx_page() 1450 dma_dir = page_pool_get_dma_dir(rx->page_pool); in tsnep_rx_poll() 1627 page = page_pool_dev_alloc_pages(rx->page_pool); in tsnep_rx_poll_zc() [all …]
|
| /drivers/net/ethernet/tehuti/ |
| A D | tn40.h | 146 struct page_pool *page_pool; member
|
| /drivers/net/ethernet/realtek/rtase/ |
| A D | rtase_main.c | 378 page = page_pool_dev_alloc_pages(tp->page_pool); in rtase_alloc_rx_data_buf() 425 static void rtase_rx_ring_clear(struct page_pool *page_pool, in rtase_rx_ring_clear() argument 437 page_pool_put_full_page(page_pool, page, true); in rtase_rx_ring_clear() 615 rtase_rx_ring_clear(tp->page_pool, &tp->rx_ring[i]); in rtase_rx_clear() 617 page_pool_destroy(tp->page_pool); in rtase_rx_clear() 618 tp->page_pool = NULL; in rtase_rx_clear() 625 struct page_pool *page_pool; in rtase_init_ring() local 638 page_pool = page_pool_create(&pp_params); in rtase_init_ring() 639 if (IS_ERR(page_pool)) { in rtase_init_ring() 644 tp->page_pool = page_pool; in rtase_init_ring()
|
| A D | rtase.h | 329 struct page_pool *page_pool; member
|
| /drivers/net/ethernet/marvell/mvpp2/ |
| A D | mvpp2_main.c | 104 static struct page_pool * 361 struct page_pool *page_pool) in mvpp2_frag_alloc() argument 363 if (page_pool) in mvpp2_frag_alloc() 373 struct page_pool *page_pool, void *data) in mvpp2_frag_free() argument 375 if (page_pool) in mvpp2_frag_free() 495 struct page_pool *pp = NULL; in mvpp2_bm_bufs_free() 662 priv->page_pool[i] = in mvpp2_bm_init() 763 struct page_pool *page_pool, in mvpp2_buf_alloc() argument 776 if (page_pool) { in mvpp2_buf_alloc() 3591 struct page_pool *page_pool, int pool) in mvpp2_rx_refill() argument [all …]
|