Lines Matching refs:netdev

154 		err = pool->netdev->netdev_ops->ndo_bpf(pool->netdev, &bpf);  in xp_disable_drv_zc()
166 struct net_device *netdev, u16 queue_id, u16 flags) in xp_assign_dev() argument
180 if (xsk_get_pool_from_qid(netdev, queue_id)) in xp_assign_dev()
183 pool->netdev = netdev; in xp_assign_dev()
185 err = xsk_reg_pool_at_qid(netdev, pool, queue_id); in xp_assign_dev()
200 dev_hold(netdev); in xp_assign_dev()
206 if ((netdev->xdp_features & NETDEV_XDP_ACT_ZC) != NETDEV_XDP_ACT_ZC) { in xp_assign_dev()
211 if (netdev->xdp_zc_max_segs == 1 && (flags & XDP_USE_SG)) { in xp_assign_dev()
216 if (dev_get_min_mp_channel_count(netdev)) { in xp_assign_dev()
225 netdev_ops_assert_locked(netdev); in xp_assign_dev()
226 err = netdev->netdev_ops->ndo_bpf(netdev, &bpf); in xp_assign_dev()
236 pool->xdp_zc_max_segs = netdev->xdp_zc_max_segs; in xp_assign_dev()
245 xsk_clear_pool_at_qid(netdev, queue_id); in xp_assign_dev()
246 dev_put(netdev); in xp_assign_dev()
270 struct net_device *netdev = pool->netdev; in xp_clear_dev() local
272 if (!pool->netdev) in xp_clear_dev()
275 netdev_lock_ops(netdev); in xp_clear_dev()
277 xsk_clear_pool_at_qid(pool->netdev, pool->queue_id); in xp_clear_dev()
278 pool->netdev = NULL; in xp_clear_dev()
279 netdev_unlock_ops(netdev); in xp_clear_dev()
280 dev_put(netdev); in xp_clear_dev()
330 if (dma_map->netdev == pool->netdev) in xp_find_dma_map()
337 static struct xsk_dma_map *xp_create_dma_map(struct device *dev, struct net_device *netdev, in xp_create_dma_map() argument
352 dma_map->netdev = netdev; in xp_create_dma_map()
464 dma_map = xp_create_dma_map(dev, pool->netdev, nr_pages, pool->umem); in xp_dma_map()