Lines Matching refs:net_dev

202 static int dpaa_netdev_init(struct net_device *net_dev,  in dpaa_netdev_init()  argument
206 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_netdev_init()
207 struct device *dev = net_dev->dev.parent; in dpaa_netdev_init()
218 percpu_priv->net_dev = net_dev; in dpaa_netdev_init()
221 net_dev->netdev_ops = dpaa_ops; in dpaa_netdev_init()
224 net_dev->mem_start = (unsigned long)priv->mac_dev->res->start; in dpaa_netdev_init()
225 net_dev->mem_end = (unsigned long)priv->mac_dev->res->end; in dpaa_netdev_init()
227 net_dev->min_mtu = ETH_MIN_MTU; in dpaa_netdev_init()
228 net_dev->max_mtu = dpaa_get_max_mtu(); in dpaa_netdev_init()
230 net_dev->hw_features |= (NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM | in dpaa_netdev_init()
233 net_dev->hw_features |= NETIF_F_SG | NETIF_F_HIGHDMA; in dpaa_netdev_init()
237 net_dev->features |= NETIF_F_GSO; in dpaa_netdev_init()
238 net_dev->features |= NETIF_F_RXCSUM; in dpaa_netdev_init()
240 net_dev->priv_flags |= IFF_LIVE_ADDR_CHANGE; in dpaa_netdev_init()
242 net_dev->priv_flags &= ~IFF_TX_SKB_SHARING; in dpaa_netdev_init()
244 net_dev->features |= net_dev->hw_features; in dpaa_netdev_init()
245 net_dev->vlan_features = net_dev->features; in dpaa_netdev_init()
247 net_dev->xdp_features = NETDEV_XDP_ACT_BASIC | in dpaa_netdev_init()
252 memcpy(net_dev->perm_addr, mac_addr, net_dev->addr_len); in dpaa_netdev_init()
253 eth_hw_addr_set(net_dev, mac_addr); in dpaa_netdev_init()
255 eth_hw_addr_random(net_dev); in dpaa_netdev_init()
257 (const enet_addr_t *)net_dev->dev_addr); in dpaa_netdev_init()
263 net_dev->dev_addr); in dpaa_netdev_init()
266 net_dev->ethtool_ops = &dpaa_ethtool_ops; in dpaa_netdev_init()
268 net_dev->needed_headroom = priv->tx_headroom; in dpaa_netdev_init()
269 net_dev->watchdog_timeo = msecs_to_jiffies(tx_timeout); in dpaa_netdev_init()
272 mac_dev->phylink_config.dev = &net_dev->dev; in dpaa_netdev_init()
286 netif_carrier_off(net_dev); in dpaa_netdev_init()
288 err = register_netdev(net_dev); in dpaa_netdev_init()
298 static int dpaa_stop(struct net_device *net_dev) in dpaa_stop() argument
304 priv = netdev_priv(net_dev); in dpaa_stop()
307 netif_tx_stop_all_queues(net_dev); in dpaa_stop()
323 net_dev->phydev = NULL; in dpaa_stop()
330 static void dpaa_tx_timeout(struct net_device *net_dev, unsigned int txqueue) in dpaa_tx_timeout() argument
335 priv = netdev_priv(net_dev); in dpaa_tx_timeout()
338 netif_crit(priv, timer, net_dev, "Transmit timeout latency: %u ms\n", in dpaa_tx_timeout()
339 jiffies_to_msecs(jiffies - dev_trans_start(net_dev))); in dpaa_tx_timeout()
347 static void dpaa_get_stats64(struct net_device *net_dev, in dpaa_get_stats64() argument
351 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_get_stats64()
368 static int dpaa_setup_tc(struct net_device *net_dev, enum tc_setup_type type, in dpaa_setup_tc() argument
371 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_setup_tc()
386 netdev_reset_tc(net_dev); in dpaa_setup_tc()
391 netdev_err(net_dev, "Too many traffic classes: max %d supported.\n", in dpaa_setup_tc()
396 netdev_set_num_tc(net_dev, num_tc); in dpaa_setup_tc()
399 netdev_set_tc_queue(net_dev, i, DPAA_TC_TXQ_NUM, in dpaa_setup_tc()
404 netif_set_real_num_tx_queues(net_dev, priv->num_tc * DPAA_TC_TXQ_NUM); in dpaa_setup_tc()
429 static int dpaa_set_mac_address(struct net_device *net_dev, void *addr) in dpaa_set_mac_address() argument
436 priv = netdev_priv(net_dev); in dpaa_set_mac_address()
438 memcpy(old_addr.sa_data, net_dev->dev_addr, ETH_ALEN); in dpaa_set_mac_address()
440 err = eth_mac_addr(net_dev, addr); in dpaa_set_mac_address()
442 netif_err(priv, drv, net_dev, "eth_mac_addr() = %d\n", err); in dpaa_set_mac_address()
449 (const enet_addr_t *)net_dev->dev_addr); in dpaa_set_mac_address()
451 netif_err(priv, drv, net_dev, "mac_dev->change_addr() = %d\n", in dpaa_set_mac_address()
454 eth_mac_addr(net_dev, &old_addr); in dpaa_set_mac_address()
462 static void dpaa_set_rx_mode(struct net_device *net_dev) in dpaa_set_rx_mode() argument
467 priv = netdev_priv(net_dev); in dpaa_set_rx_mode()
469 if (!!(net_dev->flags & IFF_PROMISC) != priv->mac_dev->promisc) { in dpaa_set_rx_mode()
474 netif_err(priv, drv, net_dev, in dpaa_set_rx_mode()
479 if (!!(net_dev->flags & IFF_ALLMULTI) != priv->mac_dev->allmulti) { in dpaa_set_rx_mode()
484 netif_err(priv, drv, net_dev, in dpaa_set_rx_mode()
489 err = priv->mac_dev->set_multi(net_dev, priv->mac_dev); in dpaa_set_rx_mode()
491 netif_err(priv, drv, net_dev, "mac_dev->set_multi() = %d\n", in dpaa_set_rx_mode()
819 netif_tx_stop_all_queues(priv->net_dev); in dpaa_eth_cgscn()
824 netif_tx_wake_all_queues(priv->net_dev); in dpaa_eth_cgscn()
882 struct net_device *net_dev = to_net_dev(mac_dev->phylink_config.dev); in dpaa_eth_cgr_set_speed() local
883 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_eth_cgr_set_speed()
902 netdev_err(net_dev, "could not update speed: %d\n", err); in dpaa_eth_cgr_set_speed()
910 fq->net_dev = priv->net_dev; in dpaa_setup_ingress()
922 fq->net_dev = priv->net_dev; in dpaa_setup_egress()
945 dev_err(priv->net_dev->dev.parent, in dpaa_fq_setup()
982 dev_warn(priv->net_dev->dev.parent, in dpaa_fq_setup()
1022 priv = netdev_priv(dpaa_fq->net_dev); in dpaa_fq_init()
1023 dev = dpaa_fq->net_dev->dev.parent; in dpaa_fq_init()
1154 err = xdp_rxq_info_reg(&dpaa_fq->xdp_rxq, dpaa_fq->net_dev, in dpaa_fq_init()
1183 priv = netdev_priv(dpaa_fq->net_dev); in dpaa_fq_free_entry()
1381 static void dpaa_fd_release(const struct net_device *net_dev, in dpaa_fd_release() argument
1410 netdev_err(net_dev, "DMA mapping failed\n"); in dpaa_fd_release()
1512 netif_alert(priv, tx_err, priv->net_dev, in dpaa_enable_tx_csum()
1529 netif_alert(priv, tx_err, priv->net_dev, in dpaa_enable_tx_csum()
1555 struct net_device *net_dev = dpaa_bp->priv->net_dev; in dpaa_bp_add_8_bufs() local
1564 netdev_err(net_dev, "dev_alloc_pages() failed\n"); in dpaa_bp_add_8_bufs()
1572 netdev_err(net_dev, "DMA map failed\n"); in dpaa_bp_add_8_bufs()
1675 struct device *dev = priv->net_dev->dev.parent; in dpaa_cleanup_tx_fd()
1751 if ((priv->net_dev->features & NETIF_F_RXCSUM) && in rx_csum_offload()
1933 struct net_device *net_dev = priv->net_dev; in skb_to_contig_fd() local
1959 netif_err(priv, tx_err, net_dev, "HW csum error: %d\n", in skb_to_contig_fd()
1973 netif_err(priv, tx_err, net_dev, "dma_map_single() failed\n"); in skb_to_contig_fd()
1986 struct net_device *net_dev = priv->net_dev; in skb_to_sg_fd() local
1999 netdev_err(net_dev, "dev_alloc_pages() failed\n"); in skb_to_sg_fd()
2013 netif_err(priv, tx_err, net_dev, "HW csum error: %d\n", in skb_to_sg_fd()
2027 netdev_err(priv->net_dev, "DMA mapping failed\n"); in skb_to_sg_fd()
2041 netdev_err(priv->net_dev, "DMA mapping failed\n"); in skb_to_sg_fd()
2067 netdev_err(priv->net_dev, "DMA mapping failed\n"); in skb_to_sg_fd()
2103 trace_dpaa_tx_fd(priv->net_dev, egress_fq, fd); in dpaa_xmit()
2123 static int dpaa_a050385_wa_skb(struct net_device *net_dev, struct sk_buff **s) in dpaa_a050385_wa_skb() argument
2125 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_a050385_wa_skb()
2158 new_skb = netdev_alloc_skb(net_dev, skb->len + DPAA_A050385_ALIGN - 1 + in dpaa_a050385_wa_skb()
2280 dpaa_start_xmit(struct sk_buff *skb, struct net_device *net_dev) in dpaa_start_xmit() argument
2292 priv = netdev_priv(net_dev); in dpaa_start_xmit()
2327 if (dpaa_a050385_wa_skb(net_dev, &skb)) in dpaa_start_xmit()
2344 txq = netdev_get_tx_queue(net_dev, queue_mapping); in dpaa_start_xmit()
2365 static void dpaa_rx_error(struct net_device *net_dev, in dpaa_rx_error() argument
2372 netif_err(priv, hw, net_dev, "Err FD status = 0x%08x\n", in dpaa_rx_error()
2386 dpaa_fd_release(net_dev, fd); in dpaa_rx_error()
2389 static void dpaa_tx_error(struct net_device *net_dev, in dpaa_tx_error() argument
2398 netif_warn(priv, hw, net_dev, "FD status = 0x%08x\n", in dpaa_tx_error()
2430 static void dpaa_tx_conf(struct net_device *net_dev, in dpaa_tx_conf() argument
2440 netif_warn(priv, hw, net_dev, "FD status = 0x%08x\n", in dpaa_tx_conf()
2476 struct net_device *net_dev; in rx_error_dqrr() local
2480 net_dev = dpaa_fq->net_dev; in rx_error_dqrr()
2481 priv = netdev_priv(net_dev); in rx_error_dqrr()
2492 dpaa_rx_error(net_dev, priv, percpu_priv, &dq->fd, fq->fqid); in rx_error_dqrr()
2497 static int dpaa_xdp_xmit_frame(struct net_device *net_dev, in dpaa_xdp_xmit_frame() argument
2500 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_xdp_xmit_frame()
2552 txq = netdev_get_tx_queue(net_dev, smp_processor_id()); in dpaa_xdp_xmit_frame()
2629 if (dpaa_xdp_xmit_frame(priv->net_dev, xdpf)) in dpaa_run_xdp()
2638 err = xdp_do_redirect(priv->net_dev, &xdp, xdp_prog); in dpaa_run_xdp()
2640 trace_xdp_exception(priv->net_dev, xdp_prog, xdp_act); in dpaa_run_xdp()
2645 bpf_warn_invalid_xdp_action(priv->net_dev, xdp_prog, xdp_act); in dpaa_run_xdp()
2648 trace_xdp_exception(priv->net_dev, xdp_prog, xdp_act); in dpaa_run_xdp()
2673 struct net_device *net_dev; in rx_default_dqrr() local
2689 net_dev = dpaa_fq->net_dev; in rx_default_dqrr()
2690 priv = netdev_priv(net_dev); in rx_default_dqrr()
2696 trace_dpaa_rx_fd(net_dev, fq, &dq->fd); in rx_default_dqrr()
2711 dpaa_fd_release(net_dev, &dq->fd); in rx_default_dqrr()
2717 netif_warn(priv, hw, net_dev, "FD status = 0x%08x\n", in rx_default_dqrr()
2721 dpaa_fd_release(net_dev, fd); in rx_default_dqrr()
2750 if (net_dev->features & NETIF_F_RXHASH && priv->keygen_in_use && in rx_default_dqrr()
2793 skb->protocol = eth_type_trans(skb, net_dev); in rx_default_dqrr()
2824 struct net_device *net_dev; in conf_error_dqrr() local
2827 net_dev = ((struct dpaa_fq *)fq)->net_dev; in conf_error_dqrr()
2828 priv = netdev_priv(net_dev); in conf_error_dqrr()
2835 dpaa_tx_error(net_dev, priv, percpu_priv, &dq->fd, fq->fqid); in conf_error_dqrr()
2846 struct net_device *net_dev; in conf_dflt_dqrr() local
2849 net_dev = ((struct dpaa_fq *)fq)->net_dev; in conf_dflt_dqrr()
2850 priv = netdev_priv(net_dev); in conf_dflt_dqrr()
2853 trace_dpaa_tx_conf_fd(net_dev, fq, &dq->fd); in conf_dflt_dqrr()
2860 dpaa_tx_conf(net_dev, priv, percpu_priv, &dq->fd, fq->fqid); in conf_dflt_dqrr()
2872 struct net_device *net_dev; in egress_ern() local
2875 net_dev = ((struct dpaa_fq *)fq)->net_dev; in egress_ern()
2876 priv = netdev_priv(net_dev); in egress_ern()
2921 static int dpaa_open(struct net_device *net_dev) in dpaa_open() argument
2927 priv = netdev_priv(net_dev); in dpaa_open()
2944 netif_err(priv, ifup, net_dev, "mac_dev->enable() = %d\n", err); in dpaa_open()
2949 netif_tx_start_all_queues(net_dev); in dpaa_open()
2964 static int dpaa_eth_stop(struct net_device *net_dev) in dpaa_eth_stop() argument
2969 err = dpaa_stop(net_dev); in dpaa_eth_stop()
2971 priv = netdev_priv(net_dev); in dpaa_eth_stop()
2985 dev_warn(priv->net_dev->dev.parent, in xdp_validate_mtu()
2994 static int dpaa_change_mtu(struct net_device *net_dev, int new_mtu) in dpaa_change_mtu() argument
2996 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_change_mtu()
3001 net_dev->mtu = new_mtu; in dpaa_change_mtu()
3005 static int dpaa_setup_xdp(struct net_device *net_dev, struct netdev_bpf *bpf) in dpaa_setup_xdp() argument
3007 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_setup_xdp()
3013 if (bpf->prog && !xdp_validate_mtu(priv, net_dev->mtu)) { in dpaa_setup_xdp()
3018 up = netif_running(net_dev); in dpaa_setup_xdp()
3021 dpaa_eth_stop(net_dev); in dpaa_setup_xdp()
3028 err = dpaa_open(net_dev); in dpaa_setup_xdp()
3038 static int dpaa_xdp(struct net_device *net_dev, struct netdev_bpf *xdp) in dpaa_xdp() argument
3042 return dpaa_setup_xdp(net_dev, xdp); in dpaa_xdp()
3048 static int dpaa_xdp_xmit(struct net_device *net_dev, int n, in dpaa_xdp_xmit() argument
3057 if (!netif_running(net_dev)) in dpaa_xdp_xmit()
3062 if (dpaa_xdp_xmit_frame(net_dev, xdpf)) in dpaa_xdp_xmit()
3109 static int dpaa_ioctl(struct net_device *net_dev, struct ifreq *rq, int cmd) in dpaa_ioctl() argument
3112 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_ioctl()
3115 if (net_dev->phydev) in dpaa_ioctl()
3121 return dpaa_ts_ioctl(net_dev, rq, cmd); in dpaa_ioctl()
3143 static int dpaa_napi_add(struct net_device *net_dev) in dpaa_napi_add() argument
3145 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_napi_add()
3152 netif_napi_add(net_dev, &percpu_priv->np.napi, dpaa_eth_poll); in dpaa_napi_add()
3158 static void dpaa_napi_del(struct net_device *net_dev) in dpaa_napi_del() argument
3160 struct dpaa_priv *priv = netdev_priv(net_dev); in dpaa_napi_del()
3285 struct net_device *net_dev = NULL; in dpaa_eth_probe() local
3330 net_dev = alloc_etherdev_mq(sizeof(*priv), DPAA_ETH_TXQ_NUM); in dpaa_eth_probe()
3331 if (!net_dev) { in dpaa_eth_probe()
3337 SET_NETDEV_DEV(net_dev, dev->parent); in dpaa_eth_probe()
3338 dev_set_drvdata(dev, net_dev); in dpaa_eth_probe()
3340 priv = netdev_priv(net_dev); in dpaa_eth_probe()
3341 priv->net_dev = net_dev; in dpaa_eth_probe()
3347 netdev_err(net_dev, "dpaa_mac_dev_get() failed\n"); in dpaa_eth_probe()
3360 netdev_err(net_dev, "dma_coerce_mask_and_coherent() failed\n"); in dpaa_eth_probe()
3371 net_dev->mtu = min(dpaa_get_max_mtu(), ETH_DATA_LEN); in dpaa_eth_probe()
3373 netdev_dbg(net_dev, "Setting initial MTU on net device: %d\n", in dpaa_eth_probe()
3374 net_dev->mtu); in dpaa_eth_probe()
3468 netif_set_real_num_tx_queues(net_dev, priv->num_tc * DPAA_TC_TXQ_NUM); in dpaa_eth_probe()
3471 err = dpaa_napi_add(net_dev); in dpaa_eth_probe()
3475 err = dpaa_netdev_init(net_dev, &dpaa_ops, tx_timeout); in dpaa_eth_probe()
3479 dpaa_eth_sysfs_init(&net_dev->dev); in dpaa_eth_probe()
3481 netif_info(priv, probe, net_dev, "Probed interface %s\n", in dpaa_eth_probe()
3482 net_dev->name); in dpaa_eth_probe()
3487 dpaa_napi_del(net_dev); in dpaa_eth_probe()
3499 free_netdev(net_dev); in dpaa_eth_probe()
3506 struct net_device *net_dev; in dpaa_remove() local
3512 net_dev = dev_get_drvdata(dev); in dpaa_remove()
3514 priv = netdev_priv(net_dev); in dpaa_remove()
3519 unregister_netdev(net_dev); in dpaa_remove()
3529 dpaa_napi_del(net_dev); in dpaa_remove()
3533 free_netdev(net_dev); in dpaa_remove()