Lines Matching refs:txq
63 struct netdev_queue *txq) in netpoll_start_xmit() argument
82 status = netdev_start_xmit(skb, dev, txq, false); in netpoll_start_xmit()
95 while ((skb = skb_dequeue(&npinfo->txq))) { in queue_process()
97 struct netdev_queue *txq; in queue_process() local
112 txq = netdev_get_tx_queue(dev, q_index); in queue_process()
113 HARD_TX_LOCK(dev, txq, smp_processor_id()); in queue_process()
114 if (netif_xmit_frozen_or_stopped(txq) || in queue_process()
115 !dev_xmit_complete(netpoll_start_xmit(skb, dev, txq))) { in queue_process()
116 skb_queue_head(&npinfo->txq, skb); in queue_process()
117 HARD_TX_UNLOCK(dev, txq); in queue_process()
123 HARD_TX_UNLOCK(dev, txq); in queue_process()
133 struct netdev_queue *txq = netdev_get_tx_queue(dev, i); in netif_local_xmit_active() local
135 if (READ_ONCE(txq->xmit_lock_owner) == smp_processor_id()) in netif_local_xmit_active()
334 if (skb_queue_len(&npinfo->txq) == 0 && !netpoll_owner_active(dev)) { in __netpoll_send_skb()
335 struct netdev_queue *txq; in __netpoll_send_skb() local
337 txq = netdev_core_pick_tx(dev, skb, NULL); in __netpoll_send_skb()
342 if (HARD_TX_TRYLOCK(dev, txq)) { in __netpoll_send_skb()
343 if (!netif_xmit_stopped(txq)) in __netpoll_send_skb()
344 status = netpoll_start_xmit(skb, dev, txq); in __netpoll_send_skb()
346 HARD_TX_UNLOCK(dev, txq); in __netpoll_send_skb()
366 skb_queue_tail(&npinfo->txq, skb); in __netpoll_send_skb()
577 skb_queue_head_init(&npinfo->txq); in __netpoll_setup()
798 skb_queue_purge(&npinfo->txq); in rcu_cleanup_netpoll_info()
804 __skb_queue_purge(&npinfo->txq); in rcu_cleanup_netpoll_info()