Lines Matching refs:txq
75 struct netdev_queue *txq) in netpoll_start_xmit() argument
94 status = netdev_start_xmit(skb, dev, txq, false); in netpoll_start_xmit()
107 while ((skb = skb_dequeue(&npinfo->txq))) { in queue_process()
109 struct netdev_queue *txq; in queue_process() local
124 txq = netdev_get_tx_queue(dev, q_index); in queue_process()
125 HARD_TX_LOCK(dev, txq, smp_processor_id()); in queue_process()
126 if (netif_xmit_frozen_or_stopped(txq) || in queue_process()
127 !dev_xmit_complete(netpoll_start_xmit(skb, dev, txq))) { in queue_process()
128 skb_queue_head(&npinfo->txq, skb); in queue_process()
129 HARD_TX_UNLOCK(dev, txq); in queue_process()
135 HARD_TX_UNLOCK(dev, txq); in queue_process()
328 if (skb_queue_len(&npinfo->txq) == 0 && !netpoll_owner_active(dev)) { in __netpoll_send_skb()
329 struct netdev_queue *txq; in __netpoll_send_skb() local
331 txq = netdev_core_pick_tx(dev, skb, NULL); in __netpoll_send_skb()
336 if (HARD_TX_TRYLOCK(dev, txq)) { in __netpoll_send_skb()
337 if (!netif_xmit_stopped(txq)) in __netpoll_send_skb()
338 status = netpoll_start_xmit(skb, dev, txq); in __netpoll_send_skb()
340 HARD_TX_UNLOCK(dev, txq); in __netpoll_send_skb()
360 skb_queue_tail(&npinfo->txq, skb); in __netpoll_send_skb()
629 skb_queue_head_init(&npinfo->txq); in __netpoll_setup()
815 skb_queue_purge(&npinfo->txq); in rcu_cleanup_netpoll_info()
821 __skb_queue_purge(&npinfo->txq); in rcu_cleanup_netpoll_info()