Lines Matching refs:netdev_queue
577 struct netdev_queue { struct
634 static inline int netdev_queue_numa_node_read(const struct netdev_queue *q) in netdev_queue_numa_node_read() argument
643 static inline void netdev_queue_numa_node_write(struct netdev_queue *q, int node) in netdev_queue_numa_node_write()
2135 struct netdev_queue __rcu *ingress_queue;
2149 struct netdev_queue *_tx ____cacheline_aligned_in_smp;
2336 struct netdev_queue *netdev_get_tx_queue(const struct net_device *dev, in netdev_get_tx_queue()
2342 static inline struct netdev_queue *skb_get_tx_queue(const struct net_device *dev, in skb_get_tx_queue()
2350 struct netdev_queue *, in netdev_for_each_tx_queue() argument
2379 struct netdev_queue *netdev_core_pick_tx(struct net_device *dev,
3410 void netif_schedule_queue(struct netdev_queue *txq);
3420 static __always_inline void netif_tx_start_queue(struct netdev_queue *dev_queue) in netif_tx_start_queue()
3441 struct netdev_queue *txq = netdev_get_tx_queue(dev, i); in netif_tx_start_all_queues()
3446 void netif_tx_wake_queue(struct netdev_queue *dev_queue);
3465 struct netdev_queue *txq = netdev_get_tx_queue(dev, i); in netif_tx_wake_all_queues()
3470 static __always_inline void netif_tx_stop_queue(struct netdev_queue *dev_queue) in netif_tx_stop_queue()
3489 static inline bool netif_tx_queue_stopped(const struct netdev_queue *dev_queue) in netif_tx_queue_stopped()
3505 static inline bool netif_xmit_stopped(const struct netdev_queue *dev_queue) in netif_xmit_stopped()
3511 netif_xmit_frozen_or_stopped(const struct netdev_queue *dev_queue) in netif_xmit_frozen_or_stopped()
3517 netif_xmit_frozen_or_drv_stopped(const struct netdev_queue *dev_queue) in netif_xmit_frozen_or_drv_stopped()
3532 static inline void netdev_queue_set_dql_min_limit(struct netdev_queue *dev_queue, in netdev_queue_set_dql_min_limit()
3547 static inline void netdev_txq_bql_enqueue_prefetchw(struct netdev_queue *dev_queue) in netdev_txq_bql_enqueue_prefetchw()
3561 static inline void netdev_txq_bql_complete_prefetchw(struct netdev_queue *dev_queue) in netdev_txq_bql_complete_prefetchw()
3568 static inline void netdev_tx_sent_queue(struct netdev_queue *dev_queue, in netdev_tx_sent_queue()
3598 static inline bool __netdev_tx_sent_queue(struct netdev_queue *dev_queue, in __netdev_tx_sent_queue()
3634 static inline void netdev_tx_completed_queue(struct netdev_queue *dev_queue, in netdev_tx_completed_queue()
3674 static inline void netdev_tx_reset_queue(struct netdev_queue *q) in netdev_tx_reset_queue()
3741 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); in netif_start_subqueue()
3755 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); in netif_stop_subqueue()
3769 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); in __netif_subqueue_stopped()
3796 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); in netif_wake_subqueue()
4087 struct netdev_queue *txq, int *ret);
4403 static inline void __netif_tx_lock(struct netdev_queue *txq, int cpu) in __netif_tx_lock()
4409 static inline bool __netif_tx_acquire(struct netdev_queue *txq) in __netif_tx_acquire()
4415 static inline void __netif_tx_release(struct netdev_queue *txq) in __netif_tx_release()
4420 static inline void __netif_tx_lock_bh(struct netdev_queue *txq) in __netif_tx_lock_bh()
4426 static inline bool __netif_tx_trylock(struct netdev_queue *txq) in __netif_tx_trylock()
4434 static inline void __netif_tx_unlock(struct netdev_queue *txq) in __netif_tx_unlock()
4440 static inline void __netif_tx_unlock_bh(struct netdev_queue *txq) in __netif_tx_unlock_bh()
4446 static inline void txq_trans_update(struct netdev_queue *txq) in txq_trans_update()
4455 struct netdev_queue *txq = netdev_get_tx_queue(dev, 0); in netif_trans_update()
4475 struct netdev_queue *txq = netdev_get_tx_queue(dev, i); in netif_tx_lock()
4500 struct netdev_queue *txq = netdev_get_tx_queue(dev, i); in netif_tx_unlock()
4548 struct netdev_queue *txq = netdev_get_tx_queue(dev, i); in netif_tx_disable()
4997 struct netdev_queue *txq, bool more) in netdev_start_xmit()