| /Linux-v5.4/net/sched/ | 
| D | sch_generic.c | 53 	const struct netdev_queue *txq = q->dev_queue;  in __skb_dequeue_bad_txq()205 	const struct netdev_queue *txq = q->dev_queue;  in dequeue_skb()
 559 	.dev_queue	=	&noop_netdev_queue,
 802 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue,  in qdisc_alloc()  argument
 812 	if (!dev_queue) {  in qdisc_alloc()
 818 	dev = dev_queue->dev;  in qdisc_alloc()
 820 			 netdev_queue_numa_node_read(dev_queue));  in qdisc_alloc()
 829 				 netdev_queue_numa_node_read(dev_queue));  in qdisc_alloc()
 862 	sch->dev_queue = dev_queue;  in qdisc_alloc()
 880 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue,  in qdisc_create_dflt()  argument
 [all …]
 
 | 
| D | sch_mq.c | 72 	struct netdev_queue *dev_queue;  in mq_init()  local89 		dev_queue = netdev_get_tx_queue(dev, ntx);  in mq_init()
 90 		qdisc = qdisc_create_dflt(dev_queue, get_default_qdisc_ops(dev, ntx),  in mq_init()
 115 		old = dev_graft_qdisc(qdisc->dev_queue, qdisc);  in mq_attach()
 192 	struct netdev_queue *dev_queue = mq_queue_get(sch, cl);  in mq_graft()  local
 199 	*old = dev_graft_qdisc(dev_queue, new);  in mq_graft()
 217 	struct netdev_queue *dev_queue = mq_queue_get(sch, cl);  in mq_leaf()  local
 219 	return dev_queue->qdisc_sleeping;  in mq_leaf()
 234 	struct netdev_queue *dev_queue = mq_queue_get(sch, cl);  in mq_dump_class()  local
 238 	tcm->tcm_info = dev_queue->qdisc_sleeping->handle;  in mq_dump_class()
 [all …]
 
 | 
| D | sch_mqprio.c | 138 	struct netdev_queue *dev_queue;  in mqprio_init()  local227 		dev_queue = netdev_get_tx_queue(dev, i);  in mqprio_init()
 228 		qdisc = qdisc_create_dflt(dev_queue,  in mqprio_init()
 299 		old = dev_graft_qdisc(qdisc->dev_queue, qdisc);  in mqprio_attach()
 324 	struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl);  in mqprio_graft()  local
 326 	if (!dev_queue)  in mqprio_graft()
 332 	*old = dev_graft_qdisc(dev_queue, new);  in mqprio_graft()
 460 	struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl);  in mqprio_leaf()  local
 462 	if (!dev_queue)  in mqprio_leaf()
 465 	return dev_queue->qdisc_sleeping;  in mqprio_leaf()
 [all …]
 
 | 
| D | sch_api.c | 1052 			struct netdev_queue *dev_queue = dev_ingress_queue(dev);  in qdisc_graft()  local1055 				dev_queue = netdev_get_tx_queue(dev, i);  in qdisc_graft()
 1057 			old = dev_graft_qdisc(dev_queue, new);  in qdisc_graft()
 1149 				  struct netdev_queue *dev_queue,  in qdisc_create()  argument
 1195 	sch = qdisc_alloc(dev_queue, ops, extack);  in qdisc_create()
 1643 		struct netdev_queue *dev_queue;  in tc_modify_qdisc()  local
 1646 			dev_queue = p->ops->cl_ops->select_queue(p, tcm);  in tc_modify_qdisc()
 1648 			dev_queue = p->dev_queue;  in tc_modify_qdisc()
 1650 			dev_queue = netdev_get_tx_queue(dev, 0);  in tc_modify_qdisc()
 1652 		q = qdisc_create(dev, dev_queue, p,  in tc_modify_qdisc()
 [all …]
 
 | 
| D | sch_plug.c | 187 		netif_schedule_queue(sch->dev_queue);  in plug_change()195 		netif_schedule_queue(sch->dev_queue);  in plug_change()
 
 | 
| D | sch_taprio.c | 1623 		struct netdev_queue *dev_queue;  in taprio_init()  local1626 		dev_queue = netdev_get_tx_queue(dev, i);  in taprio_init()
 1627 		qdisc = qdisc_create_dflt(dev_queue,  in taprio_init()
 1662 	struct netdev_queue *dev_queue = taprio_queue_get(sch, cl);  in taprio_graft()  local
 1664 	if (!dev_queue)  in taprio_graft()
 1820 	struct netdev_queue *dev_queue = taprio_queue_get(sch, cl);  in taprio_leaf()  local
 1822 	if (!dev_queue)  in taprio_leaf()
 1825 	return dev_queue->qdisc_sleeping;  in taprio_leaf()
 1840 	struct netdev_queue *dev_queue = taprio_queue_get(sch, cl);  in taprio_dump_class()  local
 1844 	tcm->tcm_info = dev_queue->qdisc_sleeping->handle;  in taprio_dump_class()
 [all …]
 
 | 
| D | sch_cbs.c | 401 	q->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in cbs_init()412 	q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0);  in cbs_init()
 487 		new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in cbs_graft()
 
 | 
| D | sch_fifo.c | 171 	q = qdisc_create_dflt(sch->dev_queue, ops, TC_H_MAKE(sch->handle, 1),  in fifo_create_dflt()
 | 
| D | sch_drr.c | 111 	cl->qdisc	   = qdisc_create_dflt(sch->dev_queue,  in drr_change_class()211 		new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in drr_graft_class()
 
 | 
| D | sch_dsmark.c | 74 		new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in dsmark_graft()390 	p->q = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, sch->handle,  in dsmark_init()
 
 | 
| D | sch_atm.c | 295 	flow->q = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, classid,  in atm_tc_change()551 	p->link.q = qdisc_create_dflt(sch->dev_queue,  in atm_tc_init()
 
 | 
| D | sch_multiq.c | 215 			child = qdisc_create_dflt(sch->dev_queue,  in multiq_tune()
 | 
| D | sch_hfsc.c | 1056 	cl->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in hfsc_change_class()1178 		new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in hfsc_graft_class()
 1410 	q->root.qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in hfsc_init_qdisc()
 
 | 
| D | sch_htb.c | 1155 	    (new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in htb_graft()1264 		new_q = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in htb_delete()
 1399 		new_q = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in htb_change_class()
 
 | 
| D | sch_prio.c | 200 		queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in prio_tune()
 | 
| D | sch_etf.c | 381 	q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0);  in etf_init()
 | 
| D | sch_cbq.c | 1195 	q->link.q = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in cbq_init()1402 		new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,  in cbq_graft()
 1635 	cl->q = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, classid,  in cbq_change_class()
 
 | 
| /Linux-v5.4/include/linux/ | 
| D | netdevice.h | 3067 static __always_inline void netif_tx_start_queue(struct netdev_queue *dev_queue)  in netif_tx_start_queue()  argument3069 	clear_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state);  in netif_tx_start_queue()
 3093 void netif_tx_wake_queue(struct netdev_queue *dev_queue);
 3117 static __always_inline void netif_tx_stop_queue(struct netdev_queue *dev_queue)  in netif_tx_stop_queue()  argument
 3119 	set_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state);  in netif_tx_stop_queue()
 3137 static inline bool netif_tx_queue_stopped(const struct netdev_queue *dev_queue)  in netif_tx_queue_stopped()  argument
 3139 	return test_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state);  in netif_tx_queue_stopped()
 3153 static inline bool netif_xmit_stopped(const struct netdev_queue *dev_queue)  in netif_xmit_stopped()  argument
 3155 	return dev_queue->state & QUEUE_STATE_ANY_XOFF;  in netif_xmit_stopped()
 3159 netif_xmit_frozen_or_stopped(const struct netdev_queue *dev_queue)  in netif_xmit_frozen_or_stopped()  argument
 [all …]
 
 | 
| /Linux-v5.4/include/net/ | 
| D | sch_generic.h | 89 	struct netdev_queue	*dev_queue;  member496 	struct Qdisc *q = rcu_dereference_rtnl(qdisc->dev_queue->qdisc);  in qdisc_root()
 503 	return rcu_dereference_bh(qdisc->dev_queue->qdisc);  in qdisc_root_bh()
 508 	return qdisc->dev_queue->qdisc_sleeping;  in qdisc_root_sleeping()
 548 	return qdisc->dev_queue->dev;  in qdisc_dev()
 631 struct Qdisc *dev_graft_qdisc(struct netdev_queue *dev_queue,
 661 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue,
 665 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue,
 
 | 
| D | pkt_sched.h | 152 	return dev_net(q->dev_queue->dev);  in qdisc_net()
 | 
| /Linux-v5.4/drivers/net/ethernet/hisilicon/hns/ | 
| D | hns_enet.c | 311 	struct netdev_queue *dev_queue;  in hns_nic_net_xmit_hw()  local362 	dev_queue = netdev_get_tx_queue(ndev, skb->queue_mapping);  in hns_nic_net_xmit_hw()
 363 	netdev_tx_sent_queue(dev_queue, skb->len);  in hns_nic_net_xmit_hw()
 954 	struct netdev_queue *dev_queue;  in hns_nic_tx_poll_one()  local
 983 	dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index);  in hns_nic_tx_poll_one()
 984 	netdev_tx_completed_queue(dev_queue, pkts, bytes);  in hns_nic_tx_poll_one()
 995 		if (netif_tx_queue_stopped(dev_queue) &&  in hns_nic_tx_poll_one()
 997 			netif_tx_wake_queue(dev_queue);  in hns_nic_tx_poll_one()
 1038 	struct netdev_queue *dev_queue;  in hns_nic_tx_clr_all_bufs()  local
 1048 	dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index);  in hns_nic_tx_clr_all_bufs()
 [all …]
 
 | 
| D | hns_ethtool.c | 377 	struct netdev_queue *dev_queue;  in __lb_other_process()  local408 		dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index);  in __lb_other_process()
 409 		netdev_tx_reset_queue(dev_queue);  in __lb_other_process()
 
 | 
| /Linux-v5.4/drivers/net/xen-netback/ | 
| D | xenbus.c | 31 	struct netdev_queue *dev_queue;  in xenvif_read_io_ring()  local92 	dev_queue = netdev_get_tx_queue(queue->vif->dev, queue->id);  in xenvif_read_io_ring()
 97 		   netif_tx_queue_stopped(dev_queue) ? "stopped" : "running");  in xenvif_read_io_ring()
 
 | 
| /Linux-v5.4/drivers/net/ethernet/hisilicon/hns3/ | 
| D | hns3_enet.c | 481 	struct netdev_queue *dev_queue;  in hns3_reset_tx_queue()  local485 		dev_queue = netdev_get_tx_queue(ndev,  in hns3_reset_tx_queue()
 487 		netdev_tx_reset_queue(dev_queue);  in hns3_reset_tx_queue()
 1323 	struct netdev_queue *dev_queue;  in hns3_nic_net_xmit()  local
 1378 	dev_queue = netdev_get_tx_queue(netdev, ring_data->queue_index);  in hns3_nic_net_xmit()
 1379 	netdev_tx_sent_queue(dev_queue, skb->len);  in hns3_nic_net_xmit()
 2325 	struct netdev_queue *dev_queue;  in hns3_clean_tx_ring()  local
 2357 	dev_queue = netdev_get_tx_queue(netdev, ring->tqp->tqp_index);  in hns3_clean_tx_ring()
 2358 	netdev_tx_completed_queue(dev_queue, pkts, bytes);  in hns3_clean_tx_ring()
 2366 		if (netif_tx_queue_stopped(dev_queue) &&  in hns3_clean_tx_ring()
 [all …]
 
 | 
| /Linux-v5.4/tools/perf/Documentation/ | 
| D | perf-trace.txt | 312 next block:*_unplug and the next three net:*dev_queue events, this last one315 …# perf trace -e sched:*switch/nr=2/,block:*_plug/nr=4/,block:*_unplug/nr=1/,net:*dev_queue/nr=3,ma…
 
 |