/Linux-v5.4/samples/bpf/ |
D | hbm.c | 168 struct hbm_queue_stats qstats = {0}; in run_bpf_prog() local 189 qstats.rate = rate; in run_bpf_prog() 190 qstats.stats = stats_flag ? 1 : 0; in run_bpf_prog() 191 qstats.loopback = loopback_flag ? 1 : 0; in run_bpf_prog() 192 qstats.no_cn = no_cn_flag ? 1 : 0; in run_bpf_prog() 193 if (bpf_map_update_elem(map_fd, &key, &qstats, BPF_ANY)) { in run_bpf_prog() 216 bpf_map_lookup_elem(map_fd, &key, &qstats); in run_bpf_prog() 224 last_cg_tx_bytes = qstats.bytes_total; in run_bpf_prog() 245 bpf_map_lookup_elem(map_fd, &key, &qstats); in run_bpf_prog() 246 new_cg_tx_bytes = qstats.bytes_total; in run_bpf_prog() [all …]
|
/Linux-v5.4/net/core/ |
D | gen_stats.c | 282 __gnet_stats_copy_queue_cpu(struct gnet_stats_queue *qstats, in __gnet_stats_copy_queue_cpu() argument 290 qstats->qlen = 0; in __gnet_stats_copy_queue_cpu() 291 qstats->backlog += qcpu->backlog; in __gnet_stats_copy_queue_cpu() 292 qstats->drops += qcpu->drops; in __gnet_stats_copy_queue_cpu() 293 qstats->requeues += qcpu->requeues; in __gnet_stats_copy_queue_cpu() 294 qstats->overlimits += qcpu->overlimits; in __gnet_stats_copy_queue_cpu() 298 void __gnet_stats_copy_queue(struct gnet_stats_queue *qstats, in __gnet_stats_copy_queue() argument 304 __gnet_stats_copy_queue_cpu(qstats, cpu); in __gnet_stats_copy_queue() 306 qstats->qlen = q->qlen; in __gnet_stats_copy_queue() 307 qstats->backlog = q->backlog; in __gnet_stats_copy_queue() [all …]
|
/Linux-v5.4/net/sched/ |
D | sch_skbprio.c | 35 struct gnet_stats_queue qstats[SKBPRIO_MAX_PRIORITY]; member 85 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue() 101 q->qstats[prio].drops++; in skbprio_enqueue() 102 q->qstats[prio].overlimits++; in skbprio_enqueue() 108 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue() 117 q->qstats[lp].backlog -= qdisc_pkt_len(to_drop); in skbprio_enqueue() 118 q->qstats[lp].drops++; in skbprio_enqueue() 119 q->qstats[lp].overlimits++; in skbprio_enqueue() 152 q->qstats[q->highest_prio].backlog -= qdisc_pkt_len(skb); in skbprio_dequeue() 186 memset(&q->qstats, 0, sizeof(q->qstats)); in skbprio_init() [all …]
|
D | sch_mq.c | 45 .qstats = &sch->qstats, in mq_offload_stats() 137 memset(&sch->qstats, 0, sizeof(sch->qstats)); in mq_dump() 153 __gnet_stats_copy_queue(&sch->qstats, in mq_dump() 155 &qdisc->qstats, qlen); in mq_dump() 160 sch->qstats.qlen += qdisc->qstats.qlen; in mq_dump() 161 sch->qstats.backlog += qdisc->qstats.backlog; in mq_dump() 162 sch->qstats.drops += qdisc->qstats.drops; in mq_dump() 163 sch->qstats.requeues += qdisc->qstats.requeues; in mq_dump() 164 sch->qstats.overlimits += qdisc->qstats.overlimits; in mq_dump()
|
D | sch_mqprio.c | 394 memset(&sch->qstats, 0, sizeof(sch->qstats)); in mqprio_dump() 411 __gnet_stats_copy_queue(&sch->qstats, in mqprio_dump() 413 &qdisc->qstats, qlen); in mqprio_dump() 418 sch->qstats.backlog += qdisc->qstats.backlog; in mqprio_dump() 419 sch->qstats.drops += qdisc->qstats.drops; in mqprio_dump() 420 sch->qstats.requeues += qdisc->qstats.requeues; in mqprio_dump() 421 sch->qstats.overlimits += qdisc->qstats.overlimits; in mqprio_dump() 515 struct gnet_stats_queue qstats = {0}; in mqprio_dump_class_stats() local 543 __gnet_stats_copy_queue(&sch->qstats, in mqprio_dump_class_stats() 545 &qdisc->qstats, in mqprio_dump_class_stats() [all …]
|
D | sch_gred.c | 117 return sch->qstats.backlog; in gred_backlog() 181 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= in gred_enqueue() 278 if (!sch->qstats.backlog) in gred_dequeue() 346 opt.set.qstats = &sch->qstats; in gred_offload() 380 table->tab[i]->backlog += hw_stats->stats.qstats[i].backlog; in gred_offload_dump_stats() 385 sch->qstats.qlen += hw_stats->stats.qstats[i].qlen; in gred_offload_dump_stats() 386 sch->qstats.backlog += hw_stats->stats.qstats[i].backlog; in gred_offload_dump_stats() 387 sch->qstats.drops += hw_stats->stats.qstats[i].drops; in gred_offload_dump_stats() 388 sch->qstats.requeues += hw_stats->stats.qstats[i].requeues; in gred_offload_dump_stats() 389 sch->qstats.overlimits += hw_stats->stats.qstats[i].overlimits; in gred_offload_dump_stats()
|
D | sch_prio.c | 88 sch->qstats.backlog += len; in prio_enqueue() 138 sch->qstats.backlog = 0; in prio_reset() 158 opt.replace_params.qstats = &sch->qstats; in prio_offload() 256 .qstats = &sch->qstats, in prio_dump_offload()
|
D | sch_fifo.c | 21 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= sch->limit)) in bfifo_enqueue() 44 prev_backlog = sch->qstats.backlog; in pfifo_tail_enqueue() 50 qdisc_tree_reduce_backlog(sch, 0, prev_backlog - sch->qstats.backlog); in pfifo_tail_enqueue()
|
D | sch_red.c | 66 child->qstats.backlog); in red_enqueue() 143 sch->qstats.backlog = 0; in red_reset() 168 opt.set.qstats = &sch->qstats; in red_offload() 292 .stats.qstats = &sch->qstats, in red_dump_offload_stats()
|
D | sch_drr.c | 23 struct gnet_stats_queue qstats; member 272 gnet_stats_copy_queue(d, cl_q->cpu_qstats, &cl_q->qstats, qlen) < 0) in drr_dump_class_stats() 361 cl->qstats.drops++; in drr_enqueue() 372 sch->qstats.backlog += len; in drr_enqueue() 446 sch->qstats.backlog = 0; in drr_reset_qdisc()
|
D | sch_fq_codel.c | 179 sch->qstats.drops += i; in fq_codel_drop() 180 sch->qstats.backlog -= len; in fq_codel_drop() 221 prev_backlog = sch->qstats.backlog; in fq_codel_enqueue() 234 prev_backlog -= sch->qstats.backlog; in fq_codel_enqueue() 269 sch->qstats.backlog -= qdisc_pkt_len(skb); in dequeue_func() 304 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->cparams, in fq_codel_dequeue() 352 sch->qstats.backlog = 0; in fq_codel_reset()
|
D | sch_tbf.c | 199 sch->qstats.backlog += len; in tbf_enqueue() 274 sch->qstats.backlog = 0; in tbf_reset() 446 sch->qstats.backlog = q->qdisc->qstats.backlog; in tbf_dump()
|
D | sch_hhf.c | 405 prev_backlog = sch->qstats.backlog; in hhf_enqueue() 414 qdisc_tree_reduce_backlog(sch, 1, prev_backlog - sch->qstats.backlog); in hhf_enqueue() 564 prev_backlog = sch->qstats.backlog; in hhf_change() 571 prev_backlog - sch->qstats.backlog); in hhf_change()
|
D | sch_codel.c | 75 sch->qstats.backlog -= qdisc_pkt_len(skb); in dequeue_func() 94 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->params, &q->vars, in codel_qdisc_dequeue()
|
D | sch_pie.c | 113 if (sch->qstats.backlog < 2 * mtu) in drop_early() 268 int qlen = sch->qstats.backlog; /* current queue size in bytes */ in pie_process_dequeue() 333 u32 qlen = sch->qstats.backlog; /* queue size in bytes */ in calculate_probability()
|
/Linux-v5.4/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_stats.h | 431 ADD_EXTEND_64(qstats->t##_hi, qstats->t##_lo, diff); \ 446 ADD_EXTEND_64(qstats->t##_hi, qstats->t##_lo, diff); \ 459 ADD_EXTEND_64(qstats->t##_hi, qstats->t##_lo, diff); \ 464 qstats->t##_lo = qstats_old->t##_lo + le32_to_cpu(s.lo); \ 465 qstats->t##_hi = qstats_old->t##_hi + le32_to_cpu(s.hi) \ 466 + ((qstats->t##_lo < qstats_old->t##_lo) ? 1 : 0); \ 471 qstats_old->f = qstats->f; \ 476 ADD_64(estats->s##_hi, qstats->s##_hi, \ 477 estats->s##_lo, qstats->s##_lo); \ 480 qstats_old->s##_hi_old = qstats->s##_hi; \ [all …]
|
D | bnx2x_stats.c | 951 struct bnx2x_eth_q_stats *qstats = in bnx2x_storm_stats_update() local 975 qstats->total_bytes_received_hi = in bnx2x_storm_stats_update() 976 qstats->total_broadcast_bytes_received_hi; in bnx2x_storm_stats_update() 977 qstats->total_bytes_received_lo = in bnx2x_storm_stats_update() 978 qstats->total_broadcast_bytes_received_lo; in bnx2x_storm_stats_update() 980 ADD_64(qstats->total_bytes_received_hi, in bnx2x_storm_stats_update() 981 qstats->total_multicast_bytes_received_hi, in bnx2x_storm_stats_update() 982 qstats->total_bytes_received_lo, in bnx2x_storm_stats_update() 983 qstats->total_multicast_bytes_received_lo); in bnx2x_storm_stats_update() 985 ADD_64(qstats->total_bytes_received_hi, in bnx2x_storm_stats_update() [all …]
|
/Linux-v5.4/include/net/ |
D | sch_generic.h | 104 struct gnet_stats_queue qstats; member 283 struct gnet_stats_queue *qstats; member 471 __u32 qlen = q->qstats.qlen; in qdisc_qlen_sum() 854 sch->qstats.backlog -= qdisc_pkt_len(skb); in qdisc_qstats_backlog_dec() 866 sch->qstats.backlog += qdisc_pkt_len(skb); in qdisc_qstats_backlog_inc() 892 sch->qstats.drops += count; in __qdisc_qstats_drop() 895 static inline void qstats_drop_inc(struct gnet_stats_queue *qstats) in qstats_drop_inc() argument 897 qstats->drops++; in qstats_drop_inc() 900 static inline void qstats_overlimit_inc(struct gnet_stats_queue *qstats) in qstats_overlimit_inc() argument 902 qstats->overlimits++; in qstats_overlimit_inc() [all …]
|
D | pkt_cls.h | 686 struct gnet_stats_queue *qstats; member 725 struct gnet_stats_queue *qstats; member 764 struct gnet_stats_queue *qstats; member 770 struct gnet_stats_queue qstats[MAX_DPs]; member 798 struct gnet_stats_queue *qstats; member
|
D | gen_stats.h | 56 void __gnet_stats_copy_queue(struct gnet_stats_queue *qstats,
|
/Linux-v5.4/drivers/infiniband/hw/hfi1/ |
D | vnic_main.c | 173 #define SUM_GRP_COUNTERS(stats, qstats, x_grp) do { \ argument 175 for (src64 = &qstats->x_grp.unicast, \ 191 struct opa_vnic_stats *qstats = &vinfo->stats[i]; in hfi1_vnic_update_stats() local 196 stats->tx_drop_state += qstats->tx_drop_state; in hfi1_vnic_update_stats() 197 stats->tx_dlid_zero += qstats->tx_dlid_zero; in hfi1_vnic_update_stats() 199 SUM_GRP_COUNTERS(stats, qstats, tx_grp); in hfi1_vnic_update_stats() 206 struct opa_vnic_stats *qstats = &vinfo->stats[i]; in hfi1_vnic_update_stats() local 211 stats->rx_drop_state += qstats->rx_drop_state; in hfi1_vnic_update_stats() 212 stats->rx_oversize += qstats->rx_oversize; in hfi1_vnic_update_stats() 213 stats->rx_runt += qstats->rx_runt; in hfi1_vnic_update_stats() [all …]
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_qdisc.c | 353 p->qstats->backlog -= backlog; in mlxsw_sp_qdisc_red_unoffload() 410 stats_ptr->qstats->overlimits += overlimits; in mlxsw_sp_qdisc_get_red_stats() 411 stats_ptr->qstats->drops += drops; in mlxsw_sp_qdisc_get_red_stats() 412 stats_ptr->qstats->backlog += in mlxsw_sp_qdisc_get_red_stats() 556 p->qstats->backlog -= backlog; in mlxsw_sp_qdisc_prio_unoffload() 585 stats_ptr->qstats->drops += drops; in mlxsw_sp_qdisc_get_prio_stats() 586 stats_ptr->qstats->backlog += in mlxsw_sp_qdisc_get_prio_stats()
|
/Linux-v5.4/Documentation/networking/ |
D | gen_stats.txt | 20 struct gnet_stats_queue qstats; 26 mystruct->qstats.backlog += skb->pkt_len; 41 gnet_stats_copy_queue(&dump, &mystruct->qstats) < 0 ||
|
/Linux-v5.4/drivers/net/ethernet/netronome/nfp/abm/ |
D | qdisc.c | 462 struct gnet_stats_queue *qstats) in nfp_abm_stats_calculate() argument 466 qstats->qlen += new->backlog_pkts - old->backlog_pkts; in nfp_abm_stats_calculate() 467 qstats->backlog += new->backlog_bytes - old->backlog_bytes; in nfp_abm_stats_calculate() 468 qstats->overlimits += new->overlimits - old->overlimits; in nfp_abm_stats_calculate() 469 qstats->drops += new->drops - old->drops; in nfp_abm_stats_calculate() 503 &stats->bstats[i], &stats->qstats[i]); in nfp_abm_gred_stats() 654 stats->bstats, stats->qstats); in nfp_abm_red_stats() 811 stats->bstats, stats->qstats); in nfp_abm_mq_stats()
|
/Linux-v5.4/drivers/net/hyperv/ |
D | netvsc_drv.c | 1406 const struct netvsc_stats *qstats; in netvsc_get_ethtool_stats() local 1424 qstats = &nvdev->chan_table[j].tx_stats; in netvsc_get_ethtool_stats() 1427 start = u64_stats_fetch_begin_irq(&qstats->syncp); in netvsc_get_ethtool_stats() 1428 packets = qstats->packets; in netvsc_get_ethtool_stats() 1429 bytes = qstats->bytes; in netvsc_get_ethtool_stats() 1430 } while (u64_stats_fetch_retry_irq(&qstats->syncp, start)); in netvsc_get_ethtool_stats() 1434 qstats = &nvdev->chan_table[j].rx_stats; in netvsc_get_ethtool_stats() 1436 start = u64_stats_fetch_begin_irq(&qstats->syncp); in netvsc_get_ethtool_stats() 1437 packets = qstats->packets; in netvsc_get_ethtool_stats() 1438 bytes = qstats->bytes; in netvsc_get_ethtool_stats() [all …]
|