| /net/core/ |
| A D | gen_stats.c | 348 qstats->qlen += qcpu->qlen; in gnet_stats_add_queue_cpu() 349 qstats->backlog += qcpu->backlog; in gnet_stats_add_queue_cpu() 350 qstats->drops += qcpu->drops; in gnet_stats_add_queue_cpu() 363 qstats->qlen += q->qlen; in gnet_stats_add_queue() 364 qstats->backlog += q->backlog; in gnet_stats_add_queue() 365 qstats->drops += q->drops; in gnet_stats_add_queue() 366 qstats->requeues += q->requeues; in gnet_stats_add_queue() 394 qstats.qlen = qlen; in gnet_stats_copy_queue() 397 d->tc_stats.drops = qstats.drops; in gnet_stats_copy_queue() 398 d->tc_stats.qlen = qstats.qlen; in gnet_stats_copy_queue() [all …]
|
| /net/sched/ |
| A D | sch_skbprio.c | 35 struct gnet_stats_queue qstats[SKBPRIO_MAX_PRIORITY]; member 87 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue() 103 q->qstats[prio].drops++; in skbprio_enqueue() 104 q->qstats[prio].overlimits++; in skbprio_enqueue() 110 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue() 119 q->qstats[lp].backlog -= qdisc_pkt_len(to_drop); in skbprio_enqueue() 120 q->qstats[lp].drops++; in skbprio_enqueue() 121 q->qstats[lp].overlimits++; in skbprio_enqueue() 188 memset(&q->qstats, 0, sizeof(q->qstats)); in skbprio_init() 218 memset(&q->qstats, 0, sizeof(q->qstats)); in skbprio_reset() [all …]
|
| A D | sch_gred.c | 118 return sch->qstats.backlog; in gred_backlog() 182 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= in gred_enqueue() 279 if (!sch->qstats.backlog) in gred_dequeue() 348 opt->set.qstats = &sch->qstats; in gred_offload() 386 table->tab[i]->backlog += hw_stats->stats.qstats[i].backlog; in gred_offload_dump_stats() 390 sch->qstats.qlen += hw_stats->stats.qstats[i].qlen; in gred_offload_dump_stats() 391 sch->qstats.backlog += hw_stats->stats.qstats[i].backlog; in gred_offload_dump_stats() 392 sch->qstats.drops += hw_stats->stats.qstats[i].drops; in gred_offload_dump_stats() 393 sch->qstats.requeues += hw_stats->stats.qstats[i].requeues; in gred_offload_dump_stats() 394 sch->qstats.overlimits += hw_stats->stats.qstats[i].overlimits; in gred_offload_dump_stats()
|
| A D | sch_mq.c | 45 .qstats = &sch->qstats, in mq_offload_stats() 136 memset(&sch->qstats, 0, sizeof(sch->qstats)); in mq_dump() 149 gnet_stats_add_queue(&sch->qstats, qdisc->cpu_qstats, in mq_dump() 150 &qdisc->qstats); in mq_dump()
|
| A D | sch_fifo.c | 22 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= in bfifo_enqueue() 49 prev_backlog = sch->qstats.backlog; in pfifo_tail_enqueue() 55 qdisc_tree_reduce_backlog(sch, 0, prev_backlog - sch->qstats.backlog); in pfifo_tail_enqueue() 95 qopt.stats.qstats = &sch->qstats; in fifo_offload_dump()
|
| A D | sch_mqprio.c | 564 memset(&sch->qstats, 0, sizeof(sch->qstats)); in mqprio_dump() 577 gnet_stats_add_queue(&sch->qstats, qdisc->cpu_qstats, in mqprio_dump() 578 &qdisc->qstats); in mqprio_dump() 669 struct gnet_stats_queue qstats = {0}; in mqprio_dump_class_stats() local 691 gnet_stats_add_queue(&qstats, qdisc->cpu_qstats, in mqprio_dump_class_stats() 692 &qdisc->qstats); in mqprio_dump_class_stats() 697 qlen = qdisc_qlen(sch) + qstats.qlen; in mqprio_dump_class_stats() 703 gnet_stats_copy_queue(d, NULL, &qstats, qlen) < 0) in mqprio_dump_class_stats()
|
| A D | sch_ets.c | 45 struct gnet_stats_queue qstats; member 132 qopt.replace_params.qstats = &sch->qstats; in ets_offload_change() 192 qopt.stats.qstats = &sch->qstats; in ets_offload_dump() 436 cl->qstats.drops++; in ets_qdisc_enqueue() 447 sch->qstats.backlog += len; in ets_qdisc_enqueue() 688 memset(&q->classes[i].qstats, 0, sizeof(q->classes[i].qstats)); in ets_qdisc_change()
|
| A D | sch_prio.c | 88 sch->qstats.backlog += len; in prio_enqueue() 156 opt.replace_params.qstats = &sch->qstats; in prio_offload() 254 .qstats = &sch->qstats, in prio_dump_offload()
|
| A D | sch_tbf.c | 156 qopt.replace_params.qstats = &sch->qstats; in tbf_offload_change() 183 qopt.stats.qstats = &sch->qstats; in tbf_offload_dump() 234 sch->qstats.backlog += len; in tbf_segment() 265 sch->qstats.backlog += len; in tbf_enqueue()
|
| A D | sch_red.c | 81 child->qstats.backlog); in red_enqueue() 135 sch->qstats.backlog += len; in red_enqueue() 207 opt.set.qstats = &sch->qstats; in red_offload() 404 .stats.qstats = &sch->qstats, in red_dump_offload_stats()
|
| A D | sch_fq_codel.c | 179 sch->qstats.drops += i; in fq_codel_drop() 180 sch->qstats.backlog -= len; in fq_codel_drop() 221 prev_backlog = sch->qstats.backlog; in fq_codel_enqueue() 234 prev_backlog -= sch->qstats.backlog; in fq_codel_enqueue() 269 sch->qstats.backlog -= qdisc_pkt_len(skb); in dequeue_func() 304 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->cparams, in fq_codel_dequeue()
|
| A D | sch_drr.c | 22 struct gnet_stats_queue qstats; member 277 gnet_stats_copy_queue(d, cl_q->cpu_qstats, &cl_q->qstats, qlen) < 0) in drr_dump_class_stats() 357 cl->qstats.drops++; in drr_enqueue() 368 sch->qstats.backlog += len; in drr_enqueue()
|
| A D | sch_hhf.c | 405 prev_backlog = sch->qstats.backlog; in hhf_enqueue() 414 qdisc_tree_reduce_backlog(sch, 1, prev_backlog - sch->qstats.backlog); in hhf_enqueue() 565 prev_backlog = sch->qstats.backlog; in hhf_change() 572 prev_backlog - sch->qstats.backlog); in hhf_change()
|
| A D | bpf_qdisc.c | 65 case offsetof(struct Qdisc, qstats) ... offsetofend(struct Qdisc, qstats) - 1: in bpf_qdisc_qdisc_access() 66 *end = offsetofend(struct Qdisc, qstats); in bpf_qdisc_qdisc_access()
|
| A D | sch_pie.c | 99 if (!pie_drop_early(sch, &q->params, &q->vars, sch->qstats.backlog, in pie_qdisc_enqueue() 434 pie_calculate_probability(&q->params, &q->vars, sch->qstats.backlog); in pie_timer() 531 pie_process_dequeue(skb, &q->params, &q->vars, sch->qstats.backlog); in pie_qdisc_dequeue()
|
| A D | sch_codel.c | 45 sch->qstats.backlog -= qdisc_pkt_len(skb); in dequeue_func() 64 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->params, &q->vars, in codel_qdisc_dequeue()
|
| A D | sch_sfb.c | 409 sch->qstats.backlog += len; in sfb_enqueue() 585 sch->qstats.backlog = q->qdisc->qstats.backlog; in sfb_dump()
|
| A D | sch_plug.c | 92 if (likely(sch->qstats.backlog + skb->len <= q->limit)) { in plug_enqueue()
|
| A D | sch_hfsc.c | 115 struct gnet_stats_queue qstats; member 1340 qdisc_qstats_qlen_backlog(cl->qdisc, &qlen, &cl->qstats.backlog); in hfsc_dump_class_stats() 1348 gnet_stats_copy_queue(d, NULL, &cl->qstats, qlen) < 0) in hfsc_dump_class_stats() 1557 cl->qstats.drops++; in hfsc_enqueue() 1563 sch->qstats.backlog += len; in hfsc_enqueue()
|
| A D | sch_qfq.c | 134 struct gnet_stats_queue qstats; member 1248 cl->qstats.drops++; in qfq_enqueue() 1258 cl->qstats.drops++; in qfq_enqueue() 1265 sch->qstats.backlog += len; in qfq_enqueue()
|
| A D | sch_fq_pie.c | 187 sch->qstats.backlog += pkt_len; in fq_pie_qdisc_enqueue() 265 sch->qstats.backlog -= pkt_len; in fq_pie_qdisc_dequeue()
|
| A D | sch_dualpi2.c | 289 if (sch->qstats.backlog < 2 * psched_mtu(qdisc_dev(sch))) in must_drop() 871 old_backlog = sch->qstats.backlog; in dualpi2_change() 881 old_backlog - sch->qstats.backlog); in dualpi2_change()
|
| A D | sch_cake.c | 1590 sch->qstats.backlog -= len; in cake_drop() 1781 sch->qstats.overlimits++; in cake_enqueue() 1818 sch->qstats.backlog += slen; in cake_enqueue() 1834 sch->qstats.drops++; in cake_enqueue() 1853 sch->qstats.backlog += len; in cake_enqueue() 1954 sch->qstats.backlog -= len; in cake_dequeue_one() 1999 sch->qstats.overlimits++; in cake_dequeue()
|
| A D | sch_cbs.c | 99 sch->qstats.backlog += len; in cbs_child_enqueue()
|
| A D | sch_netem.c | 613 sch->qstats.requeues++; in netem_enqueue() 752 sch->qstats.backlog -= pkt_len; in netem_dequeue()
|