| /net/sched/ |
| A D | sch_fifo.c | 23 READ_ONCE(sch->limit))) in bfifo_enqueue() 32 if (likely(sch->q.qlen < READ_ONCE(sch->limit))) in pfifo_enqueue() 46 if (likely(sch->q.qlen < READ_ONCE(sch->limit))) in pfifo_tail_enqueue() 51 __qdisc_queue_drop_head(sch, &sch->q, to_free); in pfifo_tail_enqueue() 52 qdisc_qstats_drop(sch); in pfifo_tail_enqueue() 55 qdisc_tree_reduce_backlog(sch, 0, prev_backlog - sch->qstats.backlog); in pfifo_tail_enqueue() 68 qopt.handle = sch->handle; in fifo_offload_init() 69 qopt.parent = sch->parent; in fifo_offload_init() 123 bypass = sch->limit >= psched_mtu(qdisc_dev(sch)); in __fifo_init() 144 fifo_offload_init(sch); in fifo_init() [all …]
|
| A D | sch_mq.c | 29 .handle = sch->handle, in mq_offload() 42 .handle = sch->handle, in mq_offload_stats() 44 .bstats = &sch->bstats, in mq_offload_stats() 45 .qstats = &sch->qstats, in mq_offload_stats() 58 mq_offload(sch, TC_MQ_DESTROY); in mq_destroy() 76 if (sch->parent != TC_H_ROOT) in mq_init() 100 sch->flags |= TCQ_F_MQROOT; in mq_init() 134 sch->q.qlen = 0; in mq_dump() 136 memset(&sch->qstats, 0, sizeof(sch->qstats)); in mq_dump() 196 qdisc_offload_graft_helper(qdisc_dev(sch), sch, new, *old, in mq_graft() [all …]
|
| A D | sch_prio.c | 80 qdisc_qstats_drop(sch); in prio_enqueue() 89 sch->q.qlen++; in prio_enqueue() 93 qdisc_qstats_drop(sch); in prio_enqueue() 122 sch->q.qlen--; in prio_dequeue() 144 .handle = sch->handle, in prio_offload() 145 .parent = sch->parent, in prio_offload() 209 sch_tree_lock(sch); in prio_tune() 222 sch_tree_unlock(sch); in prio_tune() 249 .handle = sch->handle, in prio_dump_offload() 250 .parent = sch->parent, in prio_dump_offload() [all …]
|
| A D | sch_red.c | 44 struct Qdisc *sch; member 136 sch->q.qlen++; in red_enqueue() 162 sch->q.qlen--; in red_dequeue() 191 .handle = sch->handle, in red_offload() 192 .parent = sch->parent, in red_offload() 277 sch_tree_lock(sch); in __red_change() 307 sch_tree_unlock(sch); in __red_change() 316 sch_tree_unlock(sch); in __red_change() 325 struct Qdisc *sch = q->sch; in red_adaptative_timer() local 345 q->sch = sch; in red_init() [all …]
|
| A D | sch_ets.c | 129 qopt.handle = sch->handle; in ets_offload_change() 130 qopt.parent = sch->parent; in ets_offload_change() 162 qopt.handle = sch->handle; in ets_offload_destroy() 163 qopt.parent = sch->parent; in ets_offload_destroy() 246 sch_tree_lock(sch); in ets_class_change() 248 sch_tree_unlock(sch); in ets_class_change() 250 ets_offload_change(sch); in ets_class_change() 448 sch->q.qlen++; in ets_qdisc_enqueue() 457 sch->q.qlen--; in ets_qdisc_dequeue_skb() 652 sch_tree_lock(sch); in ets_qdisc_change() [all …]
|
| A D | sch_ingress.c | 37 return ingress_find(sch, classid); in ingress_bind_filter() 51 struct ingress_sched_data *q = qdisc_priv(sch); in ingress_tcf_block() 81 struct net_device *dev = qdisc_dev(sch); in ingress_init() 86 if (sch->parent != TC_H_INGRESS) in ingress_init() 115 struct net_device *dev = qdisc_dev(sch); in ingress_destroy() 118 if (sch->parent != TC_H_INGRESS) in ingress_destroy() 196 return clsact_find(sch, classid); in clsact_bind_filter() 246 struct net_device *dev = qdisc_dev(sch); in clsact_init() 251 if (sch->parent != TC_H_CLSACT) in clsact_init() 294 struct net_device *dev = qdisc_dev(sch); in clsact_destroy() [all …]
|
| A D | sch_codel.c | 41 struct Qdisc *sch = ctx; in dequeue_func() local 56 qdisc_qstats_drop(sch); in drop_func() 64 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->params, &q->vars, in codel_qdisc_dequeue() 83 if (likely(qdisc_qlen(sch) < sch->limit)) { in codel_qdisc_enqueue() 87 q = qdisc_priv(sch); in codel_qdisc_enqueue() 114 sch_tree_lock(sch); in codel_change() 145 qlen = sch->q.qlen; in codel_change() 146 while (sch->q.qlen > sch->limit) { in codel_change() 153 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped); in codel_change() 155 sch_tree_unlock(sch); in codel_change() [all …]
|
| A D | sch_tbf.c | 152 qopt.handle = sch->handle; in tbf_offload_change() 153 qopt.parent = sch->parent; in tbf_offload_change() 192 .handle = sch->handle, in tbf_offload_graft() 193 .parent = sch->parent, in tbf_offload_graft() 198 qdisc_offload_graft_helper(qdisc_dev(sch), sch, new, old, in tbf_offload_graft() 233 sch->q.qlen += nb; in tbf_segment() 266 sch->q.qlen++; in tbf_enqueue() 311 sch->q.qlen--; in tbf_dequeue() 453 sch_tree_lock(sch); in tbf_change() 475 sch_tree_unlock(sch); in tbf_change() [all …]
|
| A D | sch_multiq.c | 71 qdisc_qstats_drop(sch); in multiq_enqueue() 79 sch->q.qlen++; in multiq_enqueue() 83 qdisc_qstats_drop(sch); in multiq_enqueue() 109 sch->q.qlen--; in multiq_dequeue() 148 multiq_reset(struct Qdisc *sch) in multiq_reset() argument 159 multiq_destroy(struct Qdisc *sch) in multiq_destroy() argument 193 sch_tree_lock(sch); in multiq_tune() 205 sch_tree_unlock(sch); in multiq_tune() 216 TC_H_MAKE(sch->handle, in multiq_tune() 219 sch_tree_lock(sch); in multiq_tune() [all …]
|
| A D | sch_drr.c | 100 sch_tree_lock(sch); in drr_change_class() 103 sch_tree_unlock(sch); in drr_change_class() 135 sch_tree_lock(sch); in drr_change_class() 137 sch_tree_unlock(sch); in drr_change_class() 163 sch_tree_lock(sch); in drr_delete_class() 168 sch_tree_unlock(sch); in drr_delete_class() 170 drr_destroy_class(sch, cl); in drr_delete_class() 349 qdisc_qstats_drop(sch); in drr_enqueue() 358 qdisc_qstats_drop(sch); in drr_enqueue() 369 sch->q.qlen++; in drr_enqueue() [all …]
|
| A D | sch_fq_codel.c | 181 sch->q.qlen -= i; in fq_codel_drop() 218 if (++sch->q.qlen <= sch->limit && !memory_limited) in fq_codel_enqueue() 268 sch->q.qlen--; in dequeue_func() 304 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->cparams, in fq_codel_dequeue() 393 sch_tree_lock(sch); in fq_codel_change() 442 while (sch->q.qlen > sch->limit || in fq_codel_change() 454 sch_tree_unlock(sch); in fq_codel_change() 474 sch->limit = 10*1024; in fq_codel_init() 516 if (sch->limit >= 1) in fq_codel_init() 593 sch_tree_lock(sch); in fq_codel_dump_stats() [all …]
|
| A D | sch_cbs.c | 99 sch->qstats.backlog += len; in cbs_child_enqueue() 100 sch->q.qlen++; in cbs_child_enqueue() 136 return q->enqueue(skb, sch, to_free); in cbs_enqueue() 169 qdisc_qstats_backlog_dec(sch, skb); in cbs_child_dequeue() 170 qdisc_bstats_update(sch, skb); in cbs_child_dequeue() 171 sch->q.qlen--; in cbs_child_dequeue() 207 skb = cbs_child_dequeue(sch, qdisc); in cbs_dequeue_soft() 243 return q->dequeue(sch); in cbs_dequeue() 413 sch->handle, extack); in cbs_init() 430 return cbs_change(sch, opt, extack); in cbs_init() [all …]
|
| A D | sch_etf.c | 192 sch->q.qlen++; in etf_enqueue_timesortedlist() 195 reset_watchdog(sch); in etf_enqueue_timesortedlist() 218 skb->dev = qdisc_dev(sch); in timesortedlist_drop() 224 qdisc_qstats_overlimit(sch); in timesortedlist_drop() 225 sch->q.qlen--; in timesortedlist_drop() 242 skb->dev = qdisc_dev(sch); in timesortedlist_remove() 250 sch->q.qlen--; in timesortedlist_remove() 291 reset_watchdog(sch); in etf_dequeue_timesortedlist() 429 sch->q.qlen--; in timesortedlist_clear() 438 if (q->watchdog.qdisc == sch) in etf_reset() [all …]
|
| A D | sch_dualpi2.c | 289 if (sch->qstats.backlog < 2 * psched_mtu(qdisc_dev(sch))) in must_drop() 391 if (unlikely(qdisc_qlen(sch) >= sch->limit) || in dualpi2_enqueue_skb() 420 ++sch->q.qlen; in dualpi2_enqueue_skb() 534 --sch->q.qlen; in dequeue_packet() 712 struct Qdisc *sch = q->sch; in dualpi2_timer() local 780 q = qdisc_priv(sch); in dualpi2_change() 781 sch_tree_lock(sch); in dualpi2_change() 872 while (qdisc_qlen(sch) > sch->limit || in dualpi2_change() 880 qdisc_tree_reduce_backlog(sch, old_qlen - qdisc_qlen(sch), in dualpi2_change() 893 q->memory_limit = get_memory_limit(sch, q->sch->limit); in dualpi2_reset_default() [all …]
|
| A D | sch_pie.c | 30 struct Qdisc *sch; member 92 if (unlikely(qdisc_qlen(sch) >= sch->limit)) { in pie_qdisc_enqueue() 99 if (!pie_drop_early(sch, &q->params, &q->vars, sch->qstats.backlog, in pie_qdisc_enqueue() 154 sch_tree_lock(sch); in pie_change() 196 qlen = sch->q.qlen; in pie_change() 197 while (sch->q.qlen > sch->limit) { in pie_change() 204 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped); in pie_change() 206 sch_tree_unlock(sch); in pie_change() 428 struct Qdisc *sch = q->sch; in pie_timer() local 452 q->sch = sch; in pie_init() [all …]
|
| A D | sch_skbprio.c | 84 if (sch->q.qlen < READ_ONCE(sch->limit)) { in skbprio_enqueue() 86 qdisc_qstats_backlog_inc(sch, skb); in skbprio_enqueue() 96 sch->q.qlen++; in skbprio_enqueue() 109 qdisc_qstats_backlog_inc(sch, skb); in skbprio_enqueue() 117 qdisc_drop(to_drop, sch, to_free); in skbprio_enqueue() 148 sch->q.qlen--; in skbprio_dequeue() 149 qdisc_qstats_backlog_dec(sch, skb); in skbprio_dequeue() 150 qdisc_bstats_update(sch, skb); in skbprio_dequeue() 174 WRITE_ONCE(sch->limit, ctl->limit); in skbprio_change() 191 sch->limit = 64; in skbprio_init() [all …]
|
| A D | sch_mqprio.c | 39 struct net_device *dev = qdisc_dev(sch); in mqprio_enable_offload() 110 mqprio_disable_offload(sch); in mqprio_destroy() 360 if (sch->parent != TC_H_ROOT) in mqprio_init() 428 sch->flags |= TCQ_F_MQROOT; in mqprio_init() 562 sch->q.qlen = 0; in mqprio_dump() 564 memset(&sch->qstats, 0, sizeof(sch->qstats)); in mqprio_dump() 579 sch->q.qlen += qdisc_qlen(qdisc); in mqprio_dump() 650 TC_H_MAKE(TC_H_MAJ(sch->handle), in mqprio_dump_class() 693 sch->q.qlen += qdisc_qlen(qdisc); in mqprio_dump_class_stats() 710 &sch->bstats, true) < 0 || in mqprio_dump_class_stats() [all …]
|
| A D | sch_gred.c | 115 struct Qdisc *sch) in gred_backlog() argument 183 sch->limit)) in gred_enqueue() 298 qdisc_reset_queue(sch); in gred_reset() 380 sch_tree_lock(sch); in gred_offload_dump_stats() 397 sch_tree_unlock(sch); in gred_offload_dump_stats() 439 sch_tree_lock(sch); in gred_change_table_def() 450 sch_tree_unlock(sch); in gred_change_table_def() 704 sch_tree_lock(sch); in gred_change() 720 sch_tree_unlock(sch); in gred_change() 727 sch_tree_unlock(sch); in gred_change() [all …]
|
| A D | sch_choke.c | 126 --sch->q.qlen; in choke_drop_by_idx() 129 qdisc_drop(skb, sch, to_free); in choke_drop_by_idx() 267 if (sch->q.qlen < q->limit) { in choke_enqueue() 270 ++sch->q.qlen; in choke_enqueue() 297 --sch->q.qlen; in choke_dequeue() 314 rtnl_qdisc_drop(skb, sch); in choke_reset() 377 sch_tree_lock(sch); in choke_change() 395 --sch->q.qlen; in choke_change() 398 qdisc_tree_reduce_backlog(sch, oqlen - sch->q.qlen, dropped); in choke_change() 406 sch_tree_lock(sch); in choke_change() [all …]
|
| A D | sch_fq_pie.c | 59 struct Qdisc *sch; member 158 if (unlikely(qdisc_qlen(sch) >= sch->limit)) { in fq_pie_qdisc_enqueue() 188 sch->q.qlen++; in fq_pie_qdisc_enqueue() 266 sch->q.qlen--; in fq_pie_qdisc_dequeue() 300 sch_tree_lock(sch); in fq_pie_change() 368 while (sch->q.qlen > sch->limit) { in fq_pie_change() 377 sch_tree_unlock(sch); in fq_pie_change() 381 sch_tree_unlock(sch); in fq_pie_change() 389 struct Qdisc *sch = q->sch; in fq_pie_timer() local 429 q->sch = sch; in fq_pie_init() [all …]
|
| A D | sch_sfq.c | 137 struct Qdisc *sch; member 303 sch->q.qlen--; in sfq_drop() 498 sch->q.qlen--; in sfq_dequeue() 604 struct Qdisc *sch = q->sch; in sfq_perturbation() local 615 sfq_rehash(sch); in sfq_perturbation() 676 sch_tree_lock(sch); in sfq_change() 714 sch_tree_unlock(sch); in sfq_change() 732 qlen = sch->q.qlen; in sfq_change() 740 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped); in sfq_change() 747 sch_tree_unlock(sch); in sfq_change() [all …]
|
| A D | sch_hhf.c | 362 sch->q.qlen--; in hhf_drop() 379 idx = hhf_classify(skb, sch); in hhf_enqueue() 402 if (++sch->q.qlen <= sch->limit) in hhf_enqueue() 414 qdisc_tree_reduce_backlog(sch, 1, prev_backlog - sch->qstats.backlog); in hhf_enqueue() 445 sch->q.qlen--; in hhf_dequeue() 534 sch_tree_lock(sch); in hhf_change() 564 qlen = sch->q.qlen; in hhf_change() 566 while (sch->q.qlen > sch->limit) { in hhf_change() 571 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, in hhf_change() 574 sch_tree_unlock(sch); in hhf_change() [all …]
|
| A D | sch_sfb.c | 297 qdisc_qstats_overlimit(sch); in sfb_enqueue() 350 qdisc_qstats_overlimit(sch); in sfb_enqueue() 376 qdisc_qstats_overlimit(sch); in sfb_enqueue() 409 sch->qstats.backlog += len; in sfb_enqueue() 410 sch->q.qlen++; in sfb_enqueue() 414 qdisc_qstats_drop(sch); in sfb_enqueue() 423 qdisc_qstats_drop(sch); in sfb_enqueue() 437 qdisc_bstats_update(sch, skb); in sfb_dequeue() 439 sch->q.qlen--; in sfb_dequeue() 524 sch_tree_lock(sch); in sfb_change() [all …]
|
| A D | sch_htb.c | 272 cl = htb_find(TC_H_MAKE(TC_H_MAJ(sch->handle), q->defcls), sch); in htb_classify() 638 qdisc_qstats_drop(sch); in htb_enqueue() 654 sch->q.qlen++; in htb_enqueue() 954 sch->q.qlen--; in htb_dequeue() 958 if (!sch->q.qlen) in htb_dequeue() 1372 return sch->dev_queue; in htb_select_queue() 1734 sch_tree_lock(sch); in htb_delete() 1753 sch_tree_unlock(sch); in htb_delete() 1947 sch_tree_lock(sch); in htb_change_class() 2018 sch_tree_lock(sch); in htb_change_class() [all …]
|
| A D | sch_netem.c | 419 sch->q.qlen++; in tfifo_enqueue() 474 qdisc_qstats_drop(sch); in netem_enqueue() 566 if (sch->q.tail) in netem_enqueue() 753 sch->q.qlen--; in netem_dequeue() 758 sch->q.qlen--; in netem_dequeue() 765 sch->q.qlen--; in netem_dequeue() 778 sch->q.qlen--; in netem_dequeue() 789 qdisc_reset_queue(sch); in netem_reset() 790 tfifo_reset(sch); in netem_reset() 1041 sch_tree_lock(sch); in netem_change() [all …]
|