Home
last modified time | relevance | path

Searched refs:sch (Results 1 – 25 of 44) sorted by relevance

12

/net/sched/
A Dsch_fifo.c23 READ_ONCE(sch->limit))) in bfifo_enqueue()
32 if (likely(sch->q.qlen < READ_ONCE(sch->limit))) in pfifo_enqueue()
46 if (likely(sch->q.qlen < READ_ONCE(sch->limit))) in pfifo_tail_enqueue()
51 __qdisc_queue_drop_head(sch, &sch->q, to_free); in pfifo_tail_enqueue()
52 qdisc_qstats_drop(sch); in pfifo_tail_enqueue()
55 qdisc_tree_reduce_backlog(sch, 0, prev_backlog - sch->qstats.backlog); in pfifo_tail_enqueue()
68 qopt.handle = sch->handle; in fifo_offload_init()
69 qopt.parent = sch->parent; in fifo_offload_init()
123 bypass = sch->limit >= psched_mtu(qdisc_dev(sch)); in __fifo_init()
144 fifo_offload_init(sch); in fifo_init()
[all …]
A Dsch_mq.c29 .handle = sch->handle, in mq_offload()
42 .handle = sch->handle, in mq_offload_stats()
44 .bstats = &sch->bstats, in mq_offload_stats()
45 .qstats = &sch->qstats, in mq_offload_stats()
58 mq_offload(sch, TC_MQ_DESTROY); in mq_destroy()
76 if (sch->parent != TC_H_ROOT) in mq_init()
100 sch->flags |= TCQ_F_MQROOT; in mq_init()
134 sch->q.qlen = 0; in mq_dump()
136 memset(&sch->qstats, 0, sizeof(sch->qstats)); in mq_dump()
196 qdisc_offload_graft_helper(qdisc_dev(sch), sch, new, *old, in mq_graft()
[all …]
A Dsch_prio.c80 qdisc_qstats_drop(sch); in prio_enqueue()
89 sch->q.qlen++; in prio_enqueue()
93 qdisc_qstats_drop(sch); in prio_enqueue()
122 sch->q.qlen--; in prio_dequeue()
144 .handle = sch->handle, in prio_offload()
145 .parent = sch->parent, in prio_offload()
209 sch_tree_lock(sch); in prio_tune()
222 sch_tree_unlock(sch); in prio_tune()
249 .handle = sch->handle, in prio_dump_offload()
250 .parent = sch->parent, in prio_dump_offload()
[all …]
A Dsch_red.c44 struct Qdisc *sch; member
136 sch->q.qlen++; in red_enqueue()
162 sch->q.qlen--; in red_dequeue()
191 .handle = sch->handle, in red_offload()
192 .parent = sch->parent, in red_offload()
277 sch_tree_lock(sch); in __red_change()
307 sch_tree_unlock(sch); in __red_change()
316 sch_tree_unlock(sch); in __red_change()
325 struct Qdisc *sch = q->sch; in red_adaptative_timer() local
345 q->sch = sch; in red_init()
[all …]
A Dsch_ets.c129 qopt.handle = sch->handle; in ets_offload_change()
130 qopt.parent = sch->parent; in ets_offload_change()
162 qopt.handle = sch->handle; in ets_offload_destroy()
163 qopt.parent = sch->parent; in ets_offload_destroy()
246 sch_tree_lock(sch); in ets_class_change()
248 sch_tree_unlock(sch); in ets_class_change()
250 ets_offload_change(sch); in ets_class_change()
448 sch->q.qlen++; in ets_qdisc_enqueue()
457 sch->q.qlen--; in ets_qdisc_dequeue_skb()
652 sch_tree_lock(sch); in ets_qdisc_change()
[all …]
A Dsch_ingress.c37 return ingress_find(sch, classid); in ingress_bind_filter()
51 struct ingress_sched_data *q = qdisc_priv(sch); in ingress_tcf_block()
81 struct net_device *dev = qdisc_dev(sch); in ingress_init()
86 if (sch->parent != TC_H_INGRESS) in ingress_init()
115 struct net_device *dev = qdisc_dev(sch); in ingress_destroy()
118 if (sch->parent != TC_H_INGRESS) in ingress_destroy()
196 return clsact_find(sch, classid); in clsact_bind_filter()
246 struct net_device *dev = qdisc_dev(sch); in clsact_init()
251 if (sch->parent != TC_H_CLSACT) in clsact_init()
294 struct net_device *dev = qdisc_dev(sch); in clsact_destroy()
[all …]
A Dsch_codel.c41 struct Qdisc *sch = ctx; in dequeue_func() local
56 qdisc_qstats_drop(sch); in drop_func()
64 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->params, &q->vars, in codel_qdisc_dequeue()
83 if (likely(qdisc_qlen(sch) < sch->limit)) { in codel_qdisc_enqueue()
87 q = qdisc_priv(sch); in codel_qdisc_enqueue()
114 sch_tree_lock(sch); in codel_change()
145 qlen = sch->q.qlen; in codel_change()
146 while (sch->q.qlen > sch->limit) { in codel_change()
153 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped); in codel_change()
155 sch_tree_unlock(sch); in codel_change()
[all …]
A Dsch_tbf.c152 qopt.handle = sch->handle; in tbf_offload_change()
153 qopt.parent = sch->parent; in tbf_offload_change()
192 .handle = sch->handle, in tbf_offload_graft()
193 .parent = sch->parent, in tbf_offload_graft()
198 qdisc_offload_graft_helper(qdisc_dev(sch), sch, new, old, in tbf_offload_graft()
233 sch->q.qlen += nb; in tbf_segment()
266 sch->q.qlen++; in tbf_enqueue()
311 sch->q.qlen--; in tbf_dequeue()
453 sch_tree_lock(sch); in tbf_change()
475 sch_tree_unlock(sch); in tbf_change()
[all …]
A Dsch_multiq.c71 qdisc_qstats_drop(sch); in multiq_enqueue()
79 sch->q.qlen++; in multiq_enqueue()
83 qdisc_qstats_drop(sch); in multiq_enqueue()
109 sch->q.qlen--; in multiq_dequeue()
148 multiq_reset(struct Qdisc *sch) in multiq_reset() argument
159 multiq_destroy(struct Qdisc *sch) in multiq_destroy() argument
193 sch_tree_lock(sch); in multiq_tune()
205 sch_tree_unlock(sch); in multiq_tune()
216 TC_H_MAKE(sch->handle, in multiq_tune()
219 sch_tree_lock(sch); in multiq_tune()
[all …]
A Dsch_drr.c100 sch_tree_lock(sch); in drr_change_class()
103 sch_tree_unlock(sch); in drr_change_class()
135 sch_tree_lock(sch); in drr_change_class()
137 sch_tree_unlock(sch); in drr_change_class()
163 sch_tree_lock(sch); in drr_delete_class()
168 sch_tree_unlock(sch); in drr_delete_class()
170 drr_destroy_class(sch, cl); in drr_delete_class()
349 qdisc_qstats_drop(sch); in drr_enqueue()
358 qdisc_qstats_drop(sch); in drr_enqueue()
369 sch->q.qlen++; in drr_enqueue()
[all …]
A Dsch_fq_codel.c181 sch->q.qlen -= i; in fq_codel_drop()
218 if (++sch->q.qlen <= sch->limit && !memory_limited) in fq_codel_enqueue()
268 sch->q.qlen--; in dequeue_func()
304 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->cparams, in fq_codel_dequeue()
393 sch_tree_lock(sch); in fq_codel_change()
442 while (sch->q.qlen > sch->limit || in fq_codel_change()
454 sch_tree_unlock(sch); in fq_codel_change()
474 sch->limit = 10*1024; in fq_codel_init()
516 if (sch->limit >= 1) in fq_codel_init()
593 sch_tree_lock(sch); in fq_codel_dump_stats()
[all …]
A Dsch_cbs.c99 sch->qstats.backlog += len; in cbs_child_enqueue()
100 sch->q.qlen++; in cbs_child_enqueue()
136 return q->enqueue(skb, sch, to_free); in cbs_enqueue()
169 qdisc_qstats_backlog_dec(sch, skb); in cbs_child_dequeue()
170 qdisc_bstats_update(sch, skb); in cbs_child_dequeue()
171 sch->q.qlen--; in cbs_child_dequeue()
207 skb = cbs_child_dequeue(sch, qdisc); in cbs_dequeue_soft()
243 return q->dequeue(sch); in cbs_dequeue()
413 sch->handle, extack); in cbs_init()
430 return cbs_change(sch, opt, extack); in cbs_init()
[all …]
A Dsch_etf.c192 sch->q.qlen++; in etf_enqueue_timesortedlist()
195 reset_watchdog(sch); in etf_enqueue_timesortedlist()
218 skb->dev = qdisc_dev(sch); in timesortedlist_drop()
224 qdisc_qstats_overlimit(sch); in timesortedlist_drop()
225 sch->q.qlen--; in timesortedlist_drop()
242 skb->dev = qdisc_dev(sch); in timesortedlist_remove()
250 sch->q.qlen--; in timesortedlist_remove()
291 reset_watchdog(sch); in etf_dequeue_timesortedlist()
429 sch->q.qlen--; in timesortedlist_clear()
438 if (q->watchdog.qdisc == sch) in etf_reset()
[all …]
A Dsch_dualpi2.c289 if (sch->qstats.backlog < 2 * psched_mtu(qdisc_dev(sch))) in must_drop()
391 if (unlikely(qdisc_qlen(sch) >= sch->limit) || in dualpi2_enqueue_skb()
420 ++sch->q.qlen; in dualpi2_enqueue_skb()
534 --sch->q.qlen; in dequeue_packet()
712 struct Qdisc *sch = q->sch; in dualpi2_timer() local
780 q = qdisc_priv(sch); in dualpi2_change()
781 sch_tree_lock(sch); in dualpi2_change()
872 while (qdisc_qlen(sch) > sch->limit || in dualpi2_change()
880 qdisc_tree_reduce_backlog(sch, old_qlen - qdisc_qlen(sch), in dualpi2_change()
893 q->memory_limit = get_memory_limit(sch, q->sch->limit); in dualpi2_reset_default()
[all …]
A Dsch_pie.c30 struct Qdisc *sch; member
92 if (unlikely(qdisc_qlen(sch) >= sch->limit)) { in pie_qdisc_enqueue()
99 if (!pie_drop_early(sch, &q->params, &q->vars, sch->qstats.backlog, in pie_qdisc_enqueue()
154 sch_tree_lock(sch); in pie_change()
196 qlen = sch->q.qlen; in pie_change()
197 while (sch->q.qlen > sch->limit) { in pie_change()
204 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped); in pie_change()
206 sch_tree_unlock(sch); in pie_change()
428 struct Qdisc *sch = q->sch; in pie_timer() local
452 q->sch = sch; in pie_init()
[all …]
A Dsch_skbprio.c84 if (sch->q.qlen < READ_ONCE(sch->limit)) { in skbprio_enqueue()
86 qdisc_qstats_backlog_inc(sch, skb); in skbprio_enqueue()
96 sch->q.qlen++; in skbprio_enqueue()
109 qdisc_qstats_backlog_inc(sch, skb); in skbprio_enqueue()
117 qdisc_drop(to_drop, sch, to_free); in skbprio_enqueue()
148 sch->q.qlen--; in skbprio_dequeue()
149 qdisc_qstats_backlog_dec(sch, skb); in skbprio_dequeue()
150 qdisc_bstats_update(sch, skb); in skbprio_dequeue()
174 WRITE_ONCE(sch->limit, ctl->limit); in skbprio_change()
191 sch->limit = 64; in skbprio_init()
[all …]
A Dsch_mqprio.c39 struct net_device *dev = qdisc_dev(sch); in mqprio_enable_offload()
110 mqprio_disable_offload(sch); in mqprio_destroy()
360 if (sch->parent != TC_H_ROOT) in mqprio_init()
428 sch->flags |= TCQ_F_MQROOT; in mqprio_init()
562 sch->q.qlen = 0; in mqprio_dump()
564 memset(&sch->qstats, 0, sizeof(sch->qstats)); in mqprio_dump()
579 sch->q.qlen += qdisc_qlen(qdisc); in mqprio_dump()
650 TC_H_MAKE(TC_H_MAJ(sch->handle), in mqprio_dump_class()
693 sch->q.qlen += qdisc_qlen(qdisc); in mqprio_dump_class_stats()
710 &sch->bstats, true) < 0 || in mqprio_dump_class_stats()
[all …]
A Dsch_gred.c115 struct Qdisc *sch) in gred_backlog() argument
183 sch->limit)) in gred_enqueue()
298 qdisc_reset_queue(sch); in gred_reset()
380 sch_tree_lock(sch); in gred_offload_dump_stats()
397 sch_tree_unlock(sch); in gred_offload_dump_stats()
439 sch_tree_lock(sch); in gred_change_table_def()
450 sch_tree_unlock(sch); in gred_change_table_def()
704 sch_tree_lock(sch); in gred_change()
720 sch_tree_unlock(sch); in gred_change()
727 sch_tree_unlock(sch); in gred_change()
[all …]
A Dsch_choke.c126 --sch->q.qlen; in choke_drop_by_idx()
129 qdisc_drop(skb, sch, to_free); in choke_drop_by_idx()
267 if (sch->q.qlen < q->limit) { in choke_enqueue()
270 ++sch->q.qlen; in choke_enqueue()
297 --sch->q.qlen; in choke_dequeue()
314 rtnl_qdisc_drop(skb, sch); in choke_reset()
377 sch_tree_lock(sch); in choke_change()
395 --sch->q.qlen; in choke_change()
398 qdisc_tree_reduce_backlog(sch, oqlen - sch->q.qlen, dropped); in choke_change()
406 sch_tree_lock(sch); in choke_change()
[all …]
A Dsch_fq_pie.c59 struct Qdisc *sch; member
158 if (unlikely(qdisc_qlen(sch) >= sch->limit)) { in fq_pie_qdisc_enqueue()
188 sch->q.qlen++; in fq_pie_qdisc_enqueue()
266 sch->q.qlen--; in fq_pie_qdisc_dequeue()
300 sch_tree_lock(sch); in fq_pie_change()
368 while (sch->q.qlen > sch->limit) { in fq_pie_change()
377 sch_tree_unlock(sch); in fq_pie_change()
381 sch_tree_unlock(sch); in fq_pie_change()
389 struct Qdisc *sch = q->sch; in fq_pie_timer() local
429 q->sch = sch; in fq_pie_init()
[all …]
A Dsch_sfq.c137 struct Qdisc *sch; member
303 sch->q.qlen--; in sfq_drop()
498 sch->q.qlen--; in sfq_dequeue()
604 struct Qdisc *sch = q->sch; in sfq_perturbation() local
615 sfq_rehash(sch); in sfq_perturbation()
676 sch_tree_lock(sch); in sfq_change()
714 sch_tree_unlock(sch); in sfq_change()
732 qlen = sch->q.qlen; in sfq_change()
740 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped); in sfq_change()
747 sch_tree_unlock(sch); in sfq_change()
[all …]
A Dsch_hhf.c362 sch->q.qlen--; in hhf_drop()
379 idx = hhf_classify(skb, sch); in hhf_enqueue()
402 if (++sch->q.qlen <= sch->limit) in hhf_enqueue()
414 qdisc_tree_reduce_backlog(sch, 1, prev_backlog - sch->qstats.backlog); in hhf_enqueue()
445 sch->q.qlen--; in hhf_dequeue()
534 sch_tree_lock(sch); in hhf_change()
564 qlen = sch->q.qlen; in hhf_change()
566 while (sch->q.qlen > sch->limit) { in hhf_change()
571 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, in hhf_change()
574 sch_tree_unlock(sch); in hhf_change()
[all …]
A Dsch_sfb.c297 qdisc_qstats_overlimit(sch); in sfb_enqueue()
350 qdisc_qstats_overlimit(sch); in sfb_enqueue()
376 qdisc_qstats_overlimit(sch); in sfb_enqueue()
409 sch->qstats.backlog += len; in sfb_enqueue()
410 sch->q.qlen++; in sfb_enqueue()
414 qdisc_qstats_drop(sch); in sfb_enqueue()
423 qdisc_qstats_drop(sch); in sfb_enqueue()
437 qdisc_bstats_update(sch, skb); in sfb_dequeue()
439 sch->q.qlen--; in sfb_dequeue()
524 sch_tree_lock(sch); in sfb_change()
[all …]
A Dsch_htb.c272 cl = htb_find(TC_H_MAKE(TC_H_MAJ(sch->handle), q->defcls), sch); in htb_classify()
638 qdisc_qstats_drop(sch); in htb_enqueue()
654 sch->q.qlen++; in htb_enqueue()
954 sch->q.qlen--; in htb_dequeue()
958 if (!sch->q.qlen) in htb_dequeue()
1372 return sch->dev_queue; in htb_select_queue()
1734 sch_tree_lock(sch); in htb_delete()
1753 sch_tree_unlock(sch); in htb_delete()
1947 sch_tree_lock(sch); in htb_change_class()
2018 sch_tree_lock(sch); in htb_change_class()
[all …]
A Dsch_netem.c419 sch->q.qlen++; in tfifo_enqueue()
474 qdisc_qstats_drop(sch); in netem_enqueue()
566 if (sch->q.tail) in netem_enqueue()
753 sch->q.qlen--; in netem_dequeue()
758 sch->q.qlen--; in netem_dequeue()
765 sch->q.qlen--; in netem_dequeue()
778 sch->q.qlen--; in netem_dequeue()
789 qdisc_reset_queue(sch); in netem_reset()
790 tfifo_reset(sch); in netem_reset()
1041 sch_tree_lock(sch); in netem_change()
[all …]

Completed in 57 milliseconds

12