/linux-6.3-rc2/net/sched/ |
A D | sch_generic.c | 660 .dev_queue = &noop_netdev_queue, 931 if (!dev_queue) { in qdisc_alloc() 937 dev = dev_queue->dev; in qdisc_alloc() 973 sch->dev_queue = dev_queue; in qdisc_alloc() 1115 dev_queue->qdisc_sleeping = qdisc; in dev_graft_qdisc() 1157 dev_queue->qdisc_sleeping = qdisc; in attach_one_default_qdisc() 1202 struct netdev_queue *dev_queue, in transition_one_qdisc() argument 1275 qdisc = dev_queue->qdisc_sleeping; in dev_reset_queue() 1300 struct netdev_queue *dev_queue; in some_qdisc_is_busy() local 1306 q = dev_queue->qdisc_sleeping; in some_qdisc_is_busy() [all …]
|
A D | sch_mq.c | 72 struct netdev_queue *dev_queue; in mq_init() local 89 dev_queue = netdev_get_tx_queue(dev, ntx); in mq_init() 115 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in mq_attach() 178 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_graft() local 185 *old = dev_graft_qdisc(dev_queue, new); in mq_graft() 203 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_leaf() local 205 return dev_queue->qdisc_sleeping; in mq_leaf() 220 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_dump_class() local 224 tcm->tcm_info = dev_queue->qdisc_sleeping->handle; in mq_dump_class() 231 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_dump_class_stats() local [all …]
|
A D | sch_mqprio.c | 229 struct netdev_queue *dev_queue; in mqprio_init() local 273 dev_queue = netdev_get_tx_queue(dev, i); in mqprio_init() 274 qdisc = qdisc_create_dflt(dev_queue, in mqprio_init() 318 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in mqprio_attach() 343 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); in mqprio_graft() local 345 if (!dev_queue) in mqprio_graft() 351 *old = dev_graft_qdisc(dev_queue, new); in mqprio_graft() 462 if (!dev_queue) in mqprio_leaf() 465 return dev_queue->qdisc_sleeping; in mqprio_leaf() 498 tcm->tcm_info = dev_queue->qdisc_sleeping->handle; in mqprio_dump_class() [all …]
|
A D | sch_htb.c | 1378 return sch->dev_queue; in htb_select_queue() 1437 cl_old->leaf.q->dev_queue = queue_new; in htb_offload_move_qdisc() 1453 struct netdev_queue *dev_queue = sch->dev_queue; in htb_graft() local 1553 struct netdev_queue *dev_queue; in htb_destroy_class_offload() local 1562 dev_queue = htb_offload_get_queue(cl); in htb_destroy_class_offload() 1594 htb_graft_helper(dev_queue, old); in htb_destroy_class_offload() 1727 struct netdev_queue *dev_queue = sch->dev_queue; in htb_delete() local 1779 struct netdev_queue *dev_queue; in htb_change_class() local 1905 dev_queue = sch->dev_queue; in htb_change_class() 1942 htb_graft_helper(dev_queue, old_q); in htb_change_class() [all …]
|
A D | sch_api.c | 1101 old = dev_graft_qdisc(dev_queue, new); in qdisc_graft() 1198 struct netdev_queue *dev_queue, in qdisc_create() argument 1671 struct netdev_queue *dev_queue; in tc_modify_qdisc() local 1676 dev_queue = p->dev_queue; in tc_modify_qdisc() 1680 q = qdisc_create(dev, dev_queue, in tc_modify_qdisc() 1777 struct netdev_queue *dev_queue; in tc_dump_qdisc() local 1790 dev_queue = dev_ingress_queue(dev); in tc_dump_qdisc() 1791 if (dev_queue && in tc_dump_qdisc() 2217 struct netdev_queue *dev_queue; in tc_dump_tclass() local 2234 dev_queue = dev_ingress_queue(dev); in tc_dump_tclass() [all …]
|
A D | sch_plug.c | 184 netif_schedule_queue(sch->dev_queue); in plug_change() 192 netif_schedule_queue(sch->dev_queue); in plug_change()
|
A D | sch_taprio.c | 2057 struct netdev_queue *dev_queue; in taprio_init() local 2060 dev_queue = netdev_get_tx_queue(dev, i); in taprio_init() 2061 qdisc = qdisc_create_dflt(dev_queue, in taprio_init() 2093 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in taprio_attach() 2095 old = dev_graft_qdisc(qdisc->dev_queue, sch); in taprio_attach() 2129 if (!dev_queue) in taprio_graft() 2136 *old = dev_graft_qdisc(dev_queue, new); in taprio_graft() 2311 if (!dev_queue) in taprio_leaf() 2314 return dev_queue->qdisc_sleeping; in taprio_leaf() 2333 tcm->tcm_info = dev_queue->qdisc_sleeping->handle; in taprio_dump_class() [all …]
|
A D | sch_cbs.c | 412 q->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in cbs_init() 423 q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0); in cbs_init() 498 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in cbs_graft()
|
A D | sch_prio.c | 198 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in prio_tune() 294 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in prio_graft()
|
A D | sch_drr.c | 111 cl->qdisc = qdisc_create_dflt(sch->dev_queue, in drr_change_class() 210 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in drr_graft_class()
|
A D | sch_fifo.c | 259 q = qdisc_create_dflt(sch->dev_queue, ops, TC_H_MAKE(sch->handle, 1), in fifo_create_dflt()
|
A D | sch_ets.c | 254 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in ets_class_graft() 637 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in ets_qdisc_change()
|
A D | sch_multiq.c | 214 child = qdisc_create_dflt(sch->dev_queue, in multiq_tune()
|
A D | sch_hfsc.c | 1054 cl->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in hfsc_change_class() 1177 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in hfsc_graft_class() 1403 q->root.qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in hfsc_init_qdisc()
|
A D | sch_etf.c | 379 q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0); in etf_init()
|
/linux-6.3-rc2/include/linux/ |
A D | netdevice.h | 3371 dev_queue->dql.min_limit = min_limit; in netdev_queue_set_dql_min_limit() 3385 prefetchw(&dev_queue->dql.num_queued); in netdev_txq_bql_enqueue_prefetchw() 3399 prefetchw(&dev_queue->dql.limit); in netdev_txq_bql_complete_prefetchw() 3417 dql_queued(&dev_queue->dql, bytes); in netdev_tx_sent_queue() 3419 if (likely(dql_avail(&dev_queue->dql) >= 0)) in netdev_tx_sent_queue() 3449 dql_queued(&dev_queue->dql, bytes); in __netdev_tx_sent_queue() 3451 return netif_tx_queue_stopped(dev_queue); in __netdev_tx_sent_queue() 3453 netdev_tx_sent_queue(dev_queue, bytes); in __netdev_tx_sent_queue() 3496 dql_completed(&dev_queue->dql, bytes); in netdev_tx_completed_queue() 3505 if (unlikely(dql_avail(&dev_queue->dql) < 0)) in netdev_tx_completed_queue() [all …]
|
/linux-6.3-rc2/include/net/ |
A D | sch_generic.h | 104 struct netdev_queue *dev_queue; member 536 struct Qdisc *q = rcu_dereference_rtnl(qdisc->dev_queue->qdisc); in qdisc_root() 543 return rcu_dereference_bh(qdisc->dev_queue->qdisc); in qdisc_root_bh() 548 return qdisc->dev_queue->qdisc_sleeping; in qdisc_root_sleeping() 561 return qdisc->dev_queue->dev; in qdisc_dev() 652 struct Qdisc *dev_graft_qdisc(struct netdev_queue *dev_queue, 685 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue, 689 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue,
|
A D | pkt_sched.h | 141 return dev_net(q->dev_queue->dev); in qdisc_net()
|
/linux-6.3-rc2/drivers/net/ethernet/hisilicon/hns/ |
A D | hns_enet.c | 310 struct netdev_queue *dev_queue; in hns_nic_net_xmit_hw() local 361 dev_queue = netdev_get_tx_queue(ndev, skb->queue_mapping); in hns_nic_net_xmit_hw() 362 netdev_tx_sent_queue(dev_queue, skb->len); in hns_nic_net_xmit_hw() 951 struct netdev_queue *dev_queue; in hns_nic_tx_poll_one() local 980 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index); in hns_nic_tx_poll_one() 981 netdev_tx_completed_queue(dev_queue, pkts, bytes); in hns_nic_tx_poll_one() 992 if (netif_tx_queue_stopped(dev_queue) && in hns_nic_tx_poll_one() 994 netif_tx_wake_queue(dev_queue); in hns_nic_tx_poll_one() 1035 struct netdev_queue *dev_queue; in hns_nic_tx_clr_all_bufs() local 1045 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index); in hns_nic_tx_clr_all_bufs() [all …]
|
A D | hns_ethtool.c | 376 struct netdev_queue *dev_queue; in __lb_other_process() local 407 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index); in __lb_other_process() 408 netdev_tx_reset_queue(dev_queue); in __lb_other_process()
|
/linux-6.3-rc2/drivers/net/ethernet/mellanox/mlx5/core/en/ |
A D | qos.c | 363 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, qid); in mlx5e_reset_qdisc() local 364 struct Qdisc *qdisc = dev_queue->qdisc_sleeping; in mlx5e_reset_qdisc()
|
/linux-6.3-rc2/drivers/net/xen-netback/ |
A D | xenbus.c | 31 struct netdev_queue *dev_queue; in xenvif_read_io_ring() local 92 dev_queue = netdev_get_tx_queue(queue->vif->dev, queue->id); in xenvif_read_io_ring() 97 netif_tx_queue_stopped(dev_queue) ? "stopped" : "running"); in xenvif_read_io_ring()
|
/linux-6.3-rc2/drivers/net/ethernet/hisilicon/hns3/ |
A D | hns3_enet.c | 813 struct netdev_queue *dev_queue; in hns3_reset_tx_queue() local 817 dev_queue = netdev_get_tx_queue(ndev, in hns3_reset_tx_queue() 819 netdev_tx_reset_queue(dev_queue); in hns3_reset_tx_queue() 2295 struct netdev_queue *dev_queue; in hns3_nic_net_xmit() local 2339 dev_queue = netdev_get_tx_queue(netdev, ring->queue_index); in hns3_nic_net_xmit() 2340 doorbell = __netdev_tx_sent_queue(dev_queue, desc_cb->send_bytes, in hns3_nic_net_xmit() 3620 struct netdev_queue *dev_queue; in hns3_clean_tx_ring() local 3637 dev_queue = netdev_get_tx_queue(netdev, ring->tqp->tqp_index); in hns3_clean_tx_ring() 3638 netdev_tx_completed_queue(dev_queue, pkts, bytes); in hns3_clean_tx_ring() 3646 if (netif_tx_queue_stopped(dev_queue) && in hns3_clean_tx_ring() [all …]
|
/linux-6.3-rc2/tools/perf/Documentation/ |
A D | perf-trace.txt | 326 next block:*_unplug and the next three net:*dev_queue events, this last one 329 …# perf trace -e sched:*switch/nr=2/,block:*_plug/nr=4/,block:*_unplug/nr=1/,net:*dev_queue/nr=3,ma…
|
/linux-6.3-rc2/drivers/net/ethernet/myricom/myri10ge/ |
A D | myri10ge.c | 1380 struct netdev_queue *dev_queue; in myri10ge_tx_done() local 1415 dev_queue = netdev_get_tx_queue(ss->dev, ss - ss->mgp->ss); in myri10ge_tx_done() 1426 __netif_tx_trylock(dev_queue)) { in myri10ge_tx_done() 1432 __netif_tx_unlock(dev_queue); in myri10ge_tx_done() 1436 if (netif_tx_queue_stopped(dev_queue) && in myri10ge_tx_done() 1440 netif_tx_wake_queue(dev_queue); in myri10ge_tx_done()
|