| /linux/tools/testing/selftests/drivers/net/ |
| A D | queues.py | 18 if queues: 19 return len([q for q in queues if q['type'] == 'rx']) 24 queues = nl_get_queues(cfg, nl) 25 if not queues: 29 ksft_eq(queues, expected) 33 queues = nl_get_queues(cfg, nl) 34 if not queues: 50 queues = nl_get_queues(cfg, nl) 51 ksft_eq(queues, expected) 55 queues = nl_get_queues(cfg, nl) [all …]
|
| /linux/drivers/gpu/drm/imagination/ |
| A D | pvr_context.c | 216 if (IS_ERR(ctx->queues.geometry)) { in pvr_context_create_queues() 217 err = PTR_ERR(ctx->queues.geometry); in pvr_context_create_queues() 218 ctx->queues.geometry = NULL; in pvr_context_create_queues() 224 if (IS_ERR(ctx->queues.fragment)) { in pvr_context_create_queues() 226 ctx->queues.fragment = NULL; in pvr_context_create_queues() 234 if (IS_ERR(ctx->queues.compute)) { in pvr_context_create_queues() 235 err = PTR_ERR(ctx->queues.compute); in pvr_context_create_queues() 236 ctx->queues.compute = NULL; in pvr_context_create_queues() 244 if (IS_ERR(ctx->queues.transfer)) { in pvr_context_create_queues() 246 ctx->queues.transfer = NULL; in pvr_context_create_queues() [all …]
|
| A D | pvr_queue.c | 523 job->ctx->queues.fragment); in pvr_queue_prepare_job() 596 mutex_lock(&pvr_dev->queues.lock); in pvr_queue_update_active_state() 598 mutex_unlock(&pvr_dev->queues.lock); in pvr_queue_update_active_state() 819 mutex_lock(&pvr_dev->queues.lock); in pvr_queue_timedout_job() 821 mutex_unlock(&pvr_dev->queues.lock); in pvr_queue_timedout_job() 835 mutex_lock(&pvr_dev->queues.lock); in pvr_queue_timedout_job() 843 mutex_unlock(&pvr_dev->queues.lock); in pvr_queue_timedout_job() 1300 mutex_lock(&pvr_dev->queues.lock); in pvr_queue_create() 1302 mutex_unlock(&pvr_dev->queues.lock); in pvr_queue_create() 1329 mutex_lock(&pvr_dev->queues.lock); in pvr_queue_device_pre_reset() [all …]
|
| A D | pvr_context.h | 87 } queues; member 98 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.geometry : NULL; in pvr_context_get_queue_for_job() 100 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.fragment : NULL; in pvr_context_get_queue_for_job() 102 return ctx->type == DRM_PVR_CTX_TYPE_COMPUTE ? ctx->queues.compute : NULL; in pvr_context_get_queue_for_job() 104 return ctx->type == DRM_PVR_CTX_TYPE_TRANSFER_FRAG ? ctx->queues.transfer : NULL; in pvr_context_get_queue_for_job()
|
| /linux/tools/testing/selftests/net/ |
| A D | ncdevmem.c | 243 struct netdev_queue_id *queues; in do_server() local 279 queues = malloc(sizeof(*queues) * num_queues); in do_server() 282 queues[i]._present.type = 1; in do_server() 283 queues[i]._present.id = 1; in do_server() 285 queues[i].id = start_queue + i; in do_server() 467 struct netdev_queue_id *queues; in run_devmem_tests() local 481 queues = calloc(num_queues, sizeof(*queues)); in run_devmem_tests() 490 queues[i]._present.type = 1; in run_devmem_tests() 491 queues[i]._present.id = 1; in run_devmem_tests() 506 queues[i]._present.type = 1; in run_devmem_tests() [all …]
|
| /linux/net/sched/ |
| A D | sch_multiq.c | 25 struct Qdisc **queues; member 54 return q->queues[0]; in multiq_classify() 56 return q->queues[band]; in multiq_classify() 166 qdisc_put(q->queues[band]); in multiq_destroy() 168 kfree(q->queues); in multiq_destroy() 220 old = q->queues[i]; in multiq_tune() 221 q->queues[i] = child; in multiq_tune() 241 q->queues = NULL; in multiq_init() 253 if (!q->queues) in multiq_init() 299 return q->queues[band]; in multiq_leaf() [all …]
|
| A D | sch_prio.c | 26 struct Qdisc *queues[TCQ_PRIO_BANDS]; member 63 return q->queues[q->prio2band[0]]; in prio_classify() 65 return q->queues[band]; in prio_classify() 137 qdisc_reset(q->queues[prio]); in prio_reset() 173 qdisc_put(q->queues[prio]); in prio_destroy() 201 if (!queues[i]) { in prio_tune() 203 qdisc_put(queues[--i]); in prio_tune() 217 q->queues[i] = queues[i]; in prio_tune() 225 qdisc_put(q->queues[i]); in prio_tune() 322 return q->queues[band]; in prio_leaf() [all …]
|
| /linux/Documentation/networking/ |
| A D | multi-pf-netdev.rst | 142 - /sys/class/net/eth2/queues/tx-0/xps_cpus:000001 143 - /sys/class/net/eth2/queues/tx-1/xps_cpus:001000 144 - /sys/class/net/eth2/queues/tx-2/xps_cpus:000002 145 - /sys/class/net/eth2/queues/tx-3/xps_cpus:002000 146 - /sys/class/net/eth2/queues/tx-4/xps_cpus:000004 147 - /sys/class/net/eth2/queues/tx-5/xps_cpus:004000 148 - /sys/class/net/eth2/queues/tx-6/xps_cpus:000008 149 - /sys/class/net/eth2/queues/tx-7/xps_cpus:008000 150 - /sys/class/net/eth2/queues/tx-8/xps_cpus:000010 151 - /sys/class/net/eth2/queues/tx-9/xps_cpus:010000 [all …]
|
| A D | tc-queue-filters.rst | 7 TC can be used for directing traffic to either a set of queues or 12 1) TC filter directing traffic to a set of queues is achieved 14 the priority maps to a traffic class (set of queues) when 23 queues and/or a single queue are supported as below: 25 1) TC flower filter directs incoming traffic to a set of queues using
|
| /linux/drivers/net/wireless/silabs/wfx/ |
| A D | queue.c | 244 queues[num_queues] = &wvif->tx_queue[i]; in wfx_tx_queues_get_skb() 248 swap(queues[j - 1], queues[j]); in wfx_tx_queues_get_skb() 256 skb = skb_dequeue(&queues[i]->offchan); in wfx_tx_queues_get_skb() 264 atomic_inc(&queues[i]->pending_frames); in wfx_tx_queues_get_skb() 265 trace_queues_stats(wdev, queues[i]); in wfx_tx_queues_get_skb() 278 skb = skb_dequeue(&queues[i]->cab); in wfx_tx_queues_get_skb() 287 atomic_inc(&queues[i]->pending_frames); in wfx_tx_queues_get_skb() 288 trace_queues_stats(wdev, queues[i]); in wfx_tx_queues_get_skb() 297 skb = skb_dequeue(&queues[i]->normal); in wfx_tx_queues_get_skb() 299 atomic_inc(&queues[i]->pending_frames); in wfx_tx_queues_get_skb() [all …]
|
| /linux/drivers/scsi/aacraid/ |
| A D | comminit.c | 373 struct aac_entry * queues; in aac_comm_init() local 399 queues += HOST_NORM_CMD_ENTRIES; in aac_comm_init() 406 queues += HOST_HIGH_CMD_ENTRIES; in aac_comm_init() 413 queues += ADAP_NORM_CMD_ENTRIES; in aac_comm_init() 420 queues += ADAP_HIGH_CMD_ENTRIES; in aac_comm_init() 426 queues += HOST_NORM_RESP_ENTRIES; in aac_comm_init() 637 if (dev->queues == NULL) { in aac_init_adapter() 643 kfree(dev->queues); in aac_init_adapter() 644 dev->queues = NULL; in aac_init_adapter() 651 kfree(dev->queues); in aac_init_adapter() [all …]
|
| /linux/Documentation/ABI/testing/ |
| A D | sysfs-class-net-queues | 1 What: /sys/class/net/<iface>/queues/rx-<queue>/rps_cpus 11 What: /sys/class/net/<iface>/queues/rx-<queue>/rps_flow_cnt 19 What: /sys/class/net/<iface>/queues/tx-<queue>/tx_timeout 27 What: /sys/class/net/<iface>/queues/tx-<queue>/tx_maxrate 35 What: /sys/class/net/<iface>/queues/tx-<queue>/xps_cpus 45 What: /sys/class/net/<iface>/queues/tx-<queue>/xps_rxqs 56 What: /sys/class/net/<iface>/queues/tx-<queue>/byte_queue_limits/hold_time 65 What: /sys/class/net/<iface>/queues/tx-<queue>/byte_queue_limits/inflight 73 What: /sys/class/net/<iface>/queues/tx-<queue>/byte_queue_limits/limit 82 What: /sys/class/net/<iface>/queues/tx-<queue>/byte_queue_limits/limit_max [all …]
|
| /linux/drivers/media/platform/nxp/imx8-isi/ |
| A D | imx8-isi-m2m.c | 58 } queues; member 85 return &ctx->queues.out; in mxc_isi_m2m_ctx_qdata() 87 return &ctx->queues.cap; in mxc_isi_m2m_ctx_qdata() 152 ctx->queues.out.info, in mxc_isi_m2m_device_run() 153 &ctx->queues.out.format); in mxc_isi_m2m_device_run() 155 ctx->queues.cap.info, in mxc_isi_m2m_device_run() 156 &ctx->queues.cap.format); in mxc_isi_m2m_device_run() 471 ctx->queues.out.format = *pix; in mxc_isi_m2m_s_fmt_vid() 472 ctx->queues.out.info = info; in mxc_isi_m2m_s_fmt_vid() 479 ctx->queues.cap.format = *pix; in mxc_isi_m2m_s_fmt_vid() [all …]
|
| /linux/drivers/nvme/target/ |
| A D | loop.c | 30 struct nvme_loop_queue *queues; member 71 return queue - queue->ctrl->queues; in nvme_loop_queue_idx() 198 iod->queue = &ctrl->queues[queue_idx]; in nvme_loop_init_iod() 275 nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); in nvme_loop_destroy_admin_queue() 292 kfree(ctrl->queues); in nvme_loop_free_ctrl() 329 ctrl->queues[i].ctrl = ctrl; in nvme_loop_init_io_queues() 362 ctrl->queues[0].ctrl = ctrl; in nvme_loop_configure_admin_queue() 403 nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); in nvme_loop_configure_admin_queue() 583 ctrl->queues = kcalloc(opts->nr_io_queues + 1, sizeof(*ctrl->queues), in nvme_loop_create_ctrl() 585 if (!ctrl->queues) in nvme_loop_create_ctrl() [all …]
|
| /linux/Documentation/devicetree/bindings/soc/ti/ |
| A D | keystone-navigator-qmss.txt | 32 -- managed-queues : the actual queues managed by each queue manager 33 instance, specified as <"base queue #" "# of queues">. 51 - qpend : pool of qpend(interruptible) queues 53 as free descriptor queues or the 54 transmit DMA queues. 58 <"base queue #" "# of queues">. 60 for interruptible queues. The driver additionally sets 67 queues looking for descriptors that have been pushed 90 monitor up to 32 queues starting at the base queue #. 132 managed-queues = <0 0x2000>; [all …]
|
| /linux/sound/virtio/ |
| A D | virtio_card.h | 64 struct virtio_snd_queue queues[VIRTIO_SND_VQ_MAX]; member 86 return &snd->queues[VIRTIO_SND_VQ_CONTROL]; in virtsnd_control_queue() 92 return &snd->queues[VIRTIO_SND_VQ_EVENT]; in virtsnd_event_queue() 98 return &snd->queues[VIRTIO_SND_VQ_TX]; in virtsnd_tx_queue() 104 return &snd->queues[VIRTIO_SND_VQ_RX]; in virtsnd_rx_queue()
|
| /linux/Documentation/networking/device_drivers/ethernet/ti/ |
| A D | cpsw.rst | 28 - CBS shapers should be used with rated queues 156 // Check if num of queues is set correctly: 175 // Leave last 2 tx queues not rated. 176 $ echo 40 > /sys/class/net/eth0/queues/tx-0/tx_maxrate 181 // Check maximum rate of tx (cpdma) queues: 182 $ cat /sys/class/net/eth0/queues/tx-*/tx_maxrate 206 map 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 queues 1@4 hw 0 387 // Check if num of queues is set correctly: 407 // Leave last 4 tx queues as not rated 415 // Check maximum rate of tx (cpdma) queues: [all …]
|
| /linux/Documentation/devicetree/bindings/misc/ |
| A D | intel,ixp4xx-ahb-queue-manager.yaml | 14 The IXP4xx AHB Queue Manager maintains queues as circular buffers in 17 IXP4xx for accelerating queues, especially for networking. Clients pick 18 queues from the queue manager with foo-queue = <&qmgr N> where the 33 - description: Interrupt for queues 0-31 34 - description: Interrupt for queues 32-63
|
| /linux/drivers/target/ |
| A D | target_core_tmr.c | 118 flush_work(&dev->queues[i].sq.work); in core_tmr_abort_task() 120 spin_lock_irqsave(&dev->queues[i].lock, flags); in core_tmr_abort_task() 121 list_for_each_entry_safe(se_cmd, next, &dev->queues[i].state_list, in core_tmr_abort_task() 148 spin_unlock_irqrestore(&dev->queues[i].lock, flags); in core_tmr_abort_task() 163 spin_unlock_irqrestore(&dev->queues[i].lock, flags); in core_tmr_abort_task() 301 flush_work(&dev->queues[i].sq.work); in core_tmr_drain_state_list() 303 spin_lock_irqsave(&dev->queues[i].lock, flags); in core_tmr_drain_state_list() 304 list_for_each_entry_safe(cmd, next, &dev->queues[i].state_list, in core_tmr_drain_state_list() 333 spin_unlock_irqrestore(&dev->queues[i].lock, flags); in core_tmr_drain_state_list()
|
| /linux/tools/perf/util/ |
| A D | intel-bts.c | 46 struct auxtrace_queues queues; member 222 if (bts->queues.new_data) { in intel_bts_update_queues() 223 bts->queues.new_data = false; in intel_bts_update_queues() 539 struct auxtrace_queues *queues = &bts->queues; in intel_bts_process_tid_exit() local 542 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_process_tid_exit() 710 struct auxtrace_queues *queues = &bts->queues; in intel_bts_free_events() local 713 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_free_events() 715 queues->queue_array[i].priv = NULL; in intel_bts_free_events() 717 auxtrace_queues__free(queues); in intel_bts_free_events() 899 if (bts->queues.populated) in intel_bts_process_auxtrace_info() [all …]
|
| A D | auxtrace.c | 223 queues->nr_queues = nr_queues; in auxtrace_queues__init_nr() 224 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); in auxtrace_queues__init_nr() 225 if (!queues->queue_array) in auxtrace_queues__init_nr() 264 queues->nr_queues = nr_queues; in auxtrace_queues__grow() 265 queues->queue_array = queue_array; in auxtrace_queues__grow() 299 if (idx >= queues->nr_queues) { in auxtrace_queues__queue_buffer() 317 queues->new_data = true; in auxtrace_queues__queue_buffer() 318 queues->populated = true; in auxtrace_queues__queue_buffer() 471 zfree(&queues->queue_array); in auxtrace_queues__free() 472 queues->nr_queues = 0; in auxtrace_queues__free() [all …]
|
| A D | s390-cpumsf.c | 170 struct auxtrace_queues queues; member 206 q = &sf->queues.queue_array[sample->cpu]; in s390_cpumcf_dumpctr() 836 if (!sf->queues.new_data) in s390_cpumsf_update_queues() 839 sf->queues.new_data = false; in s390_cpumsf_update_queues() 1011 struct auxtrace_queues *queues = &sf->queues; in s390_cpumsf_free_queues() local 1014 for (i = 0; i < queues->nr_queues; i++) { in s390_cpumsf_free_queues() 1016 queues->queue_array[i].priv; in s390_cpumsf_free_queues() 1028 zfree(&queues->queue_array[i].priv); in s390_cpumsf_free_queues() 1030 auxtrace_queues__free(queues); in s390_cpumsf_free_queues() 1160 if (sf->queues.populated) in s390_cpumsf_process_auxtrace_info() [all …]
|
| /linux/Documentation/arch/arm/keystone/ |
| A D | knav-qmss.rst | 15 management of the packet queues. Packets are queued/de-queued by writing or 24 knav qmss driver provides a set of APIs to drivers to open/close qmss queues, 25 allocate descriptor pools, map the descriptors, push/pop to queues etc. For 31 Accumulator QMSS queues using PDSP firmware 34 queue or multiple contiguous queues. drivers/soc/ti/knav_qmss_acc.c is the 37 1 or 32 queues per channel. More description on the firmware is available in 56 Use of accumulated queues requires the firmware image to be present in the 57 file system. The driver doesn't acc queues to the supported queue range if
|
| /linux/include/linux/ |
| A D | ptr_ring.h | 627 void ***queues; in ptr_ring_resize_multiple_noprof() local 630 queues = kmalloc_array_noprof(nrings, sizeof(*queues), gfp); in ptr_ring_resize_multiple_noprof() 631 if (!queues) in ptr_ring_resize_multiple_noprof() 635 queues[i] = __ptr_ring_init_queue_alloc_noprof(size, gfp); in ptr_ring_resize_multiple_noprof() 636 if (!queues[i]) in ptr_ring_resize_multiple_noprof() 643 queues[i] = __ptr_ring_swap_queue(rings[i], queues[i], in ptr_ring_resize_multiple_noprof() 650 kvfree(queues[i]); in ptr_ring_resize_multiple_noprof() 652 kfree(queues); in ptr_ring_resize_multiple_noprof() 658 kvfree(queues[i]); in ptr_ring_resize_multiple_noprof() 660 kfree(queues); in ptr_ring_resize_multiple_noprof()
|
| /linux/Documentation/block/ |
| A D | blk-mq.rst | 37 spawns multiple queues with individual entry points local to the CPU, removing 49 blk-mq has two group of queues: software staging queues and hardware dispatch 50 queues. When the request arrives at the block layer, it will try the shortest 56 Then, after the requests are processed by software queues, they will be placed 62 Software staging queues 65 The block IO subsystem adds requests in the software staging queues 71 the number of queues is defined by a per-CPU or per-node basis. 99 queue (a.k.a. run the hardware queue), the software queues mapped to that 102 Hardware dispatch queues 116 requests that were ready to be sent first. The number of hardware queues [all …]
|