Home
last modified time | relevance | path

Searched refs:queue (Results 1 – 25 of 38) sorted by relevance

12

/drivers/net/ti/
A Dicssg_queues.c17 int icssg_queue_pop(struct prueth *prueth, u8 queue) in icssg_queue_pop() argument
21 if (queue >= ICSSG_QUEUES_MAX) in icssg_queue_pop()
24 regmap_read(prueth->miig_rt, ICSSG_QUEUE_CNT_OFFSET + 4 * queue, &cnt); in icssg_queue_pop()
28 regmap_read(prueth->miig_rt, ICSSG_QUEUE_OFFSET + 4 * queue, &val); in icssg_queue_pop()
33 void icssg_queue_push(struct prueth *prueth, int queue, u16 addr) in icssg_queue_push() argument
35 if (queue >= ICSSG_QUEUES_MAX) in icssg_queue_push()
38 regmap_write(prueth->miig_rt, ICSSG_QUEUE_OFFSET + 4 * queue, addr); in icssg_queue_push()
41 u32 icssg_queue_level(struct prueth *prueth, int queue) in icssg_queue_level() argument
45 if (queue >= ICSSG_QUEUES_MAX) in icssg_queue_level()
48 regmap_read(prueth->miig_rt, ICSSG_QUEUE_CNT_OFFSET + 4 * queue, &reg); in icssg_queue_level()
A Dicssg_config.c76 int queue; member
127 int queue = 0, i, j; in icssg_miig_queues_init() local
133 queue = ICSSG_NUM_TX_QUEUES; in icssg_miig_queues_init()
136 regmap_write(miig_rt, ICSSG_QUEUE_RESET_OFFSET, queue); in icssg_miig_queues_init()
137 queue++; in icssg_miig_queues_init()
140 queue = slice ? RECYCLE_Q_SLICE1 : RECYCLE_Q_SLICE0; in icssg_miig_queues_init()
141 regmap_write(miig_rt, ICSSG_QUEUE_RESET_OFFSET, queue); in icssg_miig_queues_init()
145 hwq_map[slice][i].queue); in icssg_miig_queues_init()
174 queue = mp->queue; in icssg_miig_queues_init()
175 regmap_write(miig_rt, ICSSG_QUEUE_OFFSET + 4 * queue, in icssg_miig_queues_init()
A Dicssg_prueth.h95 int icssg_queue_pop(struct prueth *prueth, u8 queue);
96 void icssg_queue_push(struct prueth *prueth, int queue, u16 addr);
97 u32 icssg_queue_level(struct prueth *prueth, int queue);
/drivers/net/
A Dmvneta.c569 int queue; in mvneta_port_up() local
575 for (queue = 0; queue < txq_number; queue++) { in mvneta_port_up()
584 for (queue = 0; queue < rxq_number; queue++) { in mvneta_port_up()
754 int queue; in mvneta_defaults_set() local
785 for (queue = 0; queue < txq_number; queue++) { in mvneta_defaults_set()
1096 for (queue = 0; queue < txq_number; queue++) in mvneta_cleanup_txqs()
1105 for (queue = 0; queue < rxq_number; queue++) in mvneta_cleanup_rxqs()
1114 for (queue = 0; queue < rxq_number; queue++) { in mvneta_setup_rxqs()
1132 for (queue = 0; queue < txq_number; queue++) { in mvneta_setup_txqs()
1260 for (queue = 0; queue < txq_number; queue++) { in mvneta_init2()
[all …]
A Dmvpp2.c3727 for (queue = 0; queue < MVPP2_MAX_TXQ; queue++) { in mvpp2_defaults_set()
3793 int queue; in mvpp2_egress_enable() local
3798 for (queue = 0; queue < txq_number; queue++) { in mvpp2_egress_enable()
3967 for (queue = 0; queue < txq_number; queue++) { in mvpp2_txq_sent_counter_clear()
4319 for (queue = 0; queue < txq_number; queue++) { in mvpp2_cleanup_txqs()
4336 for (queue = 0; queue < rxq_number; queue++) in mvpp2_cleanup_rxqs()
4345 for (queue = 0; queue < rxq_number; queue++) { in mvpp2_setup_rxqs()
4363 for (queue = 0; queue < txq_number; queue++) { in mvpp2_setup_txqs()
4658 for (queue = 0; queue < txq_number; queue++) { in mvpp2_port_init()
4688 for (queue = 0; queue < rxq_number; queue++) { in mvpp2_port_init()
[all …]
/drivers/usb/musb-new/
A Dmusb_uboot.c127 struct int_queue *queue; in _musb_create_int_queue() local
140 queue = malloc(sizeof(*queue)); in _musb_create_int_queue()
141 if (!queue) in _musb_create_int_queue()
144 construct_urb(&queue->urb, &queue->hep, dev, USB_ENDPOINT_XFER_INT, in _musb_create_int_queue()
150 free(queue); in _musb_create_int_queue()
155 return queue; in _musb_create_int_queue()
164 if (queue->urb.status == -EINPROGRESS) in _musb_destroy_int_queue()
168 free(queue); in _musb_destroy_int_queue()
175 if (queue->urb.status != -EINPROGRESS) in _musb_poll_int_queue()
344 struct int_queue *queue) in musb_poll_int_queue() argument
[all …]
A Dmusb_host.h75 struct list_head *queue; in next_urb() local
79 queue = &qh->hep->urb_list; in next_urb()
80 if (list_empty(queue)) in next_urb()
82 return list_entry(queue->next, struct urb, urb_list); in next_urb()
A Dmusb_gadget.h82 struct list_head *queue = &ep->req_list; in next_request() local
84 if (list_empty(queue)) in next_request()
86 return container_of(queue->next, struct musb_request, list); in next_request()
/drivers/net/pfe_eth/
A Dpfe_cmd.c179 writel((tmu << 8) | queue, TMU_TEQ_CTRL); in qm_read_drop_stat()
180 writel((tmu << 8) | queue, TMU_LLM_CTRL); in qm_read_drop_stat()
182 qtotal[tmu][queue] += val; in qm_read_drop_stat()
184 *total_drops = qtotal[tmu][queue]; in qm_read_drop_stat()
186 qtotal[tmu][queue] = 0; in qm_read_drop_stat()
195 printf("%d-%02d, ", tmu, queue); in tmu_queue_stats()
197 drops = qm_read_drop_stat(tmu, queue, NULL, 0); in tmu_queue_stats()
200 writel((tmu << 8) | queue, TMU_TEQ_CTRL); in tmu_queue_stats()
201 writel((tmu << 8) | queue, TMU_LLM_CTRL); in tmu_queue_stats()
214 int queue; in tmu_queues() local
[all …]
/drivers/net/fsl-mc/
A Ddpni.c586 u8 options, const struct dpni_queue *queue) in dpni_set_queue() argument
601 cmd_params->dest_id = cpu_to_le32(queue->destination.id); in dpni_set_queue()
602 cmd_params->dest_prio = queue->destination.priority; in dpni_set_queue()
606 queue->destination.hold_active); in dpni_set_queue()
607 cmd_params->flc = cpu_to_le64(queue->flc.value); in dpni_set_queue()
609 cmd_params->cgid = queue->cgid; in dpni_set_queue()
637 struct dpni_queue *queue, struct dpni_queue_id *qid) in dpni_get_queue() argument
662 queue->destination.priority = rsp_params->dest_prio; in dpni_get_queue()
666 queue->flc.value = le64_to_cpu(rsp_params->flc); in dpni_get_queue()
671 queue->cgid = rsp_params->cgid; in dpni_get_queue()
[all …]
/drivers/usb/gadget/
A Ddwc2_udc_otg_xfer_dma.c184 if (list_empty(&ep->queue)) { in complete_rx()
192 req = list_entry(ep->queue.next, struct dwc2_request, queue); in complete_rx()
272 req = list_entry(ep->queue.next, struct dwc2_request, queue); in complete_tx()
320 req = list_entry(ep->queue.next, struct dwc2_request, queue); in complete_tx()
336 req = list_entry(ep->queue.next, struct dwc2_request, queue); in dwc2_udc_check_tx_queue()
655 list_add_tail(&req->queue, &ep->queue); in dwc2_queue()
678 list_add_tail(&req->queue, &ep->queue); in dwc2_queue()
794 req = list_entry(ep->queue.next, struct dwc2_request, queue); in dwc2_ep0_read()
831 if (list_empty(&ep->queue)) in dwc2_ep0_write()
834 req = list_entry(ep->queue.next, struct dwc2_request, queue); in dwc2_ep0_write()
[all …]
A Datmel_usba_udc.c81 if (list_empty(&ep->queue)) { in submit_next_request()
86 req = list_entry(ep->queue.next, struct usba_request, queue); in submit_next_request()
112 if (list_empty(&ep->queue)) { in receive_data()
136 list_del_init(&req->queue); in receive_data()
171 list_del_init(&req->queue); in request_complete_list()
328 INIT_LIST_HEAD(&req->queue); in usba_ep_alloc_request()
370 list_add_tail(&req->queue, &ep->queue); in usba_ep_queue()
398 list_del_init(&req->queue); in usba_ep_dequeue()
473 .queue = usba_ep_queue,
570 list_for_each_entry_safe(req, tmp_req, &ep->queue, queue) { in reset_all_endpoints()
[all …]
A Dat91_udc.c78 list_del_init(&req->queue); in done()
272 if (list_empty(&ep->queue)) in nuke()
277 req = list_entry(ep->queue.next, struct at91_request, queue); in nuke()
404 INIT_LIST_HEAD(&req->queue); in at91_ep_alloc_request()
509 list_add_tail (&req->queue, &ep->queue); in at91_ep_queue()
530 list_for_each_entry (req, &ep->queue, queue) { in at91_ep_dequeue()
591 .queue = at91_ep_queue,
653 INIT_LIST_HEAD(&ep->queue); in udc_reinit()
814 if (!list_empty(&ep->queue)) in handle_ep()
1099 if (list_empty(&ep0->queue)) in handle_ep0()
[all …]
A Dmax3420_udc.c115 struct list_head queue; member
121 struct list_head queue; member
283 list_for_each_entry_safe(req, r, &ep->queue, queue) { in max3420_ep_nuke()
284 list_del_init(&req->queue); in max3420_ep_nuke()
310 INIT_LIST_HEAD(&req->queue); in max3420_ep_alloc_request()
329 list_add_tail(&req->queue, &ep->queue); in max3420_ep_queue()
338 list_del_init(&req->queue); in max3420_ep_dequeue()
376 .queue = max3420_ep_queue,
619 if (list_empty(&ep->queue)) in do_data()
622 req = list_first_entry(&ep->queue, struct max3420_req, queue); in do_data()
[all …]
A Dci_udc.c104 .queue = ci_ep_queue,
283 INIT_LIST_HEAD(&ci_req->queue); in ci_ep_alloc_request()
488 ci_req = list_first_entry(&ci_ep->queue, struct ci_req, queue); in ci_ep_submit_next_request()
582 list_for_each_entry(ci_req, &ci_ep->queue, queue) { in ci_ep_dequeue()
590 list_del_init(&ci_req->queue); in ci_ep_dequeue()
633 list_add_tail(&ci_req->queue, &ci_ep->queue); in ci_ep_queue()
680 ci_req = list_first_entry(&ci_ep->queue, struct ci_req, queue); in handle_ep_complete()
698 list_del_init(&ci_req->queue); in handle_ep_complete()
701 if (!list_empty(&ci_ep->queue)) in handle_ep_complete()
770 list_del_init(&ci_req->queue); in handle_setup()
[all …]
A Ddwc2_udc_otg_priv.h56 struct list_head queue; member
71 struct list_head queue; member
A Dat91_udc.h92 struct list_head queue; member
144 struct list_head queue; member
A Dci_udc.h82 struct list_head queue; member
94 struct list_head queue; member
A Ddwc2_udc_otg.c146 .queue = dwc2_queue,
212 INIT_LIST_HEAD(&ep->queue); in udc_reinit()
379 list_del_init(&req->queue); in done()
430 while (!list_empty(&ep->queue)) { in nuke()
431 req = list_entry(ep->queue.next, struct dwc2_request, queue); in nuke()
723 INIT_LIST_HEAD(&req->queue); in dwc2_alloc_request()
735 WARN_ON(!list_empty(&req->queue)); in dwc2_free_request()
755 list_for_each_entry(req, &ep->queue, queue) { in dwc2_dequeue()
A Datmel_usba_udc.h261 struct list_head queue; member
276 struct list_head queue; member
/drivers/usb/host/
A Dehci-hcd.c1445 struct int_queue *queue) in _ehci_poll_int_queue() argument
1458 cur_td = &queue->tds[queue->current - queue->first]; in _ehci_poll_int_queue()
1471 queue->current++; in _ehci_poll_int_queue()
1473 queue->current = NULL; in _ehci_poll_int_queue()
1525 free(queue->tds); in _ehci_destroy_int_queue()
1526 free(queue->first); in _ehci_destroy_int_queue()
1527 free(queue); in _ehci_destroy_int_queue()
1537 struct int_queue *queue; in _ehci_submit_int_msg() local
1545 if (!queue) in _ehci_submit_int_msg()
1663 struct int_queue *queue) in ehci_poll_int_queue() argument
[all …]
A Dohci-hcd.c1609 queue = malloc(sizeof(*queue)); in _ohci_create_int_queue()
1610 if (!queue) { in _ohci_create_int_queue()
1619 if (!queue->urb[i]) in _ohci_create_int_queue()
1630 free(queue); in _ohci_create_int_queue()
1635 queue->curr_urb = 0; in _ohci_create_int_queue()
1637 return queue; in _ohci_create_int_queue()
1643 if (queue->curr_urb == queue->queuesize) in _ohci_poll_int_queue()
1649 if (queue->urb[queue->curr_urb]->finished) { in _ohci_poll_int_queue()
1650 void *ret = queue->urb[queue->curr_urb]->transfer_buffer; in _ohci_poll_int_queue()
1651 queue->curr_urb++; in _ohci_poll_int_queue()
[all …]
/drivers/virtio/
A Dvirtio_ring.c340 void *queue = NULL; in vring_create_virtqueue() local
354 queue = virtio_alloc_pages(vdev, in vring_create_virtqueue()
356 if (queue) in vring_create_virtqueue()
363 if (!queue) { in vring_create_virtqueue()
365 queue = virtio_alloc_pages(vdev, 1); in vring_create_virtqueue()
368 if (!queue) in vring_create_virtqueue()
371 memset(queue, 0, vring_size(num, vring_align)); in vring_create_virtqueue()
379 vring_init(&vring, num, queue, vring_align, bbs); in vring_create_virtqueue()
386 queue, vq, num); in vring_create_virtqueue()
393 virtio_free_pages(vdev, queue, DIV_ROUND_UP(vring.size, PAGE_SIZE)); in vring_create_virtqueue()
/drivers/dma/
A Dkeystone_nav.c15 .queue = (void *)KS2_QM_MANAGER_QUEUES_BASE,
107 writel(regd, &qm_cfg->queue[qnum].ptr_size_thresh); in qm_push()
124 uhd = readl(&qm_cfg->queue[qnum].ptr_size_thresh) & ~0xf; in qm_pop()
/drivers/nvme/
A DKconfig20 the command submission queue and the integration

Completed in 261 milliseconds

12