| /drivers/net/ti/ |
| A D | icssg_queues.c | 17 int icssg_queue_pop(struct prueth *prueth, u8 queue) in icssg_queue_pop() argument 21 if (queue >= ICSSG_QUEUES_MAX) in icssg_queue_pop() 24 regmap_read(prueth->miig_rt, ICSSG_QUEUE_CNT_OFFSET + 4 * queue, &cnt); in icssg_queue_pop() 28 regmap_read(prueth->miig_rt, ICSSG_QUEUE_OFFSET + 4 * queue, &val); in icssg_queue_pop() 33 void icssg_queue_push(struct prueth *prueth, int queue, u16 addr) in icssg_queue_push() argument 35 if (queue >= ICSSG_QUEUES_MAX) in icssg_queue_push() 38 regmap_write(prueth->miig_rt, ICSSG_QUEUE_OFFSET + 4 * queue, addr); in icssg_queue_push() 41 u32 icssg_queue_level(struct prueth *prueth, int queue) in icssg_queue_level() argument 45 if (queue >= ICSSG_QUEUES_MAX) in icssg_queue_level() 48 regmap_read(prueth->miig_rt, ICSSG_QUEUE_CNT_OFFSET + 4 * queue, ®); in icssg_queue_level()
|
| A D | icssg_config.c | 76 int queue; member 127 int queue = 0, i, j; in icssg_miig_queues_init() local 133 queue = ICSSG_NUM_TX_QUEUES; in icssg_miig_queues_init() 136 regmap_write(miig_rt, ICSSG_QUEUE_RESET_OFFSET, queue); in icssg_miig_queues_init() 137 queue++; in icssg_miig_queues_init() 140 queue = slice ? RECYCLE_Q_SLICE1 : RECYCLE_Q_SLICE0; in icssg_miig_queues_init() 141 regmap_write(miig_rt, ICSSG_QUEUE_RESET_OFFSET, queue); in icssg_miig_queues_init() 145 hwq_map[slice][i].queue); in icssg_miig_queues_init() 174 queue = mp->queue; in icssg_miig_queues_init() 175 regmap_write(miig_rt, ICSSG_QUEUE_OFFSET + 4 * queue, in icssg_miig_queues_init()
|
| A D | icssg_prueth.h | 95 int icssg_queue_pop(struct prueth *prueth, u8 queue); 96 void icssg_queue_push(struct prueth *prueth, int queue, u16 addr); 97 u32 icssg_queue_level(struct prueth *prueth, int queue);
|
| /drivers/net/ |
| A D | mvneta.c | 569 int queue; in mvneta_port_up() local 575 for (queue = 0; queue < txq_number; queue++) { in mvneta_port_up() 584 for (queue = 0; queue < rxq_number; queue++) { in mvneta_port_up() 754 int queue; in mvneta_defaults_set() local 785 for (queue = 0; queue < txq_number; queue++) { in mvneta_defaults_set() 1096 for (queue = 0; queue < txq_number; queue++) in mvneta_cleanup_txqs() 1105 for (queue = 0; queue < rxq_number; queue++) in mvneta_cleanup_rxqs() 1114 for (queue = 0; queue < rxq_number; queue++) { in mvneta_setup_rxqs() 1132 for (queue = 0; queue < txq_number; queue++) { in mvneta_setup_txqs() 1260 for (queue = 0; queue < txq_number; queue++) { in mvneta_init2() [all …]
|
| A D | mvpp2.c | 3727 for (queue = 0; queue < MVPP2_MAX_TXQ; queue++) { in mvpp2_defaults_set() 3793 int queue; in mvpp2_egress_enable() local 3798 for (queue = 0; queue < txq_number; queue++) { in mvpp2_egress_enable() 3967 for (queue = 0; queue < txq_number; queue++) { in mvpp2_txq_sent_counter_clear() 4319 for (queue = 0; queue < txq_number; queue++) { in mvpp2_cleanup_txqs() 4336 for (queue = 0; queue < rxq_number; queue++) in mvpp2_cleanup_rxqs() 4345 for (queue = 0; queue < rxq_number; queue++) { in mvpp2_setup_rxqs() 4363 for (queue = 0; queue < txq_number; queue++) { in mvpp2_setup_txqs() 4658 for (queue = 0; queue < txq_number; queue++) { in mvpp2_port_init() 4688 for (queue = 0; queue < rxq_number; queue++) { in mvpp2_port_init() [all …]
|
| /drivers/usb/musb-new/ |
| A D | musb_uboot.c | 127 struct int_queue *queue; in _musb_create_int_queue() local 140 queue = malloc(sizeof(*queue)); in _musb_create_int_queue() 141 if (!queue) in _musb_create_int_queue() 144 construct_urb(&queue->urb, &queue->hep, dev, USB_ENDPOINT_XFER_INT, in _musb_create_int_queue() 150 free(queue); in _musb_create_int_queue() 155 return queue; in _musb_create_int_queue() 164 if (queue->urb.status == -EINPROGRESS) in _musb_destroy_int_queue() 168 free(queue); in _musb_destroy_int_queue() 175 if (queue->urb.status != -EINPROGRESS) in _musb_poll_int_queue() 344 struct int_queue *queue) in musb_poll_int_queue() argument [all …]
|
| A D | musb_host.h | 75 struct list_head *queue; in next_urb() local 79 queue = &qh->hep->urb_list; in next_urb() 80 if (list_empty(queue)) in next_urb() 82 return list_entry(queue->next, struct urb, urb_list); in next_urb()
|
| A D | musb_gadget.h | 82 struct list_head *queue = &ep->req_list; in next_request() local 84 if (list_empty(queue)) in next_request() 86 return container_of(queue->next, struct musb_request, list); in next_request()
|
| /drivers/net/pfe_eth/ |
| A D | pfe_cmd.c | 179 writel((tmu << 8) | queue, TMU_TEQ_CTRL); in qm_read_drop_stat() 180 writel((tmu << 8) | queue, TMU_LLM_CTRL); in qm_read_drop_stat() 182 qtotal[tmu][queue] += val; in qm_read_drop_stat() 184 *total_drops = qtotal[tmu][queue]; in qm_read_drop_stat() 186 qtotal[tmu][queue] = 0; in qm_read_drop_stat() 195 printf("%d-%02d, ", tmu, queue); in tmu_queue_stats() 197 drops = qm_read_drop_stat(tmu, queue, NULL, 0); in tmu_queue_stats() 200 writel((tmu << 8) | queue, TMU_TEQ_CTRL); in tmu_queue_stats() 201 writel((tmu << 8) | queue, TMU_LLM_CTRL); in tmu_queue_stats() 214 int queue; in tmu_queues() local [all …]
|
| /drivers/net/fsl-mc/ |
| A D | dpni.c | 586 u8 options, const struct dpni_queue *queue) in dpni_set_queue() argument 601 cmd_params->dest_id = cpu_to_le32(queue->destination.id); in dpni_set_queue() 602 cmd_params->dest_prio = queue->destination.priority; in dpni_set_queue() 606 queue->destination.hold_active); in dpni_set_queue() 607 cmd_params->flc = cpu_to_le64(queue->flc.value); in dpni_set_queue() 609 cmd_params->cgid = queue->cgid; in dpni_set_queue() 637 struct dpni_queue *queue, struct dpni_queue_id *qid) in dpni_get_queue() argument 662 queue->destination.priority = rsp_params->dest_prio; in dpni_get_queue() 666 queue->flc.value = le64_to_cpu(rsp_params->flc); in dpni_get_queue() 671 queue->cgid = rsp_params->cgid; in dpni_get_queue() [all …]
|
| /drivers/usb/gadget/ |
| A D | dwc2_udc_otg_xfer_dma.c | 184 if (list_empty(&ep->queue)) { in complete_rx() 192 req = list_entry(ep->queue.next, struct dwc2_request, queue); in complete_rx() 272 req = list_entry(ep->queue.next, struct dwc2_request, queue); in complete_tx() 320 req = list_entry(ep->queue.next, struct dwc2_request, queue); in complete_tx() 336 req = list_entry(ep->queue.next, struct dwc2_request, queue); in dwc2_udc_check_tx_queue() 655 list_add_tail(&req->queue, &ep->queue); in dwc2_queue() 678 list_add_tail(&req->queue, &ep->queue); in dwc2_queue() 794 req = list_entry(ep->queue.next, struct dwc2_request, queue); in dwc2_ep0_read() 831 if (list_empty(&ep->queue)) in dwc2_ep0_write() 834 req = list_entry(ep->queue.next, struct dwc2_request, queue); in dwc2_ep0_write() [all …]
|
| A D | atmel_usba_udc.c | 81 if (list_empty(&ep->queue)) { in submit_next_request() 86 req = list_entry(ep->queue.next, struct usba_request, queue); in submit_next_request() 112 if (list_empty(&ep->queue)) { in receive_data() 136 list_del_init(&req->queue); in receive_data() 171 list_del_init(&req->queue); in request_complete_list() 328 INIT_LIST_HEAD(&req->queue); in usba_ep_alloc_request() 370 list_add_tail(&req->queue, &ep->queue); in usba_ep_queue() 398 list_del_init(&req->queue); in usba_ep_dequeue() 473 .queue = usba_ep_queue, 570 list_for_each_entry_safe(req, tmp_req, &ep->queue, queue) { in reset_all_endpoints() [all …]
|
| A D | at91_udc.c | 78 list_del_init(&req->queue); in done() 272 if (list_empty(&ep->queue)) in nuke() 277 req = list_entry(ep->queue.next, struct at91_request, queue); in nuke() 404 INIT_LIST_HEAD(&req->queue); in at91_ep_alloc_request() 509 list_add_tail (&req->queue, &ep->queue); in at91_ep_queue() 530 list_for_each_entry (req, &ep->queue, queue) { in at91_ep_dequeue() 591 .queue = at91_ep_queue, 653 INIT_LIST_HEAD(&ep->queue); in udc_reinit() 814 if (!list_empty(&ep->queue)) in handle_ep() 1099 if (list_empty(&ep0->queue)) in handle_ep0() [all …]
|
| A D | max3420_udc.c | 115 struct list_head queue; member 121 struct list_head queue; member 283 list_for_each_entry_safe(req, r, &ep->queue, queue) { in max3420_ep_nuke() 284 list_del_init(&req->queue); in max3420_ep_nuke() 310 INIT_LIST_HEAD(&req->queue); in max3420_ep_alloc_request() 329 list_add_tail(&req->queue, &ep->queue); in max3420_ep_queue() 338 list_del_init(&req->queue); in max3420_ep_dequeue() 376 .queue = max3420_ep_queue, 619 if (list_empty(&ep->queue)) in do_data() 622 req = list_first_entry(&ep->queue, struct max3420_req, queue); in do_data() [all …]
|
| A D | ci_udc.c | 104 .queue = ci_ep_queue, 283 INIT_LIST_HEAD(&ci_req->queue); in ci_ep_alloc_request() 488 ci_req = list_first_entry(&ci_ep->queue, struct ci_req, queue); in ci_ep_submit_next_request() 582 list_for_each_entry(ci_req, &ci_ep->queue, queue) { in ci_ep_dequeue() 590 list_del_init(&ci_req->queue); in ci_ep_dequeue() 633 list_add_tail(&ci_req->queue, &ci_ep->queue); in ci_ep_queue() 680 ci_req = list_first_entry(&ci_ep->queue, struct ci_req, queue); in handle_ep_complete() 698 list_del_init(&ci_req->queue); in handle_ep_complete() 701 if (!list_empty(&ci_ep->queue)) in handle_ep_complete() 770 list_del_init(&ci_req->queue); in handle_setup() [all …]
|
| A D | dwc2_udc_otg_priv.h | 56 struct list_head queue; member 71 struct list_head queue; member
|
| A D | at91_udc.h | 92 struct list_head queue; member 144 struct list_head queue; member
|
| A D | ci_udc.h | 82 struct list_head queue; member 94 struct list_head queue; member
|
| A D | dwc2_udc_otg.c | 146 .queue = dwc2_queue, 212 INIT_LIST_HEAD(&ep->queue); in udc_reinit() 379 list_del_init(&req->queue); in done() 430 while (!list_empty(&ep->queue)) { in nuke() 431 req = list_entry(ep->queue.next, struct dwc2_request, queue); in nuke() 723 INIT_LIST_HEAD(&req->queue); in dwc2_alloc_request() 735 WARN_ON(!list_empty(&req->queue)); in dwc2_free_request() 755 list_for_each_entry(req, &ep->queue, queue) { in dwc2_dequeue()
|
| A D | atmel_usba_udc.h | 261 struct list_head queue; member 276 struct list_head queue; member
|
| /drivers/usb/host/ |
| A D | ehci-hcd.c | 1445 struct int_queue *queue) in _ehci_poll_int_queue() argument 1458 cur_td = &queue->tds[queue->current - queue->first]; in _ehci_poll_int_queue() 1471 queue->current++; in _ehci_poll_int_queue() 1473 queue->current = NULL; in _ehci_poll_int_queue() 1525 free(queue->tds); in _ehci_destroy_int_queue() 1526 free(queue->first); in _ehci_destroy_int_queue() 1527 free(queue); in _ehci_destroy_int_queue() 1537 struct int_queue *queue; in _ehci_submit_int_msg() local 1545 if (!queue) in _ehci_submit_int_msg() 1663 struct int_queue *queue) in ehci_poll_int_queue() argument [all …]
|
| A D | ohci-hcd.c | 1609 queue = malloc(sizeof(*queue)); in _ohci_create_int_queue() 1610 if (!queue) { in _ohci_create_int_queue() 1619 if (!queue->urb[i]) in _ohci_create_int_queue() 1630 free(queue); in _ohci_create_int_queue() 1635 queue->curr_urb = 0; in _ohci_create_int_queue() 1637 return queue; in _ohci_create_int_queue() 1643 if (queue->curr_urb == queue->queuesize) in _ohci_poll_int_queue() 1649 if (queue->urb[queue->curr_urb]->finished) { in _ohci_poll_int_queue() 1650 void *ret = queue->urb[queue->curr_urb]->transfer_buffer; in _ohci_poll_int_queue() 1651 queue->curr_urb++; in _ohci_poll_int_queue() [all …]
|
| /drivers/virtio/ |
| A D | virtio_ring.c | 340 void *queue = NULL; in vring_create_virtqueue() local 354 queue = virtio_alloc_pages(vdev, in vring_create_virtqueue() 356 if (queue) in vring_create_virtqueue() 363 if (!queue) { in vring_create_virtqueue() 365 queue = virtio_alloc_pages(vdev, 1); in vring_create_virtqueue() 368 if (!queue) in vring_create_virtqueue() 371 memset(queue, 0, vring_size(num, vring_align)); in vring_create_virtqueue() 379 vring_init(&vring, num, queue, vring_align, bbs); in vring_create_virtqueue() 386 queue, vq, num); in vring_create_virtqueue() 393 virtio_free_pages(vdev, queue, DIV_ROUND_UP(vring.size, PAGE_SIZE)); in vring_create_virtqueue()
|
| /drivers/dma/ |
| A D | keystone_nav.c | 15 .queue = (void *)KS2_QM_MANAGER_QUEUES_BASE, 107 writel(regd, &qm_cfg->queue[qnum].ptr_size_thresh); in qm_push() 124 uhd = readl(&qm_cfg->queue[qnum].ptr_size_thresh) & ~0xf; in qm_pop()
|
| /drivers/nvme/ |
| A D | Kconfig | 20 the command submission queue and the integration
|