Home
last modified time | relevance | path

Searched refs:queue (Results 1 – 25 of 1424) sorted by relevance

12345678910>>...57

/linux/drivers/media/usb/uvc/
A Duvc_queue.c220 queue->queue.type = type; in uvc_queue_init()
222 queue->queue.drv_priv = queue; in uvc_queue_init()
227 queue->queue.lock = &queue->mutex; in uvc_queue_init()
234 queue->queue.io_modes |= VB2_DMABUF; in uvc_queue_init()
235 queue->queue.ops = &uvc_queue_qops; in uvc_queue_init()
239 ret = vb2_queue_init(&queue->queue); in uvc_queue_init()
254 vb2_queue_release(&queue->queue); in uvc_queue_release()
268 ret = vb2_reqbufs(&queue->queue, rb); in uvc_request_buffers()
279 ret = vb2_querybuf(&queue->queue, buf); in uvc_query_buffer()
315 ret = vb2_expbuf(&queue->queue, exp); in uvc_export_buffer()
[all …]
/linux/drivers/usb/gadget/function/
A Duvc_queue.c140 queue->queue.type = type; in uvcg_queue_init()
142 queue->queue.drv_priv = queue; in uvcg_queue_init()
144 queue->queue.ops = &uvc_queue_qops; in uvcg_queue_init()
145 queue->queue.lock = lock; in uvcg_queue_init()
155 queue->queue.dev = dev; in uvcg_queue_init()
157 ret = vb2_queue_init(&queue->queue); in uvcg_queue_init()
173 vb2_queue_release(&queue->queue); in uvcg_free_buffers()
184 ret = vb2_reqbufs(&queue->queue, rb); in uvcg_alloc_buffers()
223 return vb2_mmap(&queue->queue, vma); in uvcg_queue_mmap()
301 ret = vb2_streamon(&queue->queue, queue->queue.type); in uvcg_queue_enable()
[all …]
/linux/drivers/net/wireless/st/cw1200/
A Dqueue.c97 list_for_each_entry_safe(iter, tmp, &queue->queue, head) { in __cw1200_queue_gc()
119 if (queue->num_queued <= (queue->capacity >> 1)) { in __cw1200_queue_gc()
172 memset(queue, 0, sizeof(*queue)); in cw1200_queue_init()
177 INIT_LIST_HEAD(&queue->queue); in cw1200_queue_init()
211 list_splice_tail_init(&queue->queue, &queue->pending); in cw1200_queue_clear()
268 ret = queue->num_queued - queue->num_pending; in cw1200_queue_get_num_queued()
296 list_move_tail(&item->head, &queue->queue); in cw1200_queue_put()
343 list_for_each_entry(item, &queue->queue, head) { in cw1200_queue_get()
408 list_move(&item->head, &queue->queue); in cw1200_queue_requeue()
434 list_move(&item->head, &queue->queue); in cw1200_queue_requeue_all()
[all …]
/linux/drivers/md/dm-vdo/
A Dfunnel-workqueue.c156 queue->common.type->start(queue->private); in run_start_hook()
162 queue->common.type->finish(queue->private); in run_finish_hook()
280 vdo_free(queue); in free_simple_work_queue()
295 vdo_free(queue); in free_round_robin_work_queue()
300 if (queue == NULL) in vdo_free_work_queue()
338 vdo_free(queue); in make_simple_work_queue()
403 &queue); in vdo_make_work_queue()
410 vdo_free(queue); in vdo_make_work_queue()
421 vdo_free(queue); in vdo_make_work_queue()
614 return (queue == NULL) ? NULL : &queue->common; in vdo_get_current_work_queue()
[all …]
A Dfunnel-queue.c15 struct funnel_queue *queue; in vdo_make_funnel_queue() local
25 queue->stub.next = NULL; in vdo_make_funnel_queue()
26 queue->newest = &queue->stub; in vdo_make_funnel_queue()
27 queue->oldest = &queue->stub; in vdo_make_funnel_queue()
29 *queue_ptr = queue; in vdo_make_funnel_queue()
35 vdo_free(queue); in vdo_free_funnel_queue()
48 if (oldest == &queue->stub) { in get_oldest()
60 queue->oldest = oldest; in get_oldest()
83 vdo_funnel_queue_put(queue, &queue->stub); in get_oldest()
154 if (queue->oldest != &queue->stub) in vdo_is_funnel_queue_idle()
[all …]
/linux/drivers/net/wireless/broadcom/b43legacy/
A Dpio.c205 struct b43legacy_pioqueue *queue = packet->queue; in free_txpacket() local
219 struct b43legacy_pioqueue *queue = packet->queue; in pio_tx_packet() local
241 if (queue->tx_devq_used + octets > queue->tx_devq_size) in pio_tx_packet()
308 packet->queue = queue; in setup_txqueues()
323 queue = kzalloc(sizeof(*queue), GFP_KERNEL); in b43legacy_setup_pioqueue()
324 if (!queue) in b43legacy_setup_pioqueue()
381 if (!queue) in b43legacy_destroy_pioqueue()
413 if (!queue) in b43legacy_pio_init()
418 if (!queue) in b43legacy_pio_init()
423 if (!queue) in b43legacy_pio_init()
[all …]
/linux/drivers/iio/buffer/
A Dindustrialio-buffer-dma.c101 struct iio_dma_buffer_queue *queue = block->queue; in iio_buffer_block_release() local
195 block->queue = queue; in iio_dma_buffer_alloc_block()
234 struct iio_dma_buffer_queue *queue = block->queue; in iio_dma_buffer_block_done() local
315 return queue->fileio.enabled || !atomic_read(&queue->num_dmabufs); in iio_dma_buffer_can_use_fileio()
344 queue->fileio.enabled = iio_dma_buffer_can_use_fileio(queue); in iio_dma_buffer_request_update()
458 if (!queue->ops) in iio_dma_buffer_submit_block()
464 ret = queue->ops->submit(queue, block); in iio_dma_buffer_submit_block()
527 if (queue->ops && queue->ops->abort) in iio_dma_buffer_disable()
528 queue->ops->abort(queue); in iio_dma_buffer_disable()
738 struct iio_dma_buffer_queue *queue = block->queue; in iio_dma_can_enqueue_block() local
[all …]
/linux/drivers/nvme/target/
A Dtcp.c456 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_setup_c2h_data_pdu() local
489 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_setup_r2t_pdu() local
514 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_setup_response_pdu() local
573 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_tcp_queue_response() local
633 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_try_send_data() local
745 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_try_send_ddgst() local
1049 queue->idx, queue->state); in nvmet_tcp_done_recv_pdu()
1061 queue->cmd = nvmet_tcp_get_cmd(queue); in nvmet_tcp_done_recv_pdu()
1065 queue->idx, queue->nr_cmds, queue->send_list_len, in nvmet_tcp_done_recv_pdu()
1107 queue->cmd->req.execute(&queue->cmd->req); in nvmet_tcp_done_recv_pdu()
[all …]
A Drdma.c663 struct nvmet_rdma_queue *queue = rsp->queue; in nvmet_rdma_release_rsp() local
945 struct nvmet_rdma_queue *queue = rsp->queue; in nvmet_rdma_execute_command() local
1022 cmd->queue = queue; in nvmet_rdma_recv_done()
1033 rsp->queue = queue; in nvmet_rdma_recv_done()
1293 if (queue->port->pi_enable && queue->host_qid) in nvmet_rdma_create_queue_ib()
1301 queue->qp = queue->cm_id->qp; in nvmet_rdma_create_queue_ib()
1311 queue->cmds[i].queue = queue; in nvmet_rdma_create_queue_ib()
1346 nvmet_rdma_free_cmds(queue->dev, queue->cmds, in nvmet_rdma_free_queue()
1420 queue = kzalloc(sizeof(*queue), GFP_KERNEL); in nvmet_rdma_alloc_queue()
1462 queue->comp_vector = !queue->host_qid ? 0 : in nvmet_rdma_alloc_queue()
[all …]
/linux/drivers/net/xen-netback/
A Drx.c92 if (queue->rx_queue_len >= queue->rx_queue_max) { in xenvif_rx_queue_tail()
119 xenvif_update_needed_slots(queue, skb_peek(&queue->rx_queue)); in xenvif_rx_dequeue()
122 if (queue->rx_queue_len < queue->rx_queue_max) { in xenvif_rx_dequeue()
125 txq = netdev_get_tx_queue(queue->vif->dev, queue->id); in xenvif_rx_dequeue()
164 gnttab_batch_copy(queue->rx_copy.op, queue->rx_copy.num); in xenvif_rx_copy_flush()
204 op = &queue->rx_copy.op[queue->rx_copy.num]; in xenvif_rx_copy_add()
226 queue->rx_copy.idx[queue->rx_copy.num] = queue->rx.req_cons; in xenvif_rx_copy_add()
332 queue->rx.rsp_prod_pvt = queue->rx.req_cons; in xenvif_rx_complete()
473 req = RING_GET_REQUEST(&queue->rx, queue->rx.req_cons); in xenvif_rx_skb()
474 rsp = RING_GET_RESPONSE(&queue->rx, queue->rx.req_cons); in xenvif_rx_skb()
[all …]
A Dnetback.c163 wake_up(&queue->wq); in xenvif_kick_thread()
191 max_credit = queue->remaining_credit + queue->credit_bytes; in tx_add_credit()
201 struct xenvif_queue *queue = from_timer(queue, t, credit_timeout); in xenvif_tx_credit_callback() local
202 tx_add_credit(queue); in xenvif_tx_credit_callback()
346 queue->pages_to_map[mop-queue->tx_map_ops] = queue->mmap_pages[pending_idx]; in xenvif_tx_create_map_op()
933 if (queue->tx.sring->req_prod - queue->tx.req_cons > in xenvif_tx_build_gops()
938 queue->tx.sring->req_prod, queue->tx.req_cons, in xenvif_tx_build_gops()
1298 BUG_ON(queue->dealloc_prod - queue->dealloc_cons >= in xenvif_zerocopy_callback()
1347 queue->pages_to_unmap[gop - queue->tx_unmap_ops] = in xenvif_tx_dealloc_action()
1527 return queue->dealloc_cons != queue->dealloc_prod; in tx_dealloc_work_todo()
[all …]
A Dinterface.c314 if (queue->tx_irq != queue->rx_irq) in xenvif_up()
329 if (queue->tx_irq != queue->rx_irq) in xenvif_down()
565 queue->credit_bytes = queue->remaining_credit = ~0UL; in xenvif_init_queue()
689 unbind_from_irqhandler(queue->tx_irq, queue); in xenvif_disconnect_queue()
690 if (queue->tx_irq == queue->rx_irq) in xenvif_disconnect_queue()
696 unbind_from_irqhandler(queue->rx_irq, queue); in xenvif_disconnect_queue()
751 queue->name, queue); in xenvif_connect_data()
754 queue->tx_irq = queue->rx_irq = err; in xenvif_connect_data()
758 snprintf(queue->tx_irq_name, sizeof(queue->tx_irq_name), in xenvif_connect_data()
762 queue->tx_irq_name, queue); in xenvif_connect_data()
[all …]
/linux/drivers/md/dm-vdo/indexer/
A Dfunnel-requestqueue.c191 queue->processor(request); in request_queue_worker()
206 queue->running = true; in uds_make_request_queue()
223 &queue->thread); in uds_make_request_queue()
229 queue->started = true; in uds_make_request_queue()
230 *queue_ptr = queue; in uds_make_request_queue()
246 sub_queue = request->requeued ? queue->retry_queue : queue->main_queue; in uds_request_queue_enqueue()
254 wake_up_worker(queue); in uds_request_queue_enqueue()
259 if (queue == NULL) in uds_request_queue_finish()
271 if (queue->started) { in uds_request_queue_finish()
272 wake_up_worker(queue); in uds_request_queue_finish()
[all …]
/linux/drivers/misc/genwqe/
A Dcard_ddcb.c84 return queue->ddcb_next == queue->ddcb_act; in queue_empty()
89 if (queue->ddcb_next >= queue->ddcb_act) in queue_enqueued_ddcbs()
90 return queue->ddcb_next - queue->ddcb_act; in queue_enqueued_ddcbs()
92 return queue->ddcb_max - (queue->ddcb_act - queue->ddcb_next); in queue_enqueued_ddcbs()
332 struct ddcb_queue *queue = req->queue; in copy_ddcb_results() local
451 queue->ddcb_act = (queue->ddcb_act + 1) % queue->ddcb_max; in genwqe_check_ddcb_queue()
487 queue = req->queue; in __genwqe_wait_ddcb()
586 queue->ddcb_next = (queue->ddcb_next + 1) % queue->ddcb_max; in get_next_ddcb()
625 struct ddcb_queue *queue = req->queue; in __genwqe_purge_ddcb() local
772 queue = req->queue = &cd->queue; in __genwqe_enqueue_ddcb()
[all …]
/linux/drivers/crypto/cavium/zip/
A Dzip_device.c59 return ((zip_dev->iq[queue].sw_head - zip_dev->iq[queue].sw_tail) * in zip_cmd_queue_consumed()
81 u32 queue = 0; in zip_load_instr() local
91 queue = 0; in zip_load_instr()
93 queue = 1; in zip_load_instr()
128 zip_dev->iq[queue].sw_head = zip_dev->iq[queue].sw_tail; in zip_load_instr()
151 zip_dev->iq[queue].sw_head, zip_dev->iq[queue].sw_tail, in zip_load_instr()
152 zip_dev->iq[queue].hw_tail); in zip_load_instr()
166 return queue; in zip_load_instr()
187 zip_dev->iq[queue].hw_tail = zip_dev->iq[queue].sw_head; in zip_update_cmd_bufs()
197 zip_dev->iq[queue].sw_head, zip_dev->iq[queue].sw_tail, in zip_update_cmd_bufs()
[all …]
/linux/drivers/nvme/host/
A Dtcp.c214 return queue - queue->ctrl->queues; in nvme_tcp_queue_id()
393 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_queue_request() local
545 req->queue = queue; in nvme_tcp_init_request()
683 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_setup_h2c_data_pdu() local
911 if (queue->recv_ddgst != queue->exp_ddgst) { in nvme_tcp_recv_ddgst()
981 if (likely(queue && queue->rd_enabled) && in nvme_tcp_data_ready()
1048 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_try_send_data() local
1111 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_try_send_cmd_pdu() local
1152 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_try_send_data_pdu() local
1186 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_try_send_ddgst() local
[all …]
/linux/drivers/net/
A Dxen-netfront.c482 struct netfront_queue *queue = info->queue; in xennet_tx_setup_grant() local
619 .queue = queue, in xennet_xdp_xmit_one()
784 info.queue = queue; in xennet_start_xmit()
1821 if (queue->tx_irq && (queue->tx_irq == queue->rx_irq)) in xennet_disconnect_backend()
1823 if (queue->tx_irq && (queue->tx_irq != queue->rx_irq)) { in xennet_disconnect_backend()
1828 queue->tx_irq = queue->rx_irq = 0; in xennet_disconnect_backend()
1913 queue->rx_evtchn = queue->tx_evtchn; in setup_netfront_single()
1914 queue->rx_irq = queue->tx_irq = err; in setup_netfront_single()
1940 queue->tx_irq_name, queue); in setup_netfront_split()
1949 queue->rx_irq_name, queue); in setup_netfront_split()
[all …]
/linux/drivers/net/wireguard/
A Dqueueing.c30 memset(queue, 0, sizeof(*queue)); in wg_packet_queue_init()
31 queue->last_cpu = -1; in wg_packet_queue_init()
35 queue->worker = wg_packet_percpu_multicore_worker_alloc(function, queue); in wg_packet_queue_init()
36 if (!queue->worker) { in wg_packet_queue_init()
51 #define STUB(queue) ((struct sk_buff *)&queue->empty) argument
56 queue->head = queue->tail = STUB(queue); in wg_prev_queue_init()
57 queue->peeked = NULL; in wg_prev_queue_init()
87 queue->tail = next; in wg_prev_queue_dequeue()
92 queue->tail = next; in wg_prev_queue_dequeue()
98 __wg_prev_queue_enqueue(queue, STUB(queue)); in wg_prev_queue_dequeue()
[all …]
/linux/drivers/net/wireless/ralink/rt2x00/
A Drt2x00queue.c25 struct data_queue *queue = entry->queue; in rt2x00queue_alloc_rxskb() local
37 frame_size = queue->data_size + queue->desc_size + queue->winfo_size; in rt2x00queue_alloc_rxskb()
529 struct data_queue *queue = entry->queue; in rt2x00queue_write_tx_descriptor() local
554 queue->rt2x00dev->ops->lib->kick_queue(queue); in rt2x00queue_kick_tx_queue()
853 entry = &queue->entries[queue->index[index]]; in rt2x00queue_get_entry()
863 struct data_queue *queue = entry->queue; in rt2x00queue_index_inc() local
875 if (queue->index[index] >= queue->limit) in rt2x00queue_index_inc()
941 queue->rt2x00dev->ops->lib->kick_queue(queue); in rt2x00queue_unpause_queue()
961 queue->rt2x00dev->ops->lib->start_queue(queue); in rt2x00queue_start_queue()
980 queue->rt2x00dev->ops->lib->stop_queue(queue); in rt2x00queue_stop_queue()
[all …]
/linux/drivers/gpu/drm/imagination/
A Dpvr_queue.c270 fence->queue = queue; in pvr_queue_fence_init()
1056 queue->ctx->fw_obj, queue->ctx_offset); in pvr_queue_cleanup_fw_context()
1082 if (!queue) in pvr_queue_job_init()
1243 queue = kzalloc(sizeof(*queue), GFP_KERNEL); in pvr_queue_create()
1244 if (!queue) in pvr_queue_create()
1249 queue->ctx = ctx; in pvr_queue_create()
1263 reg_state_init, queue, &queue->reg_state_obj); in pvr_queue_create()
1304 return queue; in pvr_queue_create()
1320 kfree(queue); in pvr_queue_create()
1374 if (!queue) in pvr_queue_destroy()
[all …]
/linux/drivers/scsi/arm/
A Dqueue.c65 INIT_LIST_HEAD(&queue->head); in queue_initialise()
66 INIT_LIST_HEAD(&queue->free); in queue_initialise()
83 return queue->alloc != NULL; in queue_initialise()
95 kfree(queue->alloc); in queue_free()
115 if (list_empty(&queue->free)) in __queue_add()
118 l = queue->free.next; in __queue_add()
128 list_add(l, &queue->head); in __queue_add()
150 list_add(ent, &queue->free); in __queue_remove()
195 SCpnt = __queue_remove(queue, queue->head.next); in queue_remove()
247 __queue_remove(queue, l); in queue_remove_all_target()
[all …]
/linux/drivers/net/ethernet/ibm/ehea/
A Dehea_qmr.h208 return hw_qeit_calc(queue, queue->current_q_offset); in hw_qeit_get()
213 queue->current_q_offset += queue->qe_size; in hw_qeit_inc()
214 if (queue->current_q_offset >= queue->queue_length) { in hw_qeit_inc()
217 queue->toggle_state = (~queue->toggle_state) & 1; in hw_qeit_inc()
224 hw_qeit_inc(queue); in hw_qeit_get_inc()
236 hw_qeit_inc(queue); in hw_qeit_get_inc_valid()
237 pref = hw_qeit_calc(queue, queue->current_q_offset); in hw_qeit_get_inc_valid()
251 pref = hw_qeit_calc(queue, queue->current_q_offset); in hw_qeit_get_valid()
269 u64 last_entry_in_q = queue->queue_length - queue->qe_size; in hw_qeit_eq_get_inc()
273 queue->current_q_offset += queue->qe_size; in hw_qeit_eq_get_inc()
[all …]
/linux/drivers/crypto/hisilicon/sec/
A Dsec_drv.c703 while (test_bit(queue->expected, queue->unprocessed)) { in sec_isr_handle()
704 clear_bit(queue->expected, queue->unprocessed); in sec_isr_handle()
708 queue->shadow[queue->expected]); in sec_isr_handle()
709 queue->shadow[queue->expected] = NULL; in sec_isr_handle()
710 queue->expected = (queue->expected + 1) % in sec_isr_handle()
735 IRQF_TRIGGER_RISING, queue->name, queue); in sec_queue_irq_init()
747 free_irq(queue->task_irq, queue); in sec_queue_irq_uninit()
907 sec_queue_cmdbase_addr(queue, queue->ring_cmd.paddr); in sec_queue_hw_init()
909 sec_queue_outorder_addr(queue, queue->ring_cq.paddr); in sec_queue_hw_init()
911 sec_queue_errbase_addr(queue, queue->ring_db.paddr); in sec_queue_hw_init()
[all …]
/linux/drivers/soc/ixp4xx/
A Dixp4xx-qmgr.c35 qmgr_queue_descs[queue], queue, val); in qmgr_put_entry()
48 qmgr_queue_descs[queue], queue, val); in qmgr_get_entry()
86 if (queue >= HALF_QUEUES) in qmgr_stat_below_low_watermark()
100 if (queue >= HALF_QUEUES) in qmgr_stat_full()
136 irq_pdevs[queue] = pdev; in qmgr_set_irq()
208 int half = queue / 32; in qmgr_enable_irq()
220 int half = queue / 32; in qmgr_disable_irq()
252 BUG_ON(queue >= QUEUES); in qmgr_request_queue()
318 qmgr_queue_descs[queue], queue, addr); in qmgr_request_queue()
355 qmgr_queue_descs[queue], queue); in qmgr_release_queue()
[all …]
/linux/net/sunrpc/
A Dsched.c140 rpc_set_waitqueue_priority(queue, queue->maxpriority); in rpc_reset_waitqueue_priority()
215 queue->qlen++; in __rpc_add_wait_queue()
240 queue->qlen--; in __rpc_remove_wait_queue()
252 queue->qlen = 0; in __rpc_init_priority_wait_queue()
622 q = &queue->tasks[queue->priority]; in __rpc_find_next_queued_priority()
624 queue->nr--; in __rpc_find_next_queued_priority()
634 q = &queue->tasks[queue->maxpriority]; in __rpc_find_next_queued_priority()
641 } while (q != &queue->tasks[queue->priority]); in __rpc_find_next_queued_priority()
647 rpc_set_waitqueue_priority(queue, (unsigned int)(q - &queue->tasks[0])); in __rpc_find_next_queued_priority()
870 if (queue) in rpc_signal_task()
[all …]

Completed in 108 milliseconds

12345678910>>...57