| /drivers/net/ethernet/fungible/funeth/ |
| A D | funeth_rx.c | 276 q->rqes[q->rq_cons & q->rq_mask] = in get_buf() 281 return &q->bufs[q->rq_cons & q->rq_mask]; in get_buf() 523 if (q->rq_cons - q->rq_cons_db >= q->rq_db_thres) { in fun_rxq_napi_poll() 525 q->stats.rx_bufs += q->rq_cons - q->rq_cons_db; in fun_rxq_napi_poll() 527 q->rq_cons_db = q->rq_cons; in fun_rxq_napi_poll() 528 writel((q->rq_cons - 1) & q->rq_mask, q->rq_db); in fun_rxq_napi_poll() 561 q->cur_buf = q->bufs; in fun_rxq_alloc_bufs() 669 q->rq_dma_addr, q->bufs); in fun_rxq_create_sw() 684 q->rqes, q->rq_dma_addr, q->bufs); in fun_rxq_free_sw() 704 err = xdp_rxq_info_reg(&q->xdp_rxq, q->netdev, q->qidx, in fun_rxq_create_dev() [all …]
|
| A D | funeth_tx.c | 313 return q->mask - q->prod_cnt + q->cons_cnt; in fun_txq_avail() 431 for (head = txq_hw_head(q), reclaim_idx = q->cons_cnt & q->mask; in fun_txq_reclaim() 478 db_val = READ_ONCE(q->irq_db_val) | (q->cons_cnt & q->mask); in fun_txq_napi_poll() 545 idx = q->prod_cnt & q->mask; in fun_xdp_tx() 670 q->desc, q->dma_addr, q->info); in fun_txq_free_sw() 698 &q->hw_qid, &q->db); in fun_txq_create_dev() 709 q->ndq = netdev_get_tx_queue(q->netdev, q->qidx); in fun_txq_create_dev() 712 writel(q->irq_db_val, q->db); in fun_txq_create_dev() 719 q->ethid, q->numa_node); in fun_txq_create_dev() 740 q->irq ? "Tx" : "XDP", q->qidx, q->hw_qid, in fun_txq_free_dev() [all …]
|
| /drivers/gpu/drm/xe/ |
| A D | xe_guc_submit.c | 329 q->guc->id, q->width); in __release_guc_id() 358 q->guc->id + i, q, GFP_NOWAIT)); in alloc_guc_id() 374 __release_guc_id(guc, q, q->width); in release_guc_id() 752 struct xe_exec_queue *q = job->q; in guc_exec_queue_run_job() local 908 struct xe_exec_queue *q = ge->q; in xe_guc_exec_queue_lr_cleanup() local 1079 struct xe_exec_queue *q = job->q; in guc_exec_queue_timedout_job() local 1201 if (q->vm && q->vm->xef) { in guc_exec_queue_timedout_job() 1218 q->guc->id, q->flags); in guc_exec_queue_timedout_job() 1226 xe_gt_WARN(q->gt, q->flags & EXEC_QUEUE_FLAG_VM && !exec_queue_killed(q), in guc_exec_queue_timedout_job() 1284 struct xe_exec_queue *q = ge->q; in __guc_exec_queue_fini_async() local [all …]
|
| A D | xe_exec_queue.c | 43 xe_pxp_exec_queue_remove(gt_to_xe(q->gt)->pxp, q); in __xe_exec_queue_free() 50 kfree(q); in __xe_exec_queue_free() 67 if (!q) in __xe_exec_queue_alloc() 135 q->lrc[i] = xe_lrc_create(q->hwe, q->vm, SZ_16K, q->msix_vec, flags); in __xe_exec_queue_init() 142 err = q->ops->init(q); in __xe_exec_queue_init() 286 q->ops->fini(q); in xe_exec_queue_destroy() 689 if (q->vm && q->hwe->hw_engine_group) { in xe_exec_queue_create_ioctl() 732 args->value = q->ops->reset_status(q); in xe_exec_queue_get_property_ioctl() 842 q->xef->run_ticks[q->class] += (new_ts - old_ts) * q->width; in xe_exec_queue_update_run_ticks() 866 q->ops->kill(q); in xe_exec_queue_kill() [all …]
|
| A D | xe_sched_job.c | 65 return q->vm && (q->vm->flags & XE_VM_FLAG_MIGRATION); in xe_sched_job_is_migration() 70 struct xe_exec_queue *q = job->q; in job_free() local 106 XE_WARN_ON(!q->vm && !(q->flags & EXEC_QUEUE_FLAG_KERNEL)); in xe_sched_job_create() 112 job->q = q; in xe_sched_job_create() 117 q->xef ? q->xef->drm->client_id : 0); in xe_sched_job_create() 142 width = q->width; in xe_sched_job_create() 157 xe_exec_queue_put(q); in xe_sched_job_create() 174 struct xe_exec_queue *q = job->q; in xe_sched_job_destroy() local 180 xe_exec_queue_put(q); in xe_sched_job_destroy() 245 struct xe_exec_queue *q = job->q; in xe_sched_job_arm() local [all …]
|
| /drivers/media/common/videobuf2/ |
| A D | videobuf2-core.c | 168 ((q)->ops->op ? (q)->ops->op(args) : 0) 190 if (q && q->buf_ops && q->buf_ops->op) \ 197 if (q && q->buf_ops && q->buf_ops->op) \ 244 q->alloc_devs[plane] ? : q->dev, in __vb2_buf_mem_alloc() 850 q->bufs = kcalloc(q->max_num_buffers, sizeof(*q->bufs), GFP_KERNEL); in vb2_core_allocated_buffers_storage() 2041 call_void_qop(q, wait_prepare, q); in __vb2_wait_for_done_vb() 2054 call_void_qop(q, wait_finish, q); in __vb2_wait_for_done_vb() 2682 q->is_output ? "out" : "cap", q); in vb2_core_queue_init() 2764 if (q->is_output && q->fileio && q->queued_count < vb2_get_num_buffers(q)) in vb2_core_poll() 2965 ret = vb2_core_streamon(q, q->type); in __vb2_init_fileio() [all …]
|
| A D | videobuf2-v4l2.c | 147 if (q->is_output) { in __copy_timestamp() 697 if (q->allow_cache_hints && q->io_modes & VB2_MMAP) in vb2_set_flags_and_caps() 928 q->is_multiplanar = V4L2_TYPE_IS_MULTIPLANAR(q->type); in vb2_queue_init_name() 929 q->is_output = V4L2_TYPE_IS_OUTPUT(q->type); in vb2_queue_init_name() 930 q->copy_timestamp = (q->timestamp_flags & V4L2_BUF_FLAG_TIMESTAMP_MASK) in vb2_queue_init_name() 940 strscpy(q->name, name, sizeof(q->name)); in vb2_queue_init_name() 942 q->name[0] = '\0'; in vb2_queue_init_name() 965 if (vb2_is_busy(q)) in vb2_queue_change_type() 968 q->type = type; in vb2_queue_change_type() 1229 struct mutex *lock = q->lock ? q->lock : vdev->lock; in vb2_fop_poll() [all …]
|
| /drivers/spi/ |
| A D | spi-fsl-qspi.c | 345 reg = qspi_readl(q, q->iobase + QUADSPI_FR); in fsl_qspi_irq_handler() 346 qspi_writel(q, reg, q->iobase + QUADSPI_FR); in fsl_qspi_irq_handler() 511 reg = qspi_readl(q, q->iobase + QUADSPI_MCR); in fsl_qspi_invalidate() 513 qspi_writel(q, reg, q->iobase + QUADSPI_MCR); in fsl_qspi_invalidate() 522 qspi_writel(q, reg, q->iobase + QUADSPI_MCR); in fsl_qspi_invalidate() 555 q->ahb_addr + q->selected * q->devtype_data->ahb_buf_size, in fsl_qspi_read_ahb() 665 qspi_writel(q, qspi_readl(q, base + QUADSPI_MCR) | in fsl_qspi_exec_op() 754 qspi_writel(q, qspi_readl(q, base + QUADSPI_FLSHCR) & in fsl_qspi_default_setup() 802 qspi_writel(q, 0xffffffff, q->iobase + QUADSPI_FR); in fsl_qspi_default_setup() 853 qspi_writel(q, 0x0, q->iobase + QUADSPI_RSER); in fsl_qspi_disable() [all …]
|
| /drivers/net/wireless/broadcom/b43/ |
| A D | pio.c | 76 return q; in parse_cookie() 130 q = kzalloc(sizeof(*q), GFP_KERNEL); in b43_setup_pioqueue_tx() 131 if (!q) in b43_setup_pioqueue_tx() 156 return q; in b43_setup_pioqueue_tx() 164 q = kzalloc(sizeof(*q), GFP_KERNEL); in b43_setup_pioqueue_rx() 165 if (!q) in b43_setup_pioqueue_rx() 175 return q; in b43_setup_pioqueue_rx() 195 if (!q) in b43_destroy_pioqueue_tx() 204 if (!q) in b43_destroy_pioqueue_rx() 523 B43_WARN_ON(q->buffer_used > q->buffer_size); in b43_pio_tx() [all …]
|
| /drivers/s390/cio/ |
| A D | qdio_main.c | 149 q->handler(q->irq_ptr->cdev, QDIO_ERROR_GET_BUF_STATE, q->nr, in qdio_do_eqbs() 150 q->first_to_check, count, q->irq_ptr->int_parm); in qdio_do_eqbs() 195 q->handler(q->irq_ptr->cdev, QDIO_ERROR_SET_BUF_STATE, q->nr, in qdio_do_sqbs() 196 q->first_to_check, count, q->irq_ptr->int_parm); in qdio_do_sqbs() 307 return qdio_siga_sync(q, 0, q->mask); in qdio_sync_input_queue() 312 return qdio_siga_sync(q, q->mask, 0); in qdio_sync_output_queue() 358 "%4x cc2 BB1:%1d", SCH_NO(q), q->nr); in qdio_siga_output() 466 inbound_handle_work(q, start, count, is_qebsm(q)); in get_inbound_buffer_frontier() 653 DBF_ERROR("%4x cc2 BBC:%1d", SCH_NO(q), q->nr); in qdio_kick_outbound_q() 667 DBF_ERROR("%4x cc2 BB2:%1d", SCH_NO(q), q->nr); in qdio_kick_outbound_q() [all …]
|
| /drivers/infiniband/sw/rxe/ |
| A D | rxe_queue.h | 112 prod = q->index; in queue_get_producer() 135 cons = q->index; in queue_get_consumer() 193 prod = q->index; in queue_advance_producer() 195 q->index = prod; in queue_advance_producer() 223 cons = (q->index + 1) & q->index_mask; in queue_advance_consumer() 224 q->index = cons; in queue_advance_consumer() 255 return q->buf->data + (prod << q->log2_elem_size); in queue_producer_addr() 263 return q->buf->data + (cons << q->log2_elem_size); in queue_consumer_addr() 268 return q->buf->data + ((index & q->index_mask) in queue_addr_from_index() 275 return (((u8 *)addr - q->buf->data) >> q->log2_elem_size) in queue_index_from_addr() [all …]
|
| A D | rxe_queue.c | 66 q = kzalloc(sizeof(*q), GFP_KERNEL); in rxe_queue_init() 67 if (!q) in rxe_queue_init() 90 if (!q->buf) in rxe_queue_init() 93 q->buf->log2_elem_size = q->log2_elem_size; in rxe_queue_init() 94 q->buf->index_mask = q->index_mask; in rxe_queue_init() 99 return q; in rxe_queue_init() 102 kfree(q); in rxe_queue_init() 118 if (!queue_empty(q, q->type) && (num_elem < queue_count(q, type))) in resize_finish() 139 q->index = q->buf->consumer_index; in resize_finish() 195 if (q->ip) in rxe_queue_cleanup() [all …]
|
| /drivers/net/ethernet/chelsio/cxgb3/ |
| A D | sge.c | 543 if (++q->pidx == q->size) { in refill_fl() 580 q->sdesc[q->pidx] = q->sdesc[idx]; in recycle_rx_buf() 587 if (++q->pidx == q->size) { in recycle_rx_buf() 647 memset(q, 0, sizeof(*q)); in t3_reset_qset() 1323 if (q->pidx >= q->size) { in t3_eth_xmit() 1324 q->pidx -= q->size; in t3_eth_xmit() 1500 write_imm(&q->desc[q->pidx], skb, skb->len, q->gen); in ctrl_xmit() 1533 write_imm(&q->desc[q->pidx], skb, skb->len, q->gen); in restart_ctrlq() 1723 if (q->pidx >= q->size) { in ofld_xmit() 1724 q->pidx -= q->size; in ofld_xmit() [all …]
|
| /drivers/net/wireless/mediatek/mt76/ |
| A D | dma.c | 191 Q_WRITE(q, desc_base, q->desc_dma); in mt76_dma_sync_idx() 195 Q_WRITE(q, ring_size, q->ndesc); in mt76_dma_sync_idx() 196 q->head = Q_READ(q, dma_idx); in mt76_dma_sync_idx() 197 q->tail = q->head; in mt76_dma_sync_idx() 203 if (!q || !q->ndesc) in __mt76_dma_queue_reset() 245 desc = &q->desc[q->head]; in mt76_dma_add_rx_buf() 278 q->head = (q->head + 1) % q->ndesc; in mt76_dma_add_rx_buf() 303 next = (q->head + 1) % q->ndesc; in mt76_dma_add_buf() 378 Q_WRITE(q, cpu_idx, q->head); in mt76_dma_kick_queue() 387 if (!q || !q->ndesc) in mt76_dma_tx_cleanup() [all …]
|
| A D | wed.c | 50 buf = mt76_get_page_pool_buf(q, &offset, q->buf_size); in mt76_wed_init_rx_buf() 98 if (!q || !q->ndesc) in mt76_wed_dma_setup() 102 if (!q->wed || !mtk_wed_device_active(q->wed)) in mt76_wed_dma_setup() 116 q->wed_regs = q->wed->tx_ring[ring].reg_base; in mt76_wed_dma_setup() 126 q->wed_regs = q->wed->txfree_ring.reg_base; in mt76_wed_dma_setup() 132 q->wed_regs = q->wed->rx_ring[ring].reg_base; in mt76_wed_dma_setup() 138 q->head = q->ndesc - 1; in mt76_wed_dma_setup() 139 q->queued = q->head; in mt76_wed_dma_setup() 145 q->head = q->ndesc - 1; in mt76_wed_dma_setup() 146 q->queued = q->head; in mt76_wed_dma_setup() [all …]
|
| A D | usb.c | 445 urb = q->entry[q->tail].urb; in mt76u_get_next_rx_entry() 446 q->tail = (q->tail + 1) % q->ndesc; in mt76u_get_next_rx_entry() 581 q->head = (q->head + 1) % q->ndesc; in mt76u_complete_rx() 653 q->head = q->tail = 0; in mt76u_submit_rx_buffers() 779 entry = q->entry[q->tail]; in mt76u_status_worker() 864 if (q->queued == q->ndesc) in mt76u_tx_queue_skb() 879 q->head = (q->head + 1) % q->ndesc; in mt76u_tx_queue_skb() 892 while (q->first != q->head) { in mt76u_tx_kick() 905 q->first = (q->first + 1) % q->ndesc; in mt76u_tx_kick() 925 q->ep = q->hw_idx + 1; in mt76u_ac_to_hwq() [all …]
|
| /drivers/accel/habanalabs/common/ |
| A D | hw_queue.c | 33 int delta = (q->pi - queue_ci_get(&q->ci, queue_len)); in queue_free_slots() 104 q->pi = hl_queue_inc_ptr(q->pi); in hl_hw_queue_submit_bd() 190 free_slots_cnt = queue_free_slots(q, q->int_queue_len); in int_queue_sanity_checks() 328 q->shadow_queue[hl_pi_2_offset(q->pi)] = job; in ext_queue_schedule_job() 362 pi = q->kernel_address + (q->pi & (q->int_queue_len - 1)) * sizeof(bd); in int_queue_schedule_job() 364 q->pi++; in int_queue_schedule_job() 365 q->pi &= ((q->int_queue_len << 1) - 1); in int_queue_schedule_job() 851 q->pi = 0; in ext_and_cpu_queue_init() 870 &q->bus_address, &q->int_queue_len); in int_queue_init() 879 q->pi = 0; in int_queue_init() [all …]
|
| /drivers/net/ethernet/pensando/ionic/ |
| A D | ionic_txrx.c | 51 if (q->tail_idx == q->head_idx) { in ionic_txq_poke_doorbell() 62 q->dbval | q->head_idx); in ionic_txq_poke_doorbell() 78 if (q->tail_idx == q->head_idx) in ionic_rxq_poke_doorbell() 87 q->dbval | q->head_idx); in ionic_rxq_poke_doorbell() 313 desc = &q->txq[q->head_idx]; in ionic_xdp_post_frame() 741 q->tail_idx = (q->tail_idx + 1) & (q->num_descs - 1); in __ionic_rx_service() 811 desc = &q->rxq[q->head_idx]; in ionic_rx_fill() 878 q->dbval | q->head_idx); in ionic_rx_fill() 1255 q->tail_idx = (q->tail_idx + 1) & (q->num_descs - 1); in ionic_tx_service() 1323 q->tail_idx = (q->tail_idx + 1) & (q->num_descs - 1); in ionic_tx_empty() [all …]
|
| /drivers/net/ethernet/chelsio/cxgb4/ |
| A D | sge.c | 1221 q->q.stops++; in eth_txq_stop() 1592 wr = (void *)&q->q.desc[q->q.pidx]; in cxgb4_eth_xmit() 1593 eowr = (void *)&q->q.desc[q->q.pidx]; in cxgb4_eth_xmit() 2645 q->q.stops++; in ctrlq_check_stop() 2683 wr = (void *)&q->q.desc[q->q.pidx]; in cxgb4_selftest_lb_pkt() 2757 wr = (struct fw_wr_hdr *)&q->q.desc[q->q.pidx]; in ctrl_xmit() 2795 wr = (struct fw_wr_hdr *)&q->q.desc[q->q.pidx]; in restart_ctrlq() 2897 q->q.stops++; in txq_stop_maperr() 2913 q->q.stops++; in ofldtxq_stop() 2975 pos = (u64 *)&q->q.desc[q->q.pidx]; in service_ofldq() [all …]
|
| /drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_device_queue_manager.c | 445 q->doorbell_id = q->properties.queue_id; in allocate_doorbell() 661 q->pipe, q->queue); in create_queue_nocpsch() 685 mqd_mgr->restore_mqd(mqd_mgr, &q->mqd, q->mqd_mem_obj, &q->gart_mqd_addr, in create_queue_nocpsch() 690 &q->gart_mqd_addr, &q->properties); in create_queue_nocpsch() 703 q->queue, &q->properties, current->mm); in create_queue_nocpsch() 885 q->pipe, q->queue); in destroy_queue_nocpsch_locked() 1043 q->pipe, q->queue, in update_queue() 1311 q->queue, &q->properties, mm); in restore_process_queues_nocpsch() 1603 q->properties.sdma_queue_id = q->sdma_id / in allocate_sdma_queue() 2017 mqd_mgr->restore_mqd(mqd_mgr, &q->mqd, q->mqd_mem_obj, &q->gart_mqd_addr, in create_queue_cpsch() [all …]
|
| A D | kfd_process_queue_manager.c | 39 if ((pqn->q && pqn->q->properties.queue_id == qid) || in get_queue_by_qid() 119 if (pqn->q) in pqm_set_gws() 303 *q = NULL; in init_user_queue() 333 q = NULL; in pqm_create_queue() 402 pqn->q = q; in pqm_create_queue() 423 pqn->q = q; in pqm_create_queue() 483 if (q) { in pqm_create_queue() 526 if (pqn->q) in pqm_destroy_queue() 594 struct queue *q = pqn->q; in pqm_update_queue_properties() local 1118 q = pqn->q; in pqm_debugfs_mqds() [all …]
|
| /drivers/net/ethernet/mediatek/ |
| A D | mtk_wed_wo.c | 103 int index = (q->tail + 1) % q->n_desc; in mtk_wed_wo_dequeue() 140 while (q->queued < q->n_desc) { in mtk_wed_wo_queue_refill() 145 buf = page_frag_alloc(&q->cache, q->buf_size, in mtk_wed_wo_queue_refill() 156 q->head = (q->head + 1) % q->n_desc; in mtk_wed_wo_queue_refill() 157 entry = &q->entry[q->head]; in mtk_wed_wo_queue_refill() 160 q->entry[q->head].buf = buf; in mtk_wed_wo_queue_refill() 218 u32 index = (q->head - 1) % q->n_desc; in mtk_wed_wo_rx_run_queue() 282 dma_free_coherent(wo->hw->dev, q->n_desc * sizeof(*q->desc), q->desc, in mtk_wed_wo_queue_free() 337 q->tail = mtk_wed_mmio_r32(wo, q->regs.dma_idx); in mtk_wed_wo_queue_tx_skb() 338 index = (q->head + 1) % q->n_desc; in mtk_wed_wo_queue_tx_skb() [all …]
|
| /drivers/misc/uacce/ |
| A D | uacce.c | 23 return q->state == UACCE_Q_INIT || q->state == UACCE_Q_STARTED; in uacce_queue_is_valid() 34 ret = q->uacce->ops->start_queue(q); in uacce_start_queue() 50 if ((q->state == UACCE_Q_INIT || q->state == UACCE_Q_STARTED) && in uacce_put_queue() 125 q->pasid = pasid; in uacce_bind_queue() 131 if (!q->handle) in uacce_unbind_queue() 134 q->handle = NULL; in uacce_unbind_queue() 148 if (!q) in uacce_fops_open() 162 q->uacce = uacce; in uacce_fops_open() 165 ret = uacce->ops->get_queue(uacce, q->pasid, q); in uacce_fops_open() 183 kfree(q); in uacce_fops_open() [all …]
|
| /drivers/net/ |
| A D | tap.c | 88 if (q->enabled) in tap_enable_queue() 112 q->file = file; in tap_set_queue() 130 if (!q->enabled) in tap_disable_queue() 167 if (q->enabled) in tap_put_queue() 269 if (!q) in tap_handle_frame() 461 if (!q) in tap_open() 473 sock_init_data_uid(&q->sock, &q->sk, current_fsuid()); in tap_open() 525 if (!q) in tap_poll() 940 q->flags = (q->flags & ~TAP_IFFEATURES) | u; in tap_ioctl() 1173 if (!q) in tap_get_socket() [all …]
|
| /drivers/net/ethernet/mellanox/mlxsw/ |
| A D | pci.c | 189 int index = q->producer_counter & (q->count - 1); in mlxsw_pci_queue_elem_info_producer_get() 191 if ((u16) (q->producer_counter - q->consumer_counter) == q->count) in mlxsw_pci_queue_elem_info_producer_get() 199 int index = q->consumer_counter & (q->count - 1); in mlxsw_pci_queue_elem_info_consumer_get() 211 return owner_bit != !!(q->consumer_counter & q->count); in mlxsw_pci_elem_hw_owned() 279 q->consumer_counter + q->count); in mlxsw_pci_queue_doorbell_consumer_ring() 314 cq_num = q->num; in mlxsw_pci_sdq_init() 330 cq->u.cq.dq = q; in mlxsw_pci_sdq_init() 1253 q->count = q_ops->elem_count_f ? q_ops->elem_count_f(q) : in mlxsw_pci_queue_init() 1255 q->elem_size = q_ops->elem_size_f ? q_ops->elem_size_f(q) : in mlxsw_pci_queue_init() 1267 q->elem_info = kcalloc(q->count, sizeof(*q->elem_info), GFP_KERNEL); in mlxsw_pci_queue_init() [all …]
|