Home
last modified time | relevance | path

Searched refs:cmdq (Results 1 – 25 of 59) sorted by relevance

123

/linux-6.3-rc2/drivers/infiniband/hw/erdma/
A Derdma_cmdq.c90 struct erdma_cmdq *cmdq = &dev->cmdq; in erdma_cmdq_sq_init() local
122 struct erdma_cmdq *cmdq = &dev->cmdq; in erdma_cmdq_cq_init() local
153 struct erdma_cmdq *cmdq = &dev->cmdq; in erdma_cmdq_eq_init() local
187 struct erdma_cmdq *cmdq = &dev->cmdq; in erdma_cmdq_init() local
237 cmdq->eq.qbuf, cmdq->eq.qbuf_dma_addr); in erdma_cmdq_init()
243 cmdq->cq.qbuf, cmdq->cq.qbuf_dma_addr); in erdma_cmdq_init()
249 cmdq->sq.qbuf, cmdq->sq.qbuf_dma_addr); in erdma_cmdq_init()
263 struct erdma_cmdq *cmdq = &dev->cmdq; in erdma_cmdq_destroy() local
301 wqe = get_queue_entry(cmdq->sq.qbuf, cmdq->sq.pi, cmdq->sq.depth, in push_cmdq_sqe()
305 cmdq->sq.pi += cmdq->sq.wqebb_cnt; in push_cmdq_sqe()
[all …]
A Derdma.h206 struct erdma_cmdq cmdq; member
276 int erdma_post_cmd_wait(struct erdma_cmdq *cmdq, void *req, u32 req_size,
278 void erdma_cmdq_completion_handler(struct erdma_cmdq *cmdq);
/linux-6.3-rc2/drivers/gpu/drm/nouveau/nvkm/falcon/
A Dcmdq.c28 u32 head = nvkm_falcon_rd32(cmdq->qmgr->falcon, cmdq->head_reg); in nvkm_falcon_cmdq_has_room()
29 u32 tail = nvkm_falcon_rd32(cmdq->qmgr->falcon, cmdq->tail_reg); in nvkm_falcon_cmdq_has_room()
35 free = cmdq->offset + cmdq->size - head; in nvkm_falcon_cmdq_has_room()
67 cmdq->position = cmdq->offset; in nvkm_falcon_cmdq_rewind()
84 cmdq->position = nvkm_falcon_rd32(falcon, cmdq->head_reg); in nvkm_falcon_cmdq_open()
95 nvkm_falcon_wr32(cmdq->qmgr->falcon, cmdq->head_reg, cmdq->position); in nvkm_falcon_cmdq_close()
180 cmdq->head_reg = func->cmdq.head + index * func->cmdq.stride; in nvkm_falcon_cmdq_init()
181 cmdq->tail_reg = func->cmdq.tail + index * func->cmdq.stride; in nvkm_falcon_cmdq_init()
187 index, cmdq->offset, cmdq->size); in nvkm_falcon_cmdq_init()
194 if (cmdq) { in nvkm_falcon_cmdq_del()
[all …]
A DKbuild3 nvkm-y += nvkm/falcon/cmdq.o
/linux-6.3-rc2/drivers/mailbox/
A Dmtk-cmdq-mailbox.c67 struct cmdq *cmdq; member
107 struct cmdq *cmdq = container_of(chan->mbox, struct cmdq, mbox); in cmdq_get_shift_pa() local
138 static void cmdq_init(struct cmdq *cmdq) in cmdq_init() argument
227 struct cmdq *cmdq = task->cmdq; in cmdq_task_handle_error() local
294 struct cmdq *cmdq = dev; in cmdq_irq_handler() local
315 struct cmdq *cmdq = dev_get_drvdata(dev); in cmdq_suspend() local
343 struct cmdq *cmdq = dev_get_drvdata(dev); in cmdq_resume() local
356 struct cmdq *cmdq = platform_get_drvdata(pdev); in cmdq_remove() local
380 task->cmdq = cmdq; in cmdq_mbox_send_data()
545 struct cmdq *cmdq; in cmdq_probe() local
[all …]
/linux-6.3-rc2/drivers/crypto/cavium/nitrox/
A Dnitrox_lib.c29 cmdq->qsize = (ndev->qlen * cmdq->instr_size) + align_bytes; in nitrox_cmdq_init()
30 cmdq->unalign_base = dma_alloc_coherent(DEV(ndev), cmdq->qsize, in nitrox_cmdq_init()
36 cmdq->dma = PTR_ALIGN(cmdq->unalign_dma, align_bytes); in nitrox_cmdq_init()
37 cmdq->base = cmdq->unalign_base + (cmdq->dma - cmdq->unalign_dma); in nitrox_cmdq_init()
64 if (!cmdq) in nitrox_cmdq_cleanup()
74 cmdq->unalign_base, cmdq->unalign_dma); in nitrox_cmdq_cleanup()
82 cmdq->dma = 0; in nitrox_cmdq_cleanup()
83 cmdq->qsize = 0; in nitrox_cmdq_cleanup()
106 cmdq = kzalloc_node(sizeof(*cmdq), GFP_KERNEL, ndev->node); in nitrox_alloc_aqm_queues()
107 if (!cmdq) { in nitrox_alloc_aqm_queues()
[all …]
A Dnitrox_reqmgr.c296 idx = cmdq->write_idx; in post_se_instr()
298 ent = cmdq->base + (idx * cmdq->instr_size); in post_se_instr()
302 response_list_add(sr, cmdq); in post_se_instr()
342 post_se_instr(sr, cmdq); in post_backlog_cmds()
351 struct nitrox_cmdq *cmdq = sr->cmdq; in nitrox_enqueue_request() local
355 post_backlog_cmds(cmdq); in nitrox_enqueue_request()
364 backlog_list_add(sr, cmdq); in nitrox_enqueue_request()
367 post_se_instr(sr, cmdq); in nitrox_enqueue_request()
507 struct nitrox_cmdq *cmdq; in backlog_qflush_work() local
510 post_backlog_cmds(cmdq); in backlog_qflush_work()
[all …]
A Dnitrox_isr.c32 struct nitrox_cmdq *cmdq = qvec->cmdq; in nps_pkt_slc_isr() local
34 slc_cnts.value = readq(cmdq->compl_cnt_csr_addr); in nps_pkt_slc_isr()
337 qvec->cmdq = &ndev->pkt_inq[qvec->ring]; in nitrox_register_interrupts()
/linux-6.3-rc2/drivers/net/ethernet/brocade/bna/
A Dbfa_msgq.c56 cmdq->flags = 0; in cmdq_sm_stopped_entry()
57 cmdq->token = 0; in cmdq_sm_stopped_entry()
58 cmdq->offset = 0; in cmdq_sm_stopped_entry()
195 if (!bfa_nw_ioc_mbox_queue(cmdq->msgq->ioc, &cmdq->dbell_mb, in bfa_msgq_cmdq_dbell()
218 BFA_MSGQ_INDX_ADD(cmdq->producer_index, 1, cmdq->depth); in __cmd_copy()
268 cmdq->token = 0; in bfa_msgq_cmdq_copy_req()
290 cmdq->token++; in bfa_msgq_cmdq_copy_rsp()
294 if (!bfa_nw_ioc_mbox_queue(cmdq->msgq->ioc, &cmdq->copy_mb, in bfa_msgq_cmdq_copy_rsp()
305 cmdq->msgq = msgq; in bfa_msgq_cmdq_attach()
515 bfa_dma_be_addr_set(msgq_cfg->cmdq.addr, msgq->cmdq.addr.pa); in bfa_msgq_init()
[all …]
/linux-6.3-rc2/drivers/net/ethernet/huawei/hinic/
A Dhinic_hw_cmdq.c78 #define cmdq_to_cmdqs(cmdq) container_of((cmdq) - (cmdq)->cmdq_type, \ argument
364 cmdq->wrapped = !cmdq->wrapped; in cmdq_sync_cmd_direct_resp()
443 cmdq->wrapped = !cmdq->wrapped; in cmdq_set_arm_bit()
510 struct hinic_cmdq *cmdq = &cmdqs->cmdq[HINIC_CMDQ_SYNC]; in hinic_set_arm_bit() local
643 struct hinic_cmdq *cmdq = &cmdqs->cmdq[cmdq_type]; in cmdq_ceq_handler() local
743 cmdq->wq = wq; in init_cmdq()
745 cmdq->wrapped = 1; in init_cmdq()
749 cmdq->done = vzalloc(array_size(sizeof(*cmdq->done), wq->q_depth)); in init_cmdq()
750 if (!cmdq->done) in init_cmdq()
753 cmdq->errcode = vzalloc(array_size(sizeof(*cmdq->errcode), in init_cmdq()
[all …]
A Dhinic_hw_io.c533 enum hinic_cmdq_type cmdq, type; in hinic_io_init() local
565 for (cmdq = HINIC_CMDQ_SYNC; cmdq < HINIC_MAX_CMDQ_TYPES; cmdq++) { in hinic_io_init()
573 func_to_io->cmdq_db_area[cmdq] = db_area; in hinic_io_init()
600 for (type = HINIC_CMDQ_SYNC; type < cmdq; type++) in hinic_io_init()
619 enum hinic_cmdq_type cmdq; in hinic_io_free() local
628 for (cmdq = HINIC_CMDQ_SYNC; cmdq < HINIC_MAX_CMDQ_TYPES; cmdq++) in hinic_io_free()
629 return_db_area(func_to_io, func_to_io->cmdq_db_area[cmdq]); in hinic_io_free()
/linux-6.3-rc2/drivers/accel/ivpu/
A Divpu_job.c42 cmdq = kzalloc(sizeof(*cmdq), GFP_KERNEL); in ivpu_cmdq_alloc()
43 if (!cmdq) in ivpu_cmdq_alloc()
54 cmdq->jobq = (struct vpu_job_queue *)cmdq->mem->kvaddr; in ivpu_cmdq_alloc()
70 if (!cmdq) in ivpu_cmdq_free()
80 struct ivpu_cmdq *cmdq = file_priv->cmdq[engine]; in ivpu_cmdq_acquire() local
87 if (!cmdq) in ivpu_cmdq_acquire()
89 file_priv->cmdq[engine] = cmdq; in ivpu_cmdq_acquire()
96 cmdq->mem->vpu_addr, cmdq->mem->base.size); in ivpu_cmdq_acquire()
107 struct ivpu_cmdq *cmdq = file_priv->cmdq[engine]; in ivpu_cmdq_release_locked() local
111 if (cmdq) { in ivpu_cmdq_release_locked()
[all …]
A Divpu_mmu.c305 struct ivpu_mmu_queue *q = &mmu->cmdq; in ivpu_mmu_cmdq_alloc()
398 struct ivpu_mmu_queue *cmdq = &vdev->mmu->cmdq; in ivpu_mmu_cmdq_wait_for_cons() local
400 return REGV_POLL(MTL_VPU_HOST_MMU_CMDQ_CONS, cmdq->cons, (cmdq->prod == cmdq->cons), in ivpu_mmu_cmdq_wait_for_cons()
406 struct ivpu_mmu_queue *q = &vdev->mmu->cmdq; in ivpu_mmu_cmdq_cmd_write()
426 struct ivpu_mmu_queue *q = &vdev->mmu->cmdq; in ivpu_mmu_cmdq_sync()
478 memset(mmu->cmdq.base, 0, IVPU_MMU_CMDQ_SIZE); in ivpu_mmu_reset()
479 clflush_cache_range(mmu->cmdq.base, IVPU_MMU_CMDQ_SIZE); in ivpu_mmu_reset()
480 mmu->cmdq.prod = 0; in ivpu_mmu_reset()
481 mmu->cmdq.cons = 0; in ivpu_mmu_reset()
503 REGV_WR64(MTL_VPU_HOST_MMU_CMDQ_BASE, mmu->cmdq.dma_q); in ivpu_mmu_reset()
/linux-6.3-rc2/drivers/infiniband/hw/bnxt_re/
A Dqplib_rcfw.c62 cmdq = &rcfw->cmdq; in __wait_for_resp()
76 cmdq = &rcfw->cmdq; in __block_for_resp()
91 struct bnxt_qplib_cmdq_ctx *cmdq = &rcfw->cmdq; in __send_message() local
182 cmdq->seq_num++; in __send_message()
577 cmdq = &rcfw->cmdq; in bnxt_qplib_alloc_rcfw_channel()
616 if (!cmdq->cmdq_bitmap) in bnxt_qplib_alloc_rcfw_channel()
659 cmdq = &rcfw->cmdq; in bnxt_qplib_disable_rcfw_channel()
781 cmdq = &rcfw->cmdq; in bnxt_qplib_start_rcfw()
783 mbox = &cmdq->cmdq_mbox; in bnxt_qplib_start_rcfw()
807 cmdq = &rcfw->cmdq; in bnxt_qplib_enable_rcfw_channel()
[all …]
/linux-6.3-rc2/drivers/net/ethernet/hisilicon/hns3/hns3_common/
A Dhclge_comm_cmd.c504 spin_lock_bh(&cmdq->csq.lock); in hclge_comm_cmd_uninit()
505 spin_lock(&cmdq->crq.lock); in hclge_comm_cmd_uninit()
507 spin_unlock(&cmdq->crq.lock); in hclge_comm_cmd_uninit()
523 cmdq->csq.pdev = pdev; in hclge_comm_cmd_queue_init()
524 cmdq->crq.pdev = pdev; in hclge_comm_cmd_queue_init()
560 spin_lock(&cmdq->crq.lock); in hclge_comm_cmd_init()
562 cmdq->csq.next_to_clean = 0; in hclge_comm_cmd_init()
563 cmdq->csq.next_to_use = 0; in hclge_comm_cmd_init()
564 cmdq->crq.next_to_clean = 0; in hclge_comm_cmd_init()
565 cmdq->crq.next_to_use = 0; in hclge_comm_cmd_init()
[all …]
/linux-6.3-rc2/drivers/gpu/drm/nouveau/nvkm/engine/sec2/
A Dbase.c45 struct nvkm_falcon_cmdq *cmdq = sec2->cmdq; in nvkm_sec2_fini() local
56 ret = nvkm_falcon_cmdq_send(cmdq, &cmd, nvkm_sec2_finimsg, sec2, in nvkm_sec2_fini()
68 nvkm_falcon_cmdq_fini(cmdq); in nvkm_sec2_fini()
119 nvkm_falcon_cmdq_del(&sec2->cmdq); in nvkm_sec2_dtor()
159 (ret = nvkm_falcon_cmdq_new(sec2->qmgr, "cmdq", &sec2->cmdq)) || in nvkm_sec2_new_()
A Dga102.c49 nvkm_falcon_cmdq_init(sec2->cmdq, msg.queue_info[i].index, in ga102_sec2_initmsg()
103 return nvkm_falcon_cmdq_send(sec2->cmdq, &cmd.cmd.hdr, in ga102_sec2_acr_bootstrap_falcon()
136 .cmdq = { 0xc00, 0xc04, 8 },
A Dgp102.c71 return nvkm_falcon_cmdq_send(sec2->cmdq, &cmd.cmd.hdr, in gp102_sec2_acr_bootstrap_falcon()
142 nvkm_falcon_cmdq_init(sec2->cmdq, in gp102_sec2_initmsg()
214 .cmdq = { 0xa00, 0xa04, 8 },
/linux-6.3-rc2/drivers/iommu/arm/arm-smmu-v3/
A Darm-smmu-v3.c338 return &smmu->cmdq; in arm_smmu_get_cmdq()
457 if (atomic_read(&cmdq->lock) == 1) in arm_smmu_cmdq_shared_tryunlock()
460 arm_smmu_cmdq_shared_unlock(cmdq); in arm_smmu_cmdq_shared_tryunlock()
540 ptr = &cmdq->valid_map[swidx]; in __arm_smmu_cmdq_poll_set_valid_map()
594 WRITE_ONCE(cmdq->q.llq.cons, readl_relaxed(cmdq->q.cons_reg)); in arm_smmu_cmdq_poll_until_not_full()
788 arm_smmu_cmdq_shared_lock(cmdq); in arm_smmu_cmdq_issue_cmdlist()
802 &cmdq->q.llq.atomic.prod); in arm_smmu_cmdq_issue_cmdlist()
2919 struct arm_smmu_cmdq *cmdq = &smmu->cmdq; in arm_smmu_cmdq_init() local
2922 atomic_set(&cmdq->owner_prod, 0); in arm_smmu_cmdq_init()
2923 atomic_set(&cmdq->lock, 0); in arm_smmu_cmdq_init()
[all …]
/linux-6.3-rc2/drivers/atm/
A Dfore200e.c557 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_pca_prom_read()
562 FORE200E_NEXT_ENTRY(cmdq->head, QUEUE_SIZE_CMD); in fore200e_pca_prom_read()
1226 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_activate_vcin()
1670 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_getstats()
1716 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ];
1756 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_set_oc3()
2240 &cmdq->status, in fore200e_init_cmd_queue()
2253 cmdq->host_entry[ i ].status = in fore200e_init_cmd_queue()
2255 cmdq->host_entry[ i ].cp_entry = &cp_entry[ i ]; in fore200e_init_cmd_queue()
2257 *cmdq->host_entry[ i ].status = STATUS_FREE; in fore200e_init_cmd_queue()
[all …]
/linux-6.3-rc2/Documentation/devicetree/bindings/iommu/
A Darm,smmu-v3.yaml45 - cmdq-sync # CMD_SYNC complete
91 interrupt-names = "eventq", "gerror", "priq", "cmdq-sync";
/linux-6.3-rc2/drivers/media/platform/mediatek/mdp3/
A DMakefile4 mtk-mdp3-y += mtk-mdp3-comp.o mtk-mdp3-cmdq.o
/linux-6.3-rc2/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/
A Dgp102.c38 .cmdq = { 0x4a0, 0x4b0, 4 },
/linux-6.3-rc2/drivers/gpu/drm/nouveau/include/nvkm/engine/
A Dsec2.h17 struct nvkm_falcon_cmdq *cmdq; member
/linux-6.3-rc2/drivers/soc/mediatek/
A DMakefile2 obj-$(CONFIG_MTK_CMDQ) += mtk-cmdq-helper.o

Completed in 77 milliseconds

123