| /linux/drivers/net/ethernet/fungible/funcore/ |
| A D | fun_queue.c | 100 sqid, cqid, sqe_size_log2, in fun_sq_create() 145 cqid, rqid, cqe_size_log2, in fun_cq_create() 331 funq->cqid, new_cqes, max, funq->cq_head, funq->cq_phase); in __fun_process_cq() 420 funq->cqid = 2 * qid; in fun_alloc_queue() 423 funq->rqid = funq->cqid; in fun_alloc_queue() 431 funq->cqid = qid; in fun_alloc_queue() 475 if (funq->cqid == 0) in fun_alloc_queue() 494 rc = fun_cq_create(fdev, funq->cq_flags, funq->cqid, rqid, in fun_create_cq() 498 &funq->cqid, &funq->cq_db); in fun_create_cq() 500 dev_dbg(fdev->dev, "created CQ %u\n", funq->cqid); in fun_create_cq() [all …]
|
| A D | fun_queue.h | 41 u32 cqid; member 132 int fun_sq_create(struct fun_dev *fdev, u16 flags, u32 sqid, u32 cqid, 137 int fun_cq_create(struct fun_dev *fdev, u16 flags, u32 cqid, u32 rqid, 151 #define fun_destroy_cq(fdev, cqid) \ argument 152 fun_res_destroy((fdev), FUN_ADMIN_OP_EPCQ, 0, (cqid))
|
| A D | fun_dev.h | 119 unsigned int cqid) in fun_cq_db_addr() argument 121 return fun_db_addr(fdev, cqid * 2 + 1); in fun_cq_db_addr()
|
| /linux/include/uapi/rdma/ |
| A D | mana-abi.h | 31 __u32 cqid; member 43 __u32 cqid; member 78 __u32 cqid; member
|
| A D | cxgb4-abi.h | 61 __u32 cqid; member
|
| A D | bnxt_re-abi.h | 111 __u32 cqid; member
|
| /linux/drivers/infiniband/hw/cxgb4/ |
| A D | ev.c | 124 u32 cqid; in c4iw_ev_dispatch() local 139 cqid = qhp->attr.scq; in c4iw_ev_dispatch() 141 cqid = qhp->attr.rcq; in c4iw_ev_dispatch() 142 chp = get_chp(dev, cqid); in c4iw_ev_dispatch() 145 cqid, CQE_QPID(err_cqe), in c4iw_ev_dispatch()
|
| A D | cq.c | 58 res->u.cq.iqid = cpu_to_be32(cq->cqid); in destroy_cq() 67 c4iw_put_cqid(rdev, cq->cqid, uctx); in destroy_cq() 85 cq->cqid = c4iw_get_cqid(rdev, uctx); in create_cq() 86 if (!cq->cqid) { in create_cq() 136 res->u.cq.iqid = cpu_to_be32(cq->cqid); in create_cq() 168 pci_name(rdev->lldi.pdev), cq->cqid); in create_cq() 179 c4iw_put_cqid(rdev, cq->cqid, uctx); in create_cq() 342 pr_debug("cqid 0x%x\n", chp->cq.cqid); in c4iw_flush_hw_cq() 744 cq, cq->cqid, cq->sw_cidx); in poll_cq() 748 cq, cq->cqid, cq->cidx); in poll_cq() [all …]
|
| A D | t4.h | 688 u32 cqid; member 709 writel(val | INGRESSQID_V(cq->cqid), cq->gts); in write_gts() 738 __func__, cq->cqid); in t4_swcq_produce() 793 pr_err("cq overflow cqid %u\n", cq->cqid); in t4_next_hw_cqe()
|
| A D | iw_cxgb4.h | 343 static inline struct c4iw_cq *get_chp(struct c4iw_dev *rhp, u32 cqid) in get_chp() argument 345 return xa_load(&rhp->cqs, cqid); in get_chp()
|
| A D | qp.c | 338 FW_RI_RES_WR_IQID_V(scq->cqid)); in create_qp() 368 FW_RI_RES_WR_IQID_V(rcq->cqid)); in create_qp() 2131 schp = get_chp(rhp, ((struct c4iw_cq *)attrs->send_cq)->cq.cqid); in c4iw_create_qp() 2132 rchp = get_chp(rhp, ((struct c4iw_cq *)attrs->recv_cq)->cq.cqid); in c4iw_create_qp() 2188 qhp->attr.scq = ((struct c4iw_cq *) attrs->send_cq)->cq.cqid; in c4iw_create_qp() 2189 qhp->attr.rcq = ((struct c4iw_cq *) attrs->recv_cq)->cq.cqid; in c4iw_create_qp()
|
| A D | restrack.c | 273 if (rdma_nl_put_driver_u32(msg, "cqid", cq->cqid)) in fill_cq()
|
| /linux/drivers/nvme/host/ |
| A D | trace.c | 27 u16 cqid = get_unaligned_le16(cdw10 + 6); in nvme_trace_create_sq() local 31 sqid, qsize, sq_flags, cqid); in nvme_trace_create_sq() 40 u16 cqid = get_unaligned_le16(cdw10); in nvme_trace_delete_cq() local 42 trace_seq_printf(p, "cqid=%u", cqid); in nvme_trace_delete_cq() 51 u16 cqid = get_unaligned_le16(cdw10); in nvme_trace_create_cq() local 57 cqid, qsize, cq_flags, irq_vector); in nvme_trace_create_cq()
|
| A D | apple.c | 686 c.create_cq.cqid = cpu_to_le16(1); in apple_nvme_create_cq() 717 c.create_sq.cqid = cpu_to_le16(1); in apple_nvme_create_sq()
|
| A D | pci.c | 1207 c.create_cq.cqid = cpu_to_le16(qid); in adapter_alloc_cq() 1239 c.create_sq.cqid = cpu_to_le16(qid); in adapter_alloc_sq() 1244 static int adapter_delete_cq(struct nvme_dev *dev, u16 cqid) in adapter_delete_cq() argument 1246 return adapter_delete_queue(dev, nvme_admin_delete_cq, cqid); in adapter_delete_cq()
|
| /linux/drivers/infiniband/hw/mana/ |
| A D | cq.c | 69 resp.cqid = cq->queue.id; in mana_ib_create_cq()
|
| A D | qp.c | 214 resp.entries[i].cqid = cq->queue.id; in mana_ib_create_qp_rss() 373 resp.cqid = send_cq->queue.id; in mana_ib_create_qp_raw()
|
| /linux/drivers/scsi/lpfc/ |
| A D | lpfc_sli.c | 594 int cqid = 0; in lpfc_sli4_eqcq_flush() local 604 if (childq->queue_id == cqid) { in lpfc_sli4_eqcq_flush() 11193 cqid, ia, wqec); in lpfc_sli_prep_abort_xri() 12930 cqid = lpfc_cmd->hdwq->io_cq_map; in lpfc_sli_abort_taskmgmt() 12951 iocbq->iocb.ulpClass, cqid, in lpfc_sli_abort_taskmgmt() 14813 uint16_t cqid; in lpfc_sli4_sp_handle_eqe() local 14820 if (childq->queue_id == cqid) { in lpfc_sli4_sp_handle_eqe() 15428 uint16_t cqid, id; in lpfc_sli4_hba_handle_eqe() local 15453 if ((cqid >= id) && (cqid < (id + phba->cfg_nvmet_mrq))) { in lpfc_sli4_hba_handle_eqe() 15478 cqid, cq->queue_id); in lpfc_sli4_hba_handle_eqe() [all …]
|
| A D | lpfc_crtn.h | 380 u16 ulp_context, u16 iotag, u8 ulp_class, u16 cqid,
|
| A D | lpfc.h | 1006 u8 ulp_class, u16 cqid, bool ia,
|
| /linux/include/linux/ |
| A D | nvme.h | 1319 __le16 cqid; member 1336 __le16 cqid; member
|
| /linux/drivers/infiniband/hw/ocrdma/ |
| A D | ocrdma_hw.c | 688 u16 cqid = cqe->cqvalid_cqid & OCRDMA_AE_MCQE_CQID_MASK; in ocrdma_dispatch_ibevent() local 706 if (cqid < dev->attr.max_cq) in ocrdma_dispatch_ibevent() 707 cq = dev->cq_tbl[cqid]; in ocrdma_dispatch_ibevent() 710 dev->id, cqid); in ocrdma_dispatch_ibevent()
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| A D | ib_verbs.c | 3101 resp.cqid = cq->qplib_cq.id; in bnxt_re_create_cq()
|