Home
last modified time | relevance | path

Searched refs:cq_base (Results 1 – 18 of 18) sorted by relevance

/linux/drivers/net/ethernet/amd/pds_core/
A Dcore.c152 if (qcq->cq_base) in pdsc_qcq_free()
154 qcq->cq_base, qcq->cq_base_pa); in pdsc_qcq_free()
192 void *q_base, *cq_base; in pdsc_qcq_alloc() local
248 cq_base = PTR_ALIGN(q_base + in pdsc_qcq_alloc()
271 qcq->cq_base = dma_alloc_coherent(dev, qcq->cq_size, in pdsc_qcq_alloc()
274 if (!qcq->cq_base) { in pdsc_qcq_alloc()
278 cq_base = PTR_ALIGN(qcq->cq_base, PDS_PAGE_SIZE); in pdsc_qcq_alloc()
282 pdsc_cq_map(&qcq->cq, cq_base, cq_base_pa); in pdsc_qcq_alloc()
A Dcore.h119 void *cq_base; member
/linux/drivers/infiniband/hw/irdma/
A Duser.h291 void irdma_uk_cq_resize(struct irdma_cq_uk *cq, void *cq_base, int size);
352 struct irdma_cqe *cq_base; member
391 struct irdma_cqe *cq_base; member
A Duk.c893 void irdma_uk_cq_resize(struct irdma_cq_uk *cq, void *cq_base, int cq_size) in irdma_uk_cq_resize() argument
895 cq->cq_base = cq_base; in irdma_uk_cq_resize()
1008 ext_cqe = cq->cq_base[peek_head].buf; in irdma_uk_cq_poll_cmpl()
1475 cq->cq_base = info->cq_base; in irdma_uk_cq_init()
1502 cqe = ((struct irdma_extended_cqe *)(cq->cq_base))[cq_head].buf; in irdma_uk_clean_cq()
1504 cqe = cq->cq_base[cq_head].buf; in irdma_uk_clean_cq()
A Ddefs.h944 (_cq)->cq_base[IRDMA_RING_CURRENT_HEAD((_cq)->cq_ring)].buf \
949 ((_cq)->cq_base))[IRDMA_RING_CURRENT_HEAD((_cq)->cq_ring)].buf \
A Dtype.h625 struct irdma_cqe *cq_base; member
771 struct irdma_cqe *cq_base; member
A Dpuda.c241 ext_cqe = cq_uk->cq_base[peek_head].buf; in irdma_puda_poll_info()
801 init_info->cq_base = mem->va; in irdma_puda_cq_create()
A Dctrl.c2645 irdma_uk_cq_resize(&cq->cq_uk, info->cq_base, info->cq_size); in irdma_sc_cq_resize()
4183 cq->cq_uk.cq_base = info->cq_base; in irdma_sc_ccq_init()
A Dverbs.c1968 info.cq_base = kmem_buf.va; in irdma_resize_cq()
2185 ukinfo->cq_base = iwcq->kmem.va; in irdma_create_cq()
A Dhw.c1059 info.cq_base = ccq->mem_cq.va; in irdma_create_ccq()
/linux/drivers/accel/habanalabs/common/
A Dirq.c136 struct hl_cq_entry *cq_entry, *cq_base; in hl_irq_handler_cq() local
146 cq_base = cq->kernel_address; in hl_irq_handler_cq()
149 cq_entry = (struct hl_cq_entry *) &cq_base[cq->ci]; in hl_irq_handler_cq()
/linux/drivers/net/ethernet/pensando/ionic/
A Dionic_lif.c431 if (qcq->cq_base) { in ionic_qcq_free()
432 dma_free_coherent(dev, qcq->cq_size, qcq->cq_base, qcq->cq_base_pa); in ionic_qcq_free()
433 qcq->cq_base = NULL; in ionic_qcq_free()
698 new->cq_base = dma_alloc_coherent(dev, new->cq_size, &new->cq_base_pa, in ionic_qcq_alloc()
700 if (!new->cq_base) { in ionic_qcq_alloc()
705 new->cq.base = PTR_ALIGN(new->cq_base, PAGE_SIZE); in ionic_qcq_alloc()
733 dma_free_coherent(dev, new->cq_size, new->cq_base, new->cq_base_pa); in ionic_qcq_alloc()
826 memset(qcq->cq_base, 0, qcq->cq_size); in ionic_qcq_sanitize()
2927 swap(a->cq_base, b->cq_base); in ionic_swap_queues()
A Dionic_lif.h75 void *cq_base; member
/linux/drivers/infiniband/hw/bnxt_re/
A Dqplib_fp.c1483 struct cq_base *hw_cqe; in __clean_cq()
2166 hwq_attr.stride = sizeof(struct cq_base); in bnxt_qplib_create_cq()
2246 hwq_attr.stride = sizeof(struct cq_base); in bnxt_qplib_resize_cq()
2406 struct cq_base *peek_hwcqe; in do_wa9060()
2813 struct cq_base *hw_cqe; in bnxt_qplib_is_cq_empty()
3059 struct cq_base *hw_cqe; in bnxt_qplib_poll_cq()
A Dqplib_fp.h348 #define BNXT_QPLIB_MAX_CQE_ENTRY_SIZE sizeof(struct cq_base)
A Dib_verbs.c3044 entries * sizeof(struct cq_base), in bnxt_re_create_cq()
3183 entries * sizeof(struct cq_base), in bnxt_re_resize_cq()
4249 resp.cqe_sz = sizeof(struct cq_base); in bnxt_re_alloc_ucontext()
A Droce_hsi.h3820 struct cq_base { struct
/linux/drivers/accel/habanalabs/gaudi2/
A Dgaudi2.c6906 struct hl_cq_entry *cq_base; in gaudi2_send_job_to_kdma() local
6940 cq_base = cq->kernel_address; in gaudi2_send_job_to_kdma()
6941 polling_addr = (u32 *)&cq_base[cq->ci]; in gaudi2_send_job_to_kdma()

Completed in 146 milliseconds