Home
last modified time | relevance | path

Searched refs:q_depth (Results 1 – 25 of 33) sorted by relevance

12

/linux-6.3-rc2/drivers/net/ethernet/microsoft/mana/
A Dhw_channel.c379 hwc_cq->queue_depth = q_depth; in mana_hwc_create_cq()
410 dma_buf->num_reqs = q_depth; in mana_hwc_alloc_dma_buf()
412 buf_size = PAGE_ALIGN(q_depth * max_msg_size); in mana_hwc_alloc_dma_buf()
424 for (i = 0; i < q_depth; i++) { in mana_hwc_alloc_dma_buf()
493 hwc_wq->queue_depth = q_depth; in mana_hwc_create_wq()
578 for (i = 0; i < q_depth; i++) { in mana_hwc_test_channel()
589 for (i = 0; i < q_depth; ++i) in mana_hwc_test_channel()
622 *q_depth = hwc->hwc_init_q_depth_max; in mana_hwc_establish_channel()
644 err = mana_hwc_init_inflight_msg(hwc, q_depth); in mana_hwc_init_queues()
651 err = mana_hwc_create_cq(hwc, q_depth * 2, in mana_hwc_init_queues()
[all …]
/linux-6.3-rc2/drivers/net/ethernet/huawei/hinic/
A Dhinic_hw_wq.c504 u16 wqebb_size, u32 wq_page_size, u16 q_depth, in hinic_wq_allocate() argument
523 if (q_depth & (q_depth - 1)) { in hinic_wq_allocate()
547 wq->q_depth = q_depth; in hinic_wq_allocate()
564 atomic_set(&wq->delta, q_depth); in hinic_wq_allocate()
565 wq->mask = q_depth - 1; in hinic_wq_allocate()
602 u16 q_depth, u16 max_wqe_size) in hinic_wqs_cmdq_alloc() argument
620 if (q_depth & (q_depth - 1)) { in hinic_wqs_cmdq_alloc()
650 wq[i].q_depth = q_depth; in hinic_wqs_cmdq_alloc()
668 atomic_set(&wq[i].delta, q_depth); in hinic_wqs_cmdq_alloc()
669 wq[i].mask = q_depth - 1; in hinic_wqs_cmdq_alloc()
[all …]
A Dhinic_hw_wq.h30 u16 q_depth; member
80 u16 q_depth, u16 max_wqe_size);
91 u16 wqebb_size, u32 wq_page_size, u16 q_depth,
A Dhinic_hw_cmdq.c363 if (next_prod_idx >= wq->q_depth) { in cmdq_sync_cmd_direct_resp()
365 next_prod_idx -= wq->q_depth; in cmdq_sync_cmd_direct_resp()
442 if (next_prod_idx >= wq->q_depth) { in cmdq_set_arm_bit()
444 next_prod_idx -= wq->q_depth; in cmdq_set_arm_bit()
749 cmdq->done = vzalloc(array_size(sizeof(*cmdq->done), wq->q_depth)); in init_cmdq()
754 wq->q_depth)); in init_cmdq()
A Dhinic_hw_qp.c224 skb_arr_size = wq->q_depth * sizeof(*sq->saved_skb); in alloc_sq_skb_arr()
252 skb_arr_size = wq->q_depth * sizeof(*rq->saved_skb); in alloc_rq_skb_arr()
324 cqe_size = wq->q_depth * sizeof(*rq->cqe); in alloc_rq_cqe()
329 cqe_dma_size = wq->q_depth * sizeof(*rq->cqe_dma); in alloc_rq_cqe()
334 for (i = 0; i < wq->q_depth; i++) { in alloc_rq_cqe()
367 for (i = 0; i < wq->q_depth; i++) in free_rq_cqe()
/linux-6.3-rc2/drivers/net/ethernet/amazon/ena/
A Dena_eth_com.h80 return io_sq->q_depth - 1 - cnt; in ena_com_free_q_entries()
179 need_update = unreported_comp > (io_cq->q_depth / ENA_COMP_HEAD_THRESH); in ena_com_update_dev_comp_head()
217 if (unlikely((io_cq->head & (io_cq->q_depth - 1)) == 0)) in ena_com_cq_inc_head()
228 masked_head = io_cq->head & (io_cq->q_depth - 1); in ena_com_tx_comp_req_id_get()
246 if (unlikely(*req_id >= io_cq->q_depth)) { in ena_com_tx_comp_req_id_get()
A Dena_eth_com.c15 head_masked = io_cq->head & (io_cq->q_depth - 1); in ena_com_get_next_rx_cdesc()
41 tail_masked = io_sq->tail & (io_sq->q_depth - 1); in get_sq_desc_regular_queue()
56 dst_tail_mask = io_sq->tail & (io_sq->q_depth - 1); in ena_com_write_bounce_buffer_to_dev()
84 if (unlikely((io_sq->tail & (io_sq->q_depth - 1)) == 0)) in ena_com_write_bounce_buffer_to_dev()
221 if (unlikely((io_sq->tail & (io_sq->q_depth - 1)) == 0)) in ena_com_sq_update_tail()
230 idx &= (io_cq->q_depth - 1); in ena_com_rx_cdesc_idx_to_ptr()
259 head_masked = io_cq->head & (io_cq->q_depth - 1); in ena_com_cdesc_rx_pkt_get()
548 u16 q_depth = io_cq->q_depth; in ena_com_rx_pkt() local
578 if (unlikely(ena_buf[i].req_id >= q_depth)) in ena_com_rx_pkt()
A Dena_com.c145 aenq->head = aenq->q_depth; in ena_com_admin_init_aenq()
185 command_id, admin_queue->q_depth); in get_comp_ctxt()
226 if (cnt >= admin_queue->q_depth) { in __ena_com_submit_admin_cmd()
1254 create_cmd.sq_depth = io_sq->q_depth; in ena_com_create_io_sq()
1293 io_sq->idx, io_sq->q_depth); in ena_com_create_io_sq()
1404 create_cmd.cq_depth = io_cq->q_depth; in ena_com_create_io_cq()
1441 io_cq->idx, io_cq->q_depth); in ena_com_create_io_cq()
1542 u16 depth = ena_dev->aenq.q_depth; in ena_com_admin_aenq_enable()
1704 size = ADMIN_AENQ_SIZE(aenq->q_depth); in ena_com_admin_destroy()
1891 io_cq->q_depth = ctx->queue_size; in ena_com_create_io_queue()
[all …]
A Dena_com.h130 u16 q_depth; member
174 u16 q_depth; member
223 u16 q_depth; member
253 u16 q_depth; member
/linux-6.3-rc2/drivers/net/ethernet/brocade/bna/
A Dbnad.c91 u32 q_depth, u32 index) in bnad_tx_buff_unmap() argument
114 BNA_QE_INDX_INC(index, q_depth); in bnad_tx_buff_unmap()
126 BNA_QE_INDX_INC(index, q_depth); in bnad_tx_buff_unmap()
174 q_depth = tcb->q_depth; in bnad_txcmpl_process()
341 u32 alloced, prod, q_depth; in bnad_rxq_refill_page() local
350 q_depth = rcb->q_depth; in bnad_rxq_refill_page()
397 BNA_QE_INDX_INC(prod, q_depth); in bnad_rxq_refill_page()
424 q_depth = rcb->q_depth; in bnad_rxq_refill_skb()
453 BNA_QE_INDX_INC(prod, q_depth); in bnad_rxq_refill_skb()
2917 u32 prod, q_depth, vect_id; in bnad_start_xmit() local
[all …]
A Dbna_types.h428 u32 q_depth; member
559 u32 q_depth; member
575 int q_depth; member
623 u32 q_depth; member
A Dbfa_msgq.c516 msgq_cfg->cmdq.q_depth = htons(msgq->cmdq.depth); in bfa_msgq_init()
518 msgq_cfg->rspq.q_depth = htons(msgq->rspq.depth); in bfa_msgq_init()
A Dbfi.h413 u16 q_depth; /* Total num of entries in the queue */ member
/linux-6.3-rc2/drivers/nvme/host/
A Dpci.c131 u32 q_depth; member
200 u32 q_depth; member
1469 int q_depth = dev->q_depth; in nvme_cmb_qdepth() local
1484 if (q_depth < 64) in nvme_cmb_qdepth()
1488 return q_depth; in nvme_cmb_qdepth()
1525 nvmeq->q_depth = depth; in nvme_alloc_queue()
2295 dev->q_depth = result; in nvme_setup_io_queues()
2512 dev->q_depth = 2; in nvme_pci_enable()
2515 dev->q_depth); in nvme_pci_enable()
2519 dev->q_depth = 64; in nvme_pci_enable()
[all …]
/linux-6.3-rc2/drivers/crypto/hisilicon/sec2/
A Dsec_crypto.c310 u16 q_depth = res->depth; in sec_alloc_civ_resource() local
318 for (i = 1; i < q_depth; i++) { in sec_alloc_civ_resource()
335 u16 q_depth = res->depth; in sec_alloc_aiv_resource() local
343 for (i = 1; i < q_depth; i++) { in sec_alloc_aiv_resource()
360 u16 q_depth = res->depth; in sec_alloc_mac_resource() local
368 for (i = 1; i < q_depth; i++) { in sec_alloc_mac_resource()
397 u16 q_depth = res->depth; in sec_alloc_pbuf_resource() local
398 int size = SEC_PBUF_PAGE_NUM(q_depth); in sec_alloc_pbuf_resource()
419 if (k == q_depth) in sec_alloc_pbuf_resource()
488 u16 q_depth = qp_ctx->qp->sq_depth; in sec_alloc_qp_ctx_resource() local
[all …]
/linux-6.3-rc2/drivers/scsi/bfa/
A Dbfi.h538 #define BFI_MSGQ_FULL(_q) (((_q->pi + 1) % _q->q_depth) == _q->ci)
540 #define BFI_MSGQ_UPDATE_CI(_q) (_q->ci = (_q->ci + 1) % _q->q_depth)
541 #define BFI_MSGQ_UPDATE_PI(_q) (_q->pi = (_q->pi + 1) % _q->q_depth)
544 #define BFI_MSGQ_FREE_CNT(_q) ((_q->ci - _q->pi - 1) & (_q->q_depth - 1))
585 u16 q_depth; /* Total num of entries in the queue */ member
A Dbfa_fcpim.h119 u16 q_depth; member
/linux-6.3-rc2/drivers/net/ethernet/fungible/funcore/
A Dfun_dev.h69 unsigned int q_depth; /* max queue depth supported by device */ member
A Dfun_queue.c89 if (sq_depth > fdev->q_depth) in fun_sq_create()
138 if (cq_depth > fdev->q_depth) in fun_cq_create()
A Dfun_dev.c791 fdev->q_depth = NVME_CAP_MQES(fdev->cap_reg) + 1; in fun_dev_enable()
821 fdev->q_depth, fdev->db_stride, fdev->max_qid, in fun_dev_enable()
/linux-6.3-rc2/drivers/crypto/hisilicon/zip/
A Dzip_crypto.c605 u16 q_depth = ctx->qp_ctx[0].qp->sq_depth; in hisi_zip_create_req_q() local
611 req_q->size = q_depth; in hisi_zip_create_req_q()
657 u16 q_depth = ctx->qp_ctx[0].qp->sq_depth; in hisi_zip_create_sgl_pool() local
665 tmp->sgl_pool = hisi_acc_create_sgl_pool(dev, q_depth << 1, in hisi_zip_create_sgl_pool()
/linux-6.3-rc2/net/mac80211/
A Ddebugfs_sta.c258 u32 q_depth[IEEE80211_NUM_ACS]; in sta_aql_read() local
271 q_depth[ac] = atomic_read(&sta->airtime[ac].aql_tx_pending); in sta_aql_read()
277 q_depth[0], q_depth[1], q_depth[2], q_depth[3], in sta_aql_read()
/linux-6.3-rc2/drivers/block/
A Dublk_drv.c120 int q_depth; member
345 return round_up(ubq->q_depth * sizeof(struct ublksrv_io_desc), in ublk_queue_cmd_buf_size()
1048 for (i = 0; i < ubq->q_depth; i++) { in ublk_abort_queue()
1100 return ubq->nr_io_ready == ubq->q_depth; in ublk_queue_ready()
1110 for (i = 0; i < ubq->q_depth; i++) { in ublk_cancel_queue()
1284 if (tag >= ubq->q_depth) in ublk_ch_uring_cmd()
1389 ubq->q_depth = ub->dev_info.queue_depth; in ublk_init_queue()
1967 for (i = 0; i < ubq->q_depth; i++) { in ublk_queue_reinit()
/linux-6.3-rc2/drivers/scsi/mpi3mr/
A Dmpi3mr_os.c957 int q_depth) in mpi3mr_change_queue_depth() argument
964 q_depth = 1; in mpi3mr_change_queue_depth()
965 if (q_depth > shost->can_queue) in mpi3mr_change_queue_depth()
966 q_depth = shost->can_queue; in mpi3mr_change_queue_depth()
967 else if (!q_depth) in mpi3mr_change_queue_depth()
968 q_depth = MPI3MR_DEFAULT_SDEV_QD; in mpi3mr_change_queue_depth()
1222 tg->fw_qd = tgtdev->q_depth; in mpi3mr_update_tgtdev()
1223 tg->modified_qd = tgtdev->q_depth; in mpi3mr_update_tgtdev()
1805 u16 *q_depth = (u16 *)data; in mpi3mr_update_sdev_qd() local
1807 scsi_change_queue_depth(sdev, (int)*q_depth); in mpi3mr_update_sdev_qd()
[all …]
/linux-6.3-rc2/drivers/crypto/hisilicon/
A Ddebugfs.c257 u32 *e_id, u32 *q_id, u16 q_depth) in q_dump_param_parse() argument
283 if (ret || *e_id >= q_depth) { in q_dump_param_parse()
284 dev_err(dev, "Please input sqe num (0-%u)", q_depth - 1); in q_dump_param_parse()

Completed in 74 milliseconds

12