| /linux/drivers/net/ethernet/microsoft/mana/ |
| A D | hw_channel.c | 405 hwc_cq->queue_depth = q_depth; in mana_hwc_create_cq() 436 dma_buf->num_reqs = q_depth; in mana_hwc_alloc_dma_buf() 450 for (i = 0; i < q_depth; i++) { in mana_hwc_alloc_dma_buf() 490 enum gdma_queue_type q_type, u16 q_depth, in mana_hwc_create_wq() argument 519 hwc_wq->queue_depth = q_depth; in mana_hwc_create_wq() 604 for (i = 0; i < q_depth; i++) { in mana_hwc_test_channel() 615 for (i = 0; i < q_depth; ++i) in mana_hwc_test_channel() 648 *q_depth = hwc->hwc_init_q_depth_max; in mana_hwc_establish_channel() 670 err = mana_hwc_init_inflight_msg(hwc, q_depth); in mana_hwc_init_queues() 677 err = mana_hwc_create_cq(hwc, q_depth * 2, in mana_hwc_init_queues() [all …]
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| A D | hinic_hw_wq.c | 504 u16 wqebb_size, u32 wq_page_size, u16 q_depth, in hinic_wq_allocate() argument 523 if (q_depth & (q_depth - 1)) { in hinic_wq_allocate() 547 wq->q_depth = q_depth; in hinic_wq_allocate() 564 atomic_set(&wq->delta, q_depth); in hinic_wq_allocate() 565 wq->mask = q_depth - 1; in hinic_wq_allocate() 602 u16 q_depth, u16 max_wqe_size) in hinic_wqs_cmdq_alloc() argument 620 if (q_depth & (q_depth - 1)) { in hinic_wqs_cmdq_alloc() 650 wq[i].q_depth = q_depth; in hinic_wqs_cmdq_alloc() 668 atomic_set(&wq[i].delta, q_depth); in hinic_wqs_cmdq_alloc() 669 wq[i].mask = q_depth - 1; in hinic_wqs_cmdq_alloc() [all …]
|
| A D | hinic_hw_wq.h | 30 u16 q_depth; member 80 u16 q_depth, u16 max_wqe_size); 91 u16 wqebb_size, u32 wq_page_size, u16 q_depth,
|
| A D | hinic_hw_cmdq.c | 363 if (next_prod_idx >= wq->q_depth) { in cmdq_sync_cmd_direct_resp() 365 next_prod_idx -= wq->q_depth; in cmdq_sync_cmd_direct_resp() 442 if (next_prod_idx >= wq->q_depth) { in cmdq_set_arm_bit() 444 next_prod_idx -= wq->q_depth; in cmdq_set_arm_bit() 749 cmdq->done = vzalloc(array_size(sizeof(*cmdq->done), wq->q_depth)); in init_cmdq() 754 wq->q_depth)); in init_cmdq()
|
| A D | hinic_hw_qp.c | 224 skb_arr_size = wq->q_depth * sizeof(*sq->saved_skb); in alloc_sq_skb_arr() 252 skb_arr_size = wq->q_depth * sizeof(*rq->saved_skb); in alloc_rq_skb_arr() 324 cqe_size = wq->q_depth * sizeof(*rq->cqe); in alloc_rq_cqe() 329 cqe_dma_size = wq->q_depth * sizeof(*rq->cqe_dma); in alloc_rq_cqe() 334 for (i = 0; i < wq->q_depth; i++) { in alloc_rq_cqe() 367 for (i = 0; i < wq->q_depth; i++) in free_rq_cqe()
|
| /linux/drivers/net/ethernet/amazon/ena/ |
| A D | ena_eth_com.c | 15 head_masked = io_cq->head & (io_cq->q_depth - 1); in ena_com_get_next_rx_cdesc() 40 tail_masked = io_sq->tail & (io_sq->q_depth - 1); in get_sq_desc_regular_queue() 55 dst_tail_mask = io_sq->tail & (io_sq->q_depth - 1); in ena_com_write_bounce_buffer_to_dev() 83 if (unlikely((io_sq->tail & (io_sq->q_depth - 1)) == 0)) in ena_com_write_bounce_buffer_to_dev() 217 if (unlikely((io_sq->tail & (io_sq->q_depth - 1)) == 0)) in ena_com_sq_update_tail() 226 idx &= (io_cq->q_depth - 1); in ena_com_rx_cdesc_idx_to_ptr() 266 head_masked = io_cq->head & (io_cq->q_depth - 1); in ena_com_cdesc_rx_pkt_get() 551 u16 q_depth = io_cq->q_depth; in ena_com_rx_pkt() local 583 if (unlikely(ena_buf[i].req_id >= q_depth)) in ena_com_rx_pkt()
|
| A D | ena_eth_com.h | 82 return io_sq->q_depth - 1 - cnt; in ena_com_free_q_entries() 196 if (unlikely((io_cq->head & (io_cq->q_depth - 1)) == 0)) in ena_com_cq_inc_head() 207 masked_head = io_cq->head & (io_cq->q_depth - 1); in ena_com_tx_comp_req_id_get() 225 if (unlikely(*req_id >= io_cq->q_depth)) { in ena_com_tx_comp_req_id_get()
|
| A D | ena_com.c | 144 aenq->head = aenq->q_depth; in ena_com_admin_init_aenq() 183 command_id, admin_queue->q_depth); in get_comp_ctxt() 216 queue_size_mask = admin_queue->q_depth - 1; in __ena_com_submit_admin_cmd() 222 if (cnt >= admin_queue->q_depth) { in __ena_com_submit_admin_cmd() 276 for (i = 0; i < admin_queue->q_depth; i++) { in ena_com_init_comp_ctxt() 1209 create_cmd.sq_depth = io_sq->q_depth; in ena_com_create_io_sq() 1353 create_cmd.cq_depth = io_cq->q_depth; in ena_com_create_io_cq() 1481 u16 depth = ena_dev->aenq.q_depth; in ena_com_admin_aenq_enable() 1636 size = ADMIN_AENQ_SIZE(aenq->q_depth); in ena_com_admin_destroy() 1818 io_cq->q_depth = ctx->queue_size; in ena_com_create_io_queue() [all …]
|
| A D | ena_com.h | 129 u16 q_depth; member 171 u16 q_depth; member 220 u16 q_depth; member 250 u16 q_depth; member
|
| /linux/drivers/net/ethernet/brocade/bna/ |
| A D | bnad.c | 91 u32 q_depth, u32 index) in bnad_tx_buff_unmap() argument 114 BNA_QE_INDX_INC(index, q_depth); in bnad_tx_buff_unmap() 126 BNA_QE_INDX_INC(index, q_depth); in bnad_tx_buff_unmap() 174 q_depth = tcb->q_depth; in bnad_txcmpl_process() 341 u32 alloced, prod, q_depth; in bnad_rxq_refill_page() local 350 q_depth = rcb->q_depth; in bnad_rxq_refill_page() 397 BNA_QE_INDX_INC(prod, q_depth); in bnad_rxq_refill_page() 424 q_depth = rcb->q_depth; in bnad_rxq_refill_skb() 453 BNA_QE_INDX_INC(prod, q_depth); in bnad_rxq_refill_skb() 2915 u32 prod, q_depth, vect_id; in bnad_start_xmit() local [all …]
|
| A D | bna_types.h | 428 u32 q_depth; member 559 u32 q_depth; member 575 int q_depth; member 623 u32 q_depth; member
|
| A D | bfa_msgq.c | 516 msgq_cfg->cmdq.q_depth = htons(msgq->cmdq.depth); in bfa_msgq_init() 518 msgq_cfg->rspq.q_depth = htons(msgq->rspq.depth); in bfa_msgq_init()
|
| A D | bfi.h | 413 u16 q_depth; /* Total num of entries in the queue */ member
|
| /linux/drivers/nvme/host/ |
| A D | pci.c | 130 u32 q_depth; member 199 u32 q_depth; member 1516 int q_depth = dev->q_depth; in nvme_cmb_qdepth() local 1531 if (q_depth < 64) in nvme_cmb_qdepth() 1535 return q_depth; in nvme_cmb_qdepth() 1572 nvmeq->q_depth = depth; in nvme_alloc_queue() 1796 aqa = nvmeq->q_depth - 1; in nvme_pci_configure_admin_queue() 2345 dev->q_depth = result; in nvme_setup_io_queues() 2579 dev->q_depth = 2; in nvme_pci_enable() 2583 dev->q_depth = 64; in nvme_pci_enable() [all …]
|
| /linux/drivers/net/ethernet/marvell/octeon_ep/ |
| A D | octep_ctrl_mbox.c | 228 u32 pi, ci, r_sz, buf_sz, q_depth; in octep_ctrl_mbox_recv() local 240 q_depth = octep_ctrl_mbox_circq_depth(pi, ci, q->sz); in octep_ctrl_mbox_recv() 241 if (q_depth < mbox_hdr_sz) { in octep_ctrl_mbox_recv()
|
| /linux/drivers/crypto/hisilicon/sec2/ |
| A D | sec_crypto.c | 313 u16 q_depth = res->depth; in sec_alloc_civ_resource() local 321 for (i = 1; i < q_depth; i++) { in sec_alloc_civ_resource() 338 u16 q_depth = res->depth; in sec_alloc_aiv_resource() local 346 for (i = 1; i < q_depth; i++) { in sec_alloc_aiv_resource() 363 u16 q_depth = res->depth; in sec_alloc_mac_resource() local 371 for (i = 1; i < q_depth; i++) { in sec_alloc_mac_resource() 400 u16 q_depth = res->depth; in sec_alloc_pbuf_resource() local 401 int size = SEC_PBUF_PAGE_NUM(q_depth); in sec_alloc_pbuf_resource() 422 if (k == q_depth) in sec_alloc_pbuf_resource() 492 u16 q_depth = qp_ctx->qp->sq_depth; in sec_alloc_qp_ctx_resource() local [all …]
|
| /linux/drivers/scsi/bfa/ |
| A D | bfi.h | 538 #define BFI_MSGQ_FULL(_q) (((_q->pi + 1) % _q->q_depth) == _q->ci) 540 #define BFI_MSGQ_UPDATE_CI(_q) (_q->ci = (_q->ci + 1) % _q->q_depth) 541 #define BFI_MSGQ_UPDATE_PI(_q) (_q->pi = (_q->pi + 1) % _q->q_depth) 544 #define BFI_MSGQ_FREE_CNT(_q) ((_q->ci - _q->pi - 1) & (_q->q_depth - 1)) 585 u16 q_depth; /* Total num of entries in the queue */ member
|
| /linux/drivers/crypto/hisilicon/zip/ |
| A D | zip_crypto.c | 446 u16 q_depth = ctx->qp_ctx[0].qp->sq_depth; in hisi_zip_create_req_q() local 452 req_q->size = q_depth; in hisi_zip_create_req_q() 498 u16 q_depth = ctx->qp_ctx[0].qp->sq_depth; in hisi_zip_create_sgl_pool() local 506 tmp->sgl_pool = hisi_acc_create_sgl_pool(dev, q_depth << 1, in hisi_zip_create_sgl_pool()
|
| /linux/drivers/block/ |
| A D | ublk_drv.c | 133 int q_depth; member 1227 for (i = 0; i < ubq->q_depth; i++) { in ublk_timeout() 1235 if (nr_inflight == ubq->q_depth) { in ublk_timeout() 1391 for (i = 0; i < ubq->q_depth; i++) { in ublk_abort_queue() 1477 if (WARN_ON_ONCE(pdu->tag >= ubq->q_depth)) in ublk_uring_cmd_cancel_fn() 1501 return ubq->nr_io_ready == ubq->q_depth; in ublk_queue_ready() 1508 for (i = 0; i < ubq->q_depth; i++) in ublk_cancel_queue() 1709 if (tag >= ubq->q_depth) in __ublk_ch_uring_cmd() 1926 if (tag >= ubq->q_depth) in ublk_check_and_get_req() 2013 ubq->q_depth = ub->dev_info.queue_depth; in ublk_init_queue() [all …]
|
| /linux/drivers/net/ethernet/fungible/funcore/ |
| A D | fun_dev.h | 69 unsigned int q_depth; /* max queue depth supported by device */ member
|
| A D | fun_queue.c | 89 if (sq_depth > fdev->q_depth) in fun_sq_create() 138 if (cq_depth > fdev->q_depth) in fun_cq_create()
|
| A D | fun_dev.c | 783 fdev->q_depth = NVME_CAP_MQES(fdev->cap_reg) + 1; in fun_dev_enable() 813 fdev->q_depth, fdev->db_stride, fdev->max_qid, in fun_dev_enable()
|
| /linux/net/mac80211/ |
| A D | debugfs_sta.c | 258 u32 q_depth[IEEE80211_NUM_ACS]; in sta_aql_read() local 271 q_depth[ac] = atomic_read(&sta->airtime[ac].aql_tx_pending); in sta_aql_read() 277 q_depth[0], q_depth[1], q_depth[2], q_depth[3], in sta_aql_read()
|
| /linux/drivers/scsi/mpi3mr/ |
| A D | mpi3mr_os.c | 1007 int q_depth) in mpi3mr_change_queue_depth() argument 1014 q_depth = 1; in mpi3mr_change_queue_depth() 1015 if (q_depth > shost->can_queue) in mpi3mr_change_queue_depth() 1016 q_depth = shost->can_queue; in mpi3mr_change_queue_depth() 1017 else if (!q_depth) in mpi3mr_change_queue_depth() 1018 q_depth = MPI3MR_DEFAULT_SDEV_QD; in mpi3mr_change_queue_depth() 1313 tg->fw_qd = tgtdev->q_depth; in mpi3mr_update_tgtdev() 1314 tg->modified_qd = tgtdev->q_depth; in mpi3mr_update_tgtdev() 1952 u16 *q_depth = (u16 *)data; in mpi3mr_update_sdev_qd() local 1954 scsi_change_queue_depth(sdev, (int)*q_depth); in mpi3mr_update_sdev_qd() [all …]
|
| /linux/drivers/crypto/hisilicon/ |
| A D | debugfs.c | 281 u32 *e_id, u32 *q_id, u16 q_depth) in q_dump_param_parse() argument 307 if (ret || *e_id >= q_depth) { in q_dump_param_parse() 308 dev_err(dev, "Please input sqe num (0-%u)", q_depth - 1); in q_dump_param_parse()
|