| /linux/drivers/target/ |
| A D | target_core_tpg.c | 157 struct se_node_acl *acl, u32 queue_depth) in target_set_nacl_queue_depth() argument 159 acl->queue_depth = queue_depth; in target_set_nacl_queue_depth() 161 if (!acl->queue_depth) { in target_set_nacl_queue_depth() 165 acl->queue_depth = 1; in target_set_nacl_queue_depth() 173 u32 queue_depth; in target_alloc_node_acl() local 192 queue_depth = 1; in target_alloc_node_acl() 193 target_set_nacl_queue_depth(tpg, acl, queue_depth); in target_alloc_node_acl() 217 acl->queue_depth, in target_add_node_acl() 380 u32 queue_depth) in core_tpg_set_initiator_node_queue_depth() argument 389 if (acl->queue_depth == queue_depth) in core_tpg_set_initiator_node_queue_depth() [all …]
|
| /linux/arch/um/drivers/ |
| A D | vector_kern.c | 250 int queue_depth; in vector_advancehead() local 268 queue_depth = qi->queue_depth; in vector_advancehead() 270 return queue_depth; in vector_advancehead() 280 int queue_depth; in vector_advancetail() local 287 queue_depth = qi->queue_depth; in vector_advancetail() 289 return queue_depth; in vector_advancetail() 336 int queue_depth; in vector_enqueue() local 343 queue_depth = qi->queue_depth; in vector_enqueue() 367 return queue_depth; in vector_enqueue() 416 queue_depth = qi->queue_depth; in vector_send() [all …]
|
| A D | vector_kern.h | 46 int queue_depth, head, tail, max_depth, max_iov_frags; member
|
| /linux/drivers/ata/ |
| A D | libata-sata.c | 1034 int queue_depth) in __ata_change_queue_depth() argument 1039 if (queue_depth < 1 || queue_depth == sdev->queue_depth) in __ata_change_queue_depth() 1040 return sdev->queue_depth; in __ata_change_queue_depth() 1044 return sdev->queue_depth; in __ata_change_queue_depth() 1049 if (queue_depth == 1 || !ata_ncq_enabled(dev)) { in __ata_change_queue_depth() 1051 queue_depth = 1; in __ata_change_queue_depth() 1056 queue_depth = min(queue_depth, sdev->host->can_queue); in __ata_change_queue_depth() 1057 queue_depth = min(queue_depth, ata_id_queue_depth(dev->id)); in __ata_change_queue_depth() 1058 queue_depth = min(queue_depth, ATA_MAX_QUEUE); in __ata_change_queue_depth() 1060 if (sdev->queue_depth == queue_depth) in __ata_change_queue_depth() [all …]
|
| /linux/drivers/infiniband/ulp/rtrs/ |
| A D | rtrs-clt.c | 65 size_t max_depth = clt->queue_depth; in __rtrs_get_permit() 722 q_size = sess->queue_depth; in post_recv_sess() 1439 size_t sz = clt->queue_depth; in free_permits() 1670 if (WARN_ON(!sess->queue_depth)) in create_con_cq_qp() 1678 sess->queue_depth * 3 + 1); in create_con_cq_qp() 1805 u16 version, queue_depth; in rtrs_rdma_conn_established() local 1832 queue_depth = le16_to_cpu(msg->queue_depth); in rtrs_rdma_conn_established() 1834 if (sess->queue_depth > 0 && queue_depth != sess->queue_depth) { in rtrs_rdma_conn_established() 1852 sess->queue_depth = queue_depth; in rtrs_rdma_conn_established() 1869 clt->queue_depth = sess->queue_depth; in rtrs_rdma_conn_established() [all …]
|
| A D | rtrs-srv.c | 115 for (i = 0; i < srv->queue_depth; i++) in rtrs_srv_free_ops_ids() 147 for (i = 0; i < srv->queue_depth; ++i) { in rtrs_srv_alloc_ops_ids() 581 mrs_num = srv->queue_depth; in map_cont_bufs() 605 srv->queue_depth - chunks); in map_cont_bufs() 980 q_size = srv->queue_depth; in post_recv_sess() 1318 return srv->queue_depth; in rtrs_srv_get_queue_depth() 1352 for (i = 0; i < srv->queue_depth; i++) in free_srv() 1395 srv->queue_depth = sess_queue_depth; in get_or_create_srv() 1405 for (i = 0; i < srv->queue_depth; i++) { in get_or_create_srv() 1584 .queue_depth = cpu_to_le16(srv->queue_depth), in rtrs_rdma_do_accept() [all …]
|
| A D | rtrs-clt.h | 143 size_t queue_depth; member 172 size_t queue_depth; member
|
| A D | rtrs.h | 110 u32 queue_depth; member
|
| A D | rtrs-srv.h | 104 size_t queue_depth; member
|
| /linux/drivers/target/iscsi/ |
| A D | iscsi_target_device.c | 41 sess->cmdsn_window = se_nacl->queue_depth; in iscsit_determine_maxcmdsn() 42 atomic_add(se_nacl->queue_depth - 1, &sess->max_cmd_sn); in iscsit_determine_maxcmdsn()
|
| /linux/drivers/s390/crypto/ |
| A D | ap_card.c | 47 return scnprintf(buf, PAGE_SIZE, "%d\n", ac->queue_depth); in depth_show() 222 struct ap_card *ap_card_create(int id, int queue_depth, int raw_type, in ap_card_create() argument 234 ac->queue_depth = queue_depth; in ap_card_create()
|
| A D | ap_bus.h | 163 int queue_depth; /* AP queue depth.*/ member 289 struct ap_card *ap_card_create(int id, int queue_depth, int raw_type,
|
| /linux/block/ |
| A D | blk-rq-qos.c | 128 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth() 144 rqd->queue_depth); in rq_depth_calc_max_depth() 148 unsigned int maxd = 3 * rqd->queue_depth / 4; in rq_depth_calc_max_depth()
|
| A D | blk-mq.c | 3106 set->queue_depth, flush_rq); in blk_mq_exit_hctx() 3305 set->queue_depth); in __blk_mq_alloc_map_and_rqs() 3792 set->queue_depth); in __blk_mq_alloc_rq_maps() 3827 depth = set->queue_depth; in blk_mq_alloc_set_map_and_rqs() 3833 set->queue_depth >>= 1; in blk_mq_alloc_set_map_and_rqs() 3838 } while (set->queue_depth); in blk_mq_alloc_set_map_and_rqs() 3845 if (depth != set->queue_depth) in blk_mq_alloc_set_map_and_rqs() 3847 depth, set->queue_depth); in blk_mq_alloc_set_map_and_rqs() 3932 if (!set->queue_depth) in blk_mq_alloc_tag_set() 3962 set->queue_depth = min(64U, set->queue_depth); in blk_mq_alloc_tag_set() [all …]
|
| A D | blk-mq-tag.h | 13 unsigned int queue_depth,
|
| /linux/drivers/net/ethernet/microsoft/mana/ |
| A D | hw_channel.h | 133 u16 queue_depth; member 147 u16 queue_depth; member
|
| A D | hw_channel.c | 291 comp_read = mana_gd_poll_cq(q_self, completions, hwc_cq->queue_depth); in mana_hwc_comp_event() 292 WARN_ON_ONCE(comp_read <= 0 || comp_read > hwc_cq->queue_depth); in mana_hwc_comp_event() 371 hwc_cq->queue_depth = q_depth; in mana_hwc_create_cq() 485 hwc_wq->queue_depth = q_depth; in mana_hwc_create_wq()
|
| /linux/drivers/block/null_blk/ |
| A D | null_blk.h | 31 unsigned int queue_depth; member 117 unsigned int queue_depth; member
|
| A D | main.c | 685 tag = find_first_zero_bit(nq->tag_map, nq->queue_depth); in get_tag() 686 if (tag >= nq->queue_depth) in get_tag() 1687 nq->queue_depth = nullb->queue_depth; in null_init_queue() 1793 nq->cmds = kcalloc(nq->queue_depth, sizeof(*cmd), GFP_KERNEL); in setup_commands() 1797 tag_size = ALIGN(nq->queue_depth, BITS_PER_LONG) / BITS_PER_LONG; in setup_commands() 1804 for (i = 0; i < nq->queue_depth; i++) { in setup_commands() 1824 nullb->queue_depth = nullb->dev->hw_queue_depth; in setup_queues() 1884 set->queue_depth = nullb ? nullb->dev->hw_queue_depth : in null_init_tag_set()
|
| /linux/drivers/block/ |
| A D | virtio_blk.c | 749 module_param_named(queue_depth, virtblk_queue_depth, uint, 0444); 760 unsigned int queue_depth; in virtblk_probe() local 809 queue_depth = vblk->vqs[0].vq->num_free; in virtblk_probe() 812 queue_depth /= 2; in virtblk_probe() 814 queue_depth = virtblk_queue_depth; in virtblk_probe() 819 vblk->tag_set.queue_depth = queue_depth; in virtblk_probe()
|
| /linux/drivers/scsi/ |
| A D | scsi.c | 223 sdev->queue_depth = depth; in scsi_change_queue_depth() 230 sbitmap_resize(&sdev->budget_map, sdev->queue_depth); in scsi_change_queue_depth() 232 return sdev->queue_depth; in scsi_change_queue_depth()
|
| /linux/drivers/scsi/snic/ |
| A D | snic_main.c | 105 if (qsz < sdev->queue_depth) in snic_change_queue_depth() 107 else if (qsz > sdev->queue_depth) in snic_change_queue_depth() 110 atomic64_set(&snic->s_stats.misc.last_qsz, sdev->queue_depth); in snic_change_queue_depth() 114 return sdev->queue_depth; in snic_change_queue_depth()
|
| /linux/drivers/block/rnbd/ |
| A D | rnbd-srv.h | 25 int queue_depth; member
|
| A D | rnbd-clt.h | 86 size_t queue_depth; member
|
| /linux/include/target/ |
| A D | target_core_base.h | 567 u32 queue_depth; member 706 u32 queue_depth; member 802 u32 queue_depth; member
|