| /drivers/md/ |
| A D | dm-rq.c | 544 if (!md->tag_set) in dm_mq_init_request_queue() 547 md->tag_set->ops = &dm_mq_ops; in dm_mq_init_request_queue() 550 md->tag_set->flags = BLK_MQ_F_STACKING; in dm_mq_init_request_queue() 552 md->tag_set->driver_data = md; in dm_mq_init_request_queue() 572 blk_mq_free_tag_set(md->tag_set); in dm_mq_init_request_queue() 574 kfree(md->tag_set); in dm_mq_init_request_queue() 575 md->tag_set = NULL; in dm_mq_init_request_queue() 582 if (md->tag_set) { in dm_mq_cleanup_mapped_device() 583 blk_mq_free_tag_set(md->tag_set); in dm_mq_cleanup_mapped_device() 584 kfree(md->tag_set); in dm_mq_cleanup_mapped_device() [all …]
|
| /drivers/s390/block/ |
| A D | dasd_genhd.c | 55 block->tag_set.ops = &dasd_mq_ops; in dasd_gendisk_alloc() 56 block->tag_set.cmd_size = sizeof(struct dasd_ccw_req); in dasd_gendisk_alloc() 57 block->tag_set.nr_hw_queues = nr_hw_queues; in dasd_gendisk_alloc() 58 block->tag_set.queue_depth = queue_depth; in dasd_gendisk_alloc() 59 block->tag_set.numa_node = NUMA_NO_NODE; in dasd_gendisk_alloc() 60 rc = blk_mq_alloc_tag_set(&block->tag_set); in dasd_gendisk_alloc() 64 gdp = blk_mq_alloc_disk(&block->tag_set, &lim, block); in dasd_gendisk_alloc() 66 blk_mq_free_tag_set(&block->tag_set); in dasd_gendisk_alloc() 124 blk_mq_free_tag_set(&block->tag_set); in dasd_gendisk_free()
|
| A D | scm_blk.c | 460 bdev->tag_set.ops = &scm_mq_ops; in scm_blk_dev_setup() 461 bdev->tag_set.cmd_size = sizeof(blk_status_t); in scm_blk_dev_setup() 462 bdev->tag_set.nr_hw_queues = nr_requests; in scm_blk_dev_setup() 463 bdev->tag_set.queue_depth = nr_requests_per_io * nr_requests; in scm_blk_dev_setup() 464 bdev->tag_set.numa_node = NUMA_NO_NODE; in scm_blk_dev_setup() 466 ret = blk_mq_alloc_tag_set(&bdev->tag_set); in scm_blk_dev_setup() 470 bdev->gendisk = blk_mq_alloc_disk(&bdev->tag_set, &lim, scmdev); in scm_blk_dev_setup() 502 blk_mq_free_tag_set(&bdev->tag_set); in scm_blk_dev_setup() 512 blk_mq_free_tag_set(&bdev->tag_set); in scm_blk_dev_cleanup()
|
| A D | scm_blk.h | 20 struct blk_mq_tag_set tag_set; member
|
| /drivers/block/ |
| A D | z2ram.c | 310 static struct blk_mq_tag_set tag_set; variable 321 disk = blk_mq_alloc_disk(&tag_set, NULL, NULL); in z2ram_register_disk() 352 tag_set.ops = &z2_mq_ops; in z2_init() 353 tag_set.nr_hw_queues = 1; in z2_init() 354 tag_set.nr_maps = 1; in z2_init() 355 tag_set.queue_depth = 16; in z2_init() 356 tag_set.numa_node = NUMA_NO_NODE; in z2_init() 357 ret = blk_mq_alloc_tag_set(&tag_set); in z2_init() 370 blk_mq_free_tag_set(&tag_set); in z2_init() 386 blk_mq_free_tag_set(&tag_set); in z2_exit()
|
| A D | nbd.c | 117 struct blk_mq_tag_set tag_set; member 269 blk_mq_free_tag_set(&nbd->tag_set); in nbd_dev_remove() 513 if (!nbd->tag_set.timeout) { in nbd_xmit_timeout() 1332 if (nbd->tag_set.timeout) in nbd_reconnect_socket() 1452 nbd->tag_set.timeout = 0; in nbd_config_put() 1518 if (nbd->tag_set.timeout) in nbd_start_device() 1520 nbd->tag_set.timeout; in nbd_start_device() 1912 nbd->tag_set.ops = &nbd_mq_ops; in nbd_dev_add() 1913 nbd->tag_set.nr_hw_queues = 1; in nbd_dev_add() 1914 nbd->tag_set.queue_depth = 128; in nbd_dev_add() [all …]
|
| A D | virtio_blk.c | 71 struct blk_mq_tag_set tag_set; member 1478 memset(&vblk->tag_set, 0, sizeof(vblk->tag_set)); in virtblk_probe() 1479 vblk->tag_set.ops = &virtio_mq_ops; in virtblk_probe() 1480 vblk->tag_set.queue_depth = queue_depth; in virtblk_probe() 1481 vblk->tag_set.numa_node = NUMA_NO_NODE; in virtblk_probe() 1482 vblk->tag_set.cmd_size = in virtblk_probe() 1485 vblk->tag_set.driver_data = vblk; in virtblk_probe() 1487 vblk->tag_set.nr_maps = 1; in virtblk_probe() 1489 vblk->tag_set.nr_maps = 3; in virtblk_probe() 1544 blk_mq_free_tag_set(&vblk->tag_set); in virtblk_probe() [all …]
|
| A D | ataflop.c | 305 struct blk_mq_tag_set tag_set; member 2055 blk_mq_free_tag_set(&unit[i].tag_set); in atari_floppy_cleanup() 2073 blk_mq_free_tag_set(&fs->tag_set); in atari_cleanup_floppy_disk() 2086 memset(&unit[i].tag_set, 0, sizeof(unit[i].tag_set)); in atari_floppy_init() 2087 unit[i].tag_set.ops = &ataflop_mq_ops; in atari_floppy_init() 2088 unit[i].tag_set.nr_hw_queues = 1; in atari_floppy_init() 2089 unit[i].tag_set.nr_maps = 1; in atari_floppy_init() 2090 unit[i].tag_set.queue_depth = 2; in atari_floppy_init() 2091 unit[i].tag_set.numa_node = NUMA_NO_NODE; in atari_floppy_init() 2092 ret = blk_mq_alloc_tag_set(&unit[i].tag_set); in atari_floppy_init() [all …]
|
| A D | zloop.c | 106 struct blk_mq_tag_set tag_set; member 703 blk_mq_free_tag_set(&zlo->tag_set); in zloop_free_disk() 980 zlo->tag_set.ops = &zloop_mq_ops; in zloop_ctl_add() 981 zlo->tag_set.nr_hw_queues = opts->nr_queues; in zloop_ctl_add() 982 zlo->tag_set.queue_depth = opts->queue_depth; in zloop_ctl_add() 983 zlo->tag_set.numa_node = NUMA_NO_NODE; in zloop_ctl_add() 984 zlo->tag_set.cmd_size = sizeof(struct zloop_cmd); in zloop_ctl_add() 985 zlo->tag_set.driver_data = zlo; in zloop_ctl_add() 987 ret = blk_mq_alloc_tag_set(&zlo->tag_set); in zloop_ctl_add() 993 zlo->disk = blk_mq_alloc_disk(&zlo->tag_set, &lim, zlo); in zloop_ctl_add() [all …]
|
| A D | ps3disk.c | 32 struct blk_mq_tag_set tag_set; member 437 error = blk_mq_alloc_sq_tag_set(&priv->tag_set, &ps3disk_mq_ops, 1, 0); in ps3disk_probe() 441 gendisk = blk_mq_alloc_disk(&priv->tag_set, &lim, dev); in ps3disk_probe() 474 blk_mq_free_tag_set(&priv->tag_set); in ps3disk_probe() 500 blk_mq_free_tag_set(&priv->tag_set); in ps3disk_remove()
|
| A D | xen-blkfront.c | 226 struct blk_mq_tag_set tag_set; member 1120 memset(&info->tag_set, 0, sizeof(info->tag_set)); in xlvbd_alloc_gendisk() 1121 info->tag_set.ops = &blkfront_mq_ops; in xlvbd_alloc_gendisk() 1122 info->tag_set.nr_hw_queues = info->nr_rings; in xlvbd_alloc_gendisk() 1132 info->tag_set.queue_depth = BLK_RING_SIZE(info); in xlvbd_alloc_gendisk() 1133 info->tag_set.numa_node = NUMA_NO_NODE; in xlvbd_alloc_gendisk() 1135 info->tag_set.driver_data = info; in xlvbd_alloc_gendisk() 1137 err = blk_mq_alloc_tag_set(&info->tag_set); in xlvbd_alloc_gendisk() 1177 blk_mq_free_tag_set(&info->tag_set); in xlvbd_alloc_gendisk() 2402 blk_mq_free_tag_set(&info->tag_set); in blkfront_connect() [all …]
|
| A D | amiflop.c | 205 struct blk_mq_tag_set tag_set; member 1786 disk = blk_mq_alloc_disk(&unit[drive].tag_set, &lim, NULL); in fd_alloc_disk() 1816 memset(&unit[drive].tag_set, 0, sizeof(unit[drive].tag_set)); in fd_alloc_drive() 1817 unit[drive].tag_set.ops = &amiflop_mq_ops; in fd_alloc_drive() 1818 unit[drive].tag_set.nr_hw_queues = 1; in fd_alloc_drive() 1819 unit[drive].tag_set.nr_maps = 1; in fd_alloc_drive() 1820 unit[drive].tag_set.queue_depth = 2; in fd_alloc_drive() 1821 unit[drive].tag_set.numa_node = NUMA_NO_NODE; in fd_alloc_drive() 1822 if (blk_mq_alloc_tag_set(&unit[drive].tag_set)) in fd_alloc_drive() 1832 blk_mq_free_tag_set(&unit[drive].tag_set); in fd_alloc_drive()
|
| A D | loop.c | 73 struct blk_mq_tag_set tag_set; member 2022 lo->tag_set.ops = &loop_mq_ops; in loop_add() 2023 lo->tag_set.nr_hw_queues = 1; in loop_add() 2024 lo->tag_set.queue_depth = hw_queue_depth; in loop_add() 2025 lo->tag_set.numa_node = NUMA_NO_NODE; in loop_add() 2026 lo->tag_set.cmd_size = sizeof(struct loop_cmd); in loop_add() 2028 lo->tag_set.driver_data = lo; in loop_add() 2030 err = blk_mq_alloc_tag_set(&lo->tag_set); in loop_add() 2034 disk = lo->lo_disk = blk_mq_alloc_disk(&lo->tag_set, &lim, lo); in loop_add() 2091 blk_mq_free_tag_set(&lo->tag_set); in loop_add() [all …]
|
| A D | swim.c | 191 struct blk_mq_tag_set tag_set; member 785 blk_mq_free_tag_set(&fs->tag_set); in swim_cleanup_floppy_disk() 820 err = blk_mq_alloc_sq_tag_set(&swd->unit[drive].tag_set, in swim_floppy_init() 826 blk_mq_alloc_disk(&swd->unit[drive].tag_set, &lim, in swim_floppy_init() 829 blk_mq_free_tag_set(&swd->unit[drive].tag_set); in swim_floppy_init()
|
| A D | ublk_drv.c | 218 struct blk_mq_tag_set tag_set; member 1891 blk_mq_tagset_busy_iter(&ub->tag_set, in ublk_wait_tagset_rqs_idle() 2702 blk_mq_free_tag_set(&ub->tag_set); in ublk_cdev_rel() 2750 ub->tag_set.ops = &ublk_mq_ops; in ublk_add_tag_set() 2751 ub->tag_set.nr_hw_queues = ub->dev_info.nr_hw_queues; in ublk_add_tag_set() 2752 ub->tag_set.queue_depth = ub->dev_info.queue_depth; in ublk_add_tag_set() 2753 ub->tag_set.numa_node = NUMA_NO_NODE; in ublk_add_tag_set() 2754 ub->tag_set.driver_data = ub; in ublk_add_tag_set() 2755 return blk_mq_alloc_tag_set(&ub->tag_set); in ublk_add_tag_set() 2865 disk = blk_mq_alloc_disk(&ub->tag_set, &lim, NULL); in ublk_ctrl_start_dev() [all …]
|
| /drivers/mmc/core/ |
| A D | queue.c | 432 memset(&mq->tag_set, 0, sizeof(mq->tag_set)); in mmc_init_queue() 433 mq->tag_set.ops = &mmc_mq_ops; in mmc_init_queue() 439 mq->tag_set.queue_depth = in mmc_init_queue() 442 mq->tag_set.queue_depth = MMC_QUEUE_DEPTH; in mmc_init_queue() 443 mq->tag_set.numa_node = NUMA_NO_NODE; in mmc_init_queue() 444 mq->tag_set.flags = BLK_MQ_F_BLOCKING; in mmc_init_queue() 445 mq->tag_set.nr_hw_queues = 1; in mmc_init_queue() 447 mq->tag_set.driver_data = mq; in mmc_init_queue() 461 ret = blk_mq_alloc_tag_set(&mq->tag_set); in mmc_init_queue() 468 blk_mq_free_tag_set(&mq->tag_set); in mmc_init_queue() [all …]
|
| /drivers/mtd/ubi/ |
| A D | block.c | 85 struct blk_mq_tag_set tag_set; member 383 dev->tag_set.ops = &ubiblock_mq_ops; in ubiblock_create() 384 dev->tag_set.queue_depth = 64; in ubiblock_create() 385 dev->tag_set.numa_node = NUMA_NO_NODE; in ubiblock_create() 386 dev->tag_set.flags = BLK_MQ_F_BLOCKING; in ubiblock_create() 388 dev->tag_set.driver_data = dev; in ubiblock_create() 389 dev->tag_set.nr_hw_queues = 1; in ubiblock_create() 391 ret = blk_mq_alloc_tag_set(&dev->tag_set); in ubiblock_create() 400 gd = blk_mq_alloc_disk(&dev->tag_set, &lim, dev); in ubiblock_create() 442 blk_mq_free_tag_set(&dev->tag_set); in ubiblock_create() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| A D | definer.c | 557 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler() 562 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler() 567 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler() 572 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler() 577 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler() 582 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler() 1678 if (fc[i].tag_set) in hws_definer_get_fc_size() 1700 if (!fc[i].tag_set) in hws_definer_alloc_compressed_fc() 1720 if (!fc[i].tag_set) in hws_definer_set_hl() 2116 fc->tag_set(fc, match_param, tag); in hws_definer_create_tag_mask() [all …]
|
| /drivers/block/null_blk/ |
| A D | main.c | 72 static struct blk_mq_tag_set tag_set; variable 407 set = dev->nullb->tag_set; in nullb_update_nr_hw_queues() 1836 if (tag_set.ops) in null_init_global_tag_set() 1841 tag_set.numa_node = g_home_node; in null_init_global_tag_set() 1847 tag_set.flags |= BLK_MQ_F_BLOCKING; in null_init_global_tag_set() 1851 tag_set.ops = NULL; in null_init_global_tag_set() 1858 nullb->tag_set = &tag_set; in null_setup_tagset() 1862 nullb->tag_set = &nullb->__tag_set; in null_setup_tagset() 1863 nullb->tag_set->driver_data = nullb; in null_setup_tagset() 2202 if (tag_set.ops) in null_exit() [all …]
|
| /drivers/mtd/ |
| A D | mtd_blkdevs.c | 33 blk_mq_free_tag_set(dev->tag_set); in blktrans_dev_release() 34 kfree(dev->tag_set); in blktrans_dev_release() 327 new->tag_set = kzalloc(sizeof(*new->tag_set), GFP_KERNEL); in add_mtd_blktrans_dev() 328 if (!new->tag_set) in add_mtd_blktrans_dev() 331 ret = blk_mq_alloc_sq_tag_set(new->tag_set, &mtd_mq_ops, 2, in add_mtd_blktrans_dev() 343 gd = blk_mq_alloc_disk(new->tag_set, &lim, new); in add_mtd_blktrans_dev() 396 blk_mq_free_tag_set(new->tag_set); in add_mtd_blktrans_dev() 398 kfree(new->tag_set); in add_mtd_blktrans_dev()
|
| /drivers/block/rnbd/ |
| A D | rnbd-clt.c | 737 if (sess->tag_set.tags) in destroy_mq_tags() 738 blk_mq_free_tag_set(&sess->tag_set); in destroy_mq_tags() 1206 struct blk_mq_tag_set *tag_set = &sess->tag_set; in setup_mq_tags() local 1208 memset(tag_set, 0, sizeof(*tag_set)); in setup_mq_tags() 1209 tag_set->ops = &rnbd_mq_ops; in setup_mq_tags() 1210 tag_set->queue_depth = sess->queue_depth; in setup_mq_tags() 1211 tag_set->numa_node = NUMA_NO_NODE; in setup_mq_tags() 1212 tag_set->flags = BLK_MQ_F_TAG_QUEUE_SHARED; in setup_mq_tags() 1222 tag_set->driver_data = sess; in setup_mq_tags() 1224 return blk_mq_alloc_tag_set(tag_set); in setup_mq_tags() [all …]
|
| /drivers/scsi/ |
| A D | scsi_lib.c | 2070 struct blk_mq_tag_set *tag_set = &shost->tag_set; in scsi_mq_setup_tags() local 2079 memset(tag_set, 0, sizeof(*tag_set)); in scsi_mq_setup_tags() 2081 tag_set->ops = &scsi_mq_ops; in scsi_mq_setup_tags() 2083 tag_set->ops = &scsi_mq_ops_no_commit; in scsi_mq_setup_tags() 2085 tag_set->nr_maps = shost->nr_maps ? : 1; in scsi_mq_setup_tags() 2087 tag_set->cmd_size = cmd_size; in scsi_mq_setup_tags() 2090 tag_set->flags |= BLK_MQ_F_TAG_RR; in scsi_mq_setup_tags() 2092 tag_set->flags |= BLK_MQ_F_BLOCKING; in scsi_mq_setup_tags() 2093 tag_set->driver_data = shost; in scsi_mq_setup_tags() 2097 return blk_mq_alloc_tag_set(tag_set); in scsi_mq_setup_tags() [all …]
|
| A D | hosts.c | 614 blk_mq_tagset_busy_iter(&shost->tag_set, in scsi_host_busy() 715 blk_mq_tagset_busy_iter(&shost->tag_set, complete_all_cmds_iter, in scsi_host_complete_all_commands() 751 blk_mq_tagset_busy_iter(&shost->tag_set, __scsi_host_busy_iter_fn, in scsi_host_busy_iter()
|
| /drivers/cdrom/ |
| A D | gdrom.c | 104 struct blk_mq_tag_set tag_set; member 779 err = blk_mq_alloc_sq_tag_set(&gd.tag_set, &gdrom_mq_ops, 1, in probe_gdrom() 784 gd.disk = blk_mq_alloc_disk(&gd.tag_set, &lim, NULL); in probe_gdrom() 825 blk_mq_free_tag_set(&gd.tag_set); in probe_gdrom() 837 blk_mq_free_tag_set(&gd.tag_set); in remove_gdrom()
|
| /drivers/nvme/target/ |
| A D | loop.c | 35 struct blk_mq_tag_set tag_set; member 90 return queue->ctrl->tag_set.tags[queue_idx - 1]; in nvme_loop_tagset() 213 (set == &ctrl->tag_set) ? hctx_idx + 1 : 0); in nvme_loop_init_request() 482 blk_mq_update_nr_hw_queues(&ctrl->tag_set, in nvme_loop_reset_ctrl_work() 524 ret = nvme_alloc_io_tag_set(&ctrl->ctrl, &ctrl->tag_set, in nvme_loop_create_io_queues()
|