Lines Matching refs:disk

22 		if (h->disk)  in nvme_mpath_unfreeze()
23 blk_mq_unfreeze_queue(h->disk->queue); in nvme_mpath_unfreeze()
32 if (h->disk) in nvme_mpath_wait_freeze()
33 blk_mq_freeze_queue_wait(h->disk->queue); in nvme_mpath_wait_freeze()
42 if (h->disk) in nvme_mpath_start_freeze()
43 blk_freeze_queue_start(h->disk->queue); in nvme_mpath_start_freeze()
57 if (!ns->head->disk) { in nvme_mpath_set_disk_name()
89 bio_set_dev(bio, ns->head->disk->part0); in nvme_failover_req()
108 if (!ns->head->disk) in nvme_kick_requeue_lists()
112 disk_uevent(ns->head->disk, KOBJ_CHANGE); in nvme_kick_requeue_lists()
160 sector_t capacity = get_capacity(head->disk); in nvme_mpath_revalidate_paths()
164 if (capacity != get_capacity(ns->disk)) in nvme_mpath_revalidate_paths()
325 struct device *dev = disk_to_dev(head->disk); in nvme_ns_head_submit_bio()
339 bio_set_dev(bio, ns->disk->part0); in nvme_ns_head_submit_bio()
341 trace_block_bio_remap(bio, disk_devt(ns->head->disk), in nvme_ns_head_submit_bio()
367 static void nvme_ns_head_release(struct gendisk *disk, fmode_t mode) in nvme_ns_head_release() argument
369 nvme_put_ns_head(disk->private_data); in nvme_ns_head_release()
373 static int nvme_ns_head_report_zones(struct gendisk *disk, sector_t sector, in nvme_ns_head_report_zones() argument
376 struct nvme_ns_head *head = disk->private_data; in nvme_ns_head_report_zones()
477 head->disk = blk_alloc_disk(ctrl->numa_node); in nvme_mpath_alloc_disk()
478 if (!head->disk) in nvme_mpath_alloc_disk()
480 head->disk->fops = &nvme_ns_head_ops; in nvme_mpath_alloc_disk()
481 head->disk->private_data = head; in nvme_mpath_alloc_disk()
482 sprintf(head->disk->disk_name, "nvme%dn%d", in nvme_mpath_alloc_disk()
485 blk_queue_flag_set(QUEUE_FLAG_NONROT, head->disk->queue); in nvme_mpath_alloc_disk()
486 blk_queue_flag_set(QUEUE_FLAG_NOWAIT, head->disk->queue); in nvme_mpath_alloc_disk()
495 blk_queue_flag_set(QUEUE_FLAG_POLL, head->disk->queue); in nvme_mpath_alloc_disk()
498 blk_queue_logical_block_size(head->disk->queue, 512); in nvme_mpath_alloc_disk()
499 blk_set_stacking_limits(&head->disk->queue->limits); in nvme_mpath_alloc_disk()
504 blk_queue_write_cache(head->disk->queue, vwc, vwc); in nvme_mpath_alloc_disk()
513 if (!head->disk) in nvme_mpath_set_live()
522 rc = device_add_disk(&head->subsys->dev, head->disk, in nvme_mpath_set_live()
796 if (blk_queue_stable_writes(ns->queue) && ns->head->disk) in nvme_mpath_add_disk()
798 ns->head->disk->queue); in nvme_mpath_add_disk()
800 if (blk_queue_is_zoned(ns->queue) && ns->head->disk) in nvme_mpath_add_disk()
801 ns->head->disk->queue->nr_zones = ns->queue->nr_zones; in nvme_mpath_add_disk()
807 if (!head->disk) in nvme_mpath_shutdown_disk()
812 del_gendisk(head->disk); in nvme_mpath_shutdown_disk()
818 if (!head->disk) in nvme_mpath_remove_disk()
820 blk_set_queue_dying(head->disk->queue); in nvme_mpath_remove_disk()
824 blk_cleanup_disk(head->disk); in nvme_mpath_remove_disk()