| /block/ |
| A D | sed-opal.c | 458 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_send_cmd() 465 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_recv_cmd() 574 const u8 *epos = dev->resp, *cpos = dev->resp; in opal_discovery0_end() 1174 print_buffer(dev->cmd, dev->pos); in parse_and_check_status() 1196 set_comid(dev, dev->comid); in cmd_start() 1262 ret = cmd_finalize(dev, dev->hsn, dev->tsn); in finalize_and_send() 1268 print_buffer(dev->cmd, dev->pos); in finalize_and_send() 2219 set_comid(dev, dev->comid); in lock_unlock_locking_range_sum() 2447 set_comid(dev, dev->comid); in end_opal_session() 2501 if (!dev) in free_opal_dev() [all …]
|
| A D | early-lookup.c | 49 struct device *dev = NULL; in devt_from_partuuid() local 72 if (!dev) in devt_from_partuuid() 83 *devt = dev->devt; in devt_from_partuuid() 86 put_device(dev); in devt_from_partuuid() 114 struct device *dev; in devt_from_partlabel() local 117 if (!dev) in devt_from_partlabel() 119 *devt = dev->devt; in devt_from_partlabel() 120 put_device(dev); in devt_from_partlabel() 128 struct device *dev; in blk_lookup_devt() local 275 struct device *dev; in printk_all_partitions() local [all …]
|
| A D | blk-pm.c | 31 q->dev = dev; in blk_pm_runtime_init() 33 pm_runtime_set_autosuspend_delay(q->dev, -1); in blk_pm_runtime_init() 34 pm_runtime_use_autosuspend(q->dev); in blk_pm_runtime_init() 63 if (!q->dev) in blk_pre_runtime_suspend() 97 pm_runtime_mark_last_busy(q->dev); in blk_pre_runtime_suspend() 122 if (!q->dev) in blk_post_runtime_suspend() 130 pm_runtime_mark_last_busy(q->dev); in blk_post_runtime_suspend() 152 if (!q->dev) in blk_pre_runtime_resume() 178 if (!q->dev) in blk_post_runtime_resume() 184 pm_runtime_mark_last_busy(q->dev); in blk_post_runtime_resume() [all …]
|
| A D | blk-integrity.c | 247 return &dev_to_disk(dev)->queue->limits.integrity; in dev_to_bi() 276 struct request_queue *q = dev_to_disk(dev)->queue; in flag_store() 300 struct blk_integrity *bi = dev_to_bi(dev); in flag_show() 308 struct blk_integrity *bi = dev_to_bi(dev); in format_show() 318 struct blk_integrity *bi = dev_to_bi(dev); in tag_size_show() 327 struct blk_integrity *bi = dev_to_bi(dev); in protection_interval_bytes_show() 333 static ssize_t read_verify_store(struct device *dev, in read_verify_store() argument 340 static ssize_t read_verify_show(struct device *dev, in read_verify_show() argument 343 return flag_show(dev, page, BLK_INTEGRITY_NOVERIFY); in read_verify_show() 353 static ssize_t write_generate_show(struct device *dev, in write_generate_show() argument [all …]
|
| A D | genhd.c | 539 &disk->bdi->dev->kobj, "bdi"); in __add_disk() 855 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_show() 867 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_store() 915 struct device *dev; in disk_seqf_start() local 924 dev = class_dev_iter_next(iter); in disk_seqf_start() 925 if (!dev) in disk_seqf_start() 929 return dev_to_disk(dev); in disk_seqf_start() 934 struct device *dev; in disk_seqf_next() local 938 if (dev) in disk_seqf_next() 939 return dev_to_disk(dev); in disk_seqf_next() [all …]
|
| A D | bsg-lib.c | 161 put_device(job->dev); /* release reference for the request */ in bsg_teardown_job() 232 static bool bsg_prepare_job(struct device *dev, struct request *req) in bsg_prepare_job() argument 249 job->dev = dev; in bsg_prepare_job() 251 get_device(job->dev); in bsg_prepare_job() 276 struct device *dev = q->queuedata; in bsg_queue_rq() local 285 if (!get_device(dev)) in bsg_queue_rq() 288 if (!bsg_prepare_job(dev, req)) in bsg_queue_rq() 296 put_device(dev); in bsg_queue_rq() 362 struct request_queue *bsg_setup_queue(struct device *dev, const char *name, in bsg_setup_queue() argument 388 q = blk_mq_alloc_queue(set, lim, dev); in bsg_setup_queue() [all …]
|
| A D | blk-pm.h | 11 if (!q->dev || !blk_queue_pm_only(q)) in blk_pm_resume_queue() 15 pm_request_resume(q->dev); in blk_pm_resume_queue() 21 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_mark_last_busy() 22 pm_runtime_mark_last_busy(rq->q->dev); in blk_pm_mark_last_busy()
|
| A D | bdev.c | 500 bdev->bd_dev = dev; in bdev_add() 501 inode->i_rdev = dev; in bdev_add() 502 inode->i_ino = dev; in bdev_add() 787 MAJOR(dev), MINOR(dev), in bdev_permission() 823 inode = ilookup(blockdev_superblock, dev); in blkdev_get_no_open() 825 blk_request_module(dev); in blkdev_get_no_open() 1042 ret = bdev_permission(dev, mode, holder); in bdev_file_open_by_dev() 1046 bdev = blkdev_get_no_open(dev, true); in bdev_file_open_by_dev() 1075 dev_t dev; in bdev_file_open_by_path() local 1078 error = lookup_bdev(path, &dev); in bdev_file_open_by_path() [all …]
|
| A D | blk-mq-cpumap.c | 109 struct device *dev, unsigned int offset) in blk_mq_map_hw_queues() argument 115 if (!dev->bus->irq_get_affinity) in blk_mq_map_hw_queues() 119 mask = dev->bus->irq_get_affinity(dev, queue + offset); in blk_mq_map_hw_queues()
|
| A D | blk-timeout.c | 39 ssize_t part_timeout_show(struct device *dev, struct device_attribute *attr, in part_timeout_show() argument 42 struct gendisk *disk = dev_to_disk(dev); in part_timeout_show() 48 ssize_t part_timeout_store(struct device *dev, struct device_attribute *attr, in part_timeout_store() argument 51 struct gendisk *disk = dev_to_disk(dev); in part_timeout_store()
|
| A D | disk-events.c | 341 static ssize_t disk_events_show(struct device *dev, in disk_events_show() argument 344 struct gendisk *disk = dev_to_disk(dev); in disk_events_show() 351 static ssize_t disk_events_async_show(struct device *dev, in disk_events_async_show() argument 357 static ssize_t disk_events_poll_msecs_show(struct device *dev, in disk_events_poll_msecs_show() argument 361 struct gendisk *disk = dev_to_disk(dev); in disk_events_poll_msecs_show() 368 static ssize_t disk_events_poll_msecs_store(struct device *dev, in disk_events_poll_msecs_store() argument 372 struct gendisk *disk = dev_to_disk(dev); in disk_events_poll_msecs_store()
|
| A D | blk.h | 108 struct block_device *blkdev_get_no_open(dev_t dev, bool autoload); 340 ssize_t part_size_show(struct device *dev, struct device_attribute *attr, 342 ssize_t part_stat_show(struct device *dev, struct device_attribute *attr, 344 ssize_t part_inflight_show(struct device *dev, struct device_attribute *attr, 346 ssize_t part_fail_show(struct device *dev, struct device_attribute *attr, 348 ssize_t part_fail_store(struct device *dev, struct device_attribute *attr, 557 void bdev_add(struct block_device *bdev, dev_t dev); 728 int bdev_permission(dev_t dev, blk_mode_t mode, void *holder);
|
| A D | blk-crypto-profile.c | 52 if (profile->dev) in blk_crypto_hw_enter() 53 pm_runtime_get_sync(profile->dev); in blk_crypto_hw_enter() 60 if (profile->dev) in blk_crypto_hw_exit() 61 pm_runtime_put_sync(profile->dev); in blk_crypto_hw_exit() 155 int devm_blk_crypto_profile_init(struct device *dev, in devm_blk_crypto_profile_init() argument 164 return devm_add_action_or_reset(dev, in devm_blk_crypto_profile_init()
|
| A D | bsg.c | 170 static void bsg_device_release(struct device *dev) in bsg_device_release() argument 172 struct bsg_device *bd = container_of(dev, struct bsg_device, device); in bsg_device_release() 240 static char *bsg_devnode(const struct device *dev, umode_t *mode) in bsg_devnode() argument 242 return kasprintf(GFP_KERNEL, "bsg/%s", dev_name(dev)); in bsg_devnode()
|
| A D | badblocks.c | 1491 static int __badblocks_init(struct device *dev, struct badblocks *bb, in __badblocks_init() argument 1494 bb->dev = dev; in __badblocks_init() 1500 if (dev) in __badblocks_init() 1501 bb->page = devm_kzalloc(dev, PAGE_SIZE, GFP_KERNEL); in __badblocks_init() 1528 int devm_init_badblocks(struct device *dev, struct badblocks *bb) in devm_init_badblocks() argument 1532 return __badblocks_init(dev, bb, 1); in devm_init_badblocks() 1544 if (bb->dev) in badblocks_exit() 1545 devm_kfree(bb->dev, bb->page); in badblocks_exit()
|
| A D | kyber-iosched.c | 152 dev_t dev; member 256 trace_kyber_latency(kqd->dev, kyber_domain_names[sched_domain], in calculate_percentile() 269 trace_kyber_adjust(kqd->dev, kyber_domain_names[sched_domain], in kyber_resize_domain() 365 kqd->dev = disk_devt(q->disk); in kyber_queue_data_alloc() 769 trace_kyber_throttled(kqd->dev, in kyber_dispatch_cur_domain() 782 trace_kyber_throttled(kqd->dev, in kyber_dispatch_cur_domain()
|
| A D | ioctl.c | 537 dev_t dev = I_BDEV(file->f_mapping->host)->bd_dev; in blkdev_bszset() local 551 excl_file = bdev_file_open_by_dev(dev, mode, &dev, NULL); in blkdev_bszset()
|
| A D | blk-cgroup.c | 1159 struct device *dev; in blkcg_fill_root_iostats() local 1162 while ((dev = class_dev_iter_next(&iter))) { in blkcg_fill_root_iostats() 1163 struct block_device *bdev = dev_to_bdev(dev); in blkcg_fill_root_iostats()
|
| /block/partitions/ |
| A D | core.c | 176 static ssize_t part_partition_show(struct device *dev, in part_partition_show() argument 179 return sprintf(buf, "%d\n", bdev_partno(dev_to_bdev(dev))); in part_partition_show() 182 static ssize_t part_start_show(struct device *dev, in part_start_show() argument 188 static ssize_t part_ro_show(struct device *dev, in part_ro_show() argument 194 static ssize_t part_alignment_offset_show(struct device *dev, in part_alignment_offset_show() argument 200 static ssize_t part_discard_alignment_show(struct device *dev, in part_discard_alignment_show() argument 246 static void part_release(struct device *dev) in part_release() argument 248 put_disk(dev_to_bdev(dev)->bd_disk); in part_release() 249 bdev_drop(dev_to_bdev(dev)); in part_release() 254 const struct block_device *part = dev_to_bdev(dev); in part_uevent() [all …]
|
| A D | mac.c | 16 extern void note_bootable_part(dev_t dev, int part, int goodness);
|