Home
last modified time | relevance | path

Searched refs:dev (Results 1 – 20 of 20) sorted by relevance

/block/
A Dsed-opal.c458 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_send_cmd()
465 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_recv_cmd()
574 const u8 *epos = dev->resp, *cpos = dev->resp; in opal_discovery0_end()
1174 print_buffer(dev->cmd, dev->pos); in parse_and_check_status()
1196 set_comid(dev, dev->comid); in cmd_start()
1262 ret = cmd_finalize(dev, dev->hsn, dev->tsn); in finalize_and_send()
1268 print_buffer(dev->cmd, dev->pos); in finalize_and_send()
2219 set_comid(dev, dev->comid); in lock_unlock_locking_range_sum()
2447 set_comid(dev, dev->comid); in end_opal_session()
2501 if (!dev) in free_opal_dev()
[all …]
A Dearly-lookup.c49 struct device *dev = NULL; in devt_from_partuuid() local
72 if (!dev) in devt_from_partuuid()
83 *devt = dev->devt; in devt_from_partuuid()
86 put_device(dev); in devt_from_partuuid()
114 struct device *dev; in devt_from_partlabel() local
117 if (!dev) in devt_from_partlabel()
119 *devt = dev->devt; in devt_from_partlabel()
120 put_device(dev); in devt_from_partlabel()
128 struct device *dev; in blk_lookup_devt() local
275 struct device *dev; in printk_all_partitions() local
[all …]
A Dblk-pm.c31 q->dev = dev; in blk_pm_runtime_init()
33 pm_runtime_set_autosuspend_delay(q->dev, -1); in blk_pm_runtime_init()
34 pm_runtime_use_autosuspend(q->dev); in blk_pm_runtime_init()
63 if (!q->dev) in blk_pre_runtime_suspend()
97 pm_runtime_mark_last_busy(q->dev); in blk_pre_runtime_suspend()
122 if (!q->dev) in blk_post_runtime_suspend()
130 pm_runtime_mark_last_busy(q->dev); in blk_post_runtime_suspend()
152 if (!q->dev) in blk_pre_runtime_resume()
178 if (!q->dev) in blk_post_runtime_resume()
184 pm_runtime_mark_last_busy(q->dev); in blk_post_runtime_resume()
[all …]
A Dblk-integrity.c247 return &dev_to_disk(dev)->queue->limits.integrity; in dev_to_bi()
276 struct request_queue *q = dev_to_disk(dev)->queue; in flag_store()
300 struct blk_integrity *bi = dev_to_bi(dev); in flag_show()
308 struct blk_integrity *bi = dev_to_bi(dev); in format_show()
318 struct blk_integrity *bi = dev_to_bi(dev); in tag_size_show()
327 struct blk_integrity *bi = dev_to_bi(dev); in protection_interval_bytes_show()
333 static ssize_t read_verify_store(struct device *dev, in read_verify_store() argument
340 static ssize_t read_verify_show(struct device *dev, in read_verify_show() argument
343 return flag_show(dev, page, BLK_INTEGRITY_NOVERIFY); in read_verify_show()
353 static ssize_t write_generate_show(struct device *dev, in write_generate_show() argument
[all …]
A Dgenhd.c539 &disk->bdi->dev->kobj, "bdi"); in __add_disk()
855 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_show()
867 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_store()
915 struct device *dev; in disk_seqf_start() local
924 dev = class_dev_iter_next(iter); in disk_seqf_start()
925 if (!dev) in disk_seqf_start()
929 return dev_to_disk(dev); in disk_seqf_start()
934 struct device *dev; in disk_seqf_next() local
938 if (dev) in disk_seqf_next()
939 return dev_to_disk(dev); in disk_seqf_next()
[all …]
A Dbsg-lib.c161 put_device(job->dev); /* release reference for the request */ in bsg_teardown_job()
232 static bool bsg_prepare_job(struct device *dev, struct request *req) in bsg_prepare_job() argument
249 job->dev = dev; in bsg_prepare_job()
251 get_device(job->dev); in bsg_prepare_job()
276 struct device *dev = q->queuedata; in bsg_queue_rq() local
285 if (!get_device(dev)) in bsg_queue_rq()
288 if (!bsg_prepare_job(dev, req)) in bsg_queue_rq()
296 put_device(dev); in bsg_queue_rq()
362 struct request_queue *bsg_setup_queue(struct device *dev, const char *name, in bsg_setup_queue() argument
388 q = blk_mq_alloc_queue(set, lim, dev); in bsg_setup_queue()
[all …]
A Dblk-pm.h11 if (!q->dev || !blk_queue_pm_only(q)) in blk_pm_resume_queue()
15 pm_request_resume(q->dev); in blk_pm_resume_queue()
21 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_mark_last_busy()
22 pm_runtime_mark_last_busy(rq->q->dev); in blk_pm_mark_last_busy()
A Dbdev.c500 bdev->bd_dev = dev; in bdev_add()
501 inode->i_rdev = dev; in bdev_add()
502 inode->i_ino = dev; in bdev_add()
787 MAJOR(dev), MINOR(dev), in bdev_permission()
823 inode = ilookup(blockdev_superblock, dev); in blkdev_get_no_open()
825 blk_request_module(dev); in blkdev_get_no_open()
1042 ret = bdev_permission(dev, mode, holder); in bdev_file_open_by_dev()
1046 bdev = blkdev_get_no_open(dev, true); in bdev_file_open_by_dev()
1075 dev_t dev; in bdev_file_open_by_path() local
1078 error = lookup_bdev(path, &dev); in bdev_file_open_by_path()
[all …]
A Dblk-mq-cpumap.c109 struct device *dev, unsigned int offset) in blk_mq_map_hw_queues() argument
115 if (!dev->bus->irq_get_affinity) in blk_mq_map_hw_queues()
119 mask = dev->bus->irq_get_affinity(dev, queue + offset); in blk_mq_map_hw_queues()
A Dblk-timeout.c39 ssize_t part_timeout_show(struct device *dev, struct device_attribute *attr, in part_timeout_show() argument
42 struct gendisk *disk = dev_to_disk(dev); in part_timeout_show()
48 ssize_t part_timeout_store(struct device *dev, struct device_attribute *attr, in part_timeout_store() argument
51 struct gendisk *disk = dev_to_disk(dev); in part_timeout_store()
A Ddisk-events.c341 static ssize_t disk_events_show(struct device *dev, in disk_events_show() argument
344 struct gendisk *disk = dev_to_disk(dev); in disk_events_show()
351 static ssize_t disk_events_async_show(struct device *dev, in disk_events_async_show() argument
357 static ssize_t disk_events_poll_msecs_show(struct device *dev, in disk_events_poll_msecs_show() argument
361 struct gendisk *disk = dev_to_disk(dev); in disk_events_poll_msecs_show()
368 static ssize_t disk_events_poll_msecs_store(struct device *dev, in disk_events_poll_msecs_store() argument
372 struct gendisk *disk = dev_to_disk(dev); in disk_events_poll_msecs_store()
A Dblk.h108 struct block_device *blkdev_get_no_open(dev_t dev, bool autoload);
340 ssize_t part_size_show(struct device *dev, struct device_attribute *attr,
342 ssize_t part_stat_show(struct device *dev, struct device_attribute *attr,
344 ssize_t part_inflight_show(struct device *dev, struct device_attribute *attr,
346 ssize_t part_fail_show(struct device *dev, struct device_attribute *attr,
348 ssize_t part_fail_store(struct device *dev, struct device_attribute *attr,
557 void bdev_add(struct block_device *bdev, dev_t dev);
728 int bdev_permission(dev_t dev, blk_mode_t mode, void *holder);
A Dblk-crypto-profile.c52 if (profile->dev) in blk_crypto_hw_enter()
53 pm_runtime_get_sync(profile->dev); in blk_crypto_hw_enter()
60 if (profile->dev) in blk_crypto_hw_exit()
61 pm_runtime_put_sync(profile->dev); in blk_crypto_hw_exit()
155 int devm_blk_crypto_profile_init(struct device *dev, in devm_blk_crypto_profile_init() argument
164 return devm_add_action_or_reset(dev, in devm_blk_crypto_profile_init()
A Dbsg.c170 static void bsg_device_release(struct device *dev) in bsg_device_release() argument
172 struct bsg_device *bd = container_of(dev, struct bsg_device, device); in bsg_device_release()
240 static char *bsg_devnode(const struct device *dev, umode_t *mode) in bsg_devnode() argument
242 return kasprintf(GFP_KERNEL, "bsg/%s", dev_name(dev)); in bsg_devnode()
A Dbadblocks.c1491 static int __badblocks_init(struct device *dev, struct badblocks *bb, in __badblocks_init() argument
1494 bb->dev = dev; in __badblocks_init()
1500 if (dev) in __badblocks_init()
1501 bb->page = devm_kzalloc(dev, PAGE_SIZE, GFP_KERNEL); in __badblocks_init()
1528 int devm_init_badblocks(struct device *dev, struct badblocks *bb) in devm_init_badblocks() argument
1532 return __badblocks_init(dev, bb, 1); in devm_init_badblocks()
1544 if (bb->dev) in badblocks_exit()
1545 devm_kfree(bb->dev, bb->page); in badblocks_exit()
A Dkyber-iosched.c152 dev_t dev; member
256 trace_kyber_latency(kqd->dev, kyber_domain_names[sched_domain], in calculate_percentile()
269 trace_kyber_adjust(kqd->dev, kyber_domain_names[sched_domain], in kyber_resize_domain()
365 kqd->dev = disk_devt(q->disk); in kyber_queue_data_alloc()
769 trace_kyber_throttled(kqd->dev, in kyber_dispatch_cur_domain()
782 trace_kyber_throttled(kqd->dev, in kyber_dispatch_cur_domain()
A Dioctl.c537 dev_t dev = I_BDEV(file->f_mapping->host)->bd_dev; in blkdev_bszset() local
551 excl_file = bdev_file_open_by_dev(dev, mode, &dev, NULL); in blkdev_bszset()
A Dblk-cgroup.c1159 struct device *dev; in blkcg_fill_root_iostats() local
1162 while ((dev = class_dev_iter_next(&iter))) { in blkcg_fill_root_iostats()
1163 struct block_device *bdev = dev_to_bdev(dev); in blkcg_fill_root_iostats()
/block/partitions/
A Dcore.c176 static ssize_t part_partition_show(struct device *dev, in part_partition_show() argument
179 return sprintf(buf, "%d\n", bdev_partno(dev_to_bdev(dev))); in part_partition_show()
182 static ssize_t part_start_show(struct device *dev, in part_start_show() argument
188 static ssize_t part_ro_show(struct device *dev, in part_ro_show() argument
194 static ssize_t part_alignment_offset_show(struct device *dev, in part_alignment_offset_show() argument
200 static ssize_t part_discard_alignment_show(struct device *dev, in part_discard_alignment_show() argument
246 static void part_release(struct device *dev) in part_release() argument
248 put_disk(dev_to_bdev(dev)->bd_disk); in part_release()
249 bdev_drop(dev_to_bdev(dev)); in part_release()
254 const struct block_device *part = dev_to_bdev(dev); in part_uevent()
[all …]
A Dmac.c16 extern void note_bootable_part(dev_t dev, int part, int goodness);

Completed in 52 milliseconds