| /block/ |
| A D | blk-sysfs.c | 36 return sysfs_emit(page, "%lu\n", var); in queue_var_show() 45 err = kstrtoul(page, 10, &v); in queue_var_store() 75 ret = queue_var_store(&nr, page, count); in queue_requests_store() 292 return sysfs_emit(page, "%u\n", in queue_poll_show() 300 return sysfs_emit(page, "none\n"); in queue_zoned_show() 430 return sysfs_emit(page, "%u\n", in queue_io_timeout_show() 441 err = kstrtou32(page, 10, &val); in queue_io_timeout_store() 467 !strncmp(page, "none", 4)) { in queue_wc_store() 578 err = kstrtos64(page, 10, &v); in queue_var_store64() 598 ret = sysfs_emit(page, "0\n"); in queue_wb_lat_show() [all …]
|
| A D | blk-integrity.c | 281 err = kstrtoul(page, 10, &val); in flag_store() 306 char *page) in format_show() argument 311 return sysfs_emit(page, "none\n"); in format_show() 316 char *page) in tag_size_show() argument 325 char *page) in protection_interval_bytes_show() argument 329 return sysfs_emit(page, "%u\n", in protection_interval_bytes_show() 335 const char *page, size_t count) in read_verify_store() argument 341 struct device_attribute *attr, char *page) in read_verify_show() argument 348 const char *page, size_t count) in write_generate_store() argument 361 char *page) in device_is_integrity_capable_show() argument [all …]
|
| A D | blk-crypto-sysfs.c | 21 struct blk_crypto_attr *attr, char *page); 35 struct blk_crypto_attr *attr, char *page) in hw_wrapped_keys_show() argument 38 return sysfs_emit(page, "supported\n"); in hw_wrapped_keys_show() 42 struct blk_crypto_attr *attr, char *page) in max_dun_bits_show() argument 48 struct blk_crypto_attr *attr, char *page) in num_keyslots_show() argument 50 return sysfs_emit(page, "%u\n", profile->num_slots); in num_keyslots_show() 54 struct blk_crypto_attr *attr, char *page) in raw_keys_show() argument 57 return sysfs_emit(page, "supported\n"); in raw_keys_show() 117 struct blk_crypto_attr *attr, char *page) in blk_crypto_mode_show() argument 137 struct attribute *attr, char *page) in blk_crypto_attr_show() argument [all …]
|
| A D | badblocks.c | 460 u64 *p = bb->page; in prev_by_hint() 500 p = bb->page; in prev_badblocks() 538 u64 *p = bb->page; in can_merge_front() 556 u64 *p = bb->page; in front_merge() 591 u64 *p = bb->page; in can_combine_front() 611 u64 *p = bb->page; in front_combine() 868 p = bb->page; in _badblocks_set() 1082 p = bb->page; in _badblocks_clear() 1504 if (!bb->page) { in __badblocks_init() 1547 kfree(bb->page); in badblocks_exit() [all …]
|
| A D | blk-mq-sysfs.c | 50 struct attribute *attr, char *page) in blk_mq_hw_sysfs_show() argument 65 res = entry->show(hctx, page); in blk_mq_hw_sysfs_show() 71 char *page) in blk_mq_hw_sysfs_nr_tags_show() argument 73 return sprintf(page, "%u\n", hctx->tags->nr_tags); in blk_mq_hw_sysfs_nr_tags_show() 77 char *page) in blk_mq_hw_sysfs_nr_reserved_tags_show() argument 79 return sprintf(page, "%u\n", hctx->tags->nr_reserved_tags); in blk_mq_hw_sysfs_nr_reserved_tags_show() 82 static ssize_t blk_mq_hw_sysfs_cpus_show(struct blk_mq_hw_ctx *hctx, char *page) in blk_mq_hw_sysfs_cpus_show() argument 90 ret = snprintf(pos + page, size - pos, "%u", i); in blk_mq_hw_sysfs_cpus_show() 92 ret = snprintf(pos + page, size - pos, ", %u", i); in blk_mq_hw_sysfs_cpus_show() 101 ret = snprintf(pos + page, size + 1 - pos, "\n"); in blk_mq_hw_sysfs_cpus_show()
|
| A D | blk-map.c | 135 struct page *page; in bio_copy_user_iov() local 180 page = map_data->pages[i / nr_pages]; in bio_copy_user_iov() 181 page += (i % nr_pages); in bio_copy_user_iov() 185 page = alloc_page(GFP_NOIO | gfp_mask); in bio_copy_user_iov() 186 if (!page) { in bio_copy_user_iov() 192 if (bio_add_page(bio, page, bytes, offset) < bytes) { in bio_copy_user_iov() 194 __free_page(page); in bio_copy_user_iov() 398 struct page *page; in bio_copy_kern() local 405 if (!page) in bio_copy_kern() 409 memcpy(page_address(page), p, bytes); in bio_copy_kern() [all …]
|
| A D | bio.c | 922 static bool bvec_try_merge_page(struct bio_vec *bv, struct page *page, in bvec_try_merge_page() argument 927 phys_addr_t page_addr = page_to_phys(page); in bvec_try_merge_page() 954 struct page *page, unsigned len, unsigned offset) in bvec_try_merge_hw_page() argument 977 void __bio_add_page(struct bio *bio, struct page *page, in __bio_add_page() argument 983 if (is_pci_p2pdma_page(page)) in __bio_add_page() 1018 int bio_add_page(struct bio *bio, struct page *page, in bio_add_page() argument 1040 __bio_add_page(bio, page, len, offset); in bio_add_page() 1166 struct page **pages, unsigned int i, in get_contig_folio_len() 1213 struct page **pages = (struct page **)bv; in __bio_iov_iter_get_pages() 1257 struct page *page = pages[i]; in __bio_iov_iter_get_pages() local [all …]
|
| A D | bio-integrity.c | 122 int bio_integrity_add_page(struct bio *bio, struct page *page, in bio_integrity_add_page() argument 131 if (!zone_device_pages_have_same_pgmap(bv->bv_page, page)) in bio_integrity_add_page() 134 if (bvec_try_merge_hw_page(q, bv, page, len, offset)) { in bio_integrity_add_page() 151 bvec_set_page(&bip->bip_vec[bip->bip_vcnt], page, len, offset); in bio_integrity_add_page() 232 static unsigned int bvec_from_pages(struct bio_vec *bvec, struct page **pages, in bvec_from_pages() 266 struct page *stack_pages[UIO_FASTIOV], **pages = stack_pages; in bio_integrity_map_user()
|
| A D | blk-crypto-fallback.c | 320 struct page *plaintext_page = enc_bvec->bv_page; in blk_crypto_fallback_encrypt_bio() 321 struct page *ciphertext_page = in blk_crypto_fallback_encrypt_bio() 420 struct page *page = bv.bv_page; in blk_crypto_fallback_decrypt_bio() local 422 sg_set_page(&sg, page, data_unit_size, bv.bv_offset); in blk_crypto_fallback_decrypt_bio()
|
| A D | blk.h | 117 struct page *page, unsigned len, unsigned offset); 583 static inline void bio_release_page(struct bio *bio, struct page *page) in bio_release_page() argument 586 unpin_user_page(page); in bio_release_page()
|
| A D | elevator.h | 161 ssize_t elv_iosched_show(struct gendisk *disk, char *page); 162 ssize_t elv_iosched_store(struct gendisk *disk, const char *page, size_t count);
|
| A D | blk-mq.c | 3417 struct page *page; in blk_mq_clear_rq_mapping() local 3457 struct page *page; in blk_mq_free_rqs() local 3483 page = list_first_entry(&tags->page_list, struct page, lru); in blk_mq_free_rqs() 3484 list_del_init(&page->lru); in blk_mq_free_rqs() 3490 __free_pages(page, page->private); in blk_mq_free_rqs() 3605 struct page *page; in blk_mq_alloc_rqs() local 3613 page = alloc_pages_node(node, in blk_mq_alloc_rqs() 3616 if (page) in blk_mq_alloc_rqs() 3624 if (!page) in blk_mq_alloc_rqs() 3627 page->private = this_order; in blk_mq_alloc_rqs() [all …]
|
| A D | bfq-iosched.c | 7402 return sprintf(page, "%u\n", var); in bfq_var_show() 7408 int ret = kstrtoul(page, 10, &new_val); in bfq_var_store() 7425 return bfq_var_show(__data, (page)); \ 7504 const char *page, size_t count) in bfq_max_budget_store() argument 7510 ret = bfq_var_store(&__data, (page)); in bfq_max_budget_store() 7532 const char *page, size_t count) in bfq_timeout_sync_store() argument 7538 ret = bfq_var_store(&__data, (page)); in bfq_timeout_sync_store() 7555 const char *page, size_t count) in bfq_strict_guarantees_store() argument 7561 ret = bfq_var_store(&__data, (page)); in bfq_strict_guarantees_store() 7577 const char *page, size_t count) in bfq_low_latency_store() argument [all …]
|
| A D | elevator.c | 423 elv_attr_show(struct kobject *kobj, struct attribute *attr, char *page) in elv_attr_show() argument 435 error = entry->show(e, page); in elv_attr_show() 442 const char *page, size_t length) in elv_attr_store() argument 454 error = entry->store(e, page, length); in elv_attr_store()
|
| A D | kyber-iosched.c | 853 char *page) \ 857 return sprintf(page, "%llu\n", kqd->latency_targets[domain]); \ 861 const char *page, size_t count) \ 867 ret = kstrtoull(page, 10, &nsec); \
|
| A D | genhd.c | 853 char *page) in disk_badblocks_show() argument 858 return sysfs_emit(page, "\n"); in disk_badblocks_show() 860 return badblocks_show(disk->bb, page, 0); in disk_badblocks_show() 865 const char *page, size_t len) in disk_badblocks_store() argument 872 return badblocks_store(disk->bb, page, len, 0); in disk_badblocks_store()
|
| A D | mq-deadline.c | 762 static ssize_t __FUNC(struct elevator_queue *e, char *page) \ 766 return sysfs_emit(page, "%d\n", __VAR); \ 780 static ssize_t __FUNC(struct elevator_queue *e, const char *page, size_t count) \ 785 __ret = kstrtoint(page, 0, &__data); \
|
| A D | blk-iocost.c | 869 u64 *page, u64 *seqio, u64 *randio) in calc_lcoefs() argument 873 *page = *seqio = *randio = 0; in calc_lcoefs() 879 *page = DIV64_U64_ROUND_UP(VTIME_PER_SEC, bps_pages); in calc_lcoefs() 881 *page = 1; in calc_lcoefs() 886 if (v > *page) in calc_lcoefs() 887 *seqio = v - *page; in calc_lcoefs() 892 if (v > *page) in calc_lcoefs() 893 *randio = v - *page; in calc_lcoefs()
|