| /block/ |
| A D | blk-mq-dma.c | 28 if (!iter->iter.bi_size) in blk_map_iter_next() 31 bv = mp_bvec_iter_bvec(iter->bio->bi_io_vec, iter->iter); in blk_map_iter_next() 35 bio_advance_iter_single(iter->bio, &iter->iter, bv.bv_len); in blk_map_iter_next() 42 while (!iter->iter.bi_size || !iter->iter.bi_bvec_done) { in blk_map_iter_next() 45 if (!iter->iter.bi_size) { in blk_map_iter_next() 49 iter->iter = iter->bio->bi_iter; in blk_map_iter_next() 52 next = mp_bvec_iter_bvec(iter->bio->bi_io_vec, iter->iter); in blk_map_iter_next() 58 bio_advance_iter_single(iter->bio, &iter->iter, next.bv_len); in blk_map_iter_next() 156 iter->iter.bio = req->bio; in blk_rq_dma_map_iter_start() 157 iter->iter.iter = req->bio->bi_iter; in blk_rq_dma_map_iter_start() [all …]
|
| A D | t10-pi.c | 43 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_generate() 46 pi->guard_tag = t10_pi_csum(0, iter->data_buf, iter->interval, in t10_pi_generate() 58 iter->data_buf += iter->interval; in t10_pi_generate() 60 iter->seed++; in t10_pi_generate() 70 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_verify() 92 csum = t10_pi_csum(0, iter->data_buf, iter->interval, in t10_pi_verify() 107 iter->data_buf += iter->interval; in t10_pi_verify() 236 iter->data_buf += iter->interval; in ext_pi_crc64_generate() 255 for (i = 0; i < iter->data_size; i += iter->interval) { in ext_pi_crc64_verify() 285 iter->disk_name, (unsigned long long)iter->seed, in ext_pi_crc64_verify() [all …]
|
| A D | blk-map.c | 17 struct iov_iter iter; member 32 bmd->iter = *data; in bio_alloc_map_data() 35 bmd->iter.__iov = bmd->iov; in bio_alloc_map_data() 59 iter); in bio_copy_from_iter() 61 if (!iov_iter_count(iter)) in bio_copy_from_iter() 90 &iter); in bio_copy_to_iter() 92 if (!iov_iter_count(&iter)) in bio_copy_to_iter() 279 if (!iov_iter_count(iter)) in bio_map_user_iov() 478 if (!iov_iter_count(iter) || iov_iter_count(iter) > max_bytes) in blk_rq_map_user_bvec() 539 i = *iter; in blk_rq_map_user_iov() [all …]
|
| A D | fops.c | 39 struct iov_iter *iter) in blkdev_dio_invalid() argument 42 !bdev_iter_is_aligned(bdev, iter); in blkdev_dio_invalid() 67 if (iov_iter_rw(iter) == READ) { in __blkdev_direct_IO_simple() 69 if (user_backed_iter(iter)) in __blkdev_direct_IO_simple() 86 if (iov_iter_rw(iter) == WRITE) in __blkdev_direct_IO_simple() 318 struct iov_iter *iter, in __blkdev_direct_IO_async() argument 342 if (iov_iter_is_bvec(iter)) { in __blkdev_direct_IO_async() 349 bio_iov_bvec_set(bio, iter); in __blkdev_direct_IO_async() 358 if (user_backed_iter(iter)) { in __blkdev_direct_IO_async() 398 if (!iov_iter_count(iter)) in blkdev_direct_IO() [all …]
|
| A D | early-lookup.c | 127 struct class_dev_iter iter; in blk_lookup_devt() local 130 class_dev_iter_init(&iter, &block_class, NULL, &disk_type); in blk_lookup_devt() 131 while ((dev = class_dev_iter_next(&iter))) { in blk_lookup_devt() 149 class_dev_iter_exit(&iter); in blk_lookup_devt() 274 struct class_dev_iter iter; in printk_all_partitions() local 277 class_dev_iter_init(&iter, &block_class, NULL, &disk_type); in printk_all_partitions() 278 while ((dev = class_dev_iter_next(&iter))) { in printk_all_partitions() 315 class_dev_iter_exit(&iter); in printk_all_partitions()
|
| A D | bio-integrity.c | 164 struct iov_iter iter; in bio_integrity_copy_user() local 173 iov_iter_bvec(&iter, ITER_SOURCE, bvec, nr_vecs, len); in bio_integrity_copy_user() 174 if (!copy_from_iter_full(buf, len, &iter)) { in bio_integrity_copy_user() 262 int bio_integrity_map_user(struct bio *bio, struct iov_iter *iter) in bio_integrity_map_user() argument 268 size_t offset, bytes = iter->count; in bio_integrity_map_user() 278 nr_vecs = iov_iter_npages(iter, BIO_MAX_VECS + 1); in bio_integrity_map_user() 288 copy = !iov_iter_is_aligned(iter, align, align); in bio_integrity_map_user() 289 ret = iov_iter_extract_pages(iter, &pages, bytes, nr_vecs, 0, &offset); in bio_integrity_map_user() 345 it = meta->iter; in bio_integrity_map_iter() 361 iov_iter_advance(&meta->iter, integrity_bytes); in bio_integrity_map_iter()
|
| A D | blk-integrity.c | 33 struct bvec_iter iter; in blk_rq_count_integrity_sg() local 36 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_count_integrity_sg() 142 struct bvec_iter iter; in blk_rq_map_integrity_sg() local 145 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_map_integrity_sg() 187 struct iov_iter iter; in blk_rq_integrity_map_user() local 189 iov_iter_ubuf(&iter, rq_data_dir(rq), ubuf, bytes); in blk_rq_integrity_map_user() 190 ret = bio_integrity_map_user(rq->bio, &iter); in blk_rq_integrity_map_user()
|
| A D | bio.c | 624 struct bvec_iter iter; in zero_fill_bio_iter() local 626 __bio_for_each_segment(bv, bio, iter, start) in zero_fill_bio_iter() 644 struct bvec_iter iter; in bio_truncate() local 654 bio_for_each_segment(bv, bio, iter) { in bio_truncate() 1158 bio->bi_vcnt = iter->nr_segs; in bio_iov_bvec_set() 1237 size = iov_iter_extract_pages(iter, &pages, in __bio_iov_iter_get_pages() 1247 iov_iter_revert(iter, trim); in __bio_iov_iter_get_pages() 1292 iov_iter_revert(iter, left); in __bio_iov_iter_get_pages() 1327 if (iov_iter_is_bvec(iter)) { in bio_iov_iter_get_pages() 1328 bio_iov_bvec_set(bio, iter); in bio_iov_iter_get_pages() [all …]
|
| A D | blk-merge.c | 27 struct bvec_iter iter = bio->bi_iter; in bio_get_last_bvec() local 34 bio_advance_iter(bio, &iter, iter.bi_size); in bio_get_last_bvec() 36 if (!iter.bi_bvec_done) in bio_get_last_bvec() 37 idx = iter.bi_idx - 1; in bio_get_last_bvec() 39 idx = iter.bi_idx; in bio_get_last_bvec() 47 if (iter.bi_bvec_done) in bio_get_last_bvec() 48 bv->bv_len = iter.bi_bvec_done; in bio_get_last_bvec() 297 struct bvec_iter iter; in bio_split_rw_at() local 300 bio_for_each_bvec(bv, bio, iter) { in bio_split_rw_at() 425 struct req_iterator iter; in blk_recalc_rq_segments() local [all …]
|
| A D | blk-crypto-fallback.c | 163 struct bvec_iter iter; in blk_crypto_fallback_clone_bio() local 180 bio_for_each_segment(bv, bio_src, iter) in blk_crypto_fallback_clone_bio() 218 struct bvec_iter iter; in blk_crypto_fallback_split_bio_if_needed() local 220 bio_for_each_segment(bv, bio, iter) { in blk_crypto_fallback_split_bio_if_needed() 391 struct bvec_iter iter; in blk_crypto_fallback_decrypt_bio() local 419 __bio_for_each_segment(bv, bio, iter, f_ctx->crypt_iter) { in blk_crypto_fallback_decrypt_bio()
|
| A D | genhd.c | 914 struct class_dev_iter *iter; in disk_seqf_start() local 917 iter = kmalloc(sizeof(*iter), GFP_KERNEL); in disk_seqf_start() 918 if (!iter) in disk_seqf_start() 921 seqf->private = iter; in disk_seqf_start() 922 class_dev_iter_init(iter, &block_class, NULL, &disk_type); in disk_seqf_start() 924 dev = class_dev_iter_next(iter); in disk_seqf_start() 946 struct class_dev_iter *iter = seqf->private; in disk_seqf_stop() local 949 if (iter) { in disk_seqf_stop() 950 class_dev_iter_exit(iter); in disk_seqf_stop() 951 kfree(iter); in disk_seqf_stop()
|
| A D | sed-opal.c | 1013 struct opal_resp_tok *iter; in response_parse() local 1047 iter = resp->toks; in response_parse() 1072 iter++; in response_parse() 1238 struct opal_suspend_data *iter; in add_suspend_info() local 1241 if (iter->lr == sus->lr) { in add_suspend_info() 1242 list_del(&iter->node); in add_suspend_info() 1243 kfree(iter); in add_suspend_info() 1570 int *iter, in response_get_column() argument 1575 int n = *iter; in response_get_column() 1609 *iter = n; in response_get_column() [all …]
|
| A D | blk-crypto.c | 227 struct bvec_iter iter; in bio_crypt_check_alignment() local 230 bio_for_each_segment(bv, bio, iter) { in bio_crypt_check_alignment()
|
| A D | blk-cgroup.c | 1158 struct class_dev_iter iter; in blkcg_fill_root_iostats() local 1161 class_dev_iter_init(&iter, &block_class, NULL, &disk_type); in blkcg_fill_root_iostats() 1162 while ((dev = class_dev_iter_next(&iter))) { in blkcg_fill_root_iostats() 1193 class_dev_iter_exit(&iter); in blkcg_fill_root_iostats()
|