| /block/ |
| A D | bio.c | 313 static struct bio *__bio_chain_endio(struct bio *bio) in __bio_chain_endio() argument 339 void bio_chain(struct bio *bio, struct bio *parent) in bio_chain() argument 377 struct bio *bio; in bio_alloc_rescue() local 394 struct bio *bio; in punt_bios_to_rescuer() local 449 struct bio *bio; in bio_alloc_percpu_cache() local 509 struct bio *bio; in bio_alloc_bioset() local 613 struct bio *bio; in bio_kmalloc() local 719 struct bio *bio; in __bio_alloc_cache_prune() local 864 struct bio *bio; in bio_alloc_clone() local 1389 struct bio bio; in bdev_rw_virt() local [all …]
|
| A D | blk-merge.c | 107 static struct bio *bio_submit_split(struct bio *bio, int split_sectors) in bio_submit_split() argument 370 struct bio *bio_split_zone_append(struct bio *bio, in bio_split_zone_append() argument 382 struct bio *bio_split_write_zeroes(struct bio *bio, in bio_split_write_zeroes() argument 413 struct bio *bio_split_to_limits(struct bio *bio) in bio_split_to_limits() argument 435 struct bio *bio = rq->bio; in blk_recalc_rq_segments() local 608 struct bio *bio; in blk_rq_set_mixed_merge() local 618 for (bio = rq->bio; bio; bio = bio->bi_next) { in blk_rq_set_mixed_merge() 676 struct bio *bio) in blk_atomic_write_mergeable_rq_bio() argument 908 bio->bi_next = req->bio; in bio_attempt_front_merge() 909 req->bio = bio; in bio_attempt_front_merge() [all …]
|
| A D | blk-map.c | 136 struct bio *bio; in bio_copy_user_iov() local 256 struct bio *bio; in blk_rq_map_bio_alloc() local 276 struct bio *bio; in bio_map_user_iov() local 324 struct bio *bio; in bio_map_kern() local 381 struct bio *bio; in bio_copy_kern() local 465 rq->bio = rq->biotail = bio; in blk_rq_append_bio() 475 struct bio *bio; in blk_rq_map_user_bvec() local 514 struct bio *bio = NULL; in blk_rq_map_user_iov() local 551 bio = rq->bio; in blk_rq_map_user_iov() 641 bio = bio->bi_next; in blk_rq_unmap_user() [all …]
|
| A D | blk-core.c | 522 static inline void bio_check_ro(struct bio *bio) in bio_check_ro() argument 555 static inline int bio_check_eod(struct bio *bio) in bio_check_eod() argument 575 static int blk_partition_remap(struct bio *bio) in blk_partition_remap() argument 595 struct bio *bio) in blk_check_zone_append() argument 624 static void __submit_bio(struct bio *bio) in __submit_bio() argument 728 void submit_bio_noacct_nocheck(struct bio *bio) in submit_bio_noacct_nocheck() argument 757 struct bio *bio) in blk_validate_atomic_write_op_size() argument 777 void submit_bio_noacct(struct bio *bio) in submit_bio_noacct() argument 885 static void bio_set_ioprio(struct bio *bio) in bio_set_ioprio() argument 906 void submit_bio(struct bio *bio) in submit_bio() argument [all …]
|
| A D | blk.h | 58 void bio_await_chain(struct bio *bio); 226 void bio_integrity_free(struct bio *bio); 261 struct bio *bio) in integrity_req_gap_front_merge() argument 289 struct bio *bio) in integrity_req_gap_front_merge() argument 356 struct bio *bio_split_write_zeroes(struct bio *bio, 358 struct bio *bio_split_rw(struct bio *bio, const struct queue_limits *lim, 360 struct bio *bio_split_zone_append(struct bio *bio, 395 static inline struct bio *__bio_split_to_limits(struct bio *bio, in __bio_split_to_limits() argument 483 struct bio *bio) in blk_req_bio_is_zone_append() argument 524 struct bio *bio) in blk_req_bio_is_zone_append() argument [all …]
|
| A D | blk-lib.c | 42 struct bio *bio; in blk_alloc_discard_bio() local 48 if (!bio) in blk_alloc_discard_bio() 60 return bio; in blk_alloc_discard_bio() 66 struct bio *bio; in __blkdev_issue_discard() local 88 struct bio *bio = NULL; in blkdev_issue_discard() local 128 struct bio *bio; in __blkdev_issue_write_zeroes() local 152 struct bio *bio = NULL; in blkdev_issue_write_zeroes() local 159 if (bio) { in blkdev_issue_write_zeroes() 201 struct bio *bio; in __blkdev_issue_zero_pages() local 230 struct bio *bio = NULL; in blkdev_issue_zero_pages() local [all …]
|
| A D | blk-crypto-internal.h | 32 bool bio_crypt_rq_ctx_compatible(struct request *rq, struct bio *bio); 38 struct bio *bio) in bio_crypt_ctx_back_mergeable() argument 45 struct bio *bio) in bio_crypt_ctx_front_mergeable() argument 101 struct bio *bio) in bio_crypt_rq_ctx_compatible() argument 107 struct bio *bio) in bio_crypt_ctx_front_mergeable() argument 113 struct bio *bio) in bio_crypt_ctx_back_mergeable() argument 144 void __bio_crypt_advance(struct bio *bio, unsigned int bytes); 151 void __bio_crypt_free_ctx(struct bio *bio); 152 static inline void bio_crypt_free_ctx(struct bio *bio) in bio_crypt_free_ctx() argument 159 struct bio *bio) in bio_crypt_do_front_merge() argument [all …]
|
| A D | bio-integrity.c | 23 void bio_integrity_free(struct bio *bio) in bio_integrity_free() argument 30 void bio_integrity_init(struct bio *bio, struct bio_integrity_payload *bip, in bio_integrity_init() argument 52 struct bio_integrity_payload *bio_integrity_alloc(struct bio *bio, in bio_integrity_alloc() argument 99 void bio_integrity_unmap_user(struct bio *bio) in bio_integrity_unmap_user() argument 122 int bio_integrity_add_page(struct bio *bio, struct page *page, in bio_integrity_add_page() argument 159 static int bio_integrity_copy_user(struct bio *bio, struct bio_vec *bvec, in bio_integrity_copy_user() argument 262 int bio_integrity_map_user(struct bio *bio, struct iov_iter *iter) in bio_integrity_map_user() argument 332 int bio_integrity_map_iter(struct bio *bio, struct uio_meta *meta) in bio_integrity_map_iter() argument 376 void bio_integrity_advance(struct bio *bio, unsigned int bytes_done) in bio_integrity_advance() argument 392 void bio_integrity_trim(struct bio *bio) in bio_integrity_trim() argument [all …]
|
| A D | blk-rq-qos.h | 104 void __rq_qos_cleanup(struct rq_qos *rqos, struct bio *bio); 108 void __rq_qos_throttle(struct rq_qos *rqos, struct bio *bio); 109 void __rq_qos_track(struct rq_qos *rqos, struct request *rq, struct bio *bio); 110 void __rq_qos_merge(struct rq_qos *rqos, struct request *rq, struct bio *bio); 111 void __rq_qos_done_bio(struct rq_qos *rqos, struct bio *bio); 114 static inline void rq_qos_cleanup(struct request_queue *q, struct bio *bio) in rq_qos_cleanup() argument 139 static inline void rq_qos_done_bio(struct bio *bio) in rq_qos_done_bio() argument 142 bio->bi_bdev && (bio_flagged(bio, BIO_QOS_THROTTLED) || in rq_qos_done_bio() 150 static inline void rq_qos_throttle(struct request_queue *q, struct bio *bio) in rq_qos_throttle() argument 159 struct bio *bio) in rq_qos_track() argument [all …]
|
| A D | bio-integrity-auto.c | 17 struct bio *bio; member 40 struct bio *bio = bid->bio; in bio_integrity_verify_fn() local 44 bio_endio(bio); in bio_integrity_verify_fn() 79 bool __bio_integrity_endio(struct bio *bio) in __bio_integrity_endio() argument 85 if (bio_op(bio) == REQ_OP_READ && !bio->bi_status && in __bio_integrity_endio() 107 bool bio_integrity_prep(struct bio *bio) in bio_integrity_prep() argument 119 if (!bio_sectors(bio)) in bio_integrity_prep() 123 if (bio_integrity(bio)) in bio_integrity_prep() 126 switch (bio_op(bio)) { in bio_integrity_prep() 165 bid->bio = bio; in bio_integrity_prep() [all …]
|
| A D | blk-zoned.c | 179 struct bio bio; in blkdev_zone_reset_all() local 207 struct bio *bio = NULL; in blkdev_zone_mgmt() local 596 struct bio *bio) in blk_zone_wplug_bio_io_error() argument 612 struct bio *bio; in disk_zone_wplug_abort() local 783 struct bio *bio, unsigned int nr_segs) in disk_zone_wplug_add_bio() argument 826 bio->bi_iter.bi_sector, bio_sectors(bio)); in disk_zone_wplug_add_bio() 883 struct bio *bio; in blk_zone_write_plug_init_request() local 937 struct bio *bio) in blk_zone_wplug_prepare_bio() argument 1210 void blk_zone_write_plug_bio_endio(struct bio *bio) in blk_zone_write_plug_bio_endio() argument 1288 struct bio *bio; in blk_zone_wplug_bio_work() local [all …]
|
| A D | blk-crypto-fallback.c | 52 struct bio *bio; member 165 struct bio *bio; in blk_crypto_fallback_clone_bio() local 168 if (!bio) in blk_crypto_fallback_clone_bio() 170 bio_init(bio, bio_src->bi_bdev, bio->bi_inline_vecs, nr_segs, in blk_crypto_fallback_clone_bio() 181 bio->bi_io_vec[bio->bi_vcnt++] = bv; in blk_crypto_fallback_clone_bio() 185 return bio; in blk_crypto_fallback_clone_bio() 214 struct bio *bio = *bio_ptr; in blk_crypto_fallback_split_bio_if_needed() local 382 struct bio *bio = f_ctx->bio; in blk_crypto_fallback_decrypt_bio() local 453 static void blk_crypto_fallback_decrypt_endio(struct bio *bio) in blk_crypto_fallback_decrypt_endio() argument 468 f_ctx->bio = bio; in blk_crypto_fallback_decrypt_endio() [all …]
|
| A D | fops.c | 54 struct bio bio; in __blkdev_direct_IO_simple() local 120 struct bio bio ____cacheline_aligned_in_smp; 125 static void blkdev_bio_end_io(struct bio *bio) in blkdev_bio_end_io() argument 131 if (bio->bi_status && !dio->bio.bi_status) in blkdev_bio_end_io() 132 dio->bio.bi_status = bio->bi_status; in blkdev_bio_end_io() 174 struct bio *bio; in __blkdev_direct_IO() local 184 dio = container_of(bio, struct blkdev_dio, bio); in __blkdev_direct_IO() 190 bio_get(bio); in __blkdev_direct_IO() 289 static void blkdev_bio_end_io_async(struct bio *bio) in blkdev_bio_end_io_async() argument 325 struct bio *bio; in __blkdev_direct_IO_async() local [all …]
|
| A D | blk-throttle.h | 150 static inline bool blk_throtl_bio(struct bio *bio) { return false; } in blk_throtl_bio() argument 154 bool __blk_throtl_bio(struct bio *bio); 162 static inline bool blk_should_throtl(struct bio *bio) in blk_should_throtl() argument 165 int rw = bio_data_dir(bio); in blk_should_throtl() 175 tg = blkg_to_tg(bio->bi_blkg); in blk_should_throtl() 177 if (!bio_flagged(bio, BIO_CGROUP_ACCT)) { in blk_should_throtl() 178 bio_set_flag(bio, BIO_CGROUP_ACCT); in blk_should_throtl() 180 bio->bi_iter.bi_size); in blk_should_throtl() 195 static inline bool blk_throtl_bio(struct bio *bio) in blk_throtl_bio() argument 198 if (!blk_should_throtl(bio)) in blk_throtl_bio() [all …]
|
| A D | blk-throttle.c | 196 struct bio *bio; in throtl_peek_queued() local 203 if (!bio) in throtl_peek_queued() 230 struct bio *bio; in throtl_pop_queued() local 241 if (bio) in throtl_pop_queued() 970 struct bio *bio; in tg_update_disptime() local 973 if (bio) in tg_update_disptime() 977 if (bio) in tg_update_disptime() 1009 struct bio *bio; in tg_dispatch_one_bio() local 1051 struct bio *bio; in throtl_dispatch_tg() local 1212 struct bio *bio; in blk_throtl_dispatch_work_fn() local [all …]
|
| A D | blk-crypto.c | 101 void bio_crypt_set_ctx(struct bio *bio, const struct blk_crypto_key *key, in bio_crypt_set_ctx() argument 117 bio->bi_crypt_context = bc; in bio_crypt_set_ctx() 120 void __bio_crypt_free_ctx(struct bio *bio) in __bio_crypt_free_ctx() argument 123 bio->bi_crypt_context = NULL; in __bio_crypt_free_ctx() 126 int __bio_crypt_clone(struct bio *dst, struct bio *src, gfp_t gfp_mask) in __bio_crypt_clone() 154 void __bio_crypt_advance(struct bio *bio, unsigned int bytes) in __bio_crypt_advance() argument 203 bool bio_crypt_rq_ctx_compatible(struct request *rq, struct bio *bio) in bio_crypt_rq_ctx_compatible() argument 223 static bool bio_crypt_check_alignment(struct bio *bio) in bio_crypt_check_alignment() argument 285 struct bio *bio = *bio_ptr; in __blk_crypto_bio_prep() local 290 bio->bi_status = BLK_STS_IOERR; in __blk_crypto_bio_prep() [all …]
|
| A D | t10-pi.c | 131 struct bio *bio; in t10_pi_type1_prepare() local 133 __rq_for_each_bio(bio, rq) { in t10_pi_type1_prepare() 183 struct bio *bio; in t10_pi_type1_complete() local 185 __rq_for_each_bio(bio, rq) { in t10_pi_type1_complete() 305 struct bio *bio; in ext_pi_type1_prepare() local 307 __rq_for_each_bio(bio, rq) { in ext_pi_type1_prepare() 346 struct bio *bio; in ext_pi_type1_complete() local 348 __rq_for_each_bio(bio, rq) { in ext_pi_type1_complete() 375 void blk_integrity_generate(struct bio *bio) in blk_integrity_generate() argument 407 void blk_integrity_verify_iter(struct bio *bio, struct bvec_iter *saved_iter) in blk_integrity_verify_iter() argument [all …]
|
| A D | blk-cgroup.h | 241 static inline bool bio_issue_as_root_blkg(struct bio *bio) in bio_issue_as_root_blkg() argument 243 return (bio->bi_opf & (REQ_META | REQ_SWAP)) != 0; in bio_issue_as_root_blkg() 373 static inline void blkcg_bio_issue_init(struct bio *bio) in blkcg_bio_issue_init() argument 375 bio_issue_init(&bio->bi_issue, bio_sectors(bio)); in blkcg_bio_issue_init() 456 static inline bool blk_cgroup_mergeable(struct request *rq, struct bio *bio) in blk_cgroup_mergeable() argument 458 return rq->bio->bi_blkg == bio->bi_blkg && in blk_cgroup_mergeable() 459 bio_issue_as_root_blkg(rq->bio) == bio_issue_as_root_blkg(bio); in blk_cgroup_mergeable() 462 void blk_cgroup_bio_start(struct bio *bio); 494 static inline void blkcg_bio_issue_init(struct bio *bio) { } in blkcg_bio_issue_init() argument 495 static inline void blk_cgroup_bio_start(struct bio *bio) { } in blk_cgroup_bio_start() argument [all …]
|
| A D | blk-mq-dma.c | 20 if (!iter->bio) in blk_map_iter_next() 24 iter->bio = NULL; in blk_map_iter_next() 31 bv = mp_bvec_iter_bvec(iter->bio->bi_io_vec, iter->iter); in blk_map_iter_next() 46 if (!iter->bio->bi_next) in blk_map_iter_next() 48 iter->bio = iter->bio->bi_next; in blk_map_iter_next() 49 iter->iter = iter->bio->bi_iter; in blk_map_iter_next() 156 iter->iter.bio = req->bio; in blk_rq_dma_map_iter_start() 157 iter->iter.iter = req->bio->bi_iter; in blk_rq_dma_map_iter_start() 250 .bio = rq->bio, in __blk_rq_map_sg() 256 if (iter.bio) in __blk_rq_map_sg() [all …]
|
| A D | blk-integrity.c | 28 int blk_rq_count_integrity_sg(struct request_queue *q, struct bio *bio) in blk_rq_count_integrity_sg() argument 36 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_count_integrity_sg() 140 struct bio *bio = rq->bio; in blk_rq_map_integrity_sg() local 145 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_map_integrity_sg() 190 ret = bio_integrity_map_user(rq->bio, &iter); in blk_rq_integrity_map_user() 209 if (bio_integrity(req->bio)->bip_flags != in blk_integrity_merge_rq() 210 bio_integrity(next->bio)->bip_flags) in blk_integrity_merge_rq() 217 if (integrity_req_gap_back_merge(req, next->bio)) in blk_integrity_merge_rq() 224 struct bio *bio) in blk_integrity_merge_bio() argument 234 if (bio_integrity(req->bio)->bip_flags != bio_integrity(bio)->bip_flags) in blk_integrity_merge_bio() [all …]
|
| A D | blk-mq-sched.h | 10 bool blk_mq_sched_try_merge(struct request_queue *q, struct bio *bio, 12 bool blk_mq_sched_bio_merge(struct request_queue *q, struct bio *bio, 41 static inline bool bio_mergeable(struct bio *bio) in bio_mergeable() argument 43 return !(bio->bi_opf & REQ_NOMERGE_FLAGS); in bio_mergeable() 48 struct bio *bio) in blk_mq_sched_allow_merge() argument 54 return e->type->ops.allow_merge(q, rq, bio); in blk_mq_sched_allow_merge()
|
| A D | blk-rq-qos.c | 28 void __rq_qos_cleanup(struct rq_qos *rqos, struct bio *bio) in __rq_qos_cleanup() argument 32 rqos->ops->cleanup(rqos, bio); in __rq_qos_cleanup() 64 void __rq_qos_throttle(struct rq_qos *rqos, struct bio *bio) in __rq_qos_throttle() argument 68 rqos->ops->throttle(rqos, bio); in __rq_qos_throttle() 73 void __rq_qos_track(struct rq_qos *rqos, struct request *rq, struct bio *bio) in __rq_qos_track() argument 77 rqos->ops->track(rqos, rq, bio); in __rq_qos_track() 82 void __rq_qos_merge(struct rq_qos *rqos, struct request *rq, struct bio *bio) in __rq_qos_merge() argument 86 rqos->ops->merge(rqos, rq, bio); in __rq_qos_merge() 91 void __rq_qos_done_bio(struct rq_qos *rqos, struct bio *bio) in __rq_qos_done_bio() argument 95 rqos->ops->done_bio(rqos, bio); in __rq_qos_done_bio()
|
| A D | blk-mq.c | 862 struct bio *bio = req->bio; in blk_complete_request() local 881 struct bio *next = bio->bi_next; in blk_complete_request() 960 struct bio *bio = req->bio; in blk_update_request() local 967 req->bio = bio->bi_next; in blk_update_request() 1066 struct bio *bio = req->bio; in blk_rq_passthrough_stats() local 2665 rq->bio = rq->biotail = bio; in blk_mq_bio_to_request() 3018 struct bio *bio) in blk_mq_get_new_requests() argument 3070 struct bio *bio) in blk_mq_use_cached_rq() argument 3308 struct bio *bio; in blk_rq_unprep_clone() local 3311 rq->bio = bio->bi_next; in blk_rq_unprep_clone() [all …]
|
| A D | blk-ioprio.c | 133 void blkcg_set_ioprio(struct bio *bio) in blkcg_set_ioprio() argument 135 struct ioprio_blkcg *blkcg = blkcg_to_ioprio_blkcg(bio->bi_blkg->blkcg); in blkcg_set_ioprio() 150 if (IOPRIO_PRIO_CLASS(bio->bi_ioprio) != IOPRIO_CLASS_RT) in blkcg_set_ioprio() 151 bio->bi_ioprio = IOPRIO_PRIO_VALUE(IOPRIO_CLASS_RT, 4); in blkcg_set_ioprio() 162 prio = max_t(u16, bio->bi_ioprio, in blkcg_set_ioprio() 164 if (prio > bio->bi_ioprio) in blkcg_set_ioprio() 165 bio->bi_ioprio = prio; in blkcg_set_ioprio()
|
| A D | blk-ioprio.h | 9 struct bio; 12 void blkcg_set_ioprio(struct bio *bio); 14 static inline void blkcg_set_ioprio(struct bio *bio) in blkcg_set_ioprio() argument
|