| /drivers/md/dm-vdo/ |
| A D | io-submitter.c | 148 vio->bios_merged.head->bi_iter.bi_sector); in get_bio_list() 150 vio->bios_merged.tail->bi_iter.bi_sector); in get_bio_list() 193 sector_t merge_sector = bio->bi_iter.bi_sector; in get_mergeable_locked() 216 return (vio_merge->bios_merged.tail->bi_iter.bi_sector == merge_sector ? in get_mergeable_locked() 220 return (vio_merge->bios_merged.head->bi_iter.bi_sector == merge_sector ? in get_mergeable_locked() 229 bio_sector = vio->bios_merged.head->bi_iter.bi_sector; in map_merged_vio() 234 bio_sector = vio->bios_merged.tail->bi_iter.bi_sector; in map_merged_vio() 241 vdo_int_map_remove(bio_map, prev_vio->bios_merged.tail->bi_iter.bi_sector); in merge_to_prev_tail() 254 vdo_int_map_remove(bio_map, next_vio->bios_merged.head->bi_iter.bi_sector); in merge_to_next_head() 291 bio->bi_iter.bi_sector, in try_bio_map_merge()
|
| /drivers/md/ |
| A D | dm-ebs-target.c | 50 sector_t end_sector = __block_mod(bio->bi_iter.bi_sector, ec->u_bs) + bio_sectors(bio); in __nr_blocks() 72 unsigned int buf_off = to_bytes(__block_mod(iter->bi_sector, ec->u_bs)); in __ebs_rw_bvec() 73 sector_t block = __sector_to_block(ec, iter->bi_sector); in __ebs_rw_bvec() 145 sector_t block, blocks, sector = bio->bi_iter.bi_sector; in __ebs_discard_bio() 169 sector_t blocks, sector = bio->bi_iter.bi_sector; in __ebs_forget_bio() 195 block1 = __sector_to_block(ec, bio->bi_iter.bi_sector); in __ebs_process_bios() 200 if (__block_mod(bio->bi_iter.bi_sector, ec->u_bs)) in __ebs_process_bios() 366 bio->bi_iter.bi_sector = ec->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in ebs_map() 375 if (likely(__block_mod(bio->bi_iter.bi_sector, ec->u_bs) || in ebs_map()
|
| A D | dm-linear.c | 82 static sector_t linear_map_sector(struct dm_target *ti, sector_t bi_sector) in linear_map_sector() argument 86 return lc->start + dm_target_offset(ti, bi_sector); in linear_map_sector() 94 bio->bi_iter.bi_sector = linear_map_sector(ti, bio->bi_iter.bi_sector); in linear_map()
|
| A D | dm-flakey.c | 345 static sector_t flakey_map_sector(struct dm_target *ti, sector_t bi_sector) in flakey_map_sector() argument 349 return fc->start + dm_target_offset(ti, bi_sector); in flakey_map_sector() 357 bio->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in flakey_map_bio() 380 (unsigned long long)start.bi_sector, in corrupt_bio_common() 446 clone->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in clone_bio()
|
| A D | md-linear.c | 236 sector_t bio_sector = bio->bi_iter.bi_sector; in linear_make_request() 275 bio->bi_iter.bi_sector = bio->bi_iter.bi_sector - in linear_make_request() 294 (unsigned long long)bio->bi_iter.bi_sector, in linear_make_request()
|
| A D | dm-stripe.c | 255 stripe_map_range_sector(sc, bio->bi_iter.bi_sector, in stripe_map_range() 261 bio->bi_iter.bi_sector = begin + in stripe_map_range() 292 stripe_map_sector(sc, bio->bi_iter.bi_sector, in stripe_map() 293 &stripe, &bio->bi_iter.bi_sector); in stripe_map() 295 bio->bi_iter.bi_sector += sc->stripe[stripe].physical_start; in stripe_map()
|
| A D | dm-io-rewind.c | 62 bip->bip_iter.bi_sector -= bio_integrity_intervals(bi, bytes_done >> 9); in dm_bio_integrity_rewind() 113 iter->bi_sector -= bytes >> 9; in dm_bio_rewind_iter()
|
| A D | dm-stats.c | 635 sector_t bi_sector, sector_t end_sector, in __dm_stat_bio() argument 642 if (end_sector <= s->start || bi_sector >= s->end) in __dm_stat_bio() 644 if (unlikely(bi_sector < s->start)) { in __dm_stat_bio() 648 rel_sector = bi_sector - s->start; in __dm_stat_bio() 649 todo = end_sector - bi_sector; in __dm_stat_bio() 673 sector_t bi_sector, unsigned int bi_sectors, bool end, in dm_stats_account_io() argument 686 end_sector = bi_sector + bi_sectors; in dm_stats_account_io() 695 (bi_sector == (READ_ONCE(last->last_sector) && in dm_stats_account_io() 714 __dm_stat_bio(s, bi_rw, bi_sector, end_sector, end, duration_jiffies, stats_aux); in dm_stats_account_io()
|
| A D | dm-dust.c | 230 bio->bi_iter.bi_sector = dd->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in dust_map() 233 r = dust_map_read(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map() 235 r = dust_map_write(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map()
|
| A D | raid10.c | 1358 bio->bi_iter.bi_sector, in raid10_write_request() 1370 bio->bi_iter.bi_sector, bio_end_sector(bio))) in raid10_write_request() 1550 r10_bio->sector = bio->bi_iter.bi_sector; in __make_request() 1656 bio_start = bio->bi_iter.bi_sector; in raid10_handle_discard() 1705 bio_start = bio->bi_iter.bi_sector; in raid10_handle_discard() 1898 (bio->bi_iter.bi_sector & in raid10_make_request() 2431 tbio->bi_iter.bi_sector = r10_bio->devs[i].addr; in sync_request_write() 2825 wbio->bi_iter.bi_sector = wsector + in narrow_write_error() 3437 bio->bi_iter.bi_sector = from_addr + in raid10_sync_request() 3459 bio->bi_iter.bi_sector = to_addr in raid10_sync_request() [all …]
|
| A D | dm-writecache.c | 1306 writecache_discard(wc, bio->bi_iter.bi_sector, in writecache_flush_thread() 1343 read_original_sector(wc, e) - bio->bi_iter.bi_sector; in writecache_map_remap_origin() 1357 if (e && read_original_sector(wc, e) == bio->bi_iter.bi_sector) { in writecache_map_read() 1367 bio->bi_iter.bi_sector = cache_sector(wc, e); in writecache_map_read() 1394 write_original_sector_seq_count(wc, f, bio->bi_iter.bi_sector + in writecache_bio_copy_ssd() 1421 bio->bi_iter.bi_sector = start_cache_sec; in writecache_bio_copy_ssd() 1447 e = writecache_find_entry(wc, bio->bi_iter.bi_sector, 0); in writecache_map_write() 1532 writecache_discard(wc, bio->bi_iter.bi_sector, bio_end_sector(bio)); in writecache_map_discard() 1554 bio->bi_iter.bi_sector = dm_target_offset(ti, bio->bi_iter.bi_sector); in writecache_map() 1559 (unsigned long long)bio->bi_iter.bi_sector, in writecache_map() [all …]
|
| A D | dm-unstripe.c | 120 sector_t sector = bio->bi_iter.bi_sector; in map_to_core() 140 bio->bi_iter.bi_sector = map_to_core(ti, bio) + uc->physical_start; in unstripe_map()
|
| A D | raid0.c | 451 sector_t start = bio->bi_iter.bi_sector; in raid0_handle_discard() 467 zone->zone_end - bio->bi_iter.bi_sector, GFP_NOIO, in raid0_handle_discard() 558 sector_t bio_sector = bio->bi_iter.bi_sector; in raid0_map_submit_bio() 584 bio->bi_iter.bi_sector = sector + zone->dev_start + in raid0_map_submit_bio() 606 sector = bio->bi_iter.bi_sector; in raid0_make_request()
|
| A D | dm-target.c | 212 static sector_t io_err_map_sector(struct dm_target *ti, sector_t bi_sector) in io_err_map_sector() argument 216 return ioec->start + dm_target_offset(ti, bi_sector); in io_err_map_sector()
|
| A D | dm-thin.c | 675 sector_t block_nr = bio->bi_iter.bi_sector; in get_bio_block() 692 sector_t b = bio->bi_iter.bi_sector; in get_bio_block_range() 717 sector_t bi_sector = bio->bi_iter.bi_sector; in remap() local 721 bio->bi_iter.bi_sector = in remap() 723 (bi_sector & (pool->sectors_per_block - 1)); in remap() 726 sector_div(bi_sector, pool->sectors_per_block); in remap() 2124 sector_t bi_sector = bio->bi_iter.bi_sector; in __thin_bio_rb_add() local 2132 if (bi_sector < thin_bio(pbd)->bi_iter.bi_sector) in __thin_bio_rb_add() 2248 if (lhs_cell->holder->bi_iter.bi_sector < rhs_cell->holder->bi_iter.bi_sector) in cmp_cells() 2251 if (lhs_cell->holder->bi_iter.bi_sector > rhs_cell->holder->bi_iter.bi_sector) in cmp_cells() [all …]
|
| A D | dm-io.c | 227 dp->context_bi.bi_sector = (sector_t)bvec.bv_len; in bio_get_page() 232 unsigned int len = (unsigned int)dp->context_bi.bi_sector; in bio_next_page() 355 bio->bi_iter.bi_sector = where->sector + (where->count - remaining); in do_region()
|
| A D | dm-stats.h | 33 sector_t bi_sector, unsigned int bi_sectors, bool end,
|
| A D | dm-log-writes.c | 223 bio->bi_iter.bi_sector = sector; in write_metadata() 276 bio->bi_iter.bi_sector = sector; in write_inline_data() 357 bio->bi_iter.bi_sector = sector; in log_one_block() 375 bio->bi_iter.bi_sector = sector; in log_one_block() 703 block->sector = bio_to_dev_sectors(lc, bio->bi_iter.bi_sector); in log_writes_map()
|
| A D | dm-zoned.h | 45 #define dmz_bio_block(bio) dmz_sect2blk((bio)->bi_iter.bi_sector) 83 #define dmz_bio_chunk(zmd, bio) ((bio)->bi_iter.bi_sector >> \
|
| A D | raid1.c | 319 (unsigned long long) bio->bi_iter.bi_sector, in raid_end_bio_io() 540 (unsigned long long) mbio->bi_iter.bi_sector, in raid1_end_write_request() 1298 r1_bio->sector = bio->bi_iter.bi_sector; in init_r1bio() 1334 if (!wait_read_barrier(conf, bio->bi_iter.bi_sector, in raid1_read_request() 1403 read_bio->bi_iter.bi_sector = r1_bio->sector + in raid1_read_request() 1480 bio->bi_iter.bi_sector, in raid1_write_request() 1493 if (!wait_barrier(conf, bio->bi_iter.bi_sector, in raid1_write_request() 1707 bio->bi_iter.bi_sector, bio_sectors(bio)); in raid1_make_request() 2278 b->bi_iter.bi_sector = r1_bio->sector + in process_checks() 2540 wbio->bi_iter.bi_sector = r1_bio->sector; in narrow_write_error() [all …]
|
| /drivers/md/bcache/ |
| A D | request.c | 123 bio->bi_iter.bi_sector += sectors; in bch_data_invalidate() 128 bio->bi_iter.bi_sector, in bch_data_invalidate() 220 SET_KEY_OFFSET(k, bio->bi_iter.bi_sector); in CLOSURE_CALLBACK() 431 if (i->last == bio->bi_iter.bi_sector && in check_should_bypass() 538 KEY_START(k) > bio->bi_iter.bi_sector) { in cache_lookup_fn() 542 KEY_START(k) - bio->bi_iter.bi_sector) in cache_lookup_fn() 848 s->iop.bio->bi_iter.bi_sector = in CLOSURE_CALLBACK() 849 s->cache_miss->bi_iter.bi_sector; in CLOSURE_CALLBACK() 935 cache_bio->bi_iter.bi_sector = miss->bi_iter.bi_sector; in cached_dev_cache_miss() 1218 bio->bi_iter.bi_sector += dc->sb.data_offset; in cached_dev_submit_bio() [all …]
|
| A D | debug.c | 54 bio->bi_iter.bi_sector = PTR_OFFSET(&b->key, 0); in bch_btree_verify() 120 check->bi_iter.bi_sector = bio->bi_iter.bi_sector; in bch_data_verify() 141 (uint64_t) bio->bi_iter.bi_sector); in bch_data_verify()
|
| /drivers/nvdimm/ |
| A D | nd_virtio.c | 116 if (bio && bio->bi_iter.bi_sector != -1) { in async_pmem_flush() 124 child->bi_iter.bi_sector = -1; in async_pmem_flush()
|
| /drivers/nvme/target/ |
| A D | io-cmd-bdev.c | 211 bip_set_seed(bip, bio->bi_iter.bi_sector >> in nvmet_bdev_alloc_bip() 288 bio->bi_iter.bi_sector = sector; in nvmet_bdev_execute_rw() 313 bio->bi_iter.bi_sector = sector; in nvmet_bdev_execute_rw()
|
| /drivers/block/ |
| A D | brd.c | 112 sector_t sector = bio->bi_iter.bi_sector; in brd_rw_bvec() 186 brd_do_discard(brd, bio->bi_iter.bi_sector, in brd_submit_bio()
|