Home
last modified time | relevance | path

Searched refs:bio_sectors (Results 1 – 25 of 39) sorted by relevance

12

/linux-6.3-rc2/block/
A Dblk-core.c503 if (op_is_flush(bio->bi_opf) && !bio_sectors(bio)) in bio_check_ro()
527 unsigned int nr_sectors = bio_sectors(bio); in bio_check_eod()
550 if (bio_sectors(bio)) { in blk_partition_remap()
566 int nr_sectors = bio_sectors(bio); in blk_check_zone_append()
762 if (!bio_sectors(bio)) { in submit_bio_noacct()
838 count_vm_events(PGPGIN, bio_sectors(bio)); in submit_bio()
840 count_vm_events(PGPGOUT, bio_sectors(bio)); in submit_bio()
987 return bdev_start_io_acct(bio->bi_bdev, bio_sectors(bio), in bio_start_io_acct()
A Dblk-merge.c130 if (bio_sectors(bio) <= max_discard_sectors) in bio_split_discard()
156 if (bio_sectors(bio) <= lim->max_write_zeroes_sectors) in bio_split_write_zeroes()
645 if (blk_rq_sectors(req) + bio_sectors(bio) > in ll_back_merge_fn()
664 if (blk_rq_sectors(req) + bio_sectors(bio) > in ll_front_merge_fn()
680 if (blk_rq_sectors(req) + bio_sectors(next->bio) > in req_attempt_discard_merge()
954 else if (blk_rq_pos(rq) - bio_sectors(bio) == bio->bi_iter.bi_sector) in blk_try_merge()
1036 if (blk_rq_sectors(req) + bio_sectors(bio) > in bio_attempt_discard_merge()
A Dbio-integrity.c211 if (!bio_sectors(bio)) in bio_integrity_prep()
227 intervals = bio_integrity_intervals(bi, bio_sectors(bio)); in bio_integrity_prep()
387 bip->bip_iter.bi_size = bio_integrity_bytes(bi, bio_sectors(bio)); in bio_integrity_trim()
A Dbounce.c228 if (sectors < bio_sectors(bio_orig)) { in __blk_queue_bounce()
A Dblk-cgroup.h390 bio_issue_init(&bio->bi_issue, bio_sectors(bio)); in blkcg_bio_issue_init()
A Dblk-crypto-fallback.c223 if (num_sectors < bio_sectors(bio)) { in blk_crypto_fallback_split_bio_if_needed()
A Dbio.c1631 BUG_ON(sectors >= bio_sectors(bio)); in bio_split()
1667 offset + size > bio_sectors(bio))) in bio_trim()
/linux-6.3-rc2/drivers/md/bcache/
A Drequest.c114 bio_sectors(bio), (uint64_t) bio->bi_iter.bi_sector); in bch_data_invalidate()
116 while (bio_sectors(bio)) { in bch_data_invalidate()
117 unsigned int sectors = min(bio_sectors(bio), in bch_data_invalidate()
195 if (atomic_sub_return(bio_sectors(bio), &op->c->sectors_to_gc) < 0) in bch_data_insert_start()
222 if (!bch_alloc_sectors(op->c, k, bio_sectors(bio), in bch_data_insert_start()
398 bio_sectors(bio) & (c->cache->sb.block_size - 1)) { in check_should_bypass()
454 bch_rescale_priorities(c, bio_sectors(bio)); in check_should_bypass()
457 bch_mark_sectors_bypassed(c, dc, bio_sectors(bio)); in check_should_bypass()
525 unsigned int bio_sectors = bio_sectors(bio); in cache_lookup_fn() local
536 BUG_ON(bio_sectors <= sectors); in cache_lookup_fn()
[all …]
A Dwriteback.h117 bio_sectors(bio))) in should_writeback()
/linux-6.3-rc2/drivers/md/
A Ddm-zone.c134 return !op_is_flush(bio->bi_opf) && bio_sectors(bio); in dm_is_zone_write()
496 if (op_is_flush(bio->bi_opf) && !bio_sectors(bio)) in dm_need_zone_wp_tracking()
535 orig_bio_details.nr_sectors = bio_sectors(clone); in dm_zone_map_bio()
639 if (WARN_ON_ONCE(zwp_offset < bio_sectors(orig_bio))) in dm_zone_endio()
644 zwp_offset - bio_sectors(orig_bio); in dm_zone_endio()
A Ddm.c510 sectors = bio_sectors(bio); in dm_io_acct()
1324 unsigned int bio_sectors = bio_sectors(bio); in dm_accept_partial_bio() local
1329 BUG_ON(bio_sectors > *tio->len_ptr); in dm_accept_partial_bio()
1330 BUG_ON(n_sectors > bio_sectors); in dm_accept_partial_bio()
1332 *tio->len_ptr -= bio_sectors - n_sectors; in dm_accept_partial_bio()
1341 io->sector_offset = bio_sectors(io->orig_bio); in dm_accept_partial_bio()
1465 io->sector_offset = bio_sectors(ci->bio); in setup_split_accounting()
1722 ci->sector_count = bio_sectors(bio); in init_clone_info()
A Ddm-log-writes.c670 if (!bio_sectors(bio) && !flush_bio) in log_writes_map()
704 block->nr_sectors = bio_to_dev_sectors(lc, bio_sectors(bio)); in log_writes_map()
716 if (flush_bio && !bio_sectors(bio)) { in log_writes_map()
A Ddm-zoned.h46 #define dmz_bio_blocks(bio) dmz_sect2blk(bio_sectors(bio))
A Ddm-ebs-target.c50 sector_t end_sector = __block_mod(bio->bi_iter.bi_sector, ec->u_bs) + bio_sectors(bio); in __nr_blocks()
A Draid10.c1234 if (max_sectors < bio_sectors(bio)) { in raid10_read_request()
1538 if (r10_bio->sectors < bio_sectors(bio)) { in raid10_write_request()
1701 if (bio_sectors(bio) < stripe_size*2) in raid10_handle_discard()
1719 split_size = bio_sectors(bio) - remainder; in raid10_handle_discard()
1901 int sectors = bio_sectors(bio); in raid10_make_request()
2481 md_sync_acct(conf->mirrors[d].rdev->bdev, bio_sectors(tbio)); in sync_request_write()
2504 bio_sectors(tbio)); in sync_request_write()
2635 md_sync_acct(conf->mirrors[d].rdev->bdev, bio_sectors(wbio)); in recovery_request_write()
2641 bio_sectors(wbio2)); in recovery_request_write()
A Ddm-crypt.c1145 if (!bio_sectors(bio) || !io->cc->on_disk_tag_size) in dm_crypt_integrity_io_alloc()
1152 tag_len = io->cc->on_disk_tag_size * (bio_sectors(bio) >> io->cc->sector_shift); in dm_crypt_integrity_io_alloc()
2065 sector += bio_sectors(clone); in kcryptd_crypt_write_convert()
3391 if (bio_sectors(bio)) in crypt_map()
3418 unsigned int tag_len = cc->on_disk_tag_size * (bio_sectors(bio) >> cc->sector_shift); in crypt_map()
3426 if (bio_sectors(bio) > cc->tag_pool_max_sectors) in crypt_map()
A Draid1.c1197 r1_bio->sectors = bio_sectors(bio); in init_r1bio()
1301 if (max_sectors < bio_sectors(bio)) { in raid1_read_request()
1501 if (max_sectors < bio_sectors(bio)) { in raid1_write_request()
1614 bio->bi_iter.bi_sector, bio_sectors(bio)); in raid1_make_request()
2263 md_sync_acct(conf->mirrors[i].rdev->bdev, bio_sectors(wbio)); in sync_request_write()
A Ddm-integrity.c1641 if (likely(!bio->bi_status) && unlikely(bio_sectors(bio) != dio->range.n_sectors)) { in dec_in_flight()
1885 sector_t end_boundary = (sec + bio_sectors(bio) - 1) >> log2_max_io_len; in dm_integrity_map()
1909 if (unlikely(dio->range.logical_sector + bio_sectors(bio) > ic->provided_data_sectors)) { in dm_integrity_map()
1911 dio->range.logical_sector, bio_sectors(bio), in dm_integrity_map()
1915 …if (unlikely((dio->range.logical_sector | bio_sectors(bio)) & (unsigned int)(ic->sectors_per_block… in dm_integrity_map()
1918 dio->range.logical_sector, bio_sectors(bio)); in dm_integrity_map()
1938 unsigned int wanted_tag_size = bio_sectors(bio) >> ic->sb->log2_sectors_per_block; in dm_integrity_map()
2149 dio->range.n_sectors = bio_sectors(bio); in dm_integrity_map_continue()
A Draid0.c546 if (sectors < bio_sectors(bio)) { in raid0_make_request()
/linux-6.3-rc2/include/trace/events/
A Dblock.h273 __entry->nr_sector = bio_sectors(bio);
301 __entry->nr_sector = bio_sectors(bio);
497 __entry->nr_sector = bio_sectors(bio);
/linux-6.3-rc2/fs/btrfs/
A Draid56.h111 struct sector_ptr *bio_sectors; member
A Draid56.c81 kfree(rbio->bio_sectors); in free_raid_bio_pointers()
178 if (!rbio->bio_sectors[i].page) { in cache_rbio_pages()
192 rbio->bio_sectors[i].page, in cache_rbio_pages()
193 rbio->bio_sectors[i].pgoff, in cache_rbio_pages()
895 sector = &rbio->bio_sectors[index]; in sector_in_rbio()
936 rbio->bio_sectors = kcalloc(num_sectors, sizeof(struct sector_ptr), in alloc_rbio()
943 if (!rbio->stripe_pages || !rbio->bio_sectors || !rbio->stripe_sectors || in alloc_rbio()
1130 struct sector_ptr *sector = &rbio->bio_sectors[index]; in index_one_bio()
1421 sector = &rbio->bio_sectors[i]; in get_bio_sector_nr()
2388 rbio->bio_sectors[index].page = page; in raid56_add_scrub_pages()
[all …]
/linux-6.3-rc2/include/linux/
A Dbio.h42 #define bio_sectors(bio) bvec_iter_sectors((bio)->bi_iter) macro
396 if (sectors >= bio_sectors(bio)) in bio_next_split()
/linux-6.3-rc2/drivers/nvme/target/
A Dio-cmd-bdev.c209 bip->bip_iter.bi_size = bio_integrity_bytes(bi, bio_sectors(bio)); in nvmet_bdev_alloc_bip()
/linux-6.3-rc2/fs/ext4/
A Dpage-io.c332 (unsigned) bio_sectors(bio), in ext4_end_bio()

Completed in 68 milliseconds

12