| /linux/drivers/md/ |
| A D | raid0.c | 88 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones() 153 mddev->chunk_sectors << 9, blksize); in create_strip_zones() 354 ~(sector_t)(mddev->chunk_sectors-1)); in raid0_size() 366 if (mddev->chunk_sectors == 0) { in raid0_run() 468 mddev->chunk_sectors; in raid0_handle_discard() 470 mddev->chunk_sectors) + in raid0_handle_discard() 473 mddev->chunk_sectors; in raid0_handle_discard() 475 mddev->chunk_sectors) + in raid0_handle_discard() 484 mddev->chunk_sectors; in raid0_handle_discard() 531 chunk_sects = mddev->chunk_sectors; in raid0_make_request() [all …]
|
| A D | md-linear.c | 88 if (mddev->chunk_sectors) { in linear_conf() 90 sector_div(sectors, mddev->chunk_sectors); in linear_conf() 91 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf() 281 seq_printf(seq, " %dk rounding", mddev->chunk_sectors / 2); in linear_status()
|
| A D | raid5.c | 2928 : conf->chunk_sectors; in raid5_compute_sector() 3124 : conf->chunk_sectors; in raid5_compute_blocknr() 5311 unsigned int chunk_sectors; in in_chunk_boundary() local 5314 chunk_sectors = min(conf->chunk_sectors, conf->prev_chunk_sectors); in in_chunk_boundary() 5315 return chunk_sectors >= in in_chunk_boundary() 7048 max(conf->chunk_sectors, in alloc_scratch_buffer() 7504 int chunk_sectors; in raid5_run() local 7527 chunk_sectors = max(mddev->chunk_sectors, mddev->new_chunk_sectors); in raid5_run() 7558 here_old * chunk_sectors) in raid5_run() 8138 mddev->chunk_sectors) in check_reshape() [all …]
|
| A D | dm-zoned-target.c | 998 unsigned int chunk_sectors = dmz_zone_nr_sectors(dmz->metadata); in dmz_io_hints() local 1008 limits->max_discard_sectors = chunk_sectors; in dmz_io_hints() 1009 limits->max_hw_discard_sectors = chunk_sectors; in dmz_io_hints() 1010 limits->max_write_zeroes_sectors = chunk_sectors; in dmz_io_hints() 1013 limits->chunk_sectors = chunk_sectors; in dmz_io_hints() 1014 limits->max_sectors = chunk_sectors; in dmz_io_hints()
|
| A D | dm-raid.c | 716 mddev->new_chunk_sectors = mddev->chunk_sectors; in rs_set_cur() 729 mddev->chunk_sectors = mddev->new_chunk_sectors; in rs_set_new() 978 if (region_size < rs->md.chunk_sectors) { in validate_region_size() 1157 rs->md.new_chunk_sectors = rs->md.chunk_sectors = value; in parse_raid_params() 1487 if (rs->md.chunk_sectors) in parse_raid_params() 1488 max_io_len = rs->md.chunk_sectors; in parse_raid_params() 1877 rs->md.new_chunk_sectors != rs->md.chunk_sectors; in rs_is_layout_change() 2119 sb->stripe_sectors = cpu_to_le32(mddev->chunk_sectors); in super_sync() 2231 mddev->chunk_sectors = le32_to_cpu(sb->stripe_sectors); in super_init_validation() 2286 if (mddev->chunk_sectors != mddev->new_chunk_sectors) in super_init_validation() [all …]
|
| A D | raid5-ppl.c | 328 (data_sector >> ilog2(conf->chunk_sectors) == in ppl_log_stripe() 329 data_sector_last >> ilog2(conf->chunk_sectors)) && in ppl_log_stripe() 824 if ((pp_size >> 9) < conf->chunk_sectors) { in ppl_recover_entry() 833 (data_disks - 1) * conf->chunk_sectors + in ppl_recover_entry() 837 strip_sectors = conf->chunk_sectors; in ppl_recover_entry() 871 (disk * conf->chunk_sectors); in ppl_recover_entry()
|
| A D | dm-unstripe.c | 179 limits->chunk_sectors = uc->chunk_size; in unstripe_io_hints()
|
| A D | raid10.c | 1765 dev_start = (first_stripe_index + 1) * mddev->chunk_sectors; in raid10_handle_discard() 1767 dev_start = first_stripe_index * mddev->chunk_sectors; in raid10_handle_discard() 1772 dev_end = (last_stripe_index + 1) * mddev->chunk_sectors; in raid10_handle_discard() 1774 dev_end = last_stripe_index * mddev->chunk_sectors; in raid10_handle_discard() 1874 seq_printf(seq, " %dK chunks", mddev->chunk_sectors / 2); in raid10_status() 3172 window_size = (chunks + extra_chunk) * conf->mddev->chunk_sectors; in raid10_set_cluster_sync_high() 3896 chunk = mddev->chunk_sectors; in setup_geo() 4043 blk_queue_io_opt(conf->mddev->queue, (conf->mddev->chunk_sectors << 9) * in raid10_set_io_opt() 4091 blk_queue_io_min(mddev->queue, mddev->chunk_sectors << 9); in raid10_run() 4332 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid10_takeover_raid0() [all …]
|
| A D | md.c | 1302 mddev->chunk_sectors = sb->chunk_size >> 9; in super_90_validate() 1331 mddev->new_chunk_sectors = mddev->chunk_sectors; in super_90_validate() 1481 sb->chunk_size = mddev->chunk_sectors << 9; in super_90_sync() 4029 mddev->chunk_sectors = mddev->new_chunk_sectors; in level_store() 4234 mddev->chunk_sectors << 9); in chunk_size_show() 4265 mddev->chunk_sectors = n >> 9; in chunk_size_store() 5115 chunk = mddev->chunk_sectors; in max_sync_store() 5242 mddev->new_chunk_sectors = mddev->chunk_sectors; in reshape_position_store() 6181 mddev->chunk_sectors = 0; in md_clean() 6625 info.chunk_size = mddev->chunk_sectors << 9; in get_array_info() [all …]
|
| A D | raid5.h | 567 int chunk_sectors; member
|
| A D | md.h | 314 int chunk_sectors; member
|
| A D | dm-table.c | 1677 zone_sectors = ti_limits.chunk_sectors; in dm_calculate_queue_limits() 1719 zone_sectors = limits->chunk_sectors; in dm_calculate_queue_limits()
|
| A D | raid5-cache.c | 198 sector_div(sect, conf->chunk_sectors); in r5c_tree_index() 365 conf->chunk_sectors >> RAID5_STRIPE_SHIFT(conf))) in r5c_check_cached_full_stripe()
|
| A D | raid1.c | 3259 if (mddev->chunk_sectors != mddev->new_chunk_sectors || in raid1_reshape() 3262 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid1_reshape()
|
| /linux/block/ |
| A D | blk-settings.c | 44 lim->chunk_sectors = 0; in blk_set_default_limits() 162 void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors) in blk_queue_chunk_sectors() argument 164 q->limits.chunk_sectors = chunk_sectors; in blk_queue_chunk_sectors() 220 max_sectors = min(q->limits.chunk_sectors, max_sectors); in blk_queue_max_zone_append_sectors() 574 if (b->chunk_sectors) in blk_stack_limits() 575 t->chunk_sectors = gcd(t->chunk_sectors, b->chunk_sectors); in blk_stack_limits() 599 if ((t->chunk_sectors << 9) & (t->physical_block_size - 1)) { in blk_stack_limits() 600 t->chunk_sectors = 0; in blk_stack_limits()
|
| A D | blk.h | 325 return q->limits.chunk_sectors || bio->bi_vcnt != 1 || in blk_may_split()
|
| A D | blk-zoned.c | 643 q->limits.chunk_sectors = 0; in blk_queue_clear_zone_settings()
|
| A D | blk-sysfs.c | 155 return queue_var_show(q->limits.chunk_sectors, page); in queue_chunk_sectors_show()
|
| A D | blk-merge.c | 589 if (!q->limits.chunk_sectors || in blk_rq_get_max_sectors()
|
| A D | blk-core.c | 709 if (nr_sectors > q->limits.chunk_sectors) in blk_check_zone_append()
|
| /linux/include/linux/ |
| A D | blkdev.h | 89 unsigned int chunk_sectors; member 500 return blk_queue_is_zoned(q) ? q->limits.chunk_sectors : 0; in blk_queue_zone_sectors() 514 return sector >> ilog2(q->limits.chunk_sectors); in blk_queue_zone_no() 661 unsigned int chunk_sectors) in blk_max_size_offset() argument 663 if (!chunk_sectors) { in blk_max_size_offset() 664 if (q->limits.chunk_sectors) in blk_max_size_offset() 665 chunk_sectors = q->limits.chunk_sectors; in blk_max_size_offset() 670 if (likely(is_power_of_2(chunk_sectors))) in blk_max_size_offset() 671 chunk_sectors -= offset & (chunk_sectors - 1); in blk_max_size_offset() 673 chunk_sectors -= sector_div(offset, chunk_sectors); in blk_max_size_offset() [all …]
|
| /linux/drivers/char/ |
| A D | ps3flash.c | 26 u64 chunk_sectors; member 38 start_sector, priv->chunk_sectors, in ps3flash_read_write_sectors() 118 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read() 151 sector += priv->chunk_sectors; in ps3flash_read() 187 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write() 226 sector += priv->chunk_sectors; in ps3flash_write() 376 priv->chunk_sectors = dev->bounce_size / dev->blk_size; in ps3flash_probe()
|
| /linux/Documentation/ABI/testing/ |
| A D | sysfs-block | 326 What: /sys/block/<disk>/queue/chunk_sectors 330 chunk_sectors has different meaning depending on the type 331 of the disk. For a RAID device (dm-raid), chunk_sectors 334 host-aware or host-managed, chunk_sectors indicates the
|
| /linux/Documentation/block/ |
| A D | queue-sysfs.rst | 18 chunk_sectors (RO) 21 For a RAID device (dm-raid), chunk_sectors indicates the size in 512B sectors 23 or host-managed, chunk_sectors indicates the size in 512B sectors of the zones
|