Home
last modified time | relevance | path

Searched refs:chunk_sectors (Results 1 – 25 of 25) sorted by relevance

/linux-6.3-rc2/drivers/md/
A Draid0.c84 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones()
134 mddev->chunk_sectors << 9, blksize); in create_strip_zones()
351 ~(sector_t)(mddev->chunk_sectors-1)); in raid0_size()
376 if (mddev->chunk_sectors == 0) { in raid0_run()
473 mddev->chunk_sectors; in raid0_handle_discard()
475 mddev->chunk_sectors) + in raid0_handle_discard()
478 mddev->chunk_sectors; in raid0_handle_discard()
480 mddev->chunk_sectors) + in raid0_handle_discard()
489 mddev->chunk_sectors; in raid0_handle_discard()
536 chunk_sects = mddev->chunk_sectors; in raid0_make_request()
[all …]
A Dmd-linear.c87 if (mddev->chunk_sectors) { in linear_conf()
89 sector_div(sectors, mddev->chunk_sectors); in linear_conf()
90 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf()
270 seq_printf(seq, " %dk rounding", mddev->chunk_sectors / 2); in linear_status()
A Ddm-zoned-target.c995 unsigned int chunk_sectors = dmz_zone_nr_sectors(dmz->metadata); in dmz_io_hints() local
1005 limits->max_discard_sectors = chunk_sectors; in dmz_io_hints()
1006 limits->max_hw_discard_sectors = chunk_sectors; in dmz_io_hints()
1007 limits->max_write_zeroes_sectors = chunk_sectors; in dmz_io_hints()
1010 limits->chunk_sectors = chunk_sectors; in dmz_io_hints()
1011 limits->max_sectors = chunk_sectors; in dmz_io_hints()
A Draid5.c3006 : conf->chunk_sectors; in raid5_compute_sector()
3202 : conf->chunk_sectors; in raid5_compute_blocknr()
5417 unsigned int chunk_sectors; in in_chunk_boundary() local
5420 chunk_sectors = min(conf->chunk_sectors, conf->prev_chunk_sectors); in in_chunk_boundary()
5421 return chunk_sectors >= in in_chunk_boundary()
7317 max(conf->chunk_sectors, in alloc_scratch_buffer()
7787 int chunk_sectors; in raid5_run() local
7812 chunk_sectors = max(mddev->chunk_sectors, mddev->new_chunk_sectors); in raid5_run()
7845 here_old * chunk_sectors) in raid5_run()
8448 mddev->chunk_sectors) in check_reshape()
[all …]
A Ddm-raid.c717 mddev->new_chunk_sectors = mddev->chunk_sectors; in rs_set_cur()
730 mddev->chunk_sectors = mddev->new_chunk_sectors; in rs_set_new()
979 if (region_size < rs->md.chunk_sectors) { in validate_region_size()
1160 rs->md.new_chunk_sectors = rs->md.chunk_sectors = value; in parse_raid_params()
1490 if (rs->md.chunk_sectors) in parse_raid_params()
1491 max_io_len = rs->md.chunk_sectors; in parse_raid_params()
1880 rs->md.new_chunk_sectors != rs->md.chunk_sectors; in rs_is_layout_change()
2122 sb->stripe_sectors = cpu_to_le32(mddev->chunk_sectors); in super_sync()
2234 mddev->chunk_sectors = le32_to_cpu(sb->stripe_sectors); in super_init_validation()
2289 if (mddev->chunk_sectors != mddev->new_chunk_sectors) in super_init_validation()
[all …]
A Draid5-ppl.c329 (data_sector >> ilog2(conf->chunk_sectors) == in ppl_log_stripe()
330 data_sector_last >> ilog2(conf->chunk_sectors)) && in ppl_log_stripe()
814 if ((pp_size >> 9) < conf->chunk_sectors) { in ppl_recover_entry()
823 (data_disks - 1) * conf->chunk_sectors + in ppl_recover_entry()
827 strip_sectors = conf->chunk_sectors; in ppl_recover_entry()
861 (disk * conf->chunk_sectors); in ppl_recover_entry()
A Ddm-unstripe.c180 limits->chunk_sectors = uc->chunk_size; in unstripe_io_hints()
A Draid5.h579 int chunk_sectors; member
A Dmd.c1291 mddev->chunk_sectors = sb->chunk_size >> 9; in super_90_validate()
1320 mddev->new_chunk_sectors = mddev->chunk_sectors; in super_90_validate()
1470 sb->chunk_size = mddev->chunk_sectors << 9; in super_90_sync()
3983 mddev->chunk_sectors = mddev->new_chunk_sectors; in level_store()
4188 mddev->chunk_sectors << 9); in chunk_size_show()
4219 mddev->chunk_sectors = n >> 9; in chunk_size_store()
5081 chunk = mddev->chunk_sectors; in max_sync_store()
5208 mddev->new_chunk_sectors = mddev->chunk_sectors; in reshape_position_store()
6158 mddev->chunk_sectors = 0; in md_clean()
6601 info.chunk_size = mddev->chunk_sectors << 9; in get_array_info()
[all …]
A Draid10.c1828 dev_start = (first_stripe_index + 1) * mddev->chunk_sectors; in raid10_handle_discard()
1830 dev_start = first_stripe_index * mddev->chunk_sectors; in raid10_handle_discard()
1835 dev_end = (last_stripe_index + 1) * mddev->chunk_sectors; in raid10_handle_discard()
1837 dev_end = last_stripe_index * mddev->chunk_sectors; in raid10_handle_discard()
1939 seq_printf(seq, " %dK chunks", mddev->chunk_sectors / 2); in raid10_status()
3234 window_size = (chunks + extra_chunk) * conf->mddev->chunk_sectors; in raid10_set_cluster_sync_high()
3958 chunk = mddev->chunk_sectors; in setup_geo()
4105 blk_queue_io_opt(conf->mddev->queue, (conf->mddev->chunk_sectors << 9) * in raid10_set_io_opt()
4149 blk_queue_io_min(mddev->queue, mddev->chunk_sectors << 9); in raid10_run()
4379 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid10_takeover_raid0()
[all …]
A Dmd.h339 int chunk_sectors; member
A Ddm-table.c1688 zone_sectors = ti_limits.chunk_sectors; in dm_calculate_queue_limits()
1729 zone_sectors = limits->chunk_sectors; in dm_calculate_queue_limits()
A Draid5-cache.c198 sector_div(sect, conf->chunk_sectors); in r5c_tree_index()
365 conf->chunk_sectors >> RAID5_STRIPE_SHIFT(conf))) in r5c_check_cached_full_stripe()
A Draid1.c3270 if (mddev->chunk_sectors != mddev->new_chunk_sectors || in raid1_reshape()
3273 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid1_reshape()
/linux-6.3-rc2/block/
A Dblk-settings.c45 lim->chunk_sectors = 0; in blk_set_default_limits()
167 void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors) in blk_queue_chunk_sectors() argument
169 q->limits.chunk_sectors = chunk_sectors; in blk_queue_chunk_sectors()
225 max_sectors = min(q->limits.chunk_sectors, max_sectors); in blk_queue_max_zone_append_sectors()
612 if (b->chunk_sectors) in blk_stack_limits()
613 t->chunk_sectors = gcd(t->chunk_sectors, b->chunk_sectors); in blk_stack_limits()
637 if ((t->chunk_sectors << 9) & (t->physical_block_size - 1)) { in blk_stack_limits()
638 t->chunk_sectors = 0; in blk_stack_limits()
A Dblk-merge.c176 if (lim->chunk_sectors) { in get_max_io_size()
179 lim->chunk_sectors)); in get_max_io_size()
600 if (!q->limits.chunk_sectors || in blk_rq_get_max_sectors()
605 blk_chunk_sectors_left(offset, q->limits.chunk_sectors)); in blk_rq_get_max_sectors()
A Dblk.h322 return lim->chunk_sectors || bio->bi_vcnt != 1 || in bio_may_exceed_limits()
A Dblk-zoned.c623 q->limits.chunk_sectors = 0; in disk_clear_zone_settings()
A Dblk-core.c582 if (nr_sectors > q->limits.chunk_sectors) in blk_check_zone_append()
A Dblk-sysfs.c157 return queue_var_show(q->limits.chunk_sectors, page); in queue_chunk_sectors_show()
/linux-6.3-rc2/drivers/char/
A Dps3flash.c26 u64 chunk_sectors; member
38 start_sector, priv->chunk_sectors, in ps3flash_read_write_sectors()
118 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read()
151 sector += priv->chunk_sectors; in ps3flash_read()
187 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write()
226 sector += priv->chunk_sectors; in ps3flash_write()
376 priv->chunk_sectors = dev->bounce_size / dev->blk_size; in ps3flash_probe()
/linux-6.3-rc2/include/linux/
A Dblkdev.h289 unsigned int chunk_sectors; member
674 return sector >> ilog2(disk->queue->limits.chunk_sectors); in disk_zone_no()
901 unsigned int chunk_sectors) in blk_chunk_sectors_left() argument
903 if (unlikely(!is_power_of_2(chunk_sectors))) in blk_chunk_sectors_left()
904 return chunk_sectors - sector_div(offset, chunk_sectors); in blk_chunk_sectors_left()
905 return chunk_sectors - (offset & (chunk_sectors - 1)); in blk_chunk_sectors_left()
1314 return q->limits.chunk_sectors; in bdev_zone_sectors()
/linux-6.3-rc2/include/uapi/linux/
A Dublk_cmd.h232 __u32 chunk_sectors; member
/linux-6.3-rc2/Documentation/ABI/stable/
A Dsysfs-block145 What: /sys/block/<disk>/queue/chunk_sectors
149 [RO] chunk_sectors has different meaning depending on the type
150 of the disk. For a RAID device (dm-raid), chunk_sectors
153 host-managed, chunk_sectors indicates the size in 512B sectors
/linux-6.3-rc2/drivers/block/
A Dublk_drv.c221 blk_queue_chunk_sectors(q, p->chunk_sectors); in ublk_dev_param_basic_apply()

Completed in 259 milliseconds