| /drivers/md/dm-vdo/ |
| A D | data-vio.c | 281 (u32) (VDO_BLOCK_SIZE - data_vio->offset)), in acknowledge_data_vio() 516 for (i = 0; i < VDO_BLOCK_SIZE; i += sizeof(u64)) { in is_zero_block() 792 BUILD_BUG_ON(VDO_BLOCK_SIZE > PAGE_SIZE); in initialize_data_vio() 793 result = vdo_allocate_memory(VDO_BLOCK_SIZE, 0, "data_vio data", in initialize_data_vio() 806 result = vdo_allocate_memory(VDO_BLOCK_SIZE, 0, "vio scratch", in initialize_data_vio() 1250 VDO_BLOCK_SIZE - data_vio->offset); in finish_cleanup() 1431 fragment_size, VDO_BLOCK_SIZE); in uncompress_data_vio() 1432 if (size != VDO_BLOCK_SIZE) { in uncompress_data_vio() 1457 VDO_BLOCK_SIZE - data_vio->offset)); in modify_for_partial_write() 1519 memset(data_vio->vio.data, 0, VDO_BLOCK_SIZE); in complete_zero_read() [all …]
|
| A D | constants.h | 81 VDO_BLOCK_SIZE = 4096, enumerator 84 VDO_SECTORS_PER_BLOCK = (VDO_BLOCK_SIZE >> SECTOR_SHIFT),
|
| A D | dm-vdo-target.c | 570 if (value > (UINT_MAX / VDO_BLOCK_SIZE)) { in process_one_key_value_pair() 572 UINT_MAX / VDO_BLOCK_SIZE); in process_one_key_value_pair() 718 if ((logical_bytes % VDO_BLOCK_SIZE) != 0) { in parse_device_config() 737 config->logical_blocks = logical_bytes / VDO_BLOCK_SIZE; in parse_device_config() 883 config->physical_blocks = device_size / VDO_BLOCK_SIZE; in parse_device_config() 928 limits->physical_block_size = VDO_BLOCK_SIZE; in vdo_io_hints() 931 limits->io_min = VDO_BLOCK_SIZE; in vdo_io_hints() 933 limits->io_opt = VDO_BLOCK_SIZE; in vdo_io_hints() 955 limits->discard_granularity = VDO_BLOCK_SIZE; in vdo_io_hints() 1014 return bdev_nr_bytes(vdo_get_backing_device(vdo)) / VDO_BLOCK_SIZE; in get_underlying_device_block_count() [all …]
|
| A D | vio.c | 197 return vio_reset_bio_with_size(vio, data, vio->block_count * VDO_BLOCK_SIZE, in vio_reset_bio() 206 int vio_size = vio->block_count * VDO_BLOCK_SIZE; in vio_reset_bio_with_size() 328 size_t per_vio_size = VDO_BLOCK_SIZE * block_count; in make_vio_pool()
|
| A D | io-submitter.h | 37 operation, vio->data, vio->block_count * VDO_BLOCK_SIZE); in vdo_submit_metadata_vio()
|
| A D | packer.h | 32 VDO_COMPRESSED_BLOCK_DATA_SIZE = VDO_BLOCK_SIZE - sizeof(struct compressed_block_header),
|
| A D | slab-depot.c | 333 memcpy(block->outgoing_entries, block->entries, VDO_BLOCK_SIZE); in launch_write() 757 memcpy(pooled->vio.data, journal->block, VDO_BLOCK_SIZE); in write_slab_journal_block() 1173 slab->active_count -= vio->io_size / VDO_BLOCK_SIZE; in handle_io_error() 2304 unsigned int block_count = vio->io_size / VDO_BLOCK_SIZE; in finish_reference_block_load() 2308 for (i = 0; i < block_count; i++, block++, data += VDO_BLOCK_SIZE) { in finish_reference_block_load() 2348 REQ_OP_READ, block_count * VDO_BLOCK_SIZE); in load_reference_block_group() 2909 char *end_data = scrubber->vio.data + (end_index * VDO_BLOCK_SIZE); in apply_journal_entries() 2922 char *block_data = scrubber->vio.data + (index * VDO_BLOCK_SIZE); in apply_journal_entries() 3736 result = vdo_allocate(VDO_BLOCK_SIZE, char, "struct packed_slab_journal_block", in initialize_slab_journal() 4004 result = vdo_allocate(VDO_BLOCK_SIZE * slab_journal_size, in initialize_slab_scrubber() [all …]
|
| A D | vio.h | 21 MAX_BLOCKS_PER_VIO = (BIO_MAX_VECS << PAGE_SHIFT) / VDO_BLOCK_SIZE,
|
| A D | message-stats.c | 469 write_block_count_t("physicalSize : ", config->physical_blocks * VDO_BLOCK_SIZE, ", ", in vdo_write_config() 472 write_block_count_t("logicalSize : ", config->logical_blocks * VDO_BLOCK_SIZE, ", ", in vdo_write_config()
|
| A D | block-map.h | 208 char page_buffer[VDO_BLOCK_SIZE];
|
| A D | repair.c | 721 offset = ((vdo_get_recovery_journal_block_number(journal, sequence) * VDO_BLOCK_SIZE) + in get_sector() 1166 char *header = &data[pbn * VDO_BLOCK_SIZE]; in get_recovery_journal_block_header() 1732 result = vdo_allocate(remaining * VDO_BLOCK_SIZE, char, __func__, in vdo_repair() 1753 ptr += (blocks * VDO_BLOCK_SIZE); in vdo_repair()
|
| A D | vdo.c | 270 result = vdo_allocate(VDO_BLOCK_SIZE, u8, __func__, &block); in read_geometry_block() 725 result = vdo_allocate(VDO_BLOCK_SIZE, char, "encoded super block", in initialize_super_block() 1409 return ((size_t) vdo->device_config->cache_size) * VDO_BLOCK_SIZE; in get_block_map_cache_size() 1523 stats->block_size = VDO_BLOCK_SIZE; in get_vdo_statistics()
|
| A D | block-map.c | 160 return &cache->pages[(info - cache->infos) * VDO_BLOCK_SIZE]; in get_page_buffer() 219 u64 size = cache->page_count * (u64) VDO_BLOCK_SIZE; in allocate_cache_components() 227 result = vdo_allocate_memory(size, VDO_BLOCK_SIZE, "cache pages", &cache->pages); in allocate_cache_components() 746 memset(get_page_buffer(info), 0, VDO_BLOCK_SIZE); in handle_rebuild_read_error() 1393 memcpy(page, loaded, VDO_BLOCK_SIZE); in vdo_copy_valid_page() 1665 memcpy(vio->vio.data, tree_page->page_buffer, VDO_BLOCK_SIZE); in write_page() 2865 ((VDO_BLOCK_SIZE - sizeof(struct block_map_page)) / in vdo_decode_block_map()
|
| A D | recovery-journal.c | 666 ((VDO_BLOCK_SIZE - sizeof(struct packed_journal_header)) / in initialize_recovery_block() 673 result = vdo_allocate(VDO_BLOCK_SIZE, char, __func__, &data); in initialize_recovery_block() 965 memset(block->vio.data, 0x0, VDO_BLOCK_SIZE); in advance_tail()
|
| A D | encodings.h | 510 VDO_BLOCK_SIZE - sizeof(struct packed_slab_journal_block_header), 566 VDO_SLAB_SUMMARY_ENTRIES_PER_BLOCK = VDO_BLOCK_SIZE / sizeof(struct slab_summary_entry),
|
| A D | dedupe.c | 1126 for (i = 0; i < VDO_BLOCK_SIZE; i += sizeof(u64)) { in blocks_equal() 2190 geometry.bio_offset) * VDO_BLOCK_SIZE); in initialize_index() 2194 .size = (vdo_get_index_region_size(geometry) * VDO_BLOCK_SIZE), in initialize_index()
|
| A D | encodings.c | 341 memset(buffer, 0, VDO_BLOCK_SIZE); in vdo_format_block_map_page()
|