Lines Matching refs:SECTOR_SHIFT
125 #define JOURNAL_SECTOR_DATA ((1 << SECTOR_SHIFT) - sizeof(commit_id_t))
439 ms += offset >> (SECTOR_SHIFT + ic->log2_buffer_sectors - ic->log2_tag_size); in get_metadata_sector_and_offset()
440 mo = (offset << ic->log2_tag_size) & ((1U << SECTOR_SHIFT << ic->log2_buffer_sectors) - 1); in get_metadata_sector_and_offset()
442 ms += (__u64)offset * ic->tag_size >> (SECTOR_SHIFT + ic->log2_buffer_sectors); in get_metadata_sector_and_offset()
443 mo = (offset * ic->tag_size) & ((1U << SECTOR_SHIFT << ic->log2_buffer_sectors) - 1); in get_metadata_sector_and_offset()
496 __u8 *mac = sb + (1 << SECTOR_SHIFT) - mac_size; in sb_mac()
498 if (sizeof(struct superblock) + mac_size > 1 << SECTOR_SHIFT || in sb_mac()
727 *pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT); in page_list_location()
728 *pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); in page_list_location()
740 *n_sectors = (PAGE_SIZE - pl_offset) >> SECTOR_SHIFT; in access_page_list()
884 size_t n_bytes = (size_t)(n_sections * ic->journal_section_sectors) << SECTOR_SHIFT; in xor_journal()
1052 pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT); in rw_journal_sectors()
1053 pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); in rw_journal_sectors()
1176 pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT); in copy_from_journal()
1177 pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); in copy_from_journal()
1423 to_copy = min((1U << SECTOR_SHIFT << ic->log2_buffer_sectors) - *metadata_offset, total_size); in dm_integrity_rw_tag()
1468 if (unlikely(*metadata_offset == 1U << SECTOR_SHIFT << ic->log2_buffer_sectors)) { in dm_integrity_rw_tag()
1614 bio_advance(bio, dio->range.n_sectors << SECTOR_SHIFT); in dec_in_flight()
1667 r = crypto_shash_update(req, data, ic->sectors_per_block << SECTOR_SHIFT); in integrity_sector_checksum()
1727 alignment = dio->range.logical_sector | bio_sectors(bio) | (PAGE_SIZE >> SECTOR_SHIFT); in integrity_recheck()
1731 buffer += (sector - io_loc.sector) << SECTOR_SHIFT; in integrity_recheck()
1757 memcpy(mem + pos, buffer, ic->sectors_per_block << SECTOR_SHIFT); in integrity_recheck()
1760 pos += ic->sectors_per_block << SECTOR_SHIFT; in integrity_recheck()
1791 …checksums = kmalloc((PAGE_SIZE >> SECTOR_SHIFT >> ic->sb->log2_sectors_per_block) * ic->tag_size +… in integrity_metadata()
1812 unsigned int this_step_blocks = bi_size >> (SECTOR_SHIFT + ic->sb->log2_sectors_per_block); in integrity_metadata()
1823 bi_size -= this_step_blocks << (SECTOR_SHIFT + ic->sb->log2_sectors_per_block); in integrity_metadata()
1847 pos += ic->sectors_per_block << SECTOR_SHIFT; in integrity_metadata()
1931 if (unlikely(bv.bv_len & ((ic->sectors_per_block << SECTOR_SHIFT) - 1))) { in dm_integrity_check_limits()
2039 if (unlikely(bv.bv_len >> SECTOR_SHIFT > n_sectors)) in __journal_read_write()
2040 bv.bv_len = n_sectors << SECTOR_SHIFT; in __journal_read_write()
2041 n_sectors -= bv.bv_len >> SECTOR_SHIFT; in __journal_read_write()
2072 mem_ptr += 1 << SECTOR_SHIFT; in __journal_read_write()
2119 memcpy(js, mem + bv.bv_offset, ic->sectors_per_block << SECTOR_SHIFT); in __journal_read_write()
2149 bv.bv_offset += ic->sectors_per_block << SECTOR_SHIFT; in __journal_read_write()
2150 } while (bv.bv_len -= ic->sectors_per_block << SECTOR_SHIFT); in __journal_read_write()
2363 bio->bi_iter.bi_size = dio->range.n_sectors << SECTOR_SHIFT; in dm_integrity_map_continue()
2514 bio_advance_iter_single(bio, &dio->bio_details.bi_iter, ic->sectors_per_block << SECTOR_SHIFT); in dm_integrity_map_inline()
2562 ic->sectors_per_block << SECTOR_SHIFT); in dm_integrity_inline_recheck()
2606 memcpy(mem, outgoing_data, ic->sectors_per_block << SECTOR_SHIFT); in dm_integrity_inline_recheck()
2609 bio_advance_iter_single(bio, &dio->bio_details.bi_iter, ic->sectors_per_block << SECTOR_SHIFT); in dm_integrity_inline_recheck()
2641 bio_advance_iter_single(bio, &dio->bio_details.bi_iter, ic->sectors_per_block << SECTOR_SHIFT); in dm_integrity_end_io()
3004 recalc_buffer = __vmalloc(recalc_sectors << SECTOR_SHIFT, GFP_NOIO); in integrity_recalc()
3100 integrity_sector_checksum(ic, logical_sector + i, recalc_buffer + (i << SECTOR_SHIFT), t); in integrity_recalc()
3163 recalc_buffer = kmalloc(recalc_sectors << SECTOR_SHIFT, GFP_NOIO | __GFP_NOWARN); in integrity_recalc_inline()
3210 range.n_sectors << SECTOR_SHIFT); in integrity_recalc_inline()
3221 integrity_sector_checksum(ic, range.logical_sector + i, recalc_buffer + (i << SECTOR_SHIFT), t); in integrity_recalc_inline()
3228 range.n_sectors << SECTOR_SHIFT); in integrity_recalc_inline()
3311 bbs->idx * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), in bitmap_block_work()
3312 BITMAP_BLOCK_SIZE >> SECTOR_SHIFT, NULL); in bitmap_block_work()
3357 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL); in bitmap_flush_work()
3663 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL); in dm_integrity_resume()
3667 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL); in dm_integrity_resume()
3679 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL); in dm_integrity_resume()
3698 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL); in dm_integrity_resume()
3742 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL); in dm_integrity_resume()
3812 DMEMIT(" block_size:%u", ic->sectors_per_block << SECTOR_SHIFT); in dm_integrity_status()
3863 DMEMIT(",block_size=%u", ic->sectors_per_block << SECTOR_SHIFT); in dm_integrity_status()
3898 limits->logical_block_size = ic->sectors_per_block << SECTOR_SHIFT; in dm_integrity_io_hints()
3899 limits->physical_block_size = ic->sectors_per_block << SECTOR_SHIFT; in dm_integrity_io_hints()
3900 limits->io_min = ic->sectors_per_block << SECTOR_SHIFT; in dm_integrity_io_hints()
3902 limits->discard_granularity = ic->sectors_per_block << SECTOR_SHIFT; in dm_integrity_io_hints()
3912 ic->sb->log2_sectors_per_block + SECTOR_SHIFT; in dm_integrity_io_hints()
3953 (__u64)(METADATA_PADDING_SECTORS << SECTOR_SHIFT) : in calculate_device_limits()
3954 (__u64)(1 << SECTOR_SHIFT << METADATA_PADDING_SECTORS); in calculate_device_limits()
3957 metadata_run_padding) >> SECTOR_SHIFT; in calculate_device_limits()
3970 meta_size = (meta_size + ((1U << (ic->log2_buffer_sectors + SECTOR_SHIFT)) - 1)) in calculate_device_limits()
3971 >> (ic->log2_buffer_sectors + SECTOR_SHIFT); in calculate_device_limits()
4008 memset(ic->sb, 0, SB_SECTORS << SECTOR_SHIFT); in initialize_superblock()
4162 end = end_offset + (1 << SECTOR_SHIFT); in dm_integrity_alloc_journal_scatterlist()
4256 PAGE_SIZE >> SECTOR_SHIFT) >> (PAGE_SHIFT - SECTOR_SHIFT); in create_journal()
4442 section_req->cryptlen = (size_t)ic->journal_section_sectors << SECTOR_SHIFT; in create_journal()
4626 if (val < 1 << SECTOR_SHIFT || in dm_integrity_ctr()
4627 val > MAX_SECTORS_PER_BLOCK << SECTOR_SHIFT || in dm_integrity_ctr()
4633 ic->sectors_per_block = val >> SECTOR_SHIFT; in dm_integrity_ctr()
4691 ic->log2_buffer_sectors = min((int)__fls(buffer_sectors), 31 - SECTOR_SHIFT); in dm_integrity_ctr()
4761 if (1 << bi->interval_exp != ic->sectors_per_block << SECTOR_SHIFT) { in dm_integrity_ctr()
4862 ic->sb = alloc_pages_exact(SB_SECTORS << SECTOR_SHIFT, GFP_KERNEL); in dm_integrity_ctr()
4877 if (memchr_inv(ic->sb, 0, SB_SECTORS << SECTOR_SHIFT)) { in dm_integrity_ctr()
4972 …bits_in_journal = ((__u64)ic->journal_section_sectors * ic->journal_sections) << (SECTOR_SHIFT + 3… in dm_integrity_ctr()
5049 1U << (SECTOR_SHIFT + ic->log2_buffer_sectors), 1, 0, NULL, NULL, 0); in dm_integrity_ctr()
5098 sector = i * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT); in dm_integrity_ctr()
5099 pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT); in dm_integrity_ctr()
5100 pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); in dm_integrity_ctr()
5216 free_pages_exact(ic->sb, SB_SECTORS << SECTOR_SHIFT); in dm_integrity_dtr()