| /fs/btrfs/ |
| A D | extent-io.c | 87 state->start = other->start; in merge_state() 99 other->start = state->start; in merge_state() 120 state->start = start; in insert_state() 138 prealloc->start = orig->start; in split_state() 269 if (state->start < start) { in clear_extent_bits() 281 start = state->start; in clear_extent_bits() 381 if (state->start < start) { in set_extent_bits() 395 start = state->start; in set_extent_bits() 406 if (state->start > start) { in set_extent_bits() 534 if (state->start != start) { in set_state_private() [all …]
|
| A D | extent-cache.c | 30 if (entry->start + entry->size <= range->start) in cache_tree_comp_range() 44 range.start = entry->start; in cache_tree_comp_nodes() 77 range.start = entry->start; in cache_tree_comp_nodes2() 96 pe->start = start; in alloc_cache_extent() 133 range.start = start; in lookup_cache_extent() 151 range.start = start; in lookup_cache_extent2() 168 range.start = start; in search_cache_extent() 189 range.start = start; in search_cache_extent2() 283 } else if (start <= cache->start) { in add_merge_cache_extent() 301 next->start = start; in add_merge_cache_extent() [all …]
|
| A D | extent-io.h | 54 #define BITMAP_FIRST_BYTE_MASK(start) \ argument 55 ((BYTE_MASK << ((start) & (BITS_PER_BYTE - 1))) & BYTE_MASK) 74 u64 start; member 83 u64 start; member 100 int find_first_extent_bit(struct extent_io_tree *tree, u64 start, 102 int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, 132 u64 start); 143 unsigned long start, unsigned long len); 145 unsigned long start, unsigned long len); 147 unsigned long start, unsigned long len); [all …]
|
| A D | extent-cache.h | 21 u64 start; member 37 struct cache_extent *search_cache_extent(struct cache_tree *tree, u64 start); 45 u64 start, u64 size); 52 int add_cache_extent(struct cache_tree *tree, u64 start, u64 size); 84 u64 objectid, u64 start); 92 u64 objectid, u64 start, u64 size); 102 int add_merge_cache_extent(struct cache_tree *tree, u64 start, u64 size);
|
| A D | volumes.c | 544 if (ce && ce->start <= logical && ce->start + ce->size > logical) { in read_one_chunk() 552 map->ce.start = logical; in read_one_chunk() 584 map->ce.start, map->ce.size, ret); in read_one_chunk() 861 if (ce->start > logical || ce->start + ce->size < logical) { in btrfs_num_copies() 865 (unsigned long long)ce->start, in btrfs_num_copies() 866 (unsigned long long)ce->start + ce->size); in btrfs_num_copies() 907 cur = ce->start; in btrfs_next_bg() 910 *logical = ce->start; in btrfs_next_bg() 978 if (ce->start > logical) { in __btrfs_map_block() 979 *length = ce->start - logical; in __btrfs_map_block() [all …]
|
| A D | disk-io.c | 41 if (buf->start != btrfs_header_bytenr(buf)) in check_tree_block() 92 fprintf(stderr, "bad tree block %llu, ", eb->start); in print_tree_block_error() 104 eb->start, btrfs_header_bytenr(eb)); in print_tree_block_error() 329 (unsigned long long)buf->start, in __csum_tree_block_size() 383 (unsigned long long)eb->start, in verify_parent_transid() 412 ret = btrfs_map_block(info, READ, eb->start + offset, in read_whole_eb() 415 printk("Couldn't map the block %Lu\n", eb->start + offset); in read_whole_eb() 475 num_copies = btrfs_num_copies(fs_info, eb->start, eb->len); in read_tree_block()
|
| A D | inode.c | 624 int start, int len, char *dest) in read_and_truncate_page() argument 628 u64 aligned_start = round_down(start, fs_info->sectorsize); in read_and_truncate_page() 631 int page_off = start - aligned_start; in read_and_truncate_page() 635 ASSERT(start + len <= aligned_start + fs_info->sectorsize); in read_and_truncate_page() 653 round_down(start, fs_info->sectorsize), in read_and_truncate_page()
|
| A D | ctree.c | 142 (unsigned long long)buf->start); in btrfs_check_leaf() 563 static int leaf_space_used(struct extent_buffer *l, int start, int nr) in leaf_space_used() argument 567 int end = min(nritems, start + nr) - 1; in leaf_space_used() 571 data_len = btrfs_item_end_nr(l, start); in leaf_space_used()
|
| /fs/cbfs/ |
| A D | cbfs.c | 29 void *start; member 115 attr = start + offset; in fill_node() 121 comp = start + offset; in fill_node() 165 start += align; in file_cbfs_next_file() 191 void *start; in file_cbfs_fill_cache() local 202 start = priv->start; in file_cbfs_fill_cache() 223 start += used; in file_cbfs_fill_cache() 291 priv->start = (void *)base; in cbfs_load_header_ptr() 454 start += used; in find_uncached() 465 void *start; in file_cbfs_find_uncached() local [all …]
|
| /fs/ |
| A D | fs_internal.c | 44 if (blk_dread(blk, partition->start + sector, 1, in fs_devread() 67 blk_dread(blk, partition->start + sector, 1, in fs_devread() 73 if (blk_dread(blk, partition->start + sector, in fs_devread() 86 if (blk_dread(blk, partition->start + sector, 1, in fs_devread() 124 if (blk_dread(blk, partition->start + sector, 1, in fs_devwrite() 134 if (blk_dwrite(blk, partition->start + sector, 1, in fs_devwrite() 151 if (blk_dread(blk, partition->start + sector, 1, in fs_devwrite() 159 if (blk_dwrite(blk, partition->start + sector, 1, in fs_devwrite() 168 if (blk_dwrite(blk, partition->start + sector, in fs_devwrite() 181 if (blk_dread(blk, partition->start + sector, 1, in fs_devwrite() [all …]
|
| /fs/squashfs/ |
| A D | sqfs.c | 127 if (exp_tbl > start && exp_tbl < end) in sqfs_frag_lookup() 133 start /= ctxt.cur_dev->blksz; in sqfs_frag_lookup() 142 if (sqfs_disk_read(start, n_blks, table) < 0) { in sqfs_frag_lookup() 157 start = start_block / ctxt.cur_dev->blksz; in sqfs_frag_lookup() 285 for (i = start; i < end; i++) { in sqfs_join() 716 u64 start, n_blks, table_offset, table_size; in sqfs_read_inode_table() local 806 u64 start, n_blks, table_offset, table_size; in sqfs_read_directory_table() local 833 if (sqfs_disk_read(start, n_blks, dtb) < 0) in sqfs_read_directory_table() 1497 data_offset = finfo.start; in sqfs_read_nest() 1582 start = lldiv(frag_entry.start, ctxt.cur_dev->blksz); in sqfs_read_nest() [all …]
|
| A D | sqfs_filesystem.h | 85 u32 start; member 240 u32 start; member 245 u64 start; member 289 u64 start; member
|
| /fs/ubifs/ |
| A D | misc.h | 73 ubifs_tnc_find_child(struct ubifs_znode *znode, int start) in ubifs_tnc_find_child() argument 75 while (start < znode->child_cnt) { in ubifs_tnc_find_child() 76 if (znode->zbranch[start].znode) in ubifs_tnc_find_child() 77 return znode->zbranch[start].znode; in ubifs_tnc_find_child() 78 start += 1; in ubifs_tnc_find_child()
|
| A D | log.c | 149 c->bud_bytes += c->leb_size - bud->start; in ubifs_add_bud() 230 bud->start = offs; in ubifs_add_bud_to_log() 235 ref->offs = cpu_to_le32(bud->start); in ubifs_add_bud_to_log() 251 if (bud->start == 0) { in ubifs_add_bud_to_log() 315 c->cmt_bud_bytes += wbuf->offs - bud->start; in remove_buds() 317 bud->lnum, bud->start, dbg_jhead(bud->jhead), in remove_buds() 318 wbuf->offs - bud->start, c->cmt_bud_bytes); in remove_buds() 319 bud->start = wbuf->offs; in remove_buds() 321 c->cmt_bud_bytes += c->leb_size - bud->start; in remove_buds() 324 c->leb_size - bud->start, c->cmt_bud_bytes); in remove_buds() [all …]
|
| A D | recovery.c | 500 int start) in fix_unclean_leb() argument 502 int lnum = sleb->lnum, endpt = start; in fix_unclean_leb() 518 lnum, start, sleb->endpt); in fix_unclean_leb() 531 lnum, start, sleb->endpt); in fix_unclean_leb() 539 if (start) { in fix_unclean_leb() 541 start, 1); in fix_unclean_leb() 635 int ret = 0, err, len = c->leb_size - offs, start = offs, min_io_unit; in ubifs_recover_leb() local 781 err = fix_unclean_leb(c, sleb, start); in ubifs_recover_leb()
|
| A D | replay.c | 106 if (b->bud->start == 0 && (lp->free != c->leb_size || lp->dirty != 0)) { in set_bud_lprops() 520 err = ubifs_leb_read(c, next->lnum, (char *)&data, next->start, 4, 1); in is_last_bud() 539 int err = 0, used = 0, lnum = b->bud->lnum, offs = b->bud->start; in replay_bud() 759 bud->start = offs; in add_replay_bud() 801 if (bud->jhead == jhead && bud->start <= offs) in validate_ref()
|
| A D | ubifs.h | 1312 int start; member 2349 int ubifs_fsync(struct file *file, loff_t start, loff_t end, int datasync);
|
| /fs/ext4/ |
| A D | dev.c | 44 part_offset = info->start; in ext4fs_set_blk_dev()
|
| A D | ext4_common.c | 1655 unsigned long long start; in read_allocated_block() local 1699 start = le16_to_cpu(extent[i].ee_start_hi); in read_allocated_block() 1700 start = (start << 32) + in read_allocated_block() 1704 return (fileblock - startblock) + start; in read_allocated_block()
|
| /fs/erofs/ |
| A D | fs.c | 29 int erofs_blk_read(void *buf, erofs_blk_t start, u32 nblocks) in erofs_blk_read() argument 31 return erofs_dev_read(0, buf, erofs_pos(start), in erofs_blk_read()
|
| A D | internal.h | 324 int erofs_blk_read(void *buf, erofs_blk_t start, u32 nblocks);
|
| /fs/exfat/ |
| A D | cluster.c | 114 static cluster_t find_bit_and_set(bitmap_t* bitmap, size_t start, size_t end) in find_bit_and_set() argument 116 const size_t start_index = start / sizeof(bitmap_t) / 8; in find_bit_and_set() 127 start_bitindex = MAX(i * sizeof(bitmap_t) * 8, start); in find_bit_and_set()
|
| /fs/cramfs/ |
| A D | cramfs.c | 47 flash_info[x->dev->id->num].start[0])
|
| /fs/fat/ |
| A D | fat_write.c | 202 if (cur_part_info.start + block + nr_blocks > in disk_write() 203 cur_part_info.start + total_sector) { in disk_write() 208 ret = blk_dwrite(cur_dev, cur_part_info.start + block, nr_blocks, buf); in disk_write() 895 dentptr->start = cpu_to_le16(start_cluster & 0xffff); in set_start_cluster()
|
| A D | fat.c | 60 ret = blk_dread(cur_dev, cur_part_info.start + block, nr_blocks, buf); in disk_read() 112 info.start = 0; in fat_register_device()
|