| /fs/btrfs/ |
| A D | fiemap.c | 179 const u64 range_end = offset + len; in emit_fiemap_extent() local 209 if (range_end <= cache_end) in emit_fiemap_extent() 216 len = range_end - cache_end; in emit_fiemap_extent() 643 u64 range_end; in extent_fiemap() local 661 range_end = round_up(start + len, sectorsize); in extent_fiemap() 684 while (prev_extent_end < range_end) { in extent_fiemap() 714 const u64 hole_end = min(key.offset, range_end) - 1; in extent_fiemap() 729 if (key.offset >= range_end) { in extent_fiemap() 813 if (!stopped && prev_extent_end < range_end) { in extent_fiemap() 816 0, 0, 0, prev_extent_end, range_end - 1); in extent_fiemap() [all …]
|
| A D | compression.h | 79 static inline u32 btrfs_calc_input_length(struct folio *folio, u64 range_end, u64 cur) in btrfs_calc_input_length() argument 84 return min(range_end, folio_end(folio)) - cur; in btrfs_calc_input_length()
|
| A D | reflink.c | 63 const u64 range_end = file_offset + block_size - 1; in copy_inline_to_page() local 96 btrfs_clear_extent_bit(&inode->io_tree, file_offset, range_end, in copy_inline_to_page() 98 ret = btrfs_set_extent_delalloc(inode, file_offset, range_end, 0, NULL); in copy_inline_to_page()
|
| A D | extent_map.c | 73 static u64 range_end(u64 start, u64 len) in range_end() function 97 u64 end = range_end(em->start, em->len); in tree_insert() 517 u64 end = range_end(start, len); in lookup_extent_mapping()
|
| A D | ordered-data.c | 761 u64 range_end; in btrfs_wait_ordered_extents() local 770 range_end = range_start + range_len; in btrfs_wait_ordered_extents() 779 if (range_end <= ordered->disk_bytenr || in btrfs_wait_ordered_extents()
|
| A D | inode.c | 1065 .range_end = end, in submit_uncompressed_range() 7274 u64 range_end; in can_nocow_extent() local 7490 u64 range_end; in btrfs_invalidate_folio() local 7497 range_end = page_end; in btrfs_invalidate_folio() 7512 range_end = ordered->file_offset - 1; in btrfs_invalidate_folio() 7519 ASSERT(range_end + 1 - cur < U32_MAX); in btrfs_invalidate_folio() 7520 range_len = range_end + 1 - cur; in btrfs_invalidate_folio() 7559 cur, range_end + 1 - cur)) { in btrfs_invalidate_folio() 7591 cur = range_end + 1; in btrfs_invalidate_folio() 8751 .range_end = LLONG_MAX, in btrfs_start_delalloc_snapshot() [all …]
|
| A D | extent_io.c | 1171 const u64 range_end = min(end, ordered->file_offset + ordered->num_bytes - 1); in can_skip_ordered_extent() local 1174 while (cur < range_end) { in can_skip_ordered_extent() 2204 end = (wbc->range_end >> fs_info->nodesize_bits); in btree_write_cache_pages() 2355 end = wbc->range_end >> PAGE_SHIFT; in extent_write_cache_pages() 2356 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in extent_write_cache_pages()
|
| A D | tree-log.c | 2338 u64 range_end; in replay_dir_deletes() local 2365 range_end = 0; in replay_dir_deletes() 2368 range_end = (u64)-1; in replay_dir_deletes() 2371 &range_start, &range_end); in replay_dir_deletes() 2402 if (found_key.offset > range_end) in replay_dir_deletes() 2415 if (range_end == (u64)-1) in replay_dir_deletes() 2417 range_start = range_end + 1; in replay_dir_deletes()
|
| A D | extent-tree.c | 6521 u64 range_end = U64_MAX; in btrfs_trim_fs() local 6539 check_add_overflow(range->start, range->len, &range_end)) in btrfs_trim_fs() 6544 if (cache->start >= range_end) { in btrfs_trim_fs() 6550 end = min(range_end, cache->start + cache->length); in btrfs_trim_fs()
|
| /fs/ocfs2/ |
| A D | file.c | 889 *range_end = 0; in ocfs2_zero_extend_get_range() 942 (unsigned long long)range_end); in ocfs2_zero_extend_range() 945 while (zero_pos < range_end) { in ocfs2_zero_extend_range() 947 if (next_pos > range_end) in ocfs2_zero_extend_range() 948 next_pos = range_end; in ocfs2_zero_extend_range() 981 &range_end); in ocfs2_zero_extend() 986 if (!range_end) in ocfs2_zero_extend() 991 if (range_end > zero_to_size) in ocfs2_zero_extend() 992 range_end = zero_to_size; in ocfs2_zero_extend() 995 range_end, di_bh); in ocfs2_zero_extend() [all …]
|
| A D | alloc.h | 216 u64 range_start, u64 range_end);
|
| A D | alloc.c | 6961 u64 range_start, u64 range_end) in ocfs2_zero_range_for_truncate() argument 6981 range_end = min_t(u64, range_end, i_size_read(inode)); in ocfs2_zero_range_for_truncate() 6982 if (range_start >= range_end) in ocfs2_zero_range_for_truncate() 7008 ret = ocfs2_grab_eof_folios(inode, range_start, range_end, folios, in ocfs2_zero_range_for_truncate() 7015 ocfs2_zero_cluster_folios(inode, range_start, range_end, folios, in ocfs2_zero_range_for_truncate() 7024 range_end - 1); in ocfs2_zero_range_for_truncate()
|
| /fs/nfs/ |
| A D | nfstrace.h | 283 loff_t range_end 286 TP_ARGS(inode, range_start, range_end), 294 __field(loff_t, range_end) 305 __entry->range_end = range_end; 314 __entry->range_start, __entry->range_end 323 loff_t range_end \ 325 TP_ARGS(inode, range_start, range_end))
|
| A D | write.c | 2078 .range_end = range_start + len - 1, in nfs_wb_folio()
|
| /fs/iomap/ |
| A D | buffered-io.c | 88 struct iomap_folio_state *ifs, u64 *range_start, u64 range_end) in ifs_find_dirty_range() argument 94 offset_in_folio(folio, range_end) >> inode->i_blkbits, in ifs_find_dirty_range() 113 u64 range_end) in iomap_find_dirty_range() argument 117 if (*range_start >= range_end) in iomap_find_dirty_range() 121 return ifs_find_dirty_range(folio, ifs, range_start, range_end); in iomap_find_dirty_range() 122 return range_end - *range_start; in iomap_find_dirty_range()
|
| /fs/gfs2/ |
| A D | aops.c | 309 end = wbc->range_end >> PAGE_SHIFT; in gfs2_write_cache_jdata() 310 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in gfs2_write_cache_jdata()
|
| A D | log.c | 248 .range_end = LLONG_MAX, in gfs2_ail1_start()
|
| /fs/bcachefs/ |
| A D | fsck.c | 2900 u64 range_start, u64 range_end, u64 inum, u32 snapshot) in inc_link() argument 2906 if (inum < range_start || inum >= range_end) in inc_link() 2974 u64 range_start, u64 range_end) in check_nlinks_walk_dirents() argument 2994 inc_link(c, &s, links, range_start, range_end, in check_nlinks_walk_dirents() 3009 size_t *idx, u64 range_end) in check_nlinks_update_inode() argument 3015 if (k.k->p.offset >= range_end) in check_nlinks_update_inode() 3052 u64 range_start, u64 range_end) in check_nlinks_update_hardlinks() argument 3061 check_nlinks_update_inode(trans, &iter, k, links, &idx, range_end))); in check_nlinks_update_hardlinks()
|
| /fs/9p/ |
| A D | vfs_file.c | 493 .range_end = (loff_t)vma->vm_pgoff * PAGE_SIZE + in v9fs_mmap_vm_close()
|
| /fs/netfs/ |
| A D | buffered_write.c | 118 .range_end = iocb->ki_pos + iter->count, in netfs_perform_write()
|
| /fs/ |
| A D | fs-writeback.c | 1890 .range_end = LLONG_MAX, in writeback_sb_inodes() 2857 .range_end = LLONG_MAX, in write_inode_now()
|
| /fs/afs/ |
| A D | inode.c | 756 .range_end = LLONG_MAX, in afs_evict_inode()
|
| /fs/ceph/ |
| A D | addr.c | 1100 ceph_wbc->end = wbc->range_end >> PAGE_SHIFT; in ceph_define_writeback_range() 1101 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in ceph_define_writeback_range()
|
| /fs/nilfs2/ |
| A D | inode.c | 173 wbc->range_end); in nilfs_writepages()
|
| /fs/ext4/ |
| A D | inode.c | 2843 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in ext4_do_writepages() 2854 mpd->end_pos = wbc->range_end; in ext4_do_writepages() 3044 .range_end = jinode->i_dirty_end, in ext4_normal_submit_inode_data_buffers()
|