| /fs/xfs/scrub/ |
| A D | xfile.c | 148 len = min_t(ssize_t, count, folio_size(folio) - offset); in xfile_load() 209 len = min_t(ssize_t, count, folio_size(folio) - offset); in xfile_store() 279 if (len > folio_size(folio) - offset_in_folio(folio, pos)) { in xfile_get_folio() 308 trace_xfile_put_folio(xf, folio_pos(folio), folio_size(folio)); in xfile_put_folio()
|
| /fs/bcachefs/ |
| A D | fs-io-buffered.c | 148 BUG_ON(!bio_add_folio(bio, folio, folio_size(folio), 0)); in readpage_bio_extend() 314 BUG_ON(!bio_add_folio(&rbio->bio, folio, folio_size(folio), 0)); in bch2_readahead() 357 BUG_ON(!bio_add_folio(&rbio->bio, folio, folio_size(folio), 0)); in bch2_read_single_folio() 556 folio_size(folio)); in __bch2_writepage() 710 if (!offset && len == folio_size(folio)) in bch2_write_begin() 714 folio_zero_segment(folio, len, folio_size(folio)); in bch2_write_begin() 770 BUG_ON(offset + copied > folio_size(folio)); in bch2_write_end() 778 folio_zero_range(folio, 0, folio_size(folio)); in bch2_write_end() 856 folio_zero_range(f, 0, folio_size(f)); in __bch2_buffered_write() 920 f_copied != folio_size(f) && in __bch2_buffered_write() [all …]
|
| /fs/iomap/ |
| A D | buffered-io.c | 459 .len = folio_size(folio), in iomap_read_folio() 590 count = min(folio_size(folio) - from, count); in iomap_is_partially_uptodate() 630 folio_size(folio)); in iomap_release_folio() 653 if (offset == 0 && len == folio_size(folio)) { in iomap_invalidate_folio() 664 size_t len = folio_size(folio); in iomap_dirty_folio() 785 size_t fsize = folio_size(folio); in iomap_trim_folio_range() 1096 folio_pos(folio) + folio_size(folio) - 1); in iomap_write_delalloc_ifs_punch() 1128 folio_pos(folio) + folio_size(folio)); in iomap_write_delalloc_punch() 1168 start_byte = folio_pos(folio) + folio_size(folio); in iomap_write_delalloc_scan() 1655 folio_zero_segment(folio, poff, folio_size(folio)); in iomap_writeback_handle_eof() [all …]
|
| /fs/ |
| A D | mpage.c | 156 const unsigned blocks_per_folio = folio_size(folio) >> blkbits; in do_mpage_readpage() 269 folio_zero_segment(folio, first_hole << blkbits, folio_size(folio)); in do_mpage_readpage() 455 const unsigned blocks_per_folio = folio_size(folio) >> blkbits; in mpage_write_folio() 570 length = folio_size(folio); in mpage_write_folio() 581 folio_zero_segment(folio, length, folio_size(folio)); in mpage_write_folio() 605 wbc_account_cgroup_owner(wbc, folio, folio_size(folio)); in mpage_write_folio()
|
| A D | buffer.c | 933 offset = folio_size(folio); in folio_alloc_buffers() 1582 BUG_ON(offset >= folio_size(folio)); in folio_set_bh() 1642 BUG_ON(stop > folio_size(folio) || stop < length); in block_invalidate_folio() 1673 if (length == folio_size(folio)) in block_invalidate_folio() 2123 BUG_ON(to > folio_size(folio)); in __block_write_begin_int() 2357 to = min_t(unsigned, folio_size(folio) - from, count); in block_is_partially_uptodate() 2359 if (from < blocksize && to > folio_size(folio) - blocksize) in block_is_partially_uptodate() 2634 end = folio_size(folio); in block_page_mkwrite() 2735 if (folio_pos(folio) + folio_size(folio) <= i_size) in block_write_full_folio() 2752 folio_size(folio)); in block_write_full_folio()
|
| /fs/gfs2/ |
| A D | aops.c | 85 i_size < folio_pos(folio) + folio_size(folio)) in gfs2_write_jdata_folio() 87 folio_size(folio)); in gfs2_write_jdata_folio() 115 gfs2_trans_add_databufs(ip->i_gl, folio, 0, folio_size(folio)); in __gfs2_jdata_write_folio() 205 size += folio_size(fbatch->folios[i]); in gfs2_write_jdata_batch() 471 chunk = min(size - copied, folio_size(folio) - offset); in gfs2_internal_read() 614 int partial_page = (offset || length < folio_size(folio)); in gfs2_invalidate_folio()
|
| /fs/btrfs/ |
| A D | extent_io.c | 631 eb->folio_size = PAGE_SIZE; in alloc_eb_folio_array() 935 folio_size(folio) - zero_offset); in btrfs_do_readpage() 3217 if (folio_size(existing_folio) != eb->folio_size) { in attach_eb_folio_to_filemap() 3245 eb->folio_size = folio_size(eb->folios[i]); in attach_eb_folio_to_filemap() 3830 const int unit_size = eb->folio_size; in read_extent_buffer() 3870 const int unit_size = eb->folio_size; in read_extent_buffer_to_user_nofault() 3910 const int unit_size = eb->folio_size; in memcmp_extent_buffer() 3980 const int unit_size = eb->folio_size; in __write_extent_buffer() 4029 const int unit_size = eb->folio_size; in memset_extent_buffer() 4060 const int unit_size = src->folio_size; in copy_extent_buffer_full() [all …]
|
| A D | extent_io.h | 90 u32 folio_size; member 132 ASSERT(eb->folio_size); in offset_in_eb_folio() 133 return start & (eb->folio_size - 1); in offset_in_eb_folio()
|
| A D | accessors.c | 60 const int part = eb->folio_size - oif; \ 90 const int part = eb->folio_size - oif; \
|
| A D | misc.h | 165 return folio_pos(folio) + folio_size(folio); in folio_end()
|
| /fs/ext4/ |
| A D | readpage.c | 252 blocks_per_folio = folio_size(folio) >> blkbits; in ext4_mpage_readpages() 298 folio_size(folio)); in ext4_mpage_readpages() 332 folio_size(folio)); in ext4_mpage_readpages()
|
| /fs/ubifs/ |
| A D | file.c | 119 folio_zero_range(folio, 0, folio_size(folio)); in do_readpage() 242 if (pos == folio_pos(folio) && len >= folio_size(folio)) in write_begin_slow() 434 if (pos == folio_pos(folio) && len >= folio_size(folio)) { in ubifs_write_begin() 555 if (len == folio_size(folio)) in ubifs_write_end() 608 folio_zero_range(folio, 0, folio_size(folio)); in populate_page() 988 int err, len = folio_size(folio); in ubifs_writepage() 1030 folio_zero_segment(folio, len, folio_size(folio)); in ubifs_writepage() 1291 if (offset || length < folio_size(folio)) in ubifs_invalidate_folio()
|
| /fs/netfs/ |
| A D | buffered_read.c | 408 size_t flen = folio_size(folio); in netfs_read_gaps() 510 folio_pos(folio), folio_size(folio), in netfs_read_folio() 563 size_t plen = folio_size(folio); in netfs_skip_folio_read() 665 folio_pos(folio), folio_size(folio), in netfs_write_begin() 723 size_t flen = folio_size(folio); in netfs_prefetch_for_write()
|
| A D | read_pgpriv2.c | 22 size_t fsize = folio_size(folio), flen = fsize; in netfs_pgpriv2_copy_folio() 195 fsize = folio_size(folio); in netfs_pgpriv2_unlock_copied_folios()
|
| A D | misc.c | 215 size_t flen = folio_size(folio); in netfs_invalidate_folio() 301 end = umin(folio_pos(folio) + folio_size(folio), i_size_read(&ctx->inode)); in netfs_release_folio()
|
| /fs/nfs/ |
| A D | read.c | 57 folio_zero_segment(folio, 0, folio_size(folio)); in nfs_return_empty_folio() 292 size_t fsize = folio_size(folio); in nfs_read_add_folio() 371 size_t len = folio_size(folio); in nfs_read_folio()
|
| /fs/hfs/ |
| A D | btree.c | 86 folio_zero_range(folio, 0, folio_size(folio)); in hfs_btree_open() 91 size = folio_size(folio); in hfs_btree_open() 102 len = min_t(size_t, folio_size(folio), sb->s_blocksize); in hfs_btree_open()
|
| /fs/vboxsf/ |
| A D | file.c | 281 u32 nwrite = folio_size(folio); in vboxsf_writepages() 332 if (!folio_test_uptodate(folio) && nwritten == folio_size(folio)) in vboxsf_write_end()
|
| /fs/ceph/ |
| A D | addr.c | 149 if (offset != 0 || length != folio_size(folio)) { in ceph_invalidate_folio() 728 loff_t len = folio_size(folio); in write_folio_nounlock() 765 folio_invalidate(folio, 0, folio_size(folio)); in write_folio_nounlock() 812 WARN_ON_ONCE(len > folio_size(folio)); in write_folio_nounlock() 1174 folio_invalidate(folio, 0, folio_size(folio)); in ceph_check_page_before_write() 1380 ceph_wbc->len += folio_size(folio); in ceph_process_folio_batch() 2073 if (off + folio_size(folio) <= size) in ceph_page_mkwrite() 2074 len = folio_size(folio); in ceph_page_mkwrite() 2241 if (len > folio_size(folio)) in ceph_uninline_data() 2242 len = folio_size(folio); in ceph_uninline_data()
|
| /fs/orangefs/ |
| A D | inode.c | 45 wlen = folio_size(folio); in orangefs_writepage_locked() 95 folio_size(ow->folios[i]) - start, start); in orangefs_writepages_work() 273 bvec_set_folio(&bv, folio, folio_size(folio), 0); in orangefs_read_folio() 274 iov_iter_bvec(&iter, ITER_DEST, &bv, 1, folio_size(folio)); in orangefs_read_folio() 277 folio_size(folio), inode->i_size, NULL, NULL, file); in orangefs_read_folio()
|
| /fs/nilfs2/ |
| A D | dir.c | 115 size_t limit = folio_size(folio); in nilfs_check_folio() 456 kaddr += folio_size(folio) - reclen; in nilfs_add_link() 527 char *kaddr = (char *)((unsigned long)dir & ~(folio_size(folio) - 1)); in nilfs_delete_entry()
|
| /fs/f2fs/ |
| A D | data.c | 1013 folio_size(fio->folio)); in f2fs_submit_page_write() 1261 folio_zero_segment(folio, 0, folio_size(folio)); in f2fs_get_read_data_folio() 1372 folio_zero_segment(folio, 0, folio_size(folio)); in f2fs_get_new_data_folio() 2125 folio_zero_segment(folio, 0, folio_size(folio)); in f2fs_read_single_page() 2216 folio_zero_segment(folio, 0, folio_size(folio)); in f2fs_read_multi_pages() 2450 folio_zero_segment(folio, 0, folio_size(folio)); in f2fs_mpage_readpages() 2846 folio_zero_segment(folio, offset, folio_size(folio)); in f2fs_write_single_data_page() 3624 if (len == folio_size(folio) || folio_test_uptodate(folio)) in f2fs_write_begin() 3629 folio_zero_segment(folio, len, folio_size(folio)); in f2fs_write_begin() 3634 folio_zero_segment(folio, 0, folio_size(folio)); in f2fs_write_begin() [all …]
|
| /fs/ecryptfs/ |
| A D | mmap.c | 172 folio_size(folio), inode); in ecryptfs_read_folio() 188 folio->index, 0, folio_size(folio), in ecryptfs_read_folio()
|
| /fs/ufs/ |
| A D | dir.c | 114 unsigned limit = folio_size(folio); in ufs_check_folio() 329 kaddr += folio_size(folio) - reclen; in ufs_add_link() 552 memset(kaddr, 0, folio_size(folio)); in ufs_make_empty()
|
| /fs/erofs/ |
| A D | fscache.c | 230 folio_pos(folio), folio_size(folio)); in erofs_fscache_meta_read_folio() 242 folio_pos(folio), folio_size(folio)); in erofs_fscache_meta_read_folio() 344 folio_pos(folio), folio_size(folio)); in erofs_fscache_read_folio()
|