| /fs/netfs/ |
| A D | read_pgpriv2.c | 22 size_t fsize = folio_size(folio), flen = fsize; in netfs_pgpriv2_copy_folio() local 41 if (fpos + fsize > creq->i_size) in netfs_pgpriv2_copy_folio() 51 _debug("folio %zx %zx", flen, fsize); in netfs_pgpriv2_copy_folio() 61 cache->submit_extendable_to = fsize; in netfs_pgpriv2_copy_folio() 87 rolling_buffer_advance(&creq->buffer, fsize); in netfs_pgpriv2_copy_folio() 90 if (flen < fsize) in netfs_pgpriv2_copy_folio() 186 size_t fsize, flen; in netfs_pgpriv2_unlock_copied_folios() local 195 fsize = folio_size(folio); in netfs_pgpriv2_unlock_copied_folios() 196 flen = fsize; in netfs_pgpriv2_unlock_copied_folios() 208 creq->cleaned_to = fpos + fsize; in netfs_pgpriv2_unlock_copied_folios() [all …]
|
| A D | write_issue.c | 330 size_t fsize = folio_size(folio), flen = fsize, foff = 0; in netfs_write_folio() local 357 if (fpos + fsize > wreq->i_size) in netfs_write_folio() 375 folio_zero_segment(folio, flen, fsize); in netfs_write_folio() 509 if (fsize > iter_off) in netfs_write_folio() 511 atomic64_set(&wreq->issued_to, fpos + fsize); in netfs_write_folio() 516 if (foff + flen < fsize) in netfs_write_folio() 746 size_t fsize = folio_size(folio), flen; in netfs_write_folio_single() local 756 folio_zero_segment(folio, flen, fsize); in netfs_write_folio_single() 762 _debug("folio %zx/%zx", flen, fsize); in netfs_write_folio_single() 839 if (fsize > iter_off) in netfs_write_folio_single() [all …]
|
| A D | read_collect.c | 125 size_t fsize; in netfs_read_unlock_folios() local 138 fsize = PAGE_SIZE << order; in netfs_read_unlock_folios() 140 fend = umin(fpos + fsize, rreq->i_size); in netfs_read_unlock_folios() 149 WRITE_ONCE(rreq->cleaned_to, fpos + fsize); in netfs_read_unlock_folios() 168 if (fpos + fsize >= collected_to) in netfs_read_unlock_folios() 232 size_t fsize = PAGE_SIZE << rreq->front_folio_order; in netfs_collect_read_results() local 257 if (front->start + transferred >= rreq->cleaned_to + fsize || in netfs_collect_read_results() 479 size_t fsize = PAGE_SIZE << rreq->front_folio_order; in netfs_read_subreq_progress() local 486 if (subreq->start + subreq->transferred > rreq->cleaned_to + fsize && in netfs_read_subreq_progress()
|
| A D | write_collect.c | 141 size_t fsize, flen; in netfs_writeback_unlock_folios() local 150 fsize = folio_size(folio); in netfs_writeback_unlock_folios() 152 flen = finfo ? finfo->dirty_offset + finfo->dirty_len : fsize; in netfs_writeback_unlock_folios() 163 wreq->cleaned_to = fpos + fsize; in netfs_writeback_unlock_folios() 179 if (fpos + fsize >= collected_to) in netfs_writeback_unlock_folios()
|
| /fs/afs/ |
| A D | dir_search.c | 95 size_t fsize = folioq_folio_size(fq, slot); in afs_dir_find_block() local 97 if (blend <= fpos + fsize) { in afs_dir_find_block() 108 fpos += fsize; in afs_dir_find_block()
|
| A D | dir_edit.c | 138 size_t fsize = folioq_folio_size(fq, s); in afs_dir_get_block() local 140 if (blend <= fpos + fsize) { in afs_dir_get_block() 150 fpos += fsize; in afs_dir_get_block()
|
| /fs/nfs/ |
| A D | read.c | 292 size_t fsize = folio_size(folio); in nfs_read_add_folio() local 302 aligned_len = min_t(unsigned int, ALIGN(len, rsize), fsize); in nfs_read_add_folio() 312 if (len < fsize) in nfs_read_add_folio() 313 folio_zero_segment(folio, len, fsize); in nfs_read_add_folio()
|
| A D | file.c | 401 size_t fsize = folio_size(folio); in nfs_write_end() local 406 folio_zero_segments(folio, 0, offset, end, fsize); in nfs_write_end() 409 folio_zero_segment(folio, end, fsize); in nfs_write_end() 413 folio_zero_segment(folio, pglen, fsize); in nfs_write_end()
|
| A D | pnfs.c | 1983 loff_t fsize = i_size_read(ino); in pnfs_within_mdsthreshold() local 1995 dprintk("%s fsize %llu\n", __func__, fsize); in pnfs_within_mdsthreshold() 1997 if (fsize < t->rd_sz) in pnfs_within_mdsthreshold() 2010 dprintk("%s fsize %llu\n", __func__, fsize); in pnfs_within_mdsthreshold() 2012 if (fsize < t->wr_sz) in pnfs_within_mdsthreshold()
|
| /fs/btrfs/ |
| A D | file.c | 1865 size_t fsize = folio_size(folio); in btrfs_page_mkwrite() local 1873 reserved_space = fsize; in btrfs_page_mkwrite() 1942 ordered = btrfs_lookup_ordered_range(inode, page_start, fsize); in btrfs_page_mkwrite() 1954 if (reserved_space < fsize) { in btrfs_page_mkwrite() 1955 const u64 to_free = fsize - reserved_space; in btrfs_page_mkwrite() 1987 zero_start = fsize; in btrfs_page_mkwrite() 1989 if (zero_start != fsize) in btrfs_page_mkwrite() 1992 btrfs_folio_clear_checked(fs_info, folio, page_start, fsize); in btrfs_page_mkwrite() 2005 btrfs_delalloc_release_extents(inode, fsize); in btrfs_page_mkwrite() 2016 btrfs_delalloc_release_extents(inode, fsize); in btrfs_page_mkwrite()
|
| A D | subpage.h | 123 size_t fsize, enum btrfs_folio_type type);
|
| A D | subpage.c | 117 size_t fsize, enum btrfs_folio_type type) in btrfs_alloc_folio_state() argument 122 ASSERT(fs_info->sectorsize < fsize); in btrfs_alloc_folio_state() 126 (fsize >> fs_info->sectorsize_bits))); in btrfs_alloc_folio_state()
|
| /fs/iomap/ |
| A D | buffered-io.c | 785 size_t fsize = folio_size(folio); in iomap_trim_folio_range() local 788 WARN_ON_ONCE(pos >= folio_pos(folio) + fsize); in iomap_trim_folio_range() 791 *bytes = min(*bytes, fsize - *offset); in iomap_trim_folio_range()
|
| /fs/smb/client/ |
| A D | smbdirect.c | 2521 size_t fsize = folioq_folio_size(folioq, slot); in smb_extract_folioq_to_rdma() local 2523 if (offset < fsize) { in smb_extract_folioq_to_rdma() 2524 size_t part = umin(maxsize, fsize - offset); in smb_extract_folioq_to_rdma() 2534 if (offset >= fsize) { in smb_extract_folioq_to_rdma()
|
| A D | smb2ops.c | 4611 size_t fsize = folio_size(folio); in cifs_copy_folioq_to_iter() local 4612 size_t n, len = umin(fsize - skip, data_size); in cifs_copy_folioq_to_iter()
|
| /fs/ufs/ |
| A D | util.h | 316 ((begin) & ((uspi->fsize>>1) - 1)))
|
| /fs/ceph/ |
| A D | inode.c | 1115 u64 fsize = __le64_to_cpu(*(__le64 *)iinfo->fscrypt_file); in ceph_fill_inode() local 1117 if (size == round_up(fsize, CEPH_FSCRYPT_BLOCK_SIZE)) { in ceph_fill_inode() 1118 size = fsize; in ceph_fill_inode()
|