| /fs/gfs2/ |
| A D | aops.c | 105 struct inode *inode = folio->mapping->host; in __gfs2_jdata_write_folio() 129 struct inode *inode = mapping->host; in gfs2_jdata_writeback() 163 .inode = mapping->host, in gfs2_writepages() 196 struct inode *inode = mapping->host; in gfs2_write_jdata_batch() 219 if (unlikely(folio->mapping != mapping)) { in gfs2_write_jdata_batch() 352 mapping->writeback_index = done_index; in gfs2_write_cache_jdata() 372 ret = gfs2_write_cache_jdata(mapping, wbc); in gfs2_jdata_writepages() 424 struct inode *inode = folio->mapping->host; in gfs2_read_folio() 498 struct inode *inode = rac->mapping->host; in gfs2_readahead() 554 return block_dirty_folio(mapping, folio); in gfs2_jdata_dirty_folio() [all …]
|
| A D | meta_io.c | 89 static int gfs2_aspace_writepages(struct address_space *mapping, in gfs2_aspace_writepages() argument 95 while ((folio = writeback_iter(mapping, wbc, folio, &error))) in gfs2_aspace_writepages() 128 struct address_space *mapping = gfs2_glock2aspace(gl); in gfs2_getbuf() local 136 if (mapping == NULL) in gfs2_getbuf() 137 mapping = gfs2_aspace(sdp); in gfs2_getbuf() 144 folio = __filemap_get_folio(mapping, index, in gfs2_getbuf() 146 mapping_gfp_mask(mapping) | __GFP_NOFAIL); in gfs2_getbuf() 152 folio = __filemap_get_folio(mapping, index, in gfs2_getbuf() 346 struct address_space *mapping = bh->b_folio->mapping; in gfs2_remove_from_journal() local 347 struct gfs2_sbd *sdp = gfs2_mapping2sbd(mapping); in gfs2_remove_from_journal() [all …]
|
| /fs/ |
| A D | dax.c | 372 folio->mapping = NULL; in dax_folio_make_shared() 394 folio->mapping = NULL; in dax_folio_put() 408 new_folio->mapping = NULL; in dax_folio_put() 453 if (folio->mapping) in dax_associate_entry() 463 folio->mapping = mapping; in dax_associate_entry() 508 struct address_space *mapping = READ_ONCE(folio->mapping); in dax_lock_folio() local 511 if (!mapping || !dax_mapping(mapping)) in dax_lock_folio() 527 if (mapping != folio->mapping) { in dax_lock_folio() 549 struct address_space *mapping = folio->mapping; in dax_unlock_folio() local 764 if (!dax_mapping(mapping)) in dax_layout_busy_page_range() [all …]
|
| A D | buffer.c | 585 &mapping->i_private_list); in sync_mapping_buffers() 692 &mapping->i_private_list); in mark_buffer_dirty_inode() 693 bh->b_assoc_map = mapping; in mark_buffer_dirty_inode() 795 mapping = bh->b_assoc_map; in fsync_buffers_list() 802 bh->b_assoc_map = mapping; in fsync_buffers_list() 834 mapping = bh->b_assoc_map; in fsync_buffers_list() 842 bh->b_assoc_map = mapping; in fsync_buffers_list() 1204 mapping = folio->mapping; in mark_buffer_dirty() 1205 if (mapping) in mark_buffer_dirty() 1208 if (mapping) in mark_buffer_dirty() [all …]
|
| /fs/ecryptfs/ |
| A D | mmap.c | 43 mapping_set_error(mapping, error); in ecryptfs_writepages() 117 page_virt, folio->mapping->host); in ecryptfs_copy_up_encrypted_with_header() 165 struct inode *inode = folio->mapping->host; in ecryptfs_read_folio() 216 struct inode *inode = folio->mapping->host; in fill_zeros_to_end_of_page() 243 struct address_space *mapping, in ecryptfs_write_begin() argument 253 mapping_gfp_mask(mapping)); in ecryptfs_write_begin() 265 folio, index, 0, PAGE_SIZE, mapping->host); in ecryptfs_write_begin() 292 mapping->host); in ecryptfs_write_begin() 304 >= i_size_read(mapping->host)) { in ecryptfs_write_begin() 441 struct address_space *mapping, in ecryptfs_write_end() argument [all …]
|
| /fs/hpfs/ |
| A D | file.c | 171 static int hpfs_writepages(struct address_space *mapping, in hpfs_writepages() argument 174 return mpage_writepages(mapping, wbc, hpfs_get_block); in hpfs_writepages() 179 struct inode *inode = mapping->host; in hpfs_write_failed() 192 struct address_space *mapping, in hpfs_write_begin() argument 198 ret = cont_write_begin(iocb, mapping, pos, len, foliop, fsdata, in hpfs_write_begin() 200 &hpfs_i(mapping->host)->mmu_private); in hpfs_write_begin() 202 hpfs_write_failed(mapping, pos + len); in hpfs_write_begin() 208 struct address_space *mapping, in hpfs_write_end() argument 212 struct inode *inode = mapping->host; in hpfs_write_end() 216 hpfs_write_failed(mapping, pos + len); in hpfs_write_end() [all …]
|
| /fs/exfat/ |
| A D | inode.c | 415 struct address_space *mapping = rac->mapping; in exfat_readahead() local 416 struct inode *inode = mapping->host; in exfat_readahead() 440 struct inode *inode = mapping->host; in exfat_write_failed() 450 struct address_space *mapping, in exfat_write_begin() argument 462 exfat_write_failed(mapping, pos+len); in exfat_write_begin() 468 struct address_space *mapping, in exfat_write_end() argument 472 struct inode *inode = mapping->host; in exfat_write_end() 478 exfat_write_failed(mapping, pos+len); in exfat_write_end() 497 struct inode *inode = mapping->host; in exfat_direct_IO() 511 exfat_write_failed(mapping, size); in exfat_direct_IO() [all …]
|
| /fs/jfs/ |
| A D | inode.c | 267 static int jfs_writepages(struct address_space *mapping, in jfs_writepages() argument 270 return mpage_writepages(mapping, wbc, jfs_get_block); in jfs_writepages() 285 struct inode *inode = mapping->host; in jfs_write_failed() 294 struct address_space *mapping, in jfs_write_begin() argument 300 ret = block_write_begin(mapping, pos, len, foliop, jfs_get_block); in jfs_write_begin() 302 jfs_write_failed(mapping, pos + len); in jfs_write_begin() 308 struct address_space *mapping, in jfs_write_end() argument 316 jfs_write_failed(mapping, pos + len); in jfs_write_end() 322 return generic_block_bmap(mapping, block, jfs_get_block); in jfs_bmap() 328 struct address_space *mapping = file->f_mapping; in jfs_direct_IO() local [all …]
|
| A D | jfs_metapage.c | 171 rc = filemap_migrate_folio(mapping, dst, src, mode); in __metapage_migrate_folio() 244 rc = filemap_migrate_folio(mapping, dst, src, mode); in __metapage_migrate_folio() 395 mapping_set_error(folio->mapping, err); in last_write_complete() 428 struct inode *inode = folio->mapping->host; in metapage_write_folio() 566 struct inode *inode = folio->mapping->host; in metapage_read_folio() 689 struct address_space *mapping; in __get_metapage() local 718 mapping = inode->i_mapping; in __get_metapage() 722 folio = filemap_grab_folio(mapping, page_index); in __get_metapage() 805 struct address_space *mapping = folio->mapping; in metapage_write_one() local 827 ret = filemap_check_errors(mapping); in metapage_write_one() [all …]
|
| /fs/bfs/ |
| A D | file.c | 154 static int bfs_writepages(struct address_space *mapping, in bfs_writepages() argument 157 return mpage_writepages(mapping, wbc, bfs_get_block); in bfs_writepages() 165 static void bfs_write_failed(struct address_space *mapping, loff_t to) in bfs_write_failed() argument 167 struct inode *inode = mapping->host; in bfs_write_failed() 174 struct address_space *mapping, in bfs_write_begin() argument 180 ret = block_write_begin(mapping, pos, len, foliop, bfs_get_block); in bfs_write_begin() 182 bfs_write_failed(mapping, pos + len); in bfs_write_begin() 187 static sector_t bfs_bmap(struct address_space *mapping, sector_t block) in bfs_bmap() argument 189 return generic_block_bmap(mapping, block, bfs_get_block); in bfs_bmap()
|
| /fs/bcachefs/ |
| A D | fs-io-buffered.c | 40 struct address_space *mapping; member 50 *iter = (struct readpages_iter) { ractl->mapping }; in readpages_iter_init() 430 mapping_set_error(fi.folio->mapping, -EIO); in bch2_writepage_io_done() 699 mapping_gfp_mask(mapping)); in bch2_write_begin() 725 ret = bch2_read_single_folio(folio, mapping); in bch2_write_begin() 818 struct address_space *mapping, in __bch2_buffered_write() argument 838 mapping_gfp_mask(mapping), &fs); in __bch2_buffered_write() 846 ret = bch2_read_single_folio(f, mapping); in __bch2_buffered_write() 858 ret = bch2_read_single_folio(f, mapping); in __bch2_buffered_write() 904 if (mapping_writably_mapped(mapping)) in __bch2_buffered_write() [all …]
|
| /fs/nilfs2/ |
| A D | page.c | 47 struct address_space *mapping, in nilfs_grab_buffer() argument 56 folio = filemap_grab_folio(mapping, index); in nilfs_grab_buffer() 164 m = folio->mapping; in nilfs_folio_bug() 338 folio->mapping = NULL; in nilfs_copy_back_pages() 341 folio->mapping = dmap; in nilfs_copy_back_pages() 381 if (likely(folio->mapping == mapping)) in nilfs_clear_dirty_pages() 476 struct address_space *mapping = folio->mapping; in __nilfs_clear_folio_dirty() local 478 if (mapping) { in __nilfs_clear_folio_dirty() 479 xa_lock_irq(&mapping->i_pages); in __nilfs_clear_folio_dirty() 483 xa_unlock_irq(&mapping->i_pages); in __nilfs_clear_folio_dirty() [all …]
|
| A D | dir.c | 90 struct address_space *mapping, size_t from, size_t to) in nilfs_commit_chunk() argument 92 struct inode *dir = mapping->host; in nilfs_commit_chunk() 111 struct inode *dir = folio->mapping->host; in nilfs_check_folio() 189 struct address_space *mapping = dir->i_mapping; in nilfs_get_folio() local 408 struct address_space *mapping = folio->mapping; in nilfs_set_link() local 419 nilfs_commit_chunk(folio, mapping, from, to); in nilfs_set_link() 525 struct address_space *mapping = folio->mapping; in nilfs_delete_entry() local 526 struct inode *inode = mapping->host; in nilfs_delete_entry() 557 nilfs_commit_chunk(folio, mapping, from, to); in nilfs_delete_entry() 568 struct address_space *mapping = inode->i_mapping; in nilfs_make_empty() local [all …]
|
| /fs/hugetlbfs/ |
| A D | inode.c | 311 struct address_space *mapping, in hugetlbfs_write_begin() argument 393 struct address_space *mapping, in hugetlb_unmap_file_folio() argument 407 i_mmap_lock_write(mapping); in hugetlb_unmap_file_folio() 435 i_mmap_unlock_write(mapping); in hugetlb_unmap_file_folio() 444 i_mmap_lock_write(mapping); in hugetlb_unmap_file_folio() 512 struct address_space *mapping, in remove_inode_single_folio() argument 639 i_mmap_lock_write(mapping); in hugetlb_vmtruncate() 643 i_mmap_unlock_write(mapping); in hugetlb_vmtruncate() 648 struct address_space *mapping, in hugetlbfs_zero_partial_page() argument 692 i_mmap_lock_write(mapping); in hugetlbfs_punch_hole() [all …]
|
| /fs/nfs/ |
| A D | write.c | 182 struct address_space *mapping = folio->mapping; in nfs_folio_find_head_request() local 187 spin_lock(&mapping->i_private_lock); in nfs_folio_find_head_request() 229 nfs_zap_mapping(mapping->host, mapping); in nfs_set_pageerror() 240 struct address_space *mapping = folio->mapping; in nfs_mapping_set_error() local 242 filemap_set_wb_err(mapping, error); in nfs_mapping_set_error() 243 if (mapping->host) in nfs_mapping_set_error() 246 nfs_set_pageerror(mapping); in nfs_mapping_set_error() 761 struct address_space *mapping = folio->mapping; in nfs_inode_add_request() local 791 struct address_space *mapping = folio->mapping; in nfs_inode_remove_request() local 1337 struct address_space *mapping = folio->mapping; in nfs_update_folio() local [all …]
|
| A D | file.c | 347 struct address_space *mapping, in nfs_write_begin() argument 363 mapping_gfp_mask(mapping)); in nfs_write_begin() 384 struct address_space *mapping, in nfs_write_end() argument 423 NFS_I(mapping->host)->write_io += copied; in nfs_write_end() 426 nfs_wb_all(mapping->host); in nfs_write_end() 441 struct inode *inode = folio->mapping->host; in nfs_invalidate_folio() 478 struct address_space *mapping = folio->mapping; in nfs_check_dirty_writeback() local 485 nfsi = NFS_I(mapping->host); in nfs_check_dirty_writeback() 600 struct address_space *mapping; in nfs_vm_page_mkwrite() local 621 mapping = folio->mapping; in nfs_vm_page_mkwrite() [all …]
|
| /fs/netfs/ |
| A D | buffered_write.c | 44 if (mapping_large_folio_support(mapping)) in netfs_grab_folio_for_write() 47 return __filemap_get_folio(mapping, index, fgp_flags, in netfs_grab_folio_for_write() 48 mapping_gfp_mask(mapping)); in netfs_grab_folio_for_write() 111 struct address_space *mapping = inode->i_mapping; in netfs_perform_write() local 125 size_t max_chunk = mapping_max_folio_size(mapping); in netfs_perform_write() 130 wbc_attach_fdatawrite_inode(&wbc, mapping->host); in netfs_perform_write() 179 folio = netfs_grab_folio_for_write(mapping, pos, part); in netfs_perform_write() 220 if (mapping_writably_mapped(mapping)) in netfs_perform_write() 510 struct address_space *mapping = file->f_mapping; in netfs_page_mkwrite() local 522 if (folio->mapping != mapping) in netfs_page_mkwrite() [all …]
|
| A D | fscache_io.c | 164 struct address_space *mapping; member 173 void __fscache_clear_page_bits(struct address_space *mapping, in __fscache_clear_page_bits() argument 181 XA_STATE(xas, &mapping->i_pages, first); in __fscache_clear_page_bits() 200 fscache_clear_page_bits(wreq->mapping, wreq->start, wreq->len, in fscache_wreq_done() 210 struct address_space *mapping, in __fscache_write_to_cache() argument 229 wreq->mapping = mapping; in __fscache_write_to_cache() 251 iov_iter_xarray(&iter, ITER_SOURCE, &mapping->i_pages, start, len); in __fscache_write_to_cache() 261 fscache_clear_page_bits(mapping, start, len, cond); in __fscache_write_to_cache()
|
| A D | buffered_read.c | 400 struct address_space *mapping = folio->mapping; in netfs_read_gaps() local 402 struct netfs_inode *ctx = netfs_inode(mapping->host); in netfs_read_gaps() 497 struct address_space *mapping = folio->mapping; in netfs_read_folio() local 499 struct netfs_inode *ctx = netfs_inode(mapping->host); in netfs_read_folio() 509 rreq = netfs_alloc_request(mapping, file, in netfs_read_folio() 625 struct file *file, struct address_space *mapping, in netfs_write_begin() argument 636 mapping_gfp_mask(mapping)); in netfs_write_begin() 664 rreq = netfs_alloc_request(mapping, file, in netfs_write_begin() 720 struct address_space *mapping = folio->mapping; in netfs_prefetch_for_write() local 721 struct netfs_inode *ctx = netfs_inode(mapping->host); in netfs_prefetch_for_write() [all …]
|
| /fs/ubifs/ |
| A D | file.c | 104 struct inode *inode = folio->mapping->host; in do_readpage() 208 struct inode *inode = mapping->host; in write_begin_slow() 235 mapping_gfp_mask(mapping)); in write_begin_slow() 408 struct address_space *mapping, in ubifs_write_begin() argument 412 struct inode *inode = mapping->host; in ubifs_write_begin() 428 mapping_gfp_mask(mapping)); in ubifs_write_begin() 523 struct inode *inode = mapping->host; in ubifs_write_end() 561 filemap_dirty_folio(mapping, folio); in ubifs_write_end() 700 struct address_space *mapping = folio1->mapping; in ubifs_do_bulk_read() local 701 struct inode *inode = mapping->host; in ubifs_do_bulk_read() [all …]
|
| /fs/ceph/ |
| A D | addr.c | 83 struct inode *inode = mapping->host; in ceph_dirty_folio() 892 mapping_set_error(mapping, rc); in writepages_finish() 980 struct inode *inode = mapping->host; in is_forced_umount() 991 mapping_set_error(mapping, -EIO); in is_forced_umount() 1001 struct inode *inode = mapping->host; in ceph_define_write_size() 1071 struct inode *inode = mapping->host; in ceph_define_writeback_range() 1142 struct inode *inode = mapping->host; in ceph_check_page_before_write() 1148 if (unlikely(!folio_test_dirty(folio) || folio->mapping != mapping)) { in ceph_check_page_before_write() 1207 struct inode *inode = mapping->host; in ceph_allocate_page_array() 1646 if (is_forced_umount(mapping)) { in ceph_writepages_start() [all …]
|
| /fs/hfsplus/ |
| A D | bitmap.c | 24 struct address_space *mapping; in hfsplus_block_allocate() local 36 mapping = sbi->alloc_file->i_mapping; in hfsplus_block_allocate() 37 page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, NULL); in hfsplus_block_allocate() 81 page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, in hfsplus_block_allocate() 132 page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, in hfsplus_block_allocate() 168 struct address_space *mapping; in hfsplus_block_free() local 183 mapping = sbi->alloc_file->i_mapping; in hfsplus_block_free() 185 page = read_mapping_page(mapping, pnr, NULL); in hfsplus_block_free() 219 page = read_mapping_page(mapping, ++pnr, NULL); in hfsplus_block_free()
|
| /fs/affs/ |
| A D | file.c | 387 struct inode *inode = mapping->host; in affs_write_failed() 400 struct inode *inode = mapping->host; in affs_direct_IO() 419 struct address_space *mapping, in affs_write_begin() argument 439 struct inode *inode = mapping->host; in affs_write_end() 651 struct address_space *mapping, in affs_write_begin_ofs() argument 655 struct inode *inode = mapping->host; in affs_write_begin_ofs() 673 mapping_gfp_mask(mapping)); in affs_write_begin_ofs() 691 struct address_space *mapping, in affs_write_end_ofs() argument 695 struct inode *inode = mapping->host; in affs_write_end_ofs() 893 res = mapping->a_ops->write_begin(NULL, mapping, isize, 0, &folio, &fsdata); in affs_truncate() [all …]
|
| /fs/iomap/ |
| A D | fiemap.c | 94 iomap_bmap(struct address_space *mapping, sector_t bno, in iomap_bmap() argument 98 .inode = mapping->host, in iomap_bmap() 99 .pos = (loff_t)bno << mapping->host->i_blkbits, in iomap_bmap() 100 .len = i_blocksize(mapping->host), in iomap_bmap() 103 const unsigned int blkshift = mapping->host->i_blkbits - SECTOR_SHIFT; in iomap_bmap() 106 if (filemap_write_and_wait(mapping)) in iomap_bmap()
|
| /fs/jffs2/ |
| A D | file.c | 25 struct address_space *mapping, 29 struct address_space *mapping, 110 int ret = jffs2_do_readpage_nolock(folio->mapping->host, folio); in __jffs2_read_folio() 117 struct jffs2_inode_info *f = JFFS2_INODE_INFO(folio->mapping->host); in jffs2_read_folio() 127 struct address_space *mapping, in jffs2_write_begin() argument 132 struct inode *inode = mapping->host; in jffs2_write_begin() 210 folio = __filemap_get_folio(mapping, index, FGP_WRITEBEGIN, in jffs2_write_begin() 211 mapping_gfp_mask(mapping)); in jffs2_write_begin() 242 struct address_space *mapping, in jffs2_write_end() argument 249 struct inode *inode = mapping->host; in jffs2_write_end() [all …]
|