Lines Matching refs:mapping

2046 int balance_dirty_pages_ratelimited_flags(struct address_space *mapping,  in balance_dirty_pages_ratelimited_flags()  argument
2049 struct inode *inode = mapping->host; in balance_dirty_pages_ratelimited_flags()
2116 void balance_dirty_pages_ratelimited(struct address_space *mapping) in balance_dirty_pages_ratelimited() argument
2118 balance_dirty_pages_ratelimited_flags(mapping, 0); in balance_dirty_pages_ratelimited()
2383 void tag_pages_for_writeback(struct address_space *mapping, in tag_pages_for_writeback() argument
2386 XA_STATE(xas, &mapping->i_pages, start); in tag_pages_for_writeback()
2405 static bool folio_prepare_writeback(struct address_space *mapping, in folio_prepare_writeback() argument
2415 if (unlikely(folio->mapping != mapping)) in folio_prepare_writeback()
2451 static struct folio *writeback_get_folio(struct address_space *mapping, in writeback_get_folio() argument
2461 filemap_get_folios_tag(mapping, &wbc->index, wbc_end(wbc), in writeback_get_folio()
2469 if (unlikely(!folio_prepare_writeback(mapping, wbc, folio))) { in writeback_get_folio()
2474 trace_wbc_writepage(wbc, inode_to_bdi(mapping->host)); in writeback_get_folio()
2504 struct folio *writeback_iter(struct address_space *mapping, in writeback_iter() argument
2519 wbc->index = mapping->writeback_index; in writeback_iter()
2535 tag_pages_for_writeback(mapping, wbc->index, in writeback_iter()
2562 folio = writeback_get_folio(mapping, wbc); in writeback_iter()
2575 mapping->writeback_index = 0; in writeback_iter()
2587 mapping->writeback_index = folio_next_index(folio); in writeback_iter()
2604 int write_cache_pages(struct address_space *mapping, in write_cache_pages() argument
2611 while ((folio = writeback_iter(mapping, wbc, folio, &error))) { in write_cache_pages()
2623 int do_writepages(struct address_space *mapping, struct writeback_control *wbc) in do_writepages() argument
2630 wb = inode_to_wb_wbc(mapping->host, wbc); in do_writepages()
2633 if (mapping->a_ops->writepages) in do_writepages()
2634 ret = mapping->a_ops->writepages(mapping, wbc); in do_writepages()
2664 bool noop_dirty_folio(struct address_space *mapping, struct folio *folio) in noop_dirty_folio() argument
2678 struct address_space *mapping) in folio_account_dirtied() argument
2680 struct inode *inode = mapping->host; in folio_account_dirtied()
2682 trace_writeback_dirty_folio(folio, mapping); in folio_account_dirtied()
2684 if (mapping_can_writeback(mapping)) { in folio_account_dirtied()
2733 void __folio_mark_dirty(struct folio *folio, struct address_space *mapping, in __folio_mark_dirty() argument
2738 xa_lock_irqsave(&mapping->i_pages, flags); in __folio_mark_dirty()
2739 if (folio->mapping) { /* Race with truncate? */ in __folio_mark_dirty()
2741 folio_account_dirtied(folio, mapping); in __folio_mark_dirty()
2742 __xa_set_mark(&mapping->i_pages, folio_index(folio), in __folio_mark_dirty()
2745 xa_unlock_irqrestore(&mapping->i_pages, flags); in __folio_mark_dirty()
2767 bool filemap_dirty_folio(struct address_space *mapping, struct folio *folio) in filemap_dirty_folio() argument
2772 __folio_mark_dirty(folio, mapping, !folio_test_private(folio)); in filemap_dirty_folio()
2774 if (mapping->host) { in filemap_dirty_folio()
2776 __mark_inode_dirty(mapping->host, I_DIRTY_PAGES); in filemap_dirty_folio()
2797 struct address_space *mapping = folio->mapping; in folio_redirty_for_writepage() local
2802 ret = filemap_dirty_folio(mapping, folio); in folio_redirty_for_writepage()
2803 if (mapping && mapping_can_writeback(mapping)) { in folio_redirty_for_writepage()
2804 struct inode *inode = mapping->host; in folio_redirty_for_writepage()
2833 struct address_space *mapping = folio_mapping(folio); in folio_mark_dirty() local
2835 if (likely(mapping)) { in folio_mark_dirty()
2849 return mapping->a_ops->dirty_folio(mapping, folio); in folio_mark_dirty()
2852 return noop_dirty_folio(mapping, folio); in folio_mark_dirty()
2892 struct address_space *mapping = folio_mapping(folio); in __folio_cancel_dirty() local
2894 if (mapping_can_writeback(mapping)) { in __folio_cancel_dirty()
2895 struct inode *inode = mapping->host; in __folio_cancel_dirty()
2927 struct address_space *mapping = folio_mapping(folio); in folio_clear_dirty_for_io() local
2932 if (mapping && mapping_can_writeback(mapping)) { in folio_clear_dirty_for_io()
2933 struct inode *inode = mapping->host; in folio_clear_dirty_for_io()
3012 struct address_space *mapping = folio_mapping(folio); in __folio_end_writeback() local
3015 if (mapping && mapping_use_writeback_tags(mapping)) { in __folio_end_writeback()
3016 struct inode *inode = mapping->host; in __folio_end_writeback()
3020 xa_lock_irqsave(&mapping->i_pages, flags); in __folio_end_writeback()
3022 __xa_clear_mark(&mapping->i_pages, folio_index(folio), in __folio_end_writeback()
3029 if (!mapping_tagged(mapping, PAGECACHE_TAG_WRITEBACK)) in __folio_end_writeback()
3033 if (mapping->host && !mapping_tagged(mapping, in __folio_end_writeback()
3035 sb_clear_inode_writeback(mapping->host); in __folio_end_writeback()
3037 xa_unlock_irqrestore(&mapping->i_pages, flags); in __folio_end_writeback()
3052 struct address_space *mapping = folio_mapping(folio); in __folio_start_writeback() local
3058 if (mapping && mapping_use_writeback_tags(mapping)) { in __folio_start_writeback()
3059 XA_STATE(xas, &mapping->i_pages, folio_index(folio)); in __folio_start_writeback()
3060 struct inode *inode = mapping->host; in __folio_start_writeback()
3069 on_wblist = mapping_tagged(mapping, PAGECACHE_TAG_WRITEBACK); in __folio_start_writeback()
3085 if (mapping->host && !on_wblist) in __folio_start_writeback()
3086 sb_mark_inode_writeback(mapping->host); in __folio_start_writeback()