| /linux/include/linux/ |
| A D | writeback.h | 108 if (wbc->sync_mode == WB_SYNC_ALL) in wbc_to_write_flags() 110 else if (wbc->for_kupdate || wbc->for_background) in wbc_to_write_flags() 117 #define wbc_blkcg_css(wbc) \ argument 118 ((wbc)->wb ? (wbc)->wb->blkcg_css : blkcg_root_css) 120 #define wbc_blkcg_css(wbc) (blkcg_root_css) argument 216 void wbc_attach_and_unlock_inode(struct writeback_control *wbc, 219 void wbc_detach_inode(struct writeback_control *wbc); 271 wbc_attach_and_unlock_inode(wbc, inode); in wbc_attach_fdatawrite_inode() 292 if (wbc->wb) in wbc_init_bio() 293 bio_associate_blkg_from_css(bio, wbc->wb->blkcg_css); in wbc_init_bio() [all …]
|
| A D | backing-dev.h | 261 struct writeback_control *wbc) in inode_to_wb_wbc() argument 267 return wbc->wb ? wbc->wb : &inode_to_bdi(inode)->wb; in inode_to_wb_wbc() 345 struct writeback_control *wbc) in inode_to_wb_wbc() argument
|
| /linux/fs/ |
| A D | fs-writeback.c | 745 wbc->wb_id = wbc->wb->memcg_css->id; in wbc_attach_and_unlock_inode() 752 wb_get(wbc->wb); in wbc_attach_and_unlock_inode() 819 if (wbc->wb_bytes >= wbc->wb_lcand_bytes && in wbc_detach_inode() 820 wbc->wb_bytes >= wbc->wb_tcand_bytes) { in wbc_detach_inode() 823 } else if (wbc->wb_lcand_bytes >= wbc->wb_tcand_bytes) { in wbc_detach_inode() 885 wb_put(wbc->wb); in wbc_detach_inode() 886 wbc->wb = NULL; in wbc_detach_inode() 913 if (!wbc->wb || wbc->no_cgroup_owner) in wbc_account_cgroup_owner() 938 wbc->wb_tcand_bytes -= min(bytes, wbc->wb_tcand_bytes); in wbc_account_cgroup_owner() 1667 if (wbc->sync_mode == WB_SYNC_ALL && !wbc->for_sync) { in __writeback_single_inode() [all …]
|
| A D | mpage.c | 451 static int __mpage_writepage(struct folio *folio, struct writeback_control *wbc, in __mpage_writepage() argument 597 REQ_OP_WRITE | wbc_to_write_flags(wbc), in __mpage_writepage() 600 wbc_init_bio(wbc, bio); in __mpage_writepage() 609 wbc_account_cgroup_owner(wbc, &folio->page, folio_size(folio)); in __mpage_writepage() 639 ret = block_write_full_folio(folio, wbc, mpd->get_block); in __mpage_writepage() 657 struct writeback_control *wbc, get_block_t get_block) in mpage_writepages() argument 666 ret = write_cache_pages(mapping, wbc, __mpage_writepage, &mpd); in mpage_writepages()
|
| /linux/fs/gfs2/ |
| A D | aops.c | 93 struct writeback_control *wbc) in gfs2_write_jdata_folio() argument 111 wbc); in gfs2_write_jdata_folio() 124 struct writeback_control *wbc) in __gfs2_jdata_write_folio() argument 149 struct writeback_control *wbc) in gfs2_writepages() argument 257 if (--wbc->nr_to_write <= 0 && wbc->sync_mode == WB_SYNC_NONE) { in gfs2_write_jdata_batch() 293 if (wbc->range_cyclic) { in gfs2_write_cache_jdata() 304 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in gfs2_write_cache_jdata() 308 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in gfs2_write_cache_jdata() 314 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in gfs2_write_cache_jdata() 345 if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) in gfs2_write_cache_jdata() [all …]
|
| A D | meta_io.c | 34 struct writeback_control *wbc) in gfs2_aspace_write_folio() argument 38 blk_opf_t write_flags = REQ_META | REQ_PRIO | wbc_to_write_flags(wbc); in gfs2_aspace_write_folio() 55 if (wbc->sync_mode != WB_SYNC_NONE) { in gfs2_aspace_write_folio() 58 folio_redirty_for_writepage(wbc, folio); in gfs2_aspace_write_folio() 90 struct writeback_control *wbc) in gfs2_aspace_writepages() argument 95 while ((folio = writeback_iter(mapping, wbc, folio, &error))) in gfs2_aspace_writepages() 96 gfs2_aspace_write_folio(folio, wbc); in gfs2_aspace_writepages()
|
| /linux/include/trace/events/ |
| A D | writeback.h | 166 if (wbc->wb) in __trace_wbc_assign_cgroup() 192 TP_ARGS(inode, wbc, history), 316 TP_ARGS(inode, wbc), 345 TP_ARGS(inode, wbc) 352 TP_ARGS(inode, wbc) 454 TP_ARGS(wbc, bdi), 503 TP_ARGS(wbc, bdi)) 749 struct writeback_control *wbc, 794 struct writeback_control *wbc, 796 TP_ARGS(inode, wbc, nr_to_write) [all …]
|
| /linux/mm/ |
| A D | page-writeback.c | 2467 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in wbc_to_tag() 2474 if (wbc->range_cyclic) in wbc_end() 2489 filemap_get_folios_tag(mapping, &wbc->index, wbc_end(wbc), in writeback_get_folio() 2490 wbc_to_tag(wbc), &wbc->fbatch); in writeback_get_folio() 2546 if (wbc->range_cyclic) in writeback_iter() 2549 wbc->index = wbc->range_start >> PAGE_SHIFT; in writeback_iter() 2562 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in writeback_iter() 2564 wbc_end(wbc)); in writeback_iter() 2602 if (wbc->range_cyclic) in writeback_iter() 2609 *error = wbc->saved_err; in writeback_iter() [all …]
|
| A D | page_io.c | 288 __swap_writepage(folio, wbc); in swap_writepage() 384 if (wbc->swap_plug) in swap_writepage_fs() 385 sio = *wbc->swap_plug; in swap_writepage_fs() 404 if (sio->pages == ARRAY_SIZE(sio->bvec) || !wbc->swap_plug) { in swap_writepage_fs() 408 if (wbc->swap_plug) in swap_writepage_fs() 409 *wbc->swap_plug = sio; in swap_writepage_fs() 419 REQ_OP_WRITE | REQ_SWAP | wbc_to_write_flags(wbc)); in swap_writepage_bdev_sync() 439 REQ_OP_WRITE | REQ_SWAP | wbc_to_write_flags(wbc), in swap_writepage_bdev_async() 463 swap_writepage_fs(folio, wbc); in __swap_writepage() 470 swap_writepage_bdev_sync(folio, wbc, sis); in __swap_writepage() [all …]
|
| A D | swap.h | 22 int swap_writepage(struct page *page, struct writeback_control *wbc); 23 void __swap_writepage(struct folio *folio, struct writeback_control *wbc); 144 static inline int swap_writepage(struct page *p, struct writeback_control *wbc) in swap_writepage() argument
|
| A D | folio-compat.c | 61 bool redirty_page_for_writepage(struct writeback_control *wbc, in redirty_page_for_writepage() argument 64 return folio_redirty_for_writepage(wbc, page_folio(page)); in redirty_page_for_writepage()
|
| /linux/fs/netfs/ |
| A D | write_issue.c | 44 struct writeback_control *wbc, in netfs_kill_dirty_pages() argument 84 } while ((folio = writeback_iter(mapping, wbc, folio, &error))); in netfs_kill_dirty_pages() 312 struct writeback_control *wbc, in netfs_write_folio() argument 393 folio_redirty_for_writepage(wbc, folio); in netfs_write_folio() 539 struct writeback_control *wbc) in netfs_writepages() argument 547 if (wbc->sync_mode == WB_SYNC_NONE) { in netfs_writepages() 556 folio = writeback_iter(mapping, wbc, NULL, &error); in netfs_writepages() 581 error = netfs_write_folio(wreq, wbc, folio); in netfs_writepages() 595 netfs_kill_dirty_pages(mapping, wbc, folio); in netfs_writepages() 657 return netfs_write_folio(wreq, wbc, folio); in netfs_advance_writethrough() [all …]
|
| A D | buffered_write.c | 103 struct writeback_control wbc = { in netfs_perform_write() local 121 wbc_attach_fdatawrite_inode(&wbc, mapping->host); in netfs_perform_write() 125 wbc_detach_inode(&wbc); in netfs_perform_write() 131 wbc_detach_inode(&wbc); in netfs_perform_write() 358 netfs_advance_writethrough(wreq, &wbc, folio, copied, in netfs_perform_write() 385 ret2 = netfs_end_writethrough(wreq, &wbc, writethrough); in netfs_perform_write() 386 wbc_detach_inode(&wbc); in netfs_perform_write()
|
| /linux/fs/btrfs/ |
| A D | extent_io.c | 1120 struct writeback_control *wbc = bio_ctrl->wbc; in writepage_delalloc() local 1196 wbc); in writepage_delalloc() 1831 struct writeback_control *wbc = ctx->wbc; in submit_eb_page() local 1896 struct btrfs_eb_write_context ctx = { .wbc = wbc }; in btree_write_cache_pages() 2024 struct writeback_control *wbc = bio_ctrl->wbc; in extent_write_cache_pages() local 2062 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in extent_write_cache_pages() 2079 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in extent_write_cache_pages() 2084 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in extent_write_cache_pages() 2166 if (wbc->range_cyclic || (wbc->nr_to_write > 0 && range_whole)) in extent_write_cache_pages() 2190 .wbc = wbc, in extent_write_locked_range() [all …]
|
| A D | extent_io.h | 129 struct writeback_control *wbc; member 244 u64 start, u64 end, struct writeback_control *wbc, 246 int btrfs_writepages(struct address_space *mapping, struct writeback_control *wbc); 248 struct writeback_control *wbc);
|
| /linux/fs/nfs/ |
| A D | write.c | 298 if (wbc->sync_mode == WB_SYNC_ALL) in wb_priority() 625 struct writeback_control *wbc, in nfs_page_async_flush() argument 680 struct writeback_control *wbc) in nfs_writepage_locked() argument 732 wbc->for_background || wbc->for_sync || wbc->for_reclaim) { in nfs_writepages() 737 priority = wb_priority(wbc); in nfs_writepages() 1920 struct writeback_control *wbc) in __nfs_commit_inode() argument 1938 if (wbc && wbc->sync_mode == WB_SYNC_NONE) { in __nfs_commit_inode() 1939 if (nscan < wbc->nr_to_write) in __nfs_commit_inode() 1940 wbc->nr_to_write -= nscan; in __nfs_commit_inode() 1942 wbc->nr_to_write = 0; in __nfs_commit_inode() [all …]
|
| A D | nfs4super.c | 21 static int nfs4_write_inode(struct inode *inode, struct writeback_control *wbc); 47 static int nfs4_write_inode(struct inode *inode, struct writeback_control *wbc) in nfs4_write_inode() argument 49 int ret = nfs_write_inode(inode, wbc); in nfs4_write_inode() 53 wbc->sync_mode == WB_SYNC_ALL); in nfs4_write_inode()
|
| /linux/fs/ceph/ |
| A D | addr.c | 779 if (!wbc) in writepage_nounlock() 780 wbc = &tmp_wbc; in writepage_nounlock() 794 wbc->pages_skipped++; in writepage_nounlock() 946 struct writeback_control *wbc) in ceph_writepages_start() argument 990 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) { in ceph_writepages_start() 1010 if (wbc->range_cyclic) { in ceph_writepages_start() 1019 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in ceph_writepages_start() 1033 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in ceph_writepages_start() 1359 wbc->nr_to_write -= i; in ceph_writepages_start() 1369 if (wbc->nr_to_write <= 0 && wbc->sync_mode == WB_SYNC_NONE) in ceph_writepages_start() [all …]
|
| A D | cache.h | 47 struct writeback_control *wbc) in ceph_fscache_unpin_writeback() argument 49 return netfs_unpin_writeback(inode, wbc); in ceph_fscache_unpin_writeback() 104 struct writeback_control *wbc) in ceph_fscache_unpin_writeback() argument
|
| /linux/fs/f2fs/ |
| A D | data.c | 2812 .io_wbc = wbc, in f2fs_write_single_data_page() 2873 if (!wbc->for_reclaim) in f2fs_write_single_data_page() 2916 if (wbc->for_reclaim) { in f2fs_write_single_data_page() 2947 if (!err || wbc->for_reclaim) in f2fs_write_single_data_page() 3039 if (wbc->range_cyclic) { in f2fs_write_cache_pages() 3045 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in f2fs_write_cache_pages() 3048 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in f2fs_write_cache_pages() 3054 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in f2fs_write_cache_pages() 3184 wbc, io_type, 0, true); in f2fs_write_cache_pages() 3215 if (wbc->nr_to_write <= 0 && in f2fs_write_cache_pages() [all …]
|
| A D | checkpoint.c | 344 struct writeback_control *wbc, in __f2fs_write_meta_page() argument 369 if (wbc->for_reclaim) in __f2fs_write_meta_page() 380 redirty_page_for_writepage(wbc, page); in __f2fs_write_meta_page() 385 struct writeback_control *wbc) in f2fs_write_meta_page() argument 391 struct writeback_control *wbc) in f2fs_write_meta_pages() argument 400 if (wbc->sync_mode != WB_SYNC_ALL && in f2fs_write_meta_pages() 410 diff = nr_pages_to_write(sbi, META, wbc); in f2fs_write_meta_pages() 413 wbc->nr_to_write = max((long)0, wbc->nr_to_write - written - diff); in f2fs_write_meta_pages() 430 struct writeback_control wbc = { in f2fs_sync_meta_pages() local 1225 struct writeback_control wbc = { in block_operations() local [all …]
|
| /linux/fs/orangefs/ |
| A D | inode.c | 20 struct writeback_control *wbc) in orangefs_writepage_locked() argument 70 ret = orangefs_writepage_locked(page, wbc); in orangefs_writepage() 88 struct writeback_control *wbc) in orangefs_writepages_work() argument 150 struct writeback_control *wbc, void *data) in orangefs_writepages_callback() argument 175 orangefs_writepages_work(ow, wbc); in orangefs_writepages_callback() 189 orangefs_writepages_work(ow, wbc); in orangefs_writepages_callback() 192 ret = orangefs_writepage_locked(&folio->page, wbc); in orangefs_writepages_callback() 198 orangefs_writepages_work(ow, wbc); in orangefs_writepages_callback() 206 struct writeback_control *wbc) in orangefs_writepages() argument 229 ret = orangefs_writepages_work(ow, wbc); in orangefs_writepages() [all …]
|
| /linux/fs/9p/ |
| A D | vfs_super.c | 263 struct writeback_control *wbc) in v9fs_write_inode() argument 270 return netfs_unpin_writeback(inode, wbc); in v9fs_write_inode() 274 struct writeback_control *wbc) in v9fs_write_inode_dotl() argument 279 return netfs_unpin_writeback(inode, wbc); in v9fs_write_inode_dotl()
|
| /linux/fs/ext4/ |
| A D | page-io.c | 383 struct writeback_control *wbc) in ext4_io_submit_init() argument 385 io->io_wbc = wbc; in ext4_io_submit_init() 437 struct writeback_control *wbc = io->io_wbc; in ext4_bio_write_folio() local 485 folio_redirty_for_writepage(wbc, folio); in ext4_bio_write_folio() 528 (io->io_bio || wbc->sync_mode == WB_SYNC_ALL)) { in ext4_bio_write_folio() 540 folio_redirty_for_writepage(wbc, folio); in ext4_bio_write_folio()
|
| A D | inode.c | 2378 mpd->wbc->nr_to_write--; in mpage_journal_page_buffers() 2422 if (mpd->wbc->sync_mode == WB_SYNC_ALL || mpd->wbc->tagged_writepages) in mpage_prepare_extent_to_map() 2558 struct writeback_control *wbc = mpd->wbc; in ext4_do_writepages() local 2643 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in ext4_do_writepages() 2646 if (wbc->range_cyclic) { in ext4_do_writepages() 2659 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in ext4_do_writepages() 2784 if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) in ext4_do_writepages() 2803 .wbc = wbc, in ext4_writepages() 2836 .wbc = &wbc, in ext4_normal_submit_inode_data_buffers() 5213 if (wbc->sync_mode != WB_SYNC_ALL || wbc->for_sync) in ext4_write_inode() [all …]
|