Home
last modified time | relevance | path

Searched refs:wbc (Results 1 – 25 of 122) sorted by relevance

12345

/linux/include/linux/
A Dwriteback.h108 if (wbc->sync_mode == WB_SYNC_ALL) in wbc_to_write_flags()
110 else if (wbc->for_kupdate || wbc->for_background) in wbc_to_write_flags()
117 #define wbc_blkcg_css(wbc) \ argument
118 ((wbc)->wb ? (wbc)->wb->blkcg_css : blkcg_root_css)
120 #define wbc_blkcg_css(wbc) (blkcg_root_css) argument
216 void wbc_attach_and_unlock_inode(struct writeback_control *wbc,
219 void wbc_detach_inode(struct writeback_control *wbc);
271 wbc_attach_and_unlock_inode(wbc, inode); in wbc_attach_fdatawrite_inode()
292 if (wbc->wb) in wbc_init_bio()
293 bio_associate_blkg_from_css(bio, wbc->wb->blkcg_css); in wbc_init_bio()
[all …]
A Dbacking-dev.h261 struct writeback_control *wbc) in inode_to_wb_wbc() argument
267 return wbc->wb ? wbc->wb : &inode_to_bdi(inode)->wb; in inode_to_wb_wbc()
345 struct writeback_control *wbc) in inode_to_wb_wbc() argument
/linux/fs/
A Dfs-writeback.c745 wbc->wb_id = wbc->wb->memcg_css->id; in wbc_attach_and_unlock_inode()
752 wb_get(wbc->wb); in wbc_attach_and_unlock_inode()
819 if (wbc->wb_bytes >= wbc->wb_lcand_bytes && in wbc_detach_inode()
820 wbc->wb_bytes >= wbc->wb_tcand_bytes) { in wbc_detach_inode()
823 } else if (wbc->wb_lcand_bytes >= wbc->wb_tcand_bytes) { in wbc_detach_inode()
885 wb_put(wbc->wb); in wbc_detach_inode()
886 wbc->wb = NULL; in wbc_detach_inode()
913 if (!wbc->wb || wbc->no_cgroup_owner) in wbc_account_cgroup_owner()
938 wbc->wb_tcand_bytes -= min(bytes, wbc->wb_tcand_bytes); in wbc_account_cgroup_owner()
1667 if (wbc->sync_mode == WB_SYNC_ALL && !wbc->for_sync) { in __writeback_single_inode()
[all …]
A Dmpage.c451 static int __mpage_writepage(struct folio *folio, struct writeback_control *wbc, in __mpage_writepage() argument
597 REQ_OP_WRITE | wbc_to_write_flags(wbc), in __mpage_writepage()
600 wbc_init_bio(wbc, bio); in __mpage_writepage()
609 wbc_account_cgroup_owner(wbc, &folio->page, folio_size(folio)); in __mpage_writepage()
639 ret = block_write_full_folio(folio, wbc, mpd->get_block); in __mpage_writepage()
657 struct writeback_control *wbc, get_block_t get_block) in mpage_writepages() argument
666 ret = write_cache_pages(mapping, wbc, __mpage_writepage, &mpd); in mpage_writepages()
/linux/fs/gfs2/
A Daops.c93 struct writeback_control *wbc) in gfs2_write_jdata_folio() argument
111 wbc); in gfs2_write_jdata_folio()
124 struct writeback_control *wbc) in __gfs2_jdata_write_folio() argument
149 struct writeback_control *wbc) in gfs2_writepages() argument
257 if (--wbc->nr_to_write <= 0 && wbc->sync_mode == WB_SYNC_NONE) { in gfs2_write_jdata_batch()
293 if (wbc->range_cyclic) { in gfs2_write_cache_jdata()
304 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in gfs2_write_cache_jdata()
308 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in gfs2_write_cache_jdata()
314 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in gfs2_write_cache_jdata()
345 if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) in gfs2_write_cache_jdata()
[all …]
A Dmeta_io.c34 struct writeback_control *wbc) in gfs2_aspace_write_folio() argument
38 blk_opf_t write_flags = REQ_META | REQ_PRIO | wbc_to_write_flags(wbc); in gfs2_aspace_write_folio()
55 if (wbc->sync_mode != WB_SYNC_NONE) { in gfs2_aspace_write_folio()
58 folio_redirty_for_writepage(wbc, folio); in gfs2_aspace_write_folio()
90 struct writeback_control *wbc) in gfs2_aspace_writepages() argument
95 while ((folio = writeback_iter(mapping, wbc, folio, &error))) in gfs2_aspace_writepages()
96 gfs2_aspace_write_folio(folio, wbc); in gfs2_aspace_writepages()
/linux/include/trace/events/
A Dwriteback.h166 if (wbc->wb) in __trace_wbc_assign_cgroup()
192 TP_ARGS(inode, wbc, history),
316 TP_ARGS(inode, wbc),
345 TP_ARGS(inode, wbc)
352 TP_ARGS(inode, wbc)
454 TP_ARGS(wbc, bdi),
503 TP_ARGS(wbc, bdi))
749 struct writeback_control *wbc,
794 struct writeback_control *wbc,
796 TP_ARGS(inode, wbc, nr_to_write)
[all …]
/linux/mm/
A Dpage-writeback.c2467 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in wbc_to_tag()
2474 if (wbc->range_cyclic) in wbc_end()
2489 filemap_get_folios_tag(mapping, &wbc->index, wbc_end(wbc), in writeback_get_folio()
2490 wbc_to_tag(wbc), &wbc->fbatch); in writeback_get_folio()
2546 if (wbc->range_cyclic) in writeback_iter()
2549 wbc->index = wbc->range_start >> PAGE_SHIFT; in writeback_iter()
2562 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in writeback_iter()
2564 wbc_end(wbc)); in writeback_iter()
2602 if (wbc->range_cyclic) in writeback_iter()
2609 *error = wbc->saved_err; in writeback_iter()
[all …]
A Dpage_io.c288 __swap_writepage(folio, wbc); in swap_writepage()
384 if (wbc->swap_plug) in swap_writepage_fs()
385 sio = *wbc->swap_plug; in swap_writepage_fs()
404 if (sio->pages == ARRAY_SIZE(sio->bvec) || !wbc->swap_plug) { in swap_writepage_fs()
408 if (wbc->swap_plug) in swap_writepage_fs()
409 *wbc->swap_plug = sio; in swap_writepage_fs()
419 REQ_OP_WRITE | REQ_SWAP | wbc_to_write_flags(wbc)); in swap_writepage_bdev_sync()
439 REQ_OP_WRITE | REQ_SWAP | wbc_to_write_flags(wbc), in swap_writepage_bdev_async()
463 swap_writepage_fs(folio, wbc); in __swap_writepage()
470 swap_writepage_bdev_sync(folio, wbc, sis); in __swap_writepage()
[all …]
A Dswap.h22 int swap_writepage(struct page *page, struct writeback_control *wbc);
23 void __swap_writepage(struct folio *folio, struct writeback_control *wbc);
144 static inline int swap_writepage(struct page *p, struct writeback_control *wbc) in swap_writepage() argument
A Dfolio-compat.c61 bool redirty_page_for_writepage(struct writeback_control *wbc, in redirty_page_for_writepage() argument
64 return folio_redirty_for_writepage(wbc, page_folio(page)); in redirty_page_for_writepage()
/linux/fs/netfs/
A Dwrite_issue.c44 struct writeback_control *wbc, in netfs_kill_dirty_pages() argument
84 } while ((folio = writeback_iter(mapping, wbc, folio, &error))); in netfs_kill_dirty_pages()
312 struct writeback_control *wbc, in netfs_write_folio() argument
393 folio_redirty_for_writepage(wbc, folio); in netfs_write_folio()
539 struct writeback_control *wbc) in netfs_writepages() argument
547 if (wbc->sync_mode == WB_SYNC_NONE) { in netfs_writepages()
556 folio = writeback_iter(mapping, wbc, NULL, &error); in netfs_writepages()
581 error = netfs_write_folio(wreq, wbc, folio); in netfs_writepages()
595 netfs_kill_dirty_pages(mapping, wbc, folio); in netfs_writepages()
657 return netfs_write_folio(wreq, wbc, folio); in netfs_advance_writethrough()
[all …]
A Dbuffered_write.c103 struct writeback_control wbc = { in netfs_perform_write() local
121 wbc_attach_fdatawrite_inode(&wbc, mapping->host); in netfs_perform_write()
125 wbc_detach_inode(&wbc); in netfs_perform_write()
131 wbc_detach_inode(&wbc); in netfs_perform_write()
358 netfs_advance_writethrough(wreq, &wbc, folio, copied, in netfs_perform_write()
385 ret2 = netfs_end_writethrough(wreq, &wbc, writethrough); in netfs_perform_write()
386 wbc_detach_inode(&wbc); in netfs_perform_write()
/linux/fs/btrfs/
A Dextent_io.c1120 struct writeback_control *wbc = bio_ctrl->wbc; in writepage_delalloc() local
1196 wbc); in writepage_delalloc()
1831 struct writeback_control *wbc = ctx->wbc; in submit_eb_page() local
1896 struct btrfs_eb_write_context ctx = { .wbc = wbc }; in btree_write_cache_pages()
2024 struct writeback_control *wbc = bio_ctrl->wbc; in extent_write_cache_pages() local
2062 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in extent_write_cache_pages()
2079 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in extent_write_cache_pages()
2084 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in extent_write_cache_pages()
2166 if (wbc->range_cyclic || (wbc->nr_to_write > 0 && range_whole)) in extent_write_cache_pages()
2190 .wbc = wbc, in extent_write_locked_range()
[all …]
A Dextent_io.h129 struct writeback_control *wbc; member
244 u64 start, u64 end, struct writeback_control *wbc,
246 int btrfs_writepages(struct address_space *mapping, struct writeback_control *wbc);
248 struct writeback_control *wbc);
/linux/fs/nfs/
A Dwrite.c298 if (wbc->sync_mode == WB_SYNC_ALL) in wb_priority()
625 struct writeback_control *wbc, in nfs_page_async_flush() argument
680 struct writeback_control *wbc) in nfs_writepage_locked() argument
732 wbc->for_background || wbc->for_sync || wbc->for_reclaim) { in nfs_writepages()
737 priority = wb_priority(wbc); in nfs_writepages()
1920 struct writeback_control *wbc) in __nfs_commit_inode() argument
1938 if (wbc && wbc->sync_mode == WB_SYNC_NONE) { in __nfs_commit_inode()
1939 if (nscan < wbc->nr_to_write) in __nfs_commit_inode()
1940 wbc->nr_to_write -= nscan; in __nfs_commit_inode()
1942 wbc->nr_to_write = 0; in __nfs_commit_inode()
[all …]
A Dnfs4super.c21 static int nfs4_write_inode(struct inode *inode, struct writeback_control *wbc);
47 static int nfs4_write_inode(struct inode *inode, struct writeback_control *wbc) in nfs4_write_inode() argument
49 int ret = nfs_write_inode(inode, wbc); in nfs4_write_inode()
53 wbc->sync_mode == WB_SYNC_ALL); in nfs4_write_inode()
/linux/fs/ceph/
A Daddr.c779 if (!wbc) in writepage_nounlock()
780 wbc = &tmp_wbc; in writepage_nounlock()
794 wbc->pages_skipped++; in writepage_nounlock()
946 struct writeback_control *wbc) in ceph_writepages_start() argument
990 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) { in ceph_writepages_start()
1010 if (wbc->range_cyclic) { in ceph_writepages_start()
1019 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in ceph_writepages_start()
1033 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in ceph_writepages_start()
1359 wbc->nr_to_write -= i; in ceph_writepages_start()
1369 if (wbc->nr_to_write <= 0 && wbc->sync_mode == WB_SYNC_NONE) in ceph_writepages_start()
[all …]
A Dcache.h47 struct writeback_control *wbc) in ceph_fscache_unpin_writeback() argument
49 return netfs_unpin_writeback(inode, wbc); in ceph_fscache_unpin_writeback()
104 struct writeback_control *wbc) in ceph_fscache_unpin_writeback() argument
/linux/fs/f2fs/
A Ddata.c2812 .io_wbc = wbc, in f2fs_write_single_data_page()
2873 if (!wbc->for_reclaim) in f2fs_write_single_data_page()
2916 if (wbc->for_reclaim) { in f2fs_write_single_data_page()
2947 if (!err || wbc->for_reclaim) in f2fs_write_single_data_page()
3039 if (wbc->range_cyclic) { in f2fs_write_cache_pages()
3045 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in f2fs_write_cache_pages()
3048 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in f2fs_write_cache_pages()
3054 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in f2fs_write_cache_pages()
3184 wbc, io_type, 0, true); in f2fs_write_cache_pages()
3215 if (wbc->nr_to_write <= 0 && in f2fs_write_cache_pages()
[all …]
A Dcheckpoint.c344 struct writeback_control *wbc, in __f2fs_write_meta_page() argument
369 if (wbc->for_reclaim) in __f2fs_write_meta_page()
380 redirty_page_for_writepage(wbc, page); in __f2fs_write_meta_page()
385 struct writeback_control *wbc) in f2fs_write_meta_page() argument
391 struct writeback_control *wbc) in f2fs_write_meta_pages() argument
400 if (wbc->sync_mode != WB_SYNC_ALL && in f2fs_write_meta_pages()
410 diff = nr_pages_to_write(sbi, META, wbc); in f2fs_write_meta_pages()
413 wbc->nr_to_write = max((long)0, wbc->nr_to_write - written - diff); in f2fs_write_meta_pages()
430 struct writeback_control wbc = { in f2fs_sync_meta_pages() local
1225 struct writeback_control wbc = { in block_operations() local
[all …]
/linux/fs/orangefs/
A Dinode.c20 struct writeback_control *wbc) in orangefs_writepage_locked() argument
70 ret = orangefs_writepage_locked(page, wbc); in orangefs_writepage()
88 struct writeback_control *wbc) in orangefs_writepages_work() argument
150 struct writeback_control *wbc, void *data) in orangefs_writepages_callback() argument
175 orangefs_writepages_work(ow, wbc); in orangefs_writepages_callback()
189 orangefs_writepages_work(ow, wbc); in orangefs_writepages_callback()
192 ret = orangefs_writepage_locked(&folio->page, wbc); in orangefs_writepages_callback()
198 orangefs_writepages_work(ow, wbc); in orangefs_writepages_callback()
206 struct writeback_control *wbc) in orangefs_writepages() argument
229 ret = orangefs_writepages_work(ow, wbc); in orangefs_writepages()
[all …]
/linux/fs/9p/
A Dvfs_super.c263 struct writeback_control *wbc) in v9fs_write_inode() argument
270 return netfs_unpin_writeback(inode, wbc); in v9fs_write_inode()
274 struct writeback_control *wbc) in v9fs_write_inode_dotl() argument
279 return netfs_unpin_writeback(inode, wbc); in v9fs_write_inode_dotl()
/linux/fs/ext4/
A Dpage-io.c383 struct writeback_control *wbc) in ext4_io_submit_init() argument
385 io->io_wbc = wbc; in ext4_io_submit_init()
437 struct writeback_control *wbc = io->io_wbc; in ext4_bio_write_folio() local
485 folio_redirty_for_writepage(wbc, folio); in ext4_bio_write_folio()
528 (io->io_bio || wbc->sync_mode == WB_SYNC_ALL)) { in ext4_bio_write_folio()
540 folio_redirty_for_writepage(wbc, folio); in ext4_bio_write_folio()
A Dinode.c2378 mpd->wbc->nr_to_write--; in mpage_journal_page_buffers()
2422 if (mpd->wbc->sync_mode == WB_SYNC_ALL || mpd->wbc->tagged_writepages) in mpage_prepare_extent_to_map()
2558 struct writeback_control *wbc = mpd->wbc; in ext4_do_writepages() local
2643 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in ext4_do_writepages()
2646 if (wbc->range_cyclic) { in ext4_do_writepages()
2659 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in ext4_do_writepages()
2784 if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) in ext4_do_writepages()
2803 .wbc = wbc, in ext4_writepages()
2836 .wbc = &wbc, in ext4_normal_submit_inode_data_buffers()
5213 if (wbc->sync_mode != WB_SYNC_ALL || wbc->for_sync) in ext4_write_inode()
[all …]

Completed in 86 milliseconds

12345