/linux-6.3-rc2/include/linux/ |
A D | pagevec.h | 92 static inline void folio_batch_init(struct folio_batch *fbatch) in folio_batch_init() argument 94 fbatch->nr = 0; in folio_batch_init() 95 fbatch->percpu_pvec_drained = false; in folio_batch_init() 98 static inline void folio_batch_reinit(struct folio_batch *fbatch) in folio_batch_reinit() argument 100 fbatch->nr = 0; in folio_batch_reinit() 105 return fbatch->nr; in folio_batch_count() 110 return PAGEVEC_SIZE - fbatch->nr; in fbatch_space() 126 fbatch->folios[fbatch->nr++] = folio; in folio_batch_add() 127 return fbatch_space(fbatch); in folio_batch_add() 132 pagevec_release((struct pagevec *)fbatch); in folio_batch_release() [all …]
|
A D | pagemap.h | 739 pgoff_t end, struct folio_batch *fbatch); 741 pgoff_t *start, pgoff_t end, struct folio_batch *fbatch); 743 pgoff_t end, xa_mark_t tag, struct folio_batch *fbatch); 1111 struct folio_batch *fbatch);
|
/linux-6.3-rc2/mm/ |
A D | swap.c | 53 struct folio_batch fbatch; member 224 folios_put(fbatch->folios, folio_batch_count(fbatch)); in folio_batch_move_lru() 225 folio_batch_init(fbatch); in folio_batch_move_lru() 258 struct folio_batch *fbatch; in folio_rotate_reclaimable() local 263 fbatch = this_cpu_ptr(&lru_rotate.fbatch); in folio_rotate_reclaimable() 356 struct folio_batch *fbatch; in folio_activate() local 386 struct folio_batch *fbatch; in __lru_cache_activate_folio() local 503 struct folio_batch *fbatch; in folio_add_lru() local 654 fbatch = &per_cpu(lru_rotate.fbatch, cpu); in lru_add_drain_cpu() 692 struct folio_batch *fbatch; in deactivate_file_folio() local [all …]
|
A D | truncate.c | 88 fbatch->folios[j++] = folio; in truncate_folio_batch_exceptionals() 106 fbatch->nr = j; in truncate_folio_batch_exceptionals() 335 struct folio_batch fbatch; in truncate_inode_pages_range() local 362 folio_batch_init(&fbatch); in truncate_inode_pages_range() 365 &fbatch, indices)) { in truncate_inode_pages_range() 372 folio_batch_release(&fbatch); in truncate_inode_pages_range() 428 folio_batch_release(&fbatch); in truncate_inode_pages_range() 502 struct folio_batch fbatch; in invalidate_mapping_pagevec() local 508 folio_batch_init(&fbatch); in invalidate_mapping_pagevec() 624 struct folio_batch fbatch; in invalidate_inode_pages2_range() local [all …]
|
A D | mlock.c | 33 struct folio_batch fbatch; member 194 folio = fbatch->folios[i]; in mlock_folio_batch() 197 fbatch->folios[i] = folio; in mlock_folio_batch() 209 release_pages(fbatch->folios, fbatch->nr); in mlock_folio_batch() 210 folio_batch_reinit(fbatch); in mlock_folio_batch() 215 struct folio_batch *fbatch; in mlock_drain_local() local 218 fbatch = this_cpu_ptr(&mlock_fbatch.fbatch); in mlock_drain_local() 229 fbatch = &per_cpu(mlock_fbatch.fbatch, cpu); in mlock_drain_remote() 248 fbatch = this_cpu_ptr(&mlock_fbatch.fbatch); in mlock_folio() 274 fbatch = this_cpu_ptr(&mlock_fbatch.fbatch); in mlock_new_folio() [all …]
|
A D | filemap.c | 508 struct folio_batch fbatch; in __filemap_fdatawait_range() local 511 folio_batch_init(&fbatch); in __filemap_fdatawait_range() 528 folio_batch_release(&fbatch); in __filemap_fdatawait_range() 2612 folio = fbatch->folios[folio_batch_count(fbatch) - 1]; in filemap_get_pages() 2632 if (likely(--fbatch->nr)) in filemap_get_pages() 2666 struct folio_batch fbatch; in filemap_read() local 2677 folio_batch_init(&fbatch); in filemap_read() 2722 fbatch.folios[0])) in filemap_read() 2759 folio_batch_init(&fbatch); in filemap_read() 2886 struct folio_batch fbatch; in filemap_splice_read() local [all …]
|
A D | shmem.c | 865 struct folio_batch fbatch; in shmem_unlock_mapping() local 868 folio_batch_init(&fbatch); in shmem_unlock_mapping() 875 folio_batch_release(&fbatch); in shmem_unlock_mapping() 912 struct folio_batch fbatch; in shmem_undo_range() local 926 folio_batch_init(&fbatch); in shmem_undo_range() 929 &fbatch, indices)) { in shmem_undo_range() 931 folio = fbatch.folios[i]; in shmem_undo_range() 946 folio_batch_release(&fbatch); in shmem_undo_range() 1000 folio = fbatch.folios[i]; in shmem_undo_range() 1261 struct folio_batch fbatch; in shmem_unuse_inode() local [all …]
|
A D | internal.h | 128 pgoff_t end, struct folio_batch *fbatch, pgoff_t *indices); 130 pgoff_t end, struct folio_batch *fbatch, pgoff_t *indices);
|
A D | page-writeback.c | 2401 struct folio_batch fbatch; in write_cache_pages() local 2409 folio_batch_init(&fbatch); in write_cache_pages() 2430 tag, &fbatch); in write_cache_pages() 2436 struct folio *folio = fbatch.folios[i]; in write_cache_pages() 2513 folio_batch_release(&fbatch); in write_cache_pages()
|
/linux-6.3-rc2/fs/nilfs2/ |
A D | page.c | 243 struct folio_batch fbatch; in nilfs_copy_dirty_pages() local 248 folio_batch_init(&fbatch); in nilfs_copy_dirty_pages() 279 folio_batch_release(&fbatch); in nilfs_copy_dirty_pages() 298 struct folio_batch fbatch; in nilfs_copy_back_pages() local 302 folio_batch_init(&fbatch); in nilfs_copy_back_pages() 361 struct folio_batch fbatch; in nilfs_clear_dirty_pages() local 365 folio_batch_init(&fbatch); in nilfs_clear_dirty_pages() 487 struct folio_batch fbatch; in nilfs_find_uncommitted_extent() local 495 folio_batch_init(&fbatch); in nilfs_find_uncommitted_extent() 499 &fbatch); in nilfs_find_uncommitted_extent() [all …]
|
A D | segment.c | 683 struct folio_batch fbatch; in nilfs_lookup_dirty_data_buffers() local 697 folio_batch_init(&fbatch); in nilfs_lookup_dirty_data_buffers() 701 PAGECACHE_TAG_DIRTY, &fbatch)) in nilfs_lookup_dirty_data_buffers() 706 struct folio *folio = fbatch.folios[i]; in nilfs_lookup_dirty_data_buffers() 724 folio_batch_release(&fbatch); in nilfs_lookup_dirty_data_buffers() 730 folio_batch_release(&fbatch); in nilfs_lookup_dirty_data_buffers() 740 struct folio_batch fbatch; in nilfs_lookup_dirty_node_buffers() local 747 folio_batch_init(&fbatch); in nilfs_lookup_dirty_node_buffers() 750 (pgoff_t)-1, PAGECACHE_TAG_DIRTY, &fbatch)) { in nilfs_lookup_dirty_node_buffers() 752 bh = head = folio_buffers(fbatch.folios[i]); in nilfs_lookup_dirty_node_buffers() [all …]
|
A D | btree.c | 2153 struct folio_batch fbatch; in nilfs_btree_lookup_dirty_buffers() local 2163 folio_batch_init(&fbatch); in nilfs_btree_lookup_dirty_buffers() 2166 PAGECACHE_TAG_DIRTY, &fbatch)) { in nilfs_btree_lookup_dirty_buffers() 2167 for (i = 0; i < folio_batch_count(&fbatch); i++) { in nilfs_btree_lookup_dirty_buffers() 2168 bh = head = folio_buffers(fbatch.folios[i]); in nilfs_btree_lookup_dirty_buffers() 2175 folio_batch_release(&fbatch); in nilfs_btree_lookup_dirty_buffers()
|
/linux-6.3-rc2/fs/ramfs/ |
A D | file-nommu.c | 208 struct folio_batch fbatch; in ramfs_nommu_get_unmapped_area() local 224 folio_batch_init(&fbatch); in ramfs_nommu_get_unmapped_area() 228 ULONG_MAX, &fbatch); in ramfs_nommu_get_unmapped_area() 235 ret = (unsigned long) folio_address(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area() 236 pfn = folio_pfn(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area() 240 if (pfn + nr_pages != folio_pfn(fbatch.folios[loop])) { in ramfs_nommu_get_unmapped_area() 244 nr_pages += folio_nr_pages(fbatch.folios[loop]); in ramfs_nommu_get_unmapped_area() 250 folio_batch_release(&fbatch); in ramfs_nommu_get_unmapped_area() 256 folio_batch_release(&fbatch); in ramfs_nommu_get_unmapped_area()
|
/linux-6.3-rc2/fs/gfs2/ |
A D | aops.c | 208 struct folio_batch *fbatch, in gfs2_write_jdata_batch() argument 217 int nr_folios = folio_batch_count(fbatch); in gfs2_write_jdata_batch() 220 nr_pages += folio_nr_pages(fbatch->folios[i]); in gfs2_write_jdata_batch() 228 struct folio *folio = fbatch->folios[i]; in gfs2_write_jdata_batch() 312 struct folio_batch fbatch; in gfs2_write_cache_jdata() local 322 folio_batch_init(&fbatch); in gfs2_write_cache_jdata() 349 tag, &fbatch); in gfs2_write_cache_jdata() 353 ret = gfs2_write_jdata_batch(mapping, wbc, &fbatch, in gfs2_write_cache_jdata() 359 folio_batch_release(&fbatch); in gfs2_write_cache_jdata()
|
/linux-6.3-rc2/fs/ceph/ |
A D | addr.c | 803 struct folio_batch fbatch; in ceph_writepages_start() local 832 folio_batch_init(&fbatch); in ceph_writepages_start() 1016 fbatch.folios[i] = NULL; in ceph_writepages_start() 1028 if (!fbatch.folios[j]) in ceph_writepages_start() 1031 fbatch.folios[n] = fbatch.folios[j]; in ceph_writepages_start() 1034 fbatch.nr = n; in ceph_writepages_start() 1039 folio_batch_release(&fbatch); in ceph_writepages_start() 1177 fbatch.nr ? fbatch.folios[0] : NULL); in ceph_writepages_start() 1178 folio_batch_release(&fbatch); in ceph_writepages_start() 1198 &fbatch))) { in ceph_writepages_start() [all …]
|
/linux-6.3-rc2/fs/btrfs/tests/ |
A D | extent-io-tests.c | 24 struct folio_batch fbatch; in process_page_range() local 31 folio_batch_init(&fbatch); in process_page_range() 35 end_index, &fbatch); in process_page_range() 37 struct folio *folio = fbatch.folios[i]; in process_page_range() 47 folio_batch_release(&fbatch); in process_page_range()
|
/linux-6.3-rc2/fs/f2fs/ |
A D | node.c | 1521 struct folio_batch fbatch; in last_fsync_dnode() local 1525 folio_batch_init(&fbatch); in last_fsync_dnode() 1530 &fbatch))) { in last_fsync_dnode() 1734 struct folio_batch fbatch; in f2fs_fsync_node_pages() local 1748 folio_batch_init(&fbatch); in f2fs_fsync_node_pages() 1753 &fbatch))) { in f2fs_fsync_node_pages() 1893 struct folio_batch fbatch; in f2fs_flush_inline_data() local 1896 folio_batch_init(&fbatch); in f2fs_flush_inline_data() 1900 &fbatch))) { in f2fs_flush_inline_data() 1941 struct folio_batch fbatch; in f2fs_sync_node_pages() local [all …]
|
A D | checkpoint.c | 396 struct folio_batch fbatch; in f2fs_sync_meta_pages() local 404 folio_batch_init(&fbatch); in f2fs_sync_meta_pages() 410 PAGECACHE_TAG_DIRTY, &fbatch))) { in f2fs_sync_meta_pages() 414 struct folio *folio = fbatch.folios[i]; in f2fs_sync_meta_pages() 418 folio_nr_pages(fbatch.folios[i-1])) { in f2fs_sync_meta_pages() 419 folio_batch_release(&fbatch); in f2fs_sync_meta_pages() 451 folio_batch_release(&fbatch); in f2fs_sync_meta_pages()
|
A D | compress.c | 1898 struct folio_batch fbatch; in f2fs_invalidate_compress_pages() local 1905 folio_batch_init(&fbatch); in f2fs_invalidate_compress_pages() 1910 nr = filemap_get_folios(mapping, &index, end - 1, &fbatch); in f2fs_invalidate_compress_pages() 1915 struct folio *folio = fbatch.folios[i]; in f2fs_invalidate_compress_pages() 1931 folio_batch_release(&fbatch); in f2fs_invalidate_compress_pages()
|
/linux-6.3-rc2/fs/afs/ |
A D | write.c | 707 struct folio_batch fbatch; in afs_writepages_region() local 713 folio_batch_init(&fbatch); in afs_writepages_region() 719 PAGECACHE_TAG_DIRTY, &fbatch); in afs_writepages_region() 724 folio = fbatch.folios[i]; in afs_writepages_region() 737 folio_batch_release(&fbatch); in afs_writepages_region() 780 folio_batch_release(&fbatch); in afs_writepages_region() 787 folio_batch_release(&fbatch); in afs_writepages_region()
|
/linux-6.3-rc2/fs/hugetlbfs/ |
A D | inode.c | 621 struct folio_batch fbatch; in remove_inode_hugepages() local 626 folio_batch_init(&fbatch); in remove_inode_hugepages() 628 while (filemap_get_folios(mapping, &next, end - 1, &fbatch)) { in remove_inode_hugepages() 629 for (i = 0; i < folio_batch_count(&fbatch); ++i) { in remove_inode_hugepages() 630 struct folio *folio = fbatch.folios[i]; in remove_inode_hugepages() 646 folio_batch_release(&fbatch); in remove_inode_hugepages()
|
/linux-6.3-rc2/fs/btrfs/ |
A D | compression.c | 181 struct folio_batch fbatch; in end_compressed_writeback() local 189 folio_batch_init(&fbatch); in end_compressed_writeback() 192 &fbatch); in end_compressed_writeback() 198 struct folio *folio = fbatch.folios[i]; in end_compressed_writeback() 205 folio_batch_release(&fbatch); in end_compressed_writeback()
|
A D | extent_io.c | 297 struct folio_batch fbatch; in __process_pages_contig() local 309 folio_batch_init(&fbatch); in __process_pages_contig() 314 end_index, &fbatch); in __process_pages_contig() 334 folio_batch_release(&fbatch); in __process_pages_contig() 339 folio_batch_release(&fbatch); in __process_pages_contig() 2411 struct folio_batch fbatch; in btree_write_cache_pages() local 2418 folio_batch_init(&fbatch); in btree_write_cache_pages() 2442 tag, &fbatch))) { in btree_write_cache_pages() 2539 struct folio_batch fbatch; in extent_write_cache_pages() local 2560 folio_batch_init(&fbatch); in extent_write_cache_pages() [all …]
|
/linux-6.3-rc2/fs/ext4/ |
A D | inode.c | 1578 struct folio_batch fbatch; in mpage_release_unused_pages() local 1603 folio_batch_init(&fbatch); in mpage_release_unused_pages() 1626 folio_batch_release(&fbatch); in mpage_release_unused_pages() 2341 struct folio_batch fbatch; in mpage_map_and_submit_buffers() local 2356 folio_batch_init(&fbatch); in mpage_map_and_submit_buffers() 2378 folio_batch_release(&fbatch); in mpage_map_and_submit_buffers() 2385 folio_batch_release(&fbatch); in mpage_map_and_submit_buffers() 2598 struct folio_batch fbatch; in mpage_prepare_extent_to_map() local 2613 folio_batch_init(&fbatch); in mpage_prepare_extent_to_map() 2618 tag, &fbatch); in mpage_prepare_extent_to_map() [all …]
|
/linux-6.3-rc2/fs/ |
A D | buffer.c | 1635 struct folio_batch fbatch; in clean_bdev_aliases() local 1643 folio_batch_init(&fbatch); in clean_bdev_aliases() 1644 while (filemap_get_folios(bd_mapping, &index, end, &fbatch)) { in clean_bdev_aliases() 1645 count = folio_batch_count(&fbatch); in clean_bdev_aliases() 1647 struct folio *folio = fbatch.folios[i]; in clean_bdev_aliases() 1676 folio_batch_release(&fbatch); in clean_bdev_aliases()
|