/linux-6.3-rc2/include/linux/ |
A D | pagevec.h | 75 struct folio_batch { struct 82 static_assert(sizeof(struct pagevec) == sizeof(struct folio_batch)); argument 84 offsetof(struct folio_batch, folios)); 92 static inline void folio_batch_init(struct folio_batch *fbatch) in folio_batch_init() 98 static inline void folio_batch_reinit(struct folio_batch *fbatch) in folio_batch_reinit() 103 static inline unsigned int folio_batch_count(struct folio_batch *fbatch) in folio_batch_count() 108 static inline unsigned int fbatch_space(struct folio_batch *fbatch) in fbatch_space() 123 static inline unsigned folio_batch_add(struct folio_batch *fbatch, in folio_batch_add() 130 static inline void folio_batch_release(struct folio_batch *fbatch) in folio_batch_release() 135 void folio_batch_remove_exceptionals(struct folio_batch *fbatch);
|
A D | pagemap.h | 19 struct folio_batch; 739 pgoff_t end, struct folio_batch *fbatch); 741 pgoff_t *start, pgoff_t end, struct folio_batch *fbatch); 743 pgoff_t end, xa_mark_t tag, struct folio_batch *fbatch); 1111 struct folio_batch *fbatch);
|
A D | swap.h | 445 void check_move_unevictable_folios(struct folio_batch *fbatch);
|
/linux-6.3-rc2/mm/ |
A D | swap.c | 53 struct folio_batch fbatch; 65 struct folio_batch lru_add; 68 struct folio_batch lru_lazyfree; 70 struct folio_batch activate; 258 struct folio_batch *fbatch; in folio_rotate_reclaimable() 356 struct folio_batch *fbatch; in folio_activate() 386 struct folio_batch *fbatch; in __lru_cache_activate_folio() 503 struct folio_batch *fbatch; in folio_add_lru() 692 struct folio_batch *fbatch; in deactivate_file_folio() 717 struct folio_batch *fbatch; in folio_deactivate() [all …]
|
A D | mlock.c | 33 struct folio_batch fbatch; 186 static void mlock_folio_batch(struct folio_batch *fbatch) in mlock_folio_batch() 215 struct folio_batch *fbatch; in mlock_drain_local() 226 struct folio_batch *fbatch; in mlock_drain_remote() 245 struct folio_batch *fbatch; in mlock_folio() 270 struct folio_batch *fbatch; in mlock_new_folio() 293 struct folio_batch *fbatch; in munlock_folio()
|
A D | truncate.c | 61 struct folio_batch *fbatch, pgoff_t *indices) in truncate_folio_batch_exceptionals() 335 struct folio_batch fbatch; in truncate_inode_pages_range() 502 struct folio_batch fbatch; in invalidate_mapping_pagevec() 624 struct folio_batch fbatch; in invalidate_inode_pages2_range()
|
A D | filemap.c | 280 struct folio_batch *fbatch) in page_cache_delete_batch() 321 struct folio_batch *fbatch) in delete_from_page_cache_batch() 508 struct folio_batch fbatch; in __filemap_fdatawait_range() 2168 pgoff_t end, struct folio_batch *fbatch) in filemap_get_folios() 2231 pgoff_t *start, pgoff_t end, struct folio_batch *fbatch) in filemap_get_folios_contig() 2302 pgoff_t end, xa_mark_t tag, struct folio_batch *fbatch) in filemap_get_folios_tag() 2372 pgoff_t index, pgoff_t max, struct folio_batch *fbatch) in filemap_get_read_batch() 2522 struct folio_batch *fbatch) in filemap_create_folio() 2578 struct folio_batch *fbatch, bool need_uptodate) in filemap_get_pages() 2666 struct folio_batch fbatch; in filemap_read() [all …]
|
A D | internal.h | 16 struct folio_batch; 128 pgoff_t end, struct folio_batch *fbatch, pgoff_t *indices); 130 pgoff_t end, struct folio_batch *fbatch, pgoff_t *indices);
|
A D | shmem.c | 865 struct folio_batch fbatch; in shmem_unlock_mapping() 912 struct folio_batch fbatch; in shmem_undo_range() 1184 pgoff_t start, struct folio_batch *fbatch, in shmem_find_swap_entries() 1226 struct folio_batch *fbatch, pgoff_t *indices) in shmem_unuse_swap_entries() 1261 struct folio_batch fbatch; in shmem_unuse_inode()
|
A D | page-writeback.c | 2401 struct folio_batch fbatch; in write_cache_pages()
|
/linux-6.3-rc2/fs/nilfs2/ |
A D | page.c | 243 struct folio_batch fbatch; in nilfs_copy_dirty_pages() 298 struct folio_batch fbatch; in nilfs_copy_back_pages() 361 struct folio_batch fbatch; in nilfs_clear_dirty_pages() 487 struct folio_batch fbatch; in nilfs_find_uncommitted_extent()
|
A D | segment.c | 683 struct folio_batch fbatch; in nilfs_lookup_dirty_data_buffers() 740 struct folio_batch fbatch; in nilfs_lookup_dirty_node_buffers()
|
/linux-6.3-rc2/fs/ramfs/ |
A D | file-nommu.c | 208 struct folio_batch fbatch; in ramfs_nommu_get_unmapped_area()
|
/linux-6.3-rc2/fs/gfs2/ |
A D | aops.c | 208 struct folio_batch *fbatch, in gfs2_write_jdata_batch() 312 struct folio_batch fbatch; in gfs2_write_cache_jdata()
|
/linux-6.3-rc2/fs/btrfs/tests/ |
A D | extent-io-tests.c | 24 struct folio_batch fbatch; in process_page_range()
|
/linux-6.3-rc2/fs/f2fs/ |
A D | node.c | 1521 struct folio_batch fbatch; in last_fsync_dnode() 1734 struct folio_batch fbatch; in f2fs_fsync_node_pages() 1893 struct folio_batch fbatch; in f2fs_flush_inline_data() 1941 struct folio_batch fbatch; in f2fs_sync_node_pages()
|
A D | checkpoint.c | 396 struct folio_batch fbatch; in f2fs_sync_meta_pages()
|
A D | compress.c | 1898 struct folio_batch fbatch; in f2fs_invalidate_compress_pages()
|
/linux-6.3-rc2/fs/afs/ |
A D | write.c | 707 struct folio_batch fbatch; in afs_writepages_region()
|
/linux-6.3-rc2/fs/btrfs/ |
A D | compression.c | 181 struct folio_batch fbatch; in end_compressed_writeback()
|
A D | extent_io.c | 297 struct folio_batch fbatch; in __process_pages_contig() 2411 struct folio_batch fbatch; in btree_write_cache_pages() 2539 struct folio_batch fbatch; in extent_write_cache_pages()
|
/linux-6.3-rc2/fs/hugetlbfs/ |
A D | inode.c | 621 struct folio_batch fbatch; in remove_inode_hugepages()
|
/linux-6.3-rc2/fs/ext4/ |
A D | inode.c | 1578 struct folio_batch fbatch; in mpage_release_unused_pages() 2341 struct folio_batch fbatch; in mpage_map_and_submit_buffers() 2598 struct folio_batch fbatch; in mpage_prepare_extent_to_map()
|
/linux-6.3-rc2/fs/ceph/ |
A D | addr.c | 803 struct folio_batch fbatch; in ceph_writepages_start()
|
/linux-6.3-rc2/fs/ |
A D | buffer.c | 1635 struct folio_batch fbatch; in clean_bdev_aliases()
|