| /fs/iomap/ |
| A D | swapfile.c | 16 unsigned long nr_pages; /* number of pages collected */ member 30 unsigned long nr_pages; in iomap_swapfile_add_extent() local 37 if (unlikely(isi->nr_pages >= isi->sis->max)) in iomap_swapfile_add_extent() 39 max_pages = isi->sis->max - isi->nr_pages; in iomap_swapfile_add_extent() 52 nr_pages = next_ppage - first_ppage; in iomap_swapfile_add_extent() 53 nr_pages = min(nr_pages, max_pages); in iomap_swapfile_add_extent() 69 error = add_swap_extent(isi->sis, isi->nr_pages, nr_pages, first_ppage); in iomap_swapfile_add_extent() 73 isi->nr_pages += nr_pages; in iomap_swapfile_add_extent() 182 if (isi.nr_pages == 0) { in iomap_swapfile_activate() 188 sis->max = isi.nr_pages; in iomap_swapfile_activate() [all …]
|
| A D | trace.h | 28 TP_PROTO(struct inode *inode, int nr_pages), 29 TP_ARGS(inode, nr_pages), 33 __field(int, nr_pages) 38 __entry->nr_pages = nr_pages; 43 __entry->nr_pages) 48 TP_PROTO(struct inode *inode, int nr_pages), \ 49 TP_ARGS(inode, nr_pages))
|
| A D | direct-io.c | 336 int nr_pages, ret = 0; in iomap_dio_bio_iter() local 419 nr_pages = bio_iov_vecs_to_alloc(dio->submit.iter, BIO_MAX_VECS); in iomap_dio_bio_iter() 428 bio = iomap_dio_alloc_bio(iter, dio, nr_pages, bio_opf); in iomap_dio_bio_iter() 469 nr_pages = bio_iov_vecs_to_alloc(dio->submit.iter, in iomap_dio_bio_iter() 474 if (nr_pages) in iomap_dio_bio_iter() 478 } while (nr_pages); in iomap_dio_bio_iter()
|
| /fs/crypto/ |
| A D | bio.c | 125 unsigned int nr_pages; in fscrypt_zeroout_range() local 139 nr_pages = min_t(u64, ARRAY_SIZE(pages), in fscrypt_zeroout_range() 149 for (i = 0; i < nr_pages; i++) { in fscrypt_zeroout_range() 155 nr_pages = i; in fscrypt_zeroout_range() 156 if (WARN_ON_ONCE(nr_pages <= 0)) in fscrypt_zeroout_range() 160 bio = bio_alloc(inode->i_sb->s_bdev, nr_pages, REQ_OP_WRITE, GFP_NOFS); in fscrypt_zeroout_range() 185 } while (i != nr_pages && du_remaining != 0); in fscrypt_zeroout_range() 195 for (i = 0; i < nr_pages; i++) in fscrypt_zeroout_range()
|
| /fs/ |
| A D | fs-writeback.c | 43 long nr_pages; member 1026 nr_pages = wb_split_bdi_pages(wb, base_work->nr_pages); in bdi_split_work_to_wbs() 1031 work->nr_pages = nr_pages; in bdi_split_work_to_wbs() 1050 work->nr_pages = nr_pages; in bdi_split_work_to_wbs() 2066 .nr_pages = nr_pages, in writeback_inodes_wb() 2081 return nr_pages - work.nr_pages; in writeback_inodes_wb() 2102 long nr_pages = work->nr_pages; in wb_writeback() local 2196 return nr_pages - work->nr_pages; in wb_writeback() 2237 long nr_pages; in wb_check_old_data_flush() local 2255 .nr_pages = nr_pages, in wb_check_old_data_flush() [all …]
|
| A D | aio.c | 126 long nr_pages; member 269 inode->i_size = PAGE_SIZE * nr_pages; in aio_private_file() 336 for (i = 0; i < ctx->nr_pages; i++) { in aio_free_ring() 433 if (idx < (pgoff_t)ctx->nr_pages) { in aio_migrate_folio() 487 int nr_pages; in aio_setup_ring() local 497 nr_pages = PFN_UP(size); in aio_setup_ring() 498 if (nr_pages < 0) in aio_setup_ring() 512 if (nr_pages > AIO_RING_PAGES) { in aio_setup_ring() 521 for (i = 0; i < nr_pages; i++) { in aio_setup_ring() 536 ctx->nr_pages = i; in aio_setup_ring() [all …]
|
| A D | mpage.c | 134 unsigned int nr_pages; member 182 last_block = block_in_file + ((args->nr_pages * PAGE_SIZE) >> blkbits); in do_mpage_readpage() 287 args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), opf, in do_mpage_readpage() 370 args.nr_pages = readahead_count(rac); in mpage_readahead() 385 .nr_pages = folio_nr_pages(folio), in mpage_read_folio()
|
| A D | file_table.c | 613 unsigned long nr_pages = totalram_pages(); in files_maxfiles_init() local 614 unsigned long memreserve = (nr_pages - nr_free_pages()) * 3/2; in files_maxfiles_init() 616 memreserve = min(memreserve, nr_pages - 1); in files_maxfiles_init() 617 n = ((nr_pages - memreserve) * (PAGE_SIZE / 1024)) / 10; in files_maxfiles_init()
|
| A D | direct-io.c | 667 int ret, nr_pages; in dio_new_bio() local 673 nr_pages = bio_max_segs(sdio->pages_in_io); in dio_new_bio() 674 BUG_ON(nr_pages <= 0); in dio_new_bio() 675 dio_bio_alloc(dio, sdio, map_bh->b_bdev, sector, nr_pages); in dio_new_bio()
|
| /fs/ramfs/ |
| A D | file-nommu.c | 206 unsigned long maxpages, lpages, nr_folios, loop, ret, nr_pages, pfn; in ramfs_nommu_get_unmapped_area() local 225 nr_pages = 0; in ramfs_nommu_get_unmapped_area() 240 if (pfn + nr_pages != folio_pfn(fbatch.folios[loop])) { in ramfs_nommu_get_unmapped_area() 244 nr_pages += folio_nr_pages(fbatch.folios[loop]); in ramfs_nommu_get_unmapped_area() 245 if (nr_pages >= lpages) in ramfs_nommu_get_unmapped_area() 249 if (nr_pages < lpages) { in ramfs_nommu_get_unmapped_area()
|
| /fs/ext4/ |
| A D | readpage.c | 230 unsigned int nr_pages, folio_pages; in ext4_mpage_readpages() local 237 nr_pages = rac ? readahead_count(rac) : folio_nr_pages(folio); in ext4_mpage_readpages() 238 for (; nr_pages; nr_pages -= folio_pages) { in ext4_mpage_readpages() 256 last_block = block_in_file + nr_pages * blocks_per_page; in ext4_mpage_readpages() 359 bio = bio_alloc(bdev, bio_max_segs(nr_pages), in ext4_mpage_readpages()
|
| /fs/squashfs/ |
| A D | file.c | 574 unsigned int nr_pages = 0; in squashfs_readahead() local 598 nr_pages = __readahead_batch(ractl, pages, max_pages); in squashfs_readahead() 599 if (!nr_pages) in squashfs_readahead() 608 nr_pages, expected, start); in squashfs_readahead() 618 actor = squashfs_page_actor_init_special(msblk, pages, nr_pages, in squashfs_readahead() 636 for (i = 0; i < nr_pages; i++) { in squashfs_readahead() 642 for (i = 0; i < nr_pages; i++) { in squashfs_readahead() 654 for (i = 0; i < nr_pages; i++) { in squashfs_readahead()
|
| /fs/bcachefs/ |
| A D | compress.c | 86 unsigned nr_pages = 0; in __bio_map_or_bounce() local 111 nr_pages++; in __bio_map_or_bounce() 114 BUG_ON(DIV_ROUND_UP(start.bi_size, PAGE_SIZE) > nr_pages); in __bio_map_or_bounce() 116 pages = nr_pages > ARRAY_SIZE(stack_pages) in __bio_map_or_bounce() 117 ? kmalloc_array(nr_pages, sizeof(struct page *), GFP_NOFS) in __bio_map_or_bounce() 122 nr_pages = 0; in __bio_map_or_bounce() 124 pages[nr_pages++] = bv.bv_page; in __bio_map_or_bounce() 126 data = vmap(pages, nr_pages, VM_MAP, PAGE_KERNEL); in __bio_map_or_bounce()
|
| /fs/btrfs/ |
| A D | defrag.c | 1154 struct folio **folios, int nr_pages, in defrag_one_locked_target() argument 1176 for (int i = 0; i < nr_pages && folios[i]; i++) { in defrag_one_locked_target() 1203 const unsigned int nr_pages = ((start + len - 1) >> PAGE_SHIFT) - in defrag_one_range() local 1207 ASSERT(nr_pages <= CLUSTER_SIZE / PAGE_SIZE); in defrag_one_range() 1210 folios = kcalloc(nr_pages, sizeof(struct folio *), GFP_NOFS); in defrag_one_range() 1215 for (int i = 0; cur < start + len && i < nr_pages; i++) { in defrag_one_range() 1224 for (int i = 0; i < nr_pages; i++) { in defrag_one_range() 1248 ret = defrag_one_locked_target(inode, entry, folios, nr_pages, in defrag_one_range() 1261 for (int i = 0; i < nr_pages; i++) { in defrag_one_range()
|
| A D | raid56.h | 68 u16 nr_pages; member
|
| A D | raid56.c | 178 for (i = 0; i < rbio->nr_pages; i++) { in free_raid_bio() 309 ASSERT(page_nr < rbio->nr_pages); in full_page_sectors_uptodate() 334 ASSERT(page_index < rbio->nr_pages); in index_stripe_sectors() 392 for (i = 0; i < dest->nr_pages; i++) { in steal_rbio() 1044 rbio->nr_pages = num_pages; in alloc_rbio() 1064 ret = btrfs_alloc_page_array(rbio->nr_pages, rbio->stripe_pages, false); in alloc_rbio_pages() 1078 ret = btrfs_alloc_page_array(rbio->nr_pages - data_pages, in alloc_rbio_parity_pages()
|
| /fs/erofs/ |
| A D | dir.c | 56 pgoff_t nr_pages = DIV_ROUND_UP_POW2(dir->i_size, PAGE_SIZE); in erofs_readdir() local 74 pgoff_t pages = min(nr_pages - idx, ra_pages); in erofs_readdir()
|
| A D | zdata.c | 1132 unsigned int onstack_used, nr_pages; member 1151 DBG_BUGON((poff >> PAGE_SHIFT) >= be->nr_pages); in z_erofs_do_decompressed_bvec() 1186 DBG_BUGON(pgnr >= be->nr_pages); in z_erofs_fill_other_copies() 1281 be->nr_pages = PAGE_ALIGN(pcl->length + pcl->pageofs_out) >> PAGE_SHIFT; in z_erofs_decompress_pcluster() 1287 if (be->nr_pages <= Z_EROFS_ONSTACK_PAGES) { in z_erofs_decompress_pcluster() 1289 be->onstack_used = be->nr_pages; in z_erofs_decompress_pcluster() 1291 sizeof(struct page *) * be->nr_pages); in z_erofs_decompress_pcluster() 1299 kvcalloc(be->nr_pages, sizeof(struct page *), in z_erofs_decompress_pcluster() 1316 .outpages = be->nr_pages, in z_erofs_decompress_pcluster() 1353 for (i = 0; i < be->nr_pages; ++i) { in z_erofs_decompress_pcluster()
|
| /fs/nfs/ |
| A D | nfstrace.h | 1057 unsigned int nr_pages 1060 TP_ARGS(inode, pos, nr_pages), 1068 __field(unsigned int, nr_pages) 1079 __entry->nr_pages = nr_pages; 1087 __entry->offset, __entry->nr_pages 1094 unsigned int nr_pages, 1098 TP_ARGS(inode, nr_pages, ret), 1107 __field(unsigned int, nr_pages) 1117 __entry->nr_pages = nr_pages; 1126 __entry->nr_pages, __entry->ret
|
| A D | read.c | 410 unsigned int nr_pages = readahead_count(ractl); in nfs_readahead() local 416 trace_nfs_aop_readahead(inode, readahead_pos(ractl), nr_pages); in nfs_readahead() 450 trace_nfs_aop_readahead_done(inode, nr_pages, ret); in nfs_readahead()
|
| /fs/f2fs/ |
| A D | data.c | 2070 unsigned nr_pages, in f2fs_read_single_page() argument 2377 unsigned max_nr_pages = nr_pages; in f2fs_mpage_readpages() 2397 for (; nr_pages; nr_pages--) { in f2fs_mpage_readpages() 2969 int nr_pages; in f2fs_write_cache_pages() local 3016 nr_pages = 0; in f2fs_write_cache_pages() 3021 if (nr_pages) in f2fs_write_cache_pages() 3034 if (++nr_pages == max_pages) { in f2fs_write_cache_pages() 3045 for (i = 0; i < nr_pages; i++) { in f2fs_write_cache_pages() 3079 pages, i, nr_pages, true)) in f2fs_write_cache_pages() 3093 pages, i, nr_pages, in f2fs_write_cache_pages() [all …]
|
| /fs/fuse/ |
| A D | dax.c | 1197 long nr_pages, nr_ranges; in fuse_dax_mem_range_init() local 1209 nr_pages = dax_direct_access(fcd->dev, 0, PHYS_PFN(dax_size), in fuse_dax_mem_range_init() 1212 if (nr_pages < 0) { in fuse_dax_mem_range_init() 1213 pr_debug("dax_direct_access() returned %ld\n", nr_pages); in fuse_dax_mem_range_init() 1214 return nr_pages; in fuse_dax_mem_range_init() 1217 nr_ranges = nr_pages/FUSE_DAX_PAGES; in fuse_dax_mem_range_init() 1219 __func__, nr_pages, nr_ranges); in fuse_dax_mem_range_init()
|
| A D | virtio_fs.c | 1009 long nr_pages, enum dax_access_mode mode, in virtio_fs_direct_access() argument 1020 return nr_pages > max_nr_pages ? max_nr_pages : nr_pages; in virtio_fs_direct_access() 1024 pgoff_t pgoff, size_t nr_pages) in virtio_fs_zero_page_range() argument 1029 rc = dax_direct_access(dax_dev, pgoff, nr_pages, DAX_ACCESS, &kaddr, in virtio_fs_zero_page_range() 1034 memset(kaddr, 0, nr_pages << PAGE_SHIFT); in virtio_fs_zero_page_range() 1035 dax_flush(dax_dev, kaddr, nr_pages << PAGE_SHIFT); in virtio_fs_zero_page_range()
|
| A D | file.c | 935 unsigned int max_pages, nr_pages; in fuse_readahead() local 953 nr_pages = readahead_count(rac); in fuse_readahead() 955 while (nr_pages) { in fuse_readahead() 958 unsigned cur_pages = min(max_pages, nr_pages); in fuse_readahead() 1004 nr_pages -= pages; in fuse_readahead() 1275 unsigned int nr_pages = fuse_wr_pages(pos, iov_iter_count(ii), in fuse_perform_write() local 1284 count = fuse_fill_write_pages(&ia, mapping, ii, pos, nr_pages); in fuse_perform_write() 1489 unsigned int nr_pages = 0; in fuse_get_user_pages() local 1531 while (nbytes < *nbytesp && nr_pages < max_pages) { in fuse_get_user_pages() 1537 max_pages - nr_pages, in fuse_get_user_pages() [all …]
|
| /fs/proc/ |
| A D | kcore.c | 164 unsigned long nr_pages = ent->size >> PAGE_SHIFT; in get_sparsemem_vmemmap_info() local 170 end = ((unsigned long)pfn_to_page(pfn + nr_pages)) - 1; in get_sparsemem_vmemmap_info() 202 kclist_add_private(unsigned long pfn, unsigned long nr_pages, void *arg) in kclist_add_private() argument 217 ent->size = nr_pages << PAGE_SHIFT; in kclist_add_private()
|