Lines Matching refs:i

100 	int i;  in f2fs_drop_rpages()  local
102 for (i = 0; i < len; i++) { in f2fs_drop_rpages()
103 if (!cc->rpages[i]) in f2fs_drop_rpages()
106 unlock_page(cc->rpages[i]); in f2fs_drop_rpages()
108 put_page(cc->rpages[i]); in f2fs_drop_rpages()
125 unsigned int i; in f2fs_put_rpages_wbc() local
127 for (i = 0; i < cc->cluster_size; i++) { in f2fs_put_rpages_wbc()
128 if (!cc->rpages[i]) in f2fs_put_rpages_wbc()
131 redirty_page_for_writepage(wbc, cc->rpages[i]); in f2fs_put_rpages_wbc()
132 f2fs_put_page(cc->rpages[i], unlock); in f2fs_put_rpages_wbc()
607 int i; in f2fs_vmap() local
610 for (i = 0; i < MAX_VMAP_RETRIES; i++) { in f2fs_vmap()
627 int i, ret; in f2fs_compress_pages() local
648 for (i = 0; i < cc->nr_cpages; i++) in f2fs_compress_pages()
649 cc->cpages[i] = f2fs_compress_alloc_page(); in f2fs_compress_pages()
680 for (i = 0; i < COMPRESS_DATA_RESERVED_SIZE; i++) in f2fs_compress_pages()
681 cc->cbuf->reserved[i] = cpu_to_le32(0); in f2fs_compress_pages()
693 for (i = new_nr_cpages; i < cc->nr_cpages; i++) { in f2fs_compress_pages()
694 f2fs_compress_free_page(cc->cpages[i]); in f2fs_compress_pages()
695 cc->cpages[i] = NULL; in f2fs_compress_pages()
712 for (i = 0; i < cc->nr_cpages; i++) { in f2fs_compress_pages()
713 if (cc->cpages[i]) in f2fs_compress_pages()
714 f2fs_compress_free_page(cc->cpages[i]); in f2fs_compress_pages()
851 int i = uptodate ? 0 : 1; in f2fs_all_cluster_page_ready() local
863 for (; i < cc->cluster_size; i++) { in f2fs_all_cluster_page_ready()
864 struct folio *folio = page_folio(pages[index + i]); in f2fs_all_cluster_page_ready()
866 if (folio->index != pgidx + i) in f2fs_all_cluster_page_ready()
879 int i; in cluster_has_invalid_data() local
881 for (i = 0; i < cc->cluster_size; i++) { in cluster_has_invalid_data()
882 struct page *page = cc->rpages[i]; in cluster_has_invalid_data()
900 int i; in f2fs_sanity_check_cluster() local
912 for (i = 1, count = 1; i < cluster_size; i++, count++) { in f2fs_sanity_check_cluster()
914 dn->ofs_in_node + i); in f2fs_sanity_check_cluster()
923 cluster_end = i; in f2fs_sanity_check_cluster()
951 int count, i; in __f2fs_get_cluster_blocks() local
953 for (i = 0, count = 0; i < cluster_size; i++) { in __f2fs_get_cluster_blocks()
955 dn->ofs_in_node + i); in __f2fs_get_cluster_blocks()
1045 int i; in set_cluster_writeback() local
1047 for (i = 0; i < cc->cluster_size; i++) { in set_cluster_writeback()
1048 if (cc->rpages[i]) in set_cluster_writeback()
1049 set_page_writeback(cc->rpages[i]); in set_cluster_writeback()
1056 int i; in cancel_cluster_writeback() local
1067 for (i = 0; i < cc->cluster_size; i++) { in cancel_cluster_writeback()
1068 if (i < submitted) { in cancel_cluster_writeback()
1070 lock_page(cc->rpages[i]); in cancel_cluster_writeback()
1072 clear_page_private_gcing(cc->rpages[i]); in cancel_cluster_writeback()
1073 if (folio_test_writeback(page_folio(cc->rpages[i]))) in cancel_cluster_writeback()
1074 end_page_writeback(cc->rpages[i]); in cancel_cluster_writeback()
1080 int i; in set_cluster_dirty() local
1082 for (i = 0; i < cc->cluster_size; i++) in set_cluster_dirty()
1083 if (cc->rpages[i]) { in set_cluster_dirty()
1084 set_page_dirty(cc->rpages[i]); in set_cluster_dirty()
1085 set_page_private_gcing(cc->rpages[i]); in set_cluster_dirty()
1098 int i, ret; in prepare_compress_overwrite() local
1110 for (i = 0; i < cc->cluster_size; i++) { in prepare_compress_overwrite()
1111 folio = f2fs_filemap_get_folio(mapping, start_idx + i, in prepare_compress_overwrite()
1141 for (i = 0; i < cc->cluster_size; i++) { in prepare_compress_overwrite()
1142 f2fs_bug_on(sbi, cc->rpages[i]); in prepare_compress_overwrite()
1144 folio = filemap_lock_folio(mapping, start_idx + i); in prepare_compress_overwrite()
1157 f2fs_unlock_rpages(cc, i + 1); in prepare_compress_overwrite()
1171 f2fs_unlock_rpages(cc, i); in prepare_compress_overwrite()
1244 int i; in f2fs_truncate_partial_cluster() local
1246 for (i = cluster_size - 1; i >= 0; i--) { in f2fs_truncate_partial_cluster()
1247 struct folio *folio = page_folio(rpages[i]); in f2fs_truncate_partial_cluster()
1294 int i, err; in f2fs_write_compressed_pages() local
1320 for (i = 0; i < cc->cluster_size; i++) { in f2fs_write_compressed_pages()
1322 dn.ofs_in_node + i) == NULL_ADDR) in f2fs_write_compressed_pages()
1348 for (i = 0; i < cc->valid_nr_cpages; i++) { in f2fs_write_compressed_pages()
1349 f2fs_set_compressed_page(cc->cpages[i], inode, in f2fs_write_compressed_pages()
1350 page_folio(cc->rpages[i + 1])->index, cic); in f2fs_write_compressed_pages()
1351 fio.compressed_page = cc->cpages[i]; in f2fs_write_compressed_pages()
1354 dn.ofs_in_node + i + 1); in f2fs_write_compressed_pages()
1360 fio.page = cc->rpages[i + 1]; in f2fs_write_compressed_pages()
1364 cc->cpages[i] = fio.encrypted_page; in f2fs_write_compressed_pages()
1370 for (i = 0; i < cc->cluster_size; i++) in f2fs_write_compressed_pages()
1371 cic->rpages[i] = cc->rpages[i]; in f2fs_write_compressed_pages()
1373 for (i = 0; i < cc->cluster_size; i++, dn.ofs_in_node++) { in f2fs_write_compressed_pages()
1377 fio.page = cc->rpages[i]; in f2fs_write_compressed_pages()
1381 if (i == 0) { in f2fs_write_compressed_pages()
1393 if (i > cc->valid_nr_cpages) { in f2fs_write_compressed_pages()
1404 fio.encrypted_page = cc->cpages[i - 1]; in f2fs_write_compressed_pages()
1406 fio.compressed_page = cc->cpages[i - 1]; in f2fs_write_compressed_pages()
1408 cc->cpages[i - 1] = NULL; in f2fs_write_compressed_pages()
1412 cancel_cluster_writeback(cc, cic, i); in f2fs_write_compressed_pages()
1415 i = cc->valid_nr_cpages; in f2fs_write_compressed_pages()
1452 for (--i; i >= 0; i--) { in f2fs_write_compressed_pages()
1453 if (!cc->cpages[i]) in f2fs_write_compressed_pages()
1455 fscrypt_finalize_bounce_page(&cc->cpages[i]); in f2fs_write_compressed_pages()
1467 for (i = 0; i < cc->valid_nr_cpages; i++) { in f2fs_write_compressed_pages()
1468 f2fs_compress_free_page(cc->cpages[i]); in f2fs_write_compressed_pages()
1469 cc->cpages[i] = NULL; in f2fs_write_compressed_pages()
1483 int i; in f2fs_compress_write_end_io() local
1495 for (i = 0; i < cic->nr_rpages; i++) { in f2fs_compress_write_end_io()
1496 WARN_ON(!cic->rpages[i]); in f2fs_compress_write_end_io()
1497 clear_page_private_gcing(cic->rpages[i]); in f2fs_compress_write_end_io()
1498 end_page_writeback(cic->rpages[i]); in f2fs_compress_write_end_io()
1512 int submitted, compr_blocks, i; in f2fs_write_raw_pages() local
1517 for (i = 0; i < cc->cluster_size; i++) { in f2fs_write_raw_pages()
1518 if (!cc->rpages[i]) in f2fs_write_raw_pages()
1521 redirty_page_for_writepage(wbc, cc->rpages[i]); in f2fs_write_raw_pages()
1522 unlock_page(cc->rpages[i]); in f2fs_write_raw_pages()
1532 for (i = 0; i < cc->cluster_size; i++) { in f2fs_write_raw_pages()
1535 if (!cc->rpages[i]) in f2fs_write_raw_pages()
1537 folio = page_folio(cc->rpages[i]); in f2fs_write_raw_pages()
1636 int i; in f2fs_prepare_decomp_mem() local
1645 for (i = 0; i < dic->cluster_size; i++) { in f2fs_prepare_decomp_mem()
1646 if (dic->rpages[i]) { in f2fs_prepare_decomp_mem()
1647 dic->tpages[i] = dic->rpages[i]; in f2fs_prepare_decomp_mem()
1651 dic->tpages[i] = f2fs_compress_alloc_page(); in f2fs_prepare_decomp_mem()
1694 int i, ret; in f2fs_alloc_dic() local
1719 for (i = 0; i < dic->cluster_size; i++) in f2fs_alloc_dic()
1720 dic->rpages[i] = cc->rpages[i]; in f2fs_alloc_dic()
1729 for (i = 0; i < dic->nr_cpages; i++) { in f2fs_alloc_dic()
1734 start_idx + i + 1, dic); in f2fs_alloc_dic()
1735 dic->cpages[i] = page; in f2fs_alloc_dic()
1752 int i; in f2fs_free_dic() local
1759 for (i = 0; i < dic->cluster_size; i++) { in f2fs_free_dic()
1760 if (dic->rpages[i]) in f2fs_free_dic()
1762 if (!dic->tpages[i]) in f2fs_free_dic()
1764 f2fs_compress_free_page(dic->tpages[i]); in f2fs_free_dic()
1770 for (i = 0; i < dic->nr_cpages; i++) { in f2fs_free_dic()
1771 if (!dic->cpages[i]) in f2fs_free_dic()
1773 f2fs_compress_free_page(dic->cpages[i]); in f2fs_free_dic()
1806 int i; in f2fs_verify_cluster() local
1809 for (i = 0; i < dic->cluster_size; i++) { in f2fs_verify_cluster()
1810 struct page *rpage = dic->rpages[i]; in f2fs_verify_cluster()
1832 int i; in f2fs_decompress_end_io() local
1847 for (i = 0; i < dic->cluster_size; i++) { in f2fs_decompress_end_io()
1848 struct page *rpage = dic->rpages[i]; in f2fs_decompress_end_io()
1888 int i = compressed ? 1 : 0; in f2fs_cluster_blocks_are_contiguous() local
1890 ofs_in_node + i); in f2fs_cluster_blocks_are_contiguous()
1892 for (i += 1; i < F2FS_I(dn->inode)->i_cluster_size; i++) { in f2fs_cluster_blocks_are_contiguous()
1894 ofs_in_node + i); in f2fs_cluster_blocks_are_contiguous()
1898 if (first_blkaddr + i - (compressed ? 1 : 0) != blkaddr) in f2fs_cluster_blocks_are_contiguous()
1902 return compressed ? i - 1 : i; in f2fs_cluster_blocks_are_contiguous()
2000 unsigned int nr, i; in f2fs_invalidate_compress_pages() local
2006 for (i = 0; i < nr; i++) { in f2fs_invalidate_compress_pages()
2007 struct folio *folio = fbatch.folios[i]; in f2fs_invalidate_compress_pages()