| /fs/f2fs/ |
| A D | inode.c | 284 if (ino_of_node(node_folio) != nid_of_node(node_folio)) { in sanity_check_inode() 287 ino_of_node(node_folio), nid_of_node(node_folio)); in sanity_check_inode() 411 struct folio *node_folio; in do_read_inode() local 420 if (IS_ERR(node_folio)) in do_read_inode() 421 return PTR_ERR(node_folio); in do_read_inode() 423 ri = F2FS_INODE(node_folio); in do_read_inode() 488 folio_mark_dirty(node_folio); in do_read_inode() 669 folio_mark_dirty(node_folio); in f2fs_update_inode() 673 ri = F2FS_INODE(node_folio); in f2fs_update_inode() 763 struct folio *node_folio; in f2fs_update_inode_page() local [all …]
|
| A D | node.h | 246 static inline nid_t ino_of_node(const struct folio *node_folio) in ino_of_node() argument 248 struct f2fs_node *rn = F2FS_NODE(node_folio); in ino_of_node() 252 static inline nid_t nid_of_node(const struct folio *node_folio) in nid_of_node() argument 254 struct f2fs_node *rn = F2FS_NODE(node_folio); in nid_of_node() 258 static inline unsigned int ofs_of_node(const struct folio *node_folio) in ofs_of_node() argument 260 struct f2fs_node *rn = F2FS_NODE(node_folio); in ofs_of_node() 265 static inline __u64 cpver_of_node(const struct folio *node_folio) in cpver_of_node() argument 267 struct f2fs_node *rn = F2FS_NODE(node_folio); in cpver_of_node() 273 struct f2fs_node *rn = F2FS_NODE(node_folio); in next_blkaddr_of_node() 353 static inline bool IS_DNODE(const struct folio *node_folio) in IS_DNODE() argument [all …]
|
| A D | recovery.c | 497 struct folio *sum_folio, *node_folio; in check_index_in_prev_nodes() local 542 tdn.node_folio = dn->inode_folio; in check_index_in_prev_nodes() 551 node_folio = f2fs_get_node_folio(sbi, nid); in check_index_in_prev_nodes() 552 if (IS_ERR(node_folio)) in check_index_in_prev_nodes() 553 return PTR_ERR(node_folio); in check_index_in_prev_nodes() 555 offset = ofs_of_node(node_folio); in check_index_in_prev_nodes() 556 ino = ino_of_node(node_folio); in check_index_in_prev_nodes() 557 f2fs_folio_put(node_folio, true); in check_index_in_prev_nodes() 675 inode->i_ino, ofs_of_node(dn.node_folio), in do_recover_data() 761 copy_node_footer(dn.node_folio, folio); in do_recover_data() [all …]
|
| A D | gc.c | 1051 struct folio *node_folio; in gc_node_segment() local 1075 if (IS_ERR(node_folio)) in gc_node_segment() 1080 f2fs_folio_put(node_folio, true); in gc_node_segment() 1140 struct folio *node_folio; in is_alive() local 1149 if (IS_ERR(node_folio)) in is_alive() 1153 f2fs_folio_put(node_folio, true); in is_alive() 1164 f2fs_folio_put(node_folio, true); in is_alive() 1168 if (IS_INODE(node_folio)) { in is_alive() 1179 f2fs_folio_put(node_folio, true); in is_alive() 1183 *nofs = ofs_of_node(node_folio); in is_alive() [all …]
|
| A D | node.c | 888 dn->node_folio = nfolio[level]; in f2fs_get_dnode_of_data() 926 dn->node_folio = NULL; in f2fs_get_dnode_of_data() 970 index = dn->node_folio->index; in truncate_node() 976 dn->node_folio = NULL; in truncate_node() 1008 dn->node_folio = folio; in truncate_dnode() 1080 dn->node_folio = folio; in truncate_nodes() 1139 dn->node_folio = folios[idx]; in truncate_partial_nodes() 1801 folio_mark_dirty(node_folio); in f2fs_move_node_folio() 1815 folio_mark_dirty(node_folio); in f2fs_move_node_folio() 1818 folio_unlock(node_folio); in f2fs_move_node_folio() [all …]
|
| A D | f2fs.h | 1029 dn->node_folio = nfolio; in set_new_dnode() 2932 if (dn->node_folio) in f2fs_put_dnode() 2933 f2fs_folio_put(dn->node_folio, true); in f2fs_put_dnode() 2936 dn->node_folio = NULL; in f2fs_put_dnode() 3051 struct folio *node_folio) in get_dnode_base() argument 3053 if (!IS_INODE(node_folio)) in get_dnode_base() 3057 offset_in_addr(&F2FS_NODE(node_folio)->i); in get_dnode_base() 3061 struct folio *node_folio) in get_dnode_addr() argument 3063 return blkaddr_in_node(F2FS_NODE(node_folio)) + in get_dnode_addr() 3064 get_dnode_base(inode, node_folio); in get_dnode_addr() [all …]
|
| A D | compress.c | 913 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_sanity_check_cluster() 954 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in __f2fs_get_cluster_blocks() 1321 if (data_blkaddr(dn.inode, dn.node_folio, in f2fs_write_compressed_pages() 1353 fio.old_blkaddr = data_blkaddr(dn.inode, dn.node_folio, in f2fs_write_compressed_pages() 1886 bool compressed = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous() 1889 block_t first_blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous() 1893 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous()
|
| A D | file.c | 492 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_seek_block() 821 count = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_do_truncate_blocks() 826 if (dn.ofs_in_node || IS_INODE(dn.node_folio)) { in f2fs_do_truncate_blocks() 1240 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_truncate_hole() 1428 ADDRS_PER_PAGE(dn.node_folio, dst_inode) - in __clone_blkaddrs() 1724 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_zero_range() 3777 blkaddr = data_blkaddr(dn->inode, dn->node_folio, in release_compress_blocks() 3895 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_release_compress_blocks() 3946 blkaddr = data_blkaddr(dn->inode, dn->node_folio, in reserve_compress_blocks() 3963 blkaddr = data_blkaddr(dn->inode, dn->node_folio, in reserve_compress_blocks() [all …]
|
| A D | data.c | 1116 __le32 *addr = get_dnode_addr(dn->inode, dn->node_folio); in __set_data_blkaddr() 1130 f2fs_folio_wait_writeback(dn->node_folio, NODE, true, true); in f2fs_set_data_blkaddr() 1132 if (folio_mark_dirty(dn->node_folio)) in f2fs_set_data_blkaddr() 1160 f2fs_folio_wait_writeback(dn->node_folio, NODE, true, true); in f2fs_reserve_new_blocks() 1171 if (folio_mark_dirty(dn->node_folio)) in f2fs_reserve_new_blocks() 1596 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_map_blocks() 2250 blkaddr = from_dnode ? data_blkaddr(dn.inode, dn.node_folio, in f2fs_read_multi_pages() 2284 blkaddr = from_dnode ? data_blkaddr(dn.inode, dn.node_folio, in f2fs_read_multi_pages()
|
| A D | extent_cache.c | 937 ei.fofs = f2fs_start_bidx_of_node(ofs_of_node(dn->node_folio), dn->inode) + in __update_extent_cache()
|
| A D | segment.c | 337 blen = min((pgoff_t)ADDRS_PER_PAGE(dn.node_folio, cow_inode), in __f2fs_commit_atomic_write()
|