| /fs/notify/dnotify/ |
| A D | dnotify.c | 72 for (dn = dn_mark->dn; dn != NULL; dn = dn->dn_next) in dnotify_recalc_inode_mask() 174 if ((dn->dn_owner == id) && (dn->dn_filp == filp)) { in dnotify_flush() 233 odn = dn_mark->dn; in attach_dn() 245 dn->dn_fd = fd; in attach_dn() 247 dn->dn_owner = id; in attach_dn() 248 dn->dn_next = dn_mark->dn; in attach_dn() 249 dn_mark->dn = dn; in attach_dn() 272 dn = NULL; in fcntl_dirnotify() 306 if (!dn) { in fcntl_dirnotify() 374 dn = NULL; in fcntl_dirnotify() [all …]
|
| /fs/f2fs/ |
| A D | recovery.c | 530 max_addrs = ADDRS_PER_PAGE(dn->node_folio, dn->inode); in check_index_in_prev_nodes() 540 if (!dn->inode_folio_locked) in check_index_in_prev_nodes() 545 } else if (dn->nid == nid) { in check_index_in_prev_nodes() 573 inode = dn->inode; in check_index_in_prev_nodes() 583 if (ino == dn->inode->i_ino && dn->inode_folio_locked) in check_index_in_prev_nodes() 595 if (ino != dn->inode->i_ino) in check_index_in_prev_nodes() 604 if (dn->inode->i_ino == nid && !dn->inode_folio_locked) in check_index_in_prev_nodes() 625 struct dnode_of_data dn; in do_recover_data() local 686 dest = data_blkaddr(dn.inode, folio, dn.ofs_in_node); in do_recover_data() 762 fill_node_footer(dn.node_folio, dn.nid, ni.ino, in do_recover_data() [all …]
|
| A D | node.c | 889 dn->data_blkaddr = f2fs_data_blkaddr(dn); in f2fs_get_dnode_of_data() 909 blkaddr = data_blkaddr(dn->inode, dn->node_folio, ofs_in_node); in f2fs_get_dnode_of_data() 911 blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_get_dnode_of_data() 958 dec_valid_node_count(sbi, dn->inode, dn->nid == dn->inode->i_ino); in truncate_node() 961 if (dn->nid == dn->inode->i_ino) { in truncate_node() 977 trace_f2fs_truncate_node(dn->inode, dn->nid, ni.blk_addr); in truncate_node() 1000 dn->inode->i_ino, dn->nid, ino_of_node(folio)); in truncate_dnode() 1034 trace_f2fs_truncate_nodes_enter(dn->inode, dn->nid, dn->data_blkaddr); in truncate_nodes() 1036 folio = f2fs_get_node_folio(F2FS_I_SB(dn->inode), dn->nid); in truncate_nodes() 1404 fill_node_footer(folio, dn->nid, dn->inode->i_ino, ofs, true); in f2fs_new_node_folio() [all …]
|
| A D | data.c | 1116 __le32 *addr = get_dnode_addr(dn->inode, dn->node_folio); in __set_data_blkaddr() 1119 addr[dn->ofs_in_node] = cpu_to_le32(dn->data_blkaddr); in __set_data_blkaddr() 1157 trace_f2fs_reserve_new_blocks(dn->inode, dn->nid, in f2fs_reserve_new_blocks() 1199 f2fs_put_dnode(dn); in f2fs_reserve_block() 1232 f2fs_put_dnode(&dn); in f2fs_get_read_data_folio() 1407 dn->data_blkaddr = f2fs_data_blkaddr(dn); in __allocate_data_block() 1414 set_summary(&sum, dn->nid, dn->ofs_in_node, ni.version); in __allocate_data_block() 1424 f2fs_update_data_blkaddr(dn, dn->data_blkaddr); in __allocate_data_block() 1707 dn.ofs_in_node++; in f2fs_map_blocks() 2250 blkaddr = from_dnode ? data_blkaddr(dn.inode, dn.node_folio, in f2fs_read_multi_pages() [all …]
|
| A D | inline.c | 148 .sbi = F2FS_I_SB(dn->inode), in f2fs_convert_inline_folio() 149 .ino = dn->inode->i_ino, in f2fs_convert_inline_folio() 170 f2fs_put_dnode(dn); in f2fs_convert_inline_folio() 177 f2fs_put_dnode(dn); in f2fs_convert_inline_folio() 180 __func__, dn->inode->i_ino, dn->data_blkaddr); in f2fs_convert_inline_folio() 208 f2fs_truncate_inline_inode(dn->inode, dn->inode_folio, 0); in f2fs_convert_inline_folio() 213 f2fs_put_dnode(dn); in f2fs_convert_inline_folio() 220 struct dnode_of_data dn; in f2fs_convert_inline_inode() local 251 f2fs_put_dnode(&dn); in f2fs_convert_inline_inode() 414 struct dnode_of_data dn; in f2fs_move_inline_dirents() local [all …]
|
| A D | file.c | 139 f2fs_put_dnode(&dn); in f2fs_vm_page_mkwrite() 422 block_t first_blkaddr = data_blkaddr(dn->inode, dn->node_folio, in __found_offset() 650 addr = get_dnode_addr(dn->inode, dn->node_folio) + ofs; in f2fs_truncate_data_blocks_range() 723 trace_f2fs_truncate_data_blocks_range(dn->inode, dn->nid, in f2fs_truncate_data_blocks_range() 826 if (dn.ofs_in_node || IS_INODE(dn.node_folio)) { in f2fs_do_truncate_blocks() 1431 dn.data_blkaddr = f2fs_data_blkaddr(&dn); in __clone_blkaddrs() 1439 f2fs_replace_block(sbi, &dn, dn.data_blkaddr, in __clone_blkaddrs() 1622 dn->data_blkaddr = f2fs_data_blkaddr(dn); in f2fs_do_zero_range() 3777 blkaddr = data_blkaddr(dn->inode, dn->node_folio, in release_compress_blocks() 3946 blkaddr = data_blkaddr(dn->inode, dn->node_folio, in reserve_compress_blocks() [all …]
|
| A D | compress.c | 913 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_sanity_check_cluster() 939 dn->inode->i_ino, dn->nid, dn->ofs_in_node, reason); in f2fs_sanity_check_cluster() 954 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in __f2fs_get_cluster_blocks() 994 f2fs_put_dnode(&dn); in __f2fs_cluster_blocks() 1321 if (data_blkaddr(dn.inode, dn.node_folio, in f2fs_write_compressed_pages() 1353 fio.old_blkaddr = data_blkaddr(dn.inode, dn.node_folio, in f2fs_write_compressed_pages() 1432 f2fs_put_dnode(&dn); in f2fs_write_compressed_pages() 1460 f2fs_put_dnode(&dn); in f2fs_write_compressed_pages() 1886 bool compressed = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous() 1889 block_t first_blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous() [all …]
|
| A D | extent_cache.c | 934 if (!__may_extent_tree(dn->inode, type)) in __update_extent_cache() 937 ei.fofs = f2fs_start_bidx_of_node(ofs_of_node(dn->node_folio), dn->inode) + in __update_extent_cache() 938 dn->ofs_in_node; in __update_extent_cache() 942 if (dn->data_blkaddr == NEW_ADDR) in __update_extent_cache() 945 ei.blk = dn->data_blkaddr; in __update_extent_cache() 947 if (__get_new_block_age(dn->inode, &ei, dn->data_blkaddr)) in __update_extent_cache() 950 __update_extent_tree_range(dn->inode, &ei, type); in __update_extent_cache() 1055 return __update_extent_cache(dn, EX_READ); in f2fs_update_read_extent_cache() 1067 if (!__may_extent_tree(dn->inode, EX_READ)) in f2fs_update_read_extent_cache_range() 1093 return __update_extent_cache(dn, EX_BLOCK_AGE); in f2fs_update_age_extent_cache() [all …]
|
| A D | gc.c | 1210 struct dnode_of_data dn; in ra_data_block() local 1229 &dn.data_blkaddr)) { in ra_data_block() 1238 set_new_dnode(&dn, inode, NULL, NULL, 0); in ra_data_block() 1242 f2fs_put_dnode(&dn); in ra_data_block() 1267 dn.data_blkaddr, in ra_data_block() 1310 struct dnode_of_data dn; in move_data_block() local 1335 set_new_dnode(&dn, inode, NULL, NULL, 0); in move_data_block() 1340 if (unlikely(dn.data_blkaddr == NULL_ADDR)) { in move_data_block() 1396 set_summary(&sum, dn.nid, dn.ofs_in_node, ni.version); in move_data_block() 1436 f2fs_update_data_blkaddr(&dn, newaddr); in move_data_block() [all …]
|
| A D | f2fs.h | 1026 memset(dn, 0, sizeof(*dn)); in set_new_dnode() 1027 dn->inode = inode; in set_new_dnode() 1028 dn->inode_folio = ifolio; in set_new_dnode() 1029 dn->node_folio = nfolio; in set_new_dnode() 1030 dn->nid = nid; in set_new_dnode() 2932 if (dn->node_folio) in f2fs_put_dnode() 2933 f2fs_folio_put(dn->node_folio, true); in f2fs_put_dnode() 2934 if (dn->inode_folio && dn->node_folio != dn->inode_folio) in f2fs_put_dnode() 2936 dn->node_folio = NULL; in f2fs_put_dnode() 2937 dn->inode_folio = NULL; in f2fs_put_dnode() [all …]
|
| A D | segment.c | 228 struct dnode_of_data dn; in __replace_atomic_write_block() local 245 f2fs_put_dnode(&dn); in __replace_atomic_write_block() 257 f2fs_replace_block(sbi, &dn, dn.data_blkaddr, in __replace_atomic_write_block() 265 f2fs_put_dnode(&dn); in __replace_atomic_write_block() 273 f2fs_replace_block(sbi, &dn, dn.data_blkaddr, new_addr, in __replace_atomic_write_block() 277 f2fs_put_dnode(&dn); in __replace_atomic_write_block() 347 f2fs_put_dnode(&dn); in __f2fs_commit_atomic_write() 358 f2fs_put_dnode(&dn); in __f2fs_commit_atomic_write() 367 f2fs_put_dnode(&dn); in __f2fs_commit_atomic_write() 4015 set_summary(&sum, dn->nid, dn->ofs_in_node, fio->version); in f2fs_outplace_write_data() [all …]
|
| A D | xattr.c | 487 struct dnode_of_data dn; in write_all_xattrs() local 489 set_new_dnode(&dn, inode, NULL, NULL, new_nid); in write_all_xattrs() 490 xfolio = f2fs_new_node_folio(&dn, XATTR_NODE_OFFSET); in write_all_xattrs()
|
| /fs/ceph/ |
| A D | inode.c | 1495 dn, d_count(dn), realdn, d_count(realdn), in splice_dentry() 1497 dput(dn); in splice_dentry() 1502 d_inode(dn), ceph_vinop(d_inode(dn))); in splice_dentry() 1612 dn, d_inode(dn)); in ceph_fill_trace() 1696 req->r_old_dentry, req->r_old_dentry, dn, dn); in ceph_fill_trace() 1705 req->r_old_dentry, req->r_old_dentry, dn, dn); in ceph_fill_trace() 1752 dn, d_inode(dn), ceph_vinop(d_inode(dn)), in ceph_fill_trace() 1993 if (!dn) { in ceph_readdir_prepopulate() 2012 dn, d_inode(dn)); in ceph_readdir_prepopulate() 2024 dput(dn); in ceph_readdir_prepopulate() [all …]
|
| A D | export.c | 373 struct dentry *dn; in ceph_get_parent() local 380 dn = ERR_PTR(-EINVAL); in ceph_get_parent() 385 dn = ERR_CAST(dir); in ceph_get_parent() 396 dn = ERR_CAST(snapdir); in ceph_get_parent() 405 dn = d_obtain_root(dir); in ceph_get_parent() 407 dn = d_obtain_alias(dir); in ceph_get_parent() 409 dn = __get_parent(child->d_sb, child, 0); in ceph_get_parent() 413 ceph_vinop(inode), (long)PTR_ERR_OR_ZERO(dn)); in ceph_get_parent() 414 return dn; in ceph_get_parent()
|
| A D | dir.c | 1346 struct dentry *dn; in ceph_unlink() local 1366 dn = d_find_alias(dir); in ceph_unlink() 1367 if (!dn) { in ceph_unlink() 1378 dput(dn); in ceph_unlink() 1540 struct dentry *dn = di->dentry; in __ceph_dentry_lease_touch() local 1544 doutc(cl, "%p %p '%pd'\n", di, dn, dn); in __ceph_dentry_lease_touch() 1578 doutc(cl, "%p %p '%pd' (offset 0x%llx)\n", di, dn, dn, di->offset); in __ceph_dentry_dir_lease_touch() 2198 return dn->d_name.hash; in ceph_dentry_hash() 2201 spin_lock(&dn->d_lock); in ceph_dentry_hash() 2203 dn->d_name.name, dn->d_name.len); in ceph_dentry_hash() [all …]
|
| A D | file.c | 741 struct dentry *dn; in ceph_finish_async_create() local 760 dn = d_splice_alias(inode, dentry); in ceph_finish_async_create() 761 WARN_ON_ONCE(dn && dn != dentry); in ceph_finish_async_create() 788 struct dentry *dn; in ceph_atomic_open() local 813 dn = d_find_alias(dir); in ceph_atomic_open() 814 if (!dn) { in ceph_atomic_open() 830 dput(dn); in ceph_atomic_open() 957 if (IS_ERR(dn)) in ceph_atomic_open() 958 err = PTR_ERR(dn); in ceph_atomic_open() 961 dn = NULL; in ceph_atomic_open() [all …]
|
| A D | super.h | 1269 extern int ceph_encode_dentry_release(void **p, struct dentry *dn, 1335 extern unsigned ceph_dentry_hash(struct inode *dir, struct dentry *dn);
|
| A D | caps.c | 3432 struct dentry *dn, *prev = NULL; in invalidate_aliases() local 3445 while ((dn = d_find_alias(inode))) { in invalidate_aliases() 3446 if (dn == prev) { in invalidate_aliases() 3447 dput(dn); in invalidate_aliases() 3450 d_invalidate(dn); in invalidate_aliases() 3453 prev = dn; in invalidate_aliases()
|
| /fs/ubifs/ |
| A D | crypto.c | 38 int ubifs_encrypt(const struct inode *inode, struct ubifs_data_node *dn, in ubifs_encrypt() argument 42 void *p = &dn->data; in ubifs_encrypt() 47 dn->compr_size = cpu_to_le16(in_len); in ubifs_encrypt() 64 int ubifs_decrypt(const struct inode *inode, struct ubifs_data_node *dn, in ubifs_decrypt() argument 69 unsigned int clen = le16_to_cpu(dn->compr_size); in ubifs_decrypt() 78 err = fscrypt_decrypt_block_inplace(inode, virt_to_page(&dn->data), in ubifs_decrypt() 79 dlen, offset_in_page(&dn->data), in ubifs_decrypt()
|
| A D | file.c | 54 err = ubifs_tnc_lookup(c, &key, dn); in read_block() 62 ubifs_assert(c, le64_to_cpu(dn->ch.sqnum) > in read_block() 64 len = le32_to_cpu(dn->size); in read_block() 71 err = ubifs_decrypt(inode, dn, &dlen, block); in read_block() 103 struct ubifs_data_node *dn = NULL; in do_readpage() local 124 if (!dn) { in do_readpage() 144 int dlen = le32_to_cpu(dn->size); in do_readpage() 171 kfree(dn); in do_readpage() 620 struct ubifs_data_node *dn; in populate_page() local 627 len = le32_to_cpu(dn->size); in populate_page() [all …]
|
| A D | journal.c | 1630 out_len = le32_to_cpu(dn->size); in truncate_data_node() 1637 compr_type = le16_to_cpu(dn->compr_type); in truncate_data_node() 1662 dn->compr_size = 0; in truncate_data_node() 1666 dn->compr_type = cpu_to_le16(compr_type); in truncate_data_node() 1667 dn->size = cpu_to_le32(*new_len); in truncate_data_node() 1696 struct ubifs_data_node *dn; in ubifs_jnl_truncate() local 1733 dn = (void *)trun + UBIFS_TRUN_NODE_SZ; in ubifs_jnl_truncate() 1737 err = ubifs_tnc_lookup(c, &key, dn); in ubifs_jnl_truncate() 1743 int dn_len = le32_to_cpu(dn->size); in ubifs_jnl_truncate() 1748 ubifs_dump_node(c, dn, dn_size); in ubifs_jnl_truncate() [all …]
|
| A D | debug.c | 500 const struct ubifs_data_node *dn = node; in ubifs_dump_node() local 502 key_read(c, &dn->key, &key); in ubifs_dump_node() 505 pr_err("\tsize %u\n", le32_to_cpu(dn->size)); in ubifs_dump_node() 507 (int)le16_to_cpu(dn->compr_type)); in ubifs_dump_node() 513 (void *)&dn->data, in ubifs_dump_node() 2060 struct ubifs_data_node *dn = node; in check_leaf() local 2068 inum = key_inum_flash(c, &dn->key); in check_leaf() 2078 blk_offs = key_block_flash(c, &dn->key); in check_leaf() 2080 blk_offs += le32_to_cpu(dn->size); in check_leaf()
|
| A D | ubifs.h | 2111 struct ubifs_data_node *dn, in ubifs_encrypt() argument 2120 struct ubifs_data_node *dn, in ubifs_decrypt() argument 2129 int ubifs_encrypt(const struct inode *inode, struct ubifs_data_node *dn, 2131 int ubifs_decrypt(const struct inode *inode, struct ubifs_data_node *dn,
|
| A D | replay.c | 754 struct ubifs_data_node *dn = snod->node; in replay_bud() local 755 loff_t new_size = le32_to_cpu(dn->size) + in replay_bud()
|
| /fs/hpfs/ |
| A D | dnode.c | 797 struct quad_buffer_head *qbh, struct dnode **dn) in map_nth_dirent() argument 804 if (dn) *dn=dnode; in map_nth_dirent()
|