| /mm/ |
| A D | interval_tree.c | 30 struct rb_root_cached *root) in vma_interval_tree_insert_after() argument 57 rb_insert_augmented(&node->shared.rb, &root->rb_root, in vma_interval_tree_insert_after() 76 struct rb_root_cached *root) in INTERVAL_TREE_DEFINE() 82 __anon_vma_interval_tree_insert(node, root); in INTERVAL_TREE_DEFINE() 86 struct rb_root_cached *root) in anon_vma_interval_tree_remove() argument 88 __anon_vma_interval_tree_remove(node, root); in anon_vma_interval_tree_remove() 92 anon_vma_interval_tree_iter_first(struct rb_root_cached *root, in anon_vma_interval_tree_iter_first() argument 95 return __anon_vma_interval_tree_iter_first(root, first, last); in anon_vma_interval_tree_iter_first()
|
| A D | rmap.c | 249 root = new_root; in lock_anon_vma_root() 252 return root; in lock_anon_vma_root() 257 if (root) in unlock_anon_vma_root() 291 root = NULL; in anon_vma_clone() 297 root = lock_anon_vma_root(root, anon_vma); in anon_vma_clone() 372 anon_vma->root = pvma->anon_vma->root; in anon_vma_fork() 408 root = lock_anon_vma_root(root, anon_vma); in unlink_anon_vmas() 798 vma->anon_vma->root != anon_vma->root) in page_address_in_vma() 1400 VM_BUG_ON_FOLIO(folio_anon_vma(folio)->root != vma->anon_vma->root, in __page_check_anon_rmap() 2762 struct anon_vma *root = anon_vma->root; in __put_anon_vma() local [all …]
|
| A D | vmalloc.c | 894 struct rb_root root; member 1169 if (root) { in find_va_links() 1170 link = &root->rb_node; in find_va_links() 1292 rb_erase(&va->rb_node, root); in __unlink_va() 1301 __unlink_va(va, root, false); in unlink_va() 1307 __unlink_va(va, root, true); in unlink_va_augment() 1558 node = root->rb_node; in find_vmap_lowest_match() 1702 unlink_va_augment(va, root); in va_clip() 1862 unlink_va(va, &vn->busy.root); in free_vmap_area() 5193 vn->busy.root = RB_ROOT; in vmap_init_nodes() [all …]
|
| A D | ksm.c | 1704 root); in stable_node_dup() 1767 struct rb_root *root, in __stable_node_chain() argument 1806 struct rb_root *root; in stable_tree_search() local 1824 new = &root->rb_node; in stable_tree_search() 1950 root); in stable_tree_search() 1992 root); in stable_tree_search() 2021 struct rb_root *root; in stable_tree_insert() local 2032 new = &root->rb_node; in stable_tree_insert() 2117 struct rb_root *root; in unstable_tree_search_insert() local 2123 new = &root->rb_node; in unstable_tree_search_insert() [all …]
|
| A D | memcontrol-v1.h | 15 #define for_each_mem_cgroup_tree(iter, root) \ argument 16 for (iter = mem_cgroup_iter(root, NULL, NULL); \ 18 iter = mem_cgroup_iter(root, iter, NULL))
|
| A D | shmem_quota.c | 96 struct rb_root *root = info->dqi_priv; in shmem_free_file_info() local 101 node = rb_first(root); in shmem_free_file_info() 106 rb_erase(&entry->node, root); in shmem_free_file_info() 110 kfree(root); in shmem_free_file_info()
|
| A D | page_counter.c | 424 void page_counter_calculate_protection(struct page_counter *root, in page_counter_calculate_protection() argument 438 if (root == counter) in page_counter_calculate_protection() 445 if (parent == root) { in page_counter_calculate_protection()
|
| A D | memcontrol.c | 1013 if (!root) in mem_cgroup_iter() 1014 root = root_mem_cgroup; in mem_cgroup_iter() 1024 iter = &root->nodeinfo[nid]->iter; in mem_cgroup_iter() 1048 if (css == &root->css || css_tryget(css)) in mem_cgroup_iter() 1061 if (css && css != &root->css) in mem_cgroup_iter() 1082 if (prev && prev != root) in mem_cgroup_iter() 1096 if (!root) in mem_cgroup_iter_break() 1097 root = root_mem_cgroup; in mem_cgroup_iter_break() 1098 if (prev && prev != root) in mem_cgroup_iter_break() 4691 if (!root) in mem_cgroup_calculate_protection() [all …]
|
| A D | sparse.c | 84 unsigned long root = SECTION_NR_TO_ROOT(section_nr); in sparse_index_init() local 94 if (mem_section[root]) in sparse_index_init() 101 mem_section[root] = section; in sparse_index_init()
|
| A D | mempolicy.c | 2819 struct rb_node *n = sp->root.rb_node; in sp_lookup() 2852 struct rb_node **p = &sp->root.rb_node; in sp_insert() 2867 rb_insert_color(&new->nd, &sp->root); in sp_insert() 2877 if (!sp->root.rb_node) in mpol_shared_policy_lookup() 3025 rb_erase(&n->nd, &sp->root); in sp_delete() 3139 sp->root = RB_ROOT; /* empty tree == default mempolicy */ in mpol_shared_policy_init() 3198 if (!sp->root.rb_node) in mpol_free_shared_policy() 3201 next = rb_first(&sp->root); in mpol_free_shared_policy()
|
| A D | vma.c | 2084 if (!test_bit(0, (unsigned long *) &anon_vma->root->rb_root.rb_root.rb_node)) { in vm_lock_anon_vma() 2089 down_write_nest_lock(&anon_vma->root->rwsem, &mm->mmap_lock); in vm_lock_anon_vma() 2100 &anon_vma->root->rb_root.rb_root.rb_node)) in vm_lock_anon_vma() 2219 if (test_bit(0, (unsigned long *) &anon_vma->root->rb_root.rb_root.rb_node)) { in vm_unlock_anon_vma() 2233 &anon_vma->root->rb_root.rb_root.rb_node)) in vm_unlock_anon_vma()
|
| A D | memblock.c | 2802 struct dentry *root = debugfs_create_dir("memblock", NULL); in memblock_init_debugfs() local 2804 debugfs_create_file("memory", 0444, root, in memblock_init_debugfs() 2806 debugfs_create_file("reserved", 0444, root, in memblock_init_debugfs() 2809 debugfs_create_file("physmem", 0444, root, &physmem, in memblock_init_debugfs()
|
| A D | shmem.c | 4859 struct shmem_sb_info *sbinfo = SHMEM_SB(fc->root->d_sb); in shmem_reconfigure() 4904 !sb_any_quota_loaded(fc->root->d_sb)) { in shmem_reconfigure() 4952 static int shmem_show_options(struct seq_file *seq, struct dentry *root) in shmem_show_options() argument 4954 struct shmem_sb_info *sbinfo = SHMEM_SB(root->d_sb); in shmem_show_options() 5003 if (sb_has_quota_active(root->d_sb, USRQUOTA)) in shmem_show_options() 5005 if (sb_has_quota_active(root->d_sb, GRPQUOTA)) in shmem_show_options()
|
| A D | memory.c | 4023 static inline void unmap_mapping_range_tree(struct rb_root_cached *root, in unmap_mapping_range_tree() argument 4031 vma_interval_tree_foreach(vma, root, first_index, last_index) { in unmap_mapping_range_tree()
|
| A D | Kconfig | 751 root has set /sys/kernel/mm/ksm/run to 1 (if CONFIG_SYSFS is set).
|