| /mm/ |
| A D | page_counter.c | 27 if (!c->parent) in propagate_protected_usage() 81 for (c = counter; c; c = c->parent) { in page_counter_charge() 126 for (c = counter; c; c = c->parent) { in page_counter_try_charge() 183 for (c = counter; c; c = c->parent) in page_counter_uncharge() 242 for (c = counter; c; c = c->parent) in page_counter_set_min() 259 for (c = counter; c; c = c->parent) in page_counter_set_low() 429 struct page_counter *parent = counter->parent; in page_counter_calculate_protection() local 445 if (parent == root) { in page_counter_calculate_protection() 451 parent_usage = page_counter_read(parent); in page_counter_calculate_protection() 455 READ_ONCE(parent->emin), in page_counter_calculate_protection() [all …]
|
| A D | interval_tree.c | 33 struct vm_area_struct *parent; in vma_interval_tree_insert_after() local 39 parent = prev; in vma_interval_tree_insert_after() 42 parent = rb_entry(prev->shared.rb.rb_right, in vma_interval_tree_insert_after() 44 if (parent->shared.rb_subtree_last < last) in vma_interval_tree_insert_after() 45 parent->shared.rb_subtree_last = last; in vma_interval_tree_insert_after() 46 while (parent->shared.rb.rb_left) { in vma_interval_tree_insert_after() 47 parent = rb_entry(parent->shared.rb.rb_left, in vma_interval_tree_insert_after() 49 if (parent->shared.rb_subtree_last < last) in vma_interval_tree_insert_after() 50 parent->shared.rb_subtree_last = last; in vma_interval_tree_insert_after() 52 link = &parent->shared.rb.rb_left; in vma_interval_tree_insert_after() [all …]
|
| A D | memcontrol.c | 3302 struct mem_cgroup *parent; in memcg_offline_kmem() local 3311 if (!parent) in memcg_offline_kmem() 3346 if (!memcg->css.parent) in mem_cgroup_wb_domain() 3391 memcg = parent; in mem_cgroup_wb_stats() 3738 if (parent) in mem_cgroup_alloc() 3803 if (parent) { in mem_cgroup_css_alloc() 4029 if (parent) in flush_nmi_stats() 4038 if (parent) in flush_nmi_stats() 4081 .ppending = parent ? parent->vmstats->state_pending : NULL, in mem_cgroup_css_rstat_flush() 4092 .ppending = parent ? parent->vmstats->events_pending : NULL, in mem_cgroup_css_rstat_flush() [all …]
|
| A D | list_lru.c | 467 void memcg_reparent_list_lrus(struct mem_cgroup *memcg, struct mem_cgroup *parent) in memcg_reparent_list_lrus() argument 493 memcg_reparent_list_lru_one(lru, i, &mlru->node[i], parent); in memcg_reparent_list_lrus() 518 struct mem_cgroup *pos, *parent; in memcg_list_lru_alloc() local 536 parent = parent_mem_cgroup(pos); in memcg_list_lru_alloc() 537 while (!memcg_list_lru_allocated(parent, lru)) { in memcg_list_lru_alloc() 538 pos = parent; in memcg_list_lru_alloc() 539 parent = parent_mem_cgroup(pos); in memcg_list_lru_alloc()
|
| A D | shmem_quota.c | 169 struct rb_node *parent = NULL, *new_node = NULL; in shmem_acquire_dquot() local 181 parent = *n; in shmem_acquire_dquot() 182 entry = rb_entry(parent, struct quota_id, node); in shmem_acquire_dquot() 209 rb_link_node(new_node, parent, n); in shmem_acquire_dquot()
|
| A D | vmalloc.c | 1164 struct rb_node **parent) in find_va_links() argument 1172 *parent = NULL; in find_va_links() 1204 *parent = &tmp_va->rb_node; in find_va_links() 1213 if (unlikely(!parent)) in get_va_next_sibling() 1235 if (likely(parent)) { in __link_va() 1237 if (&parent->rb_right != link) in __link_va() 1242 rb_link_node(&va->rb_node, parent, link); in __link_va() 1384 struct rb_node *parent; in insert_vmap_area() local 1388 link_va(va, root, parent, link, head); in insert_vmap_area() 1397 struct rb_node *parent; in insert_vmap_area_augment() local [all …]
|
| A D | backing-dev.c | 787 if (!memcg_css->parent) in wb_get_lookup() 1040 struct rb_node *parent = NULL; in bdi_lookup_rb_node() local 1046 parent = *p; in bdi_lookup_rb_node() 1047 bdi = rb_entry(parent, struct backing_dev_info, rb_node); in bdi_lookup_rb_node() 1058 *parentp = parent; in bdi_lookup_rb_node() 1088 struct rb_node *parent, **p; in bdi_register_va() local 1108 p = bdi_lookup_rb_node(bdi->id, &parent); in bdi_register_va() 1109 rb_link_node(&bdi->rb_node, parent, p); in bdi_register_va()
|
| A D | ksm.c | 1808 struct rb_node *parent; in stable_tree_search() local 1825 parent = NULL; in stable_tree_search() 1850 parent = *new; in stable_tree_search() 1852 new = &parent->rb_left; in stable_tree_search() 1854 new = &parent->rb_right; in stable_tree_search() 2023 struct rb_node *parent; in stable_tree_insert() local 2031 parent = NULL; in stable_tree_insert() 2057 parent = *new; in stable_tree_insert() 2059 new = &parent->rb_left; in stable_tree_insert() 2146 parent = *new; in unstable_tree_search_insert() [all …]
|
| A D | shrinker.c | 289 struct mem_cgroup *parent; in reparent_shrinker_deferred() local 293 parent = parent_mem_cgroup(memcg); in reparent_shrinker_deferred() 294 if (!parent) in reparent_shrinker_deferred() 295 parent = root_mem_cgroup; in reparent_shrinker_deferred() 301 parent_info = shrinker_info_protected(parent, nid); in reparent_shrinker_deferred()
|
| A D | hugetlb_cgroup.c | 83 return hugetlb_cgroup_from_css(h_cg->css.parent); in parent_hugetlb_cgroup() 200 struct hugetlb_cgroup *parent = parent_hugetlb_cgroup(h_cg); in hugetlb_cgroup_move_parent() local 212 if (!parent) { in hugetlb_cgroup_move_parent() 213 parent = root_h_cgroup; in hugetlb_cgroup_move_parent() 215 page_counter_charge(&parent->hugepage[idx], nr_pages); in hugetlb_cgroup_move_parent() 221 set_hugetlb_cgroup(folio, parent); in hugetlb_cgroup_move_parent()
|
| A D | kmemleak.c | 704 struct kmemleak_object *parent; in __link_object() local 732 parent = rb_entry(rb_parent, struct kmemleak_object, rb_node); in __link_object() 733 untagged_objp = (unsigned long)kasan_reset_tag((void *)parent->pointer); in __link_object() 735 link = &parent->rb_node.rb_left; in __link_object() 736 else if (untagged_objp + parent->size <= untagged_ptr) in __link_object() 737 link = &parent->rb_node.rb_right; in __link_object() 745 dump_object_info(parent); in __link_object()
|
| A D | nommu.c | 473 struct rb_node **p, *parent; in add_nommu_region() local 477 parent = NULL; in add_nommu_region() 480 parent = *p; in add_nommu_region() 481 pregion = rb_entry(parent, struct vm_region, vm_rb); in add_nommu_region() 492 rb_link_node(®ion->vm_rb, parent, p); in add_nommu_region()
|
| A D | memcontrol-v1.c | 114 struct rb_node *parent = NULL; in __mem_cgroup_insert_exceeded() local 125 parent = *p; in __mem_cgroup_insert_exceeded() 126 mz_node = rb_entry(parent, struct mem_cgroup_per_node, in __mem_cgroup_insert_exceeded() 139 rb_link_node(&mz->tree_node, parent, p); in __mem_cgroup_insert_exceeded()
|
| A D | swapfile.c | 2445 struct rb_node **link = &sis->swap_extent_root.rb_node, *parent = NULL; in add_swap_extent() local 2454 parent = *link; in add_swap_extent() 2455 link = &parent->rb_right; in add_swap_extent() 2458 if (parent) { in add_swap_extent() 2459 se = rb_entry(parent, struct swap_extent, rb_node); in add_swap_extent() 2476 rb_link_node(&new_se->rb_node, parent, link); in add_swap_extent()
|
| A D | rmap.c | 99 anon_vma->parent = anon_vma; in anon_vma_alloc() 373 anon_vma->parent = pvma->anon_vma; in anon_vma_fork() 384 anon_vma->parent->num_children++; in anon_vma_fork() 416 anon_vma->parent->num_children--; in unlink_anon_vmas()
|
| A D | mempolicy.c | 2853 struct rb_node *parent = NULL; in sp_insert() local 2857 parent = *p; in sp_insert() 2858 nd = rb_entry(parent, struct sp_node, nd); in sp_insert() 2866 rb_link_node(&new->nd, parent, p); in sp_insert()
|
| A D | huge_memory.c | 708 static struct thpsize *thpsize_create(int order, struct kobject *parent) in thpsize_create() argument 720 ret = kobject_init_and_add(&thpsize->kobj, &thpsize_ktype, parent, in thpsize_create()
|
| A D | hugetlb.c | 4447 static int hugetlb_sysfs_add_hstate(struct hstate *h, struct kobject *parent, in hugetlb_sysfs_add_hstate() argument 4454 hstate_kobjs[hi] = kobject_create_and_add(h->name, parent); in hugetlb_sysfs_add_hstate()
|
| A D | shmem.c | 4506 struct inode *parent) in shmem_encode_fh() argument
|