| /linux/mm/ |
| A D | swap.c | 84 struct lruvec *lruvec; in __page_cache_release() local 194 void (*move_fn)(struct page *page, struct lruvec *lruvec)) in pagevec_lru_move_fn() argument 197 struct lruvec *lruvec = NULL; in pagevec_lru_move_fn() local 213 if (lruvec) in pagevec_lru_move_fn() 303 } while ((lruvec = parent_lruvec(lruvec))); in lru_note_cost() 369 struct lruvec *lruvec; in folio_activate() local 906 struct lruvec *lruvec = NULL; in release_pages() local 961 struct lruvec *prev_lruvec = lruvec; in release_pages() 963 lruvec = folio_lruvec_relock_irqsave(folio, lruvec, in release_pages() 976 if (lruvec) in release_pages() [all …]
|
| A D | workingset.c | 229 void workingset_age_nonresident(struct lruvec *lruvec, unsigned long nr_pages) in workingset_age_nonresident() argument 244 } while ((lruvec = parent_lruvec(lruvec))); in workingset_age_nonresident() 259 struct lruvec *lruvec; in workingset_eviction() local 267 lruvec = mem_cgroup_lruvec(target_memcg, pgdat); in workingset_eviction() 269 memcgid = mem_cgroup_id(lruvec_memcg(lruvec)); in workingset_eviction() 288 struct lruvec *eviction_lruvec; in workingset_refault() 294 struct lruvec *lruvec; in workingset_refault() local 353 lruvec = mem_cgroup_lruvec(memcg, pgdat); in workingset_refault() 382 workingset_age_nonresident(lruvec, nr); in workingset_refault() 494 struct lruvec *lruvec; in count_shadow_nodes() local [all …]
|
| A D | vmscan.c | 2117 struct lruvec *lruvec, struct list_head *dst, in isolate_lru_pages() argument 2246 struct lruvec *lruvec; in isolate_lru_page() local 2498 struct lruvec *lruvec, in shrink_active_list() argument 2652 struct lruvec *lruvec, struct scan_control *sc) in shrink_list() argument 3133 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, pgdat); in shrink_node_memcgs() local 3614 struct lruvec *lruvec; in do_try_to_free_pages() local 3839 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, pgdat); in mem_cgroup_shrink_node() local 3916 struct lruvec *lruvec; in age_active_anon() local 3997 struct lruvec *lruvec = mem_cgroup_lruvec(NULL, pgdat); in clear_pgdat_congested() local 4853 struct lruvec *lruvec = NULL; in check_move_unevictable_pages() local [all …]
|
| A D | mmzone.c | 75 void lruvec_init(struct lruvec *lruvec) in lruvec_init() argument 79 memset(lruvec, 0, sizeof(struct lruvec)); in lruvec_init() 80 spin_lock_init(&lruvec->lru_lock); in lruvec_init() 83 INIT_LIST_HEAD(&lruvec->lists[lru]); in lruvec_init()
|
| A D | memcontrol.c | 699 pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in __mod_memcg_lruvec_state() 738 struct lruvec *lruvec; in __mod_lruvec_page_state() local 759 struct lruvec *lruvec; in __mod_lruvec_kmem_state() local 1190 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock() local 1195 return lruvec; in folio_lruvec_lock() 1213 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock_irq() local 1218 return lruvec; in folio_lruvec_lock_irq() 1238 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock_irqsave() local 1243 return lruvec; in folio_lruvec_lock_irqsave() 2810 struct lruvec *lruvec; in mod_objcg_mlstate() local [all …]
|
| A D | mlock.c | 266 struct lruvec *lruvec = NULL; in __munlock_pagevec() local 282 lruvec = folio_lruvec_relock_irq(folio, lruvec); in __munlock_pagevec() 283 del_page_from_lru_list(page, lruvec); in __munlock_pagevec() 300 if (lruvec) { in __munlock_pagevec() 302 unlock_page_lruvec_irq(lruvec); in __munlock_pagevec()
|
| A D | compaction.c | 805 struct lruvec *lruvec; in isolate_migratepages_block() local 807 struct lruvec *locked = NULL; in isolate_migratepages_block() 1031 lruvec = folio_lruvec(page_folio(page)); in isolate_migratepages_block() 1034 if (lruvec != locked) { in isolate_migratepages_block() 1038 compact_lock_irqsave(&lruvec->lru_lock, &flags, cc); in isolate_migratepages_block() 1039 locked = lruvec; in isolate_migratepages_block() 1041 lruvec_memcg_debug(lruvec, page_folio(page)); in isolate_migratepages_block() 1067 del_page_from_lru_list(page, lruvec); in isolate_migratepages_block()
|
| A D | huge_memory.c | 2320 struct lruvec *lruvec, struct list_head *list) in lru_add_page_tail() argument 2325 lockdep_assert_held(&lruvec->lru_lock); in lru_add_page_tail() 2341 struct lruvec *lruvec, struct list_head *list) in __split_huge_page_tail() argument 2402 lru_add_page_tail(head, page_tail, lruvec, list); in __split_huge_page_tail() 2410 struct lruvec *lruvec; in __split_huge_page() local 2428 lruvec = folio_lruvec_lock(folio); in __split_huge_page() 2433 __split_huge_page_tail(head, i, lruvec, list); in __split_huge_page() 2451 unlock_page_lruvec(lruvec); in __split_huge_page()
|
| A D | migrate.c | 467 struct lruvec *old_lruvec, *new_lruvec; in folio_migrate_mapping()
|
| /linux/include/linux/ |
| A D | mm_inline.h | 33 static __always_inline void update_lru_size(struct lruvec *lruvec, in update_lru_size() argument 37 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in update_lru_size() 94 void lruvec_add_folio(struct lruvec *lruvec, struct folio *folio) in lruvec_add_folio() argument 100 list_add(&folio->lru, &lruvec->lists[lru]); in lruvec_add_folio() 104 struct lruvec *lruvec) in add_page_to_lru_list() argument 106 lruvec_add_folio(lruvec, page_folio(page)); in add_page_to_lru_list() 110 void lruvec_add_folio_tail(struct lruvec *lruvec, struct folio *folio) in lruvec_add_folio_tail() argument 120 struct lruvec *lruvec) in add_page_to_lru_list_tail() argument 126 void lruvec_del_folio(struct lruvec *lruvec, struct folio *folio) in lruvec_del_folio() argument 134 struct lruvec *lruvec) in del_page_from_lru_list() argument [all …]
|
| A D | memcontrol.h | 138 struct lruvec lruvec; member 758 struct lruvec *lruvec; in mem_cgroup_lruvec() local 769 lruvec = &mz->lruvec; in mem_cgroup_lruvec() 778 return lruvec; in mem_cgroup_lruvec() 876 mz = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in lruvec_memcg() 935 mz = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in mem_cgroup_get_zone_lru_size() 1007 pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in lruvec_page_state() 1021 pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in lruvec_page_state_local() 1536 static inline struct lruvec *parent_lruvec(struct lruvec *lruvec) in parent_lruvec() argument 1549 static inline void unlock_page_lruvec(struct lruvec *lruvec) in unlock_page_lruvec() argument [all …]
|
| A D | vmstat.h | 546 void __mod_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, 549 static inline void mod_lruvec_state(struct lruvec *lruvec, in mod_lruvec_state() argument 555 __mod_lruvec_state(lruvec, idx, val); in mod_lruvec_state() 574 static inline void __mod_lruvec_state(struct lruvec *lruvec, in __mod_lruvec_state() argument 577 __mod_node_page_state(lruvec_pgdat(lruvec), idx, val); in __mod_lruvec_state() 580 static inline void mod_lruvec_state(struct lruvec *lruvec, in mod_lruvec_state() argument 583 mod_node_page_state(lruvec_pgdat(lruvec), idx, val); in mod_lruvec_state()
|
| A D | mmzone.h | 306 struct lruvec { struct 910 struct lruvec __lruvec; 968 extern void lruvec_init(struct lruvec *lruvec); 970 static inline struct pglist_data *lruvec_pgdat(struct lruvec *lruvec) in lruvec_pgdat() argument 973 return lruvec->pgdat; in lruvec_pgdat() 975 return container_of(lruvec, struct pglist_data, __lruvec); in lruvec_pgdat()
|
| A D | swap.h | 330 void workingset_age_nonresident(struct lruvec *lruvec, unsigned long nr_pages); 350 extern void lru_note_cost(struct lruvec *lruvec, bool file,
|
| /linux/Documentation/trace/ |
| A D | events-kmem.rst | 72 contention on the lruvec->lru_lock.
|
| /linux/Documentation/admin-guide/cgroup-v1/ |
| A D | memory.rst | 294 lruvec->lru_lock. 297 lruvec->lru_lock; PG_lru bit of page->flags is cleared before 298 isolating a page from its LRU under lruvec->lru_lock.
|