Lines Matching refs:lruvec
389 unsigned long lruvec_page_state(struct lruvec *lruvec, enum node_stat_item idx) in lruvec_page_state() argument
396 return node_page_state(lruvec_pgdat(lruvec), idx); in lruvec_page_state()
402 pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in lruvec_page_state()
411 unsigned long lruvec_page_state_local(struct lruvec *lruvec, in lruvec_page_state_local() argument
419 return node_page_state(lruvec_pgdat(lruvec), idx); in lruvec_page_state_local()
425 pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in lruvec_page_state_local()
727 static void mod_memcg_lruvec_state(struct lruvec *lruvec, in mod_memcg_lruvec_state() argument
739 pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in mod_memcg_lruvec_state()
767 void __mod_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, in __mod_lruvec_state() argument
771 __mod_node_page_state(lruvec_pgdat(lruvec), idx, val); in __mod_lruvec_state()
775 mod_memcg_lruvec_state(lruvec, idx, val); in __mod_lruvec_state()
783 struct lruvec *lruvec; in __lruvec_stat_mod_folio() local
794 lruvec = mem_cgroup_lruvec(memcg, pgdat); in __lruvec_stat_mod_folio()
795 __mod_lruvec_state(lruvec, idx, val); in __lruvec_stat_mod_folio()
804 struct lruvec *lruvec; in __mod_lruvec_kmem_state() local
818 lruvec = mem_cgroup_lruvec(memcg, pgdat); in __mod_lruvec_kmem_state()
819 __mod_lruvec_state(lruvec, idx, val); in __mod_lruvec_kmem_state()
1177 void lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio) in lruvec_memcg_debug() argument
1187 VM_BUG_ON_FOLIO(!mem_cgroup_is_root(lruvec_memcg(lruvec)), folio); in lruvec_memcg_debug()
1189 VM_BUG_ON_FOLIO(lruvec_memcg(lruvec) != memcg, folio); in lruvec_memcg_debug()
1204 struct lruvec *folio_lruvec_lock(struct folio *folio) in folio_lruvec_lock()
1206 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock() local
1208 spin_lock(&lruvec->lru_lock); in folio_lruvec_lock()
1209 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock()
1211 return lruvec; in folio_lruvec_lock()
1226 struct lruvec *folio_lruvec_lock_irq(struct folio *folio) in folio_lruvec_lock_irq()
1228 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock_irq() local
1230 spin_lock_irq(&lruvec->lru_lock); in folio_lruvec_lock_irq()
1231 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock_irq()
1233 return lruvec; in folio_lruvec_lock_irq()
1249 struct lruvec *folio_lruvec_lock_irqsave(struct folio *folio, in folio_lruvec_lock_irqsave()
1252 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock_irqsave() local
1254 spin_lock_irqsave(&lruvec->lru_lock, *flags); in folio_lruvec_lock_irqsave()
1255 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock_irqsave()
1257 return lruvec; in folio_lruvec_lock_irqsave()
1270 void mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru, in mem_cgroup_update_lru_size() argument
1280 mz = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in mem_cgroup_update_lru_size()
1289 __func__, lruvec, lru, nr_pages, size)) { in mem_cgroup_update_lru_size()
2520 struct lruvec *lruvec; in account_slab_nmi_safe() local
2523 lruvec = mem_cgroup_lruvec(memcg, pgdat); in account_slab_nmi_safe()
2524 mod_memcg_lruvec_state(lruvec, idx, nr); in account_slab_nmi_safe()
2541 struct lruvec *lruvec; in account_slab_nmi_safe() local
2543 lruvec = mem_cgroup_lruvec(memcg, pgdat); in account_slab_nmi_safe()
2544 mod_memcg_lruvec_state(lruvec, idx, nr); in account_slab_nmi_safe()
3674 lruvec_init(&pn->lruvec); in alloc_mem_cgroup_per_node_info()
4501 static inline unsigned long lruvec_page_state_output(struct lruvec *lruvec, in lruvec_page_state_output() argument
4504 return lruvec_page_state(lruvec, item) * in lruvec_page_state_output()
4524 struct lruvec *lruvec; in memory_numa_stat_show() local
4526 lruvec = mem_cgroup_lruvec(memcg, NODE_DATA(nid)); in memory_numa_stat_show()
4527 size = lruvec_page_state_output(lruvec, in memory_numa_stat_show()