| /linux/mm/ |
| A D | mm_init.c | 711 pg_data_t *pgdat; in init_reserved_page() local 1260 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) { in reset_memoryless_node_totalpages() 1422 pgdat->node_id, in init_currently_empty_zone() 1528 pgdat->nr_zones = 0; in free_area_init_core_hotplug() 1628 __func__, pgdat->node_id, (unsigned long)pgdat, in alloc_node_mem_map() 1678 WARN_ON(pgdat->nr_zones || pgdat->kswapd_highest_zoneidx); in free_area_init_node() 1682 pgdat->node_id = nid; in free_area_init_node() 1836 pg_data_t *pgdat; in free_area_init() local 2113 BUG_ON(pgdat->first_deferred_pfn < pgdat->node_start_pfn); in deferred_init_memmap() 2114 BUG_ON(pgdat->first_deferred_pfn > pgdat_end_pfn(pgdat)); in deferred_init_memmap() [all …]
|
| A D | show_mem.c | 95 pg_data_t *pgdat = NODE_DATA(nid); in si_meminfo_node() local 191 pg_data_t *pgdat; in show_free_areas() local 232 for_each_online_pgdat(pgdat) { in show_free_areas() 264 pgdat->node_id, in show_free_areas() 273 K(node_page_state(pgdat, NR_FILE_DIRTY)), in show_free_areas() 274 K(node_page_state(pgdat, NR_WRITEBACK)), in show_free_areas() 275 K(node_page_state(pgdat, NR_SHMEM)), in show_free_areas() 277 K(node_page_state(pgdat, NR_SHMEM_THPS)), in show_free_areas() 279 K(node_page_state(pgdat, NR_ANON_THPS)), in show_free_areas() 284 node_page_state(pgdat, NR_KERNEL_SCS_KB), in show_free_areas() [all …]
|
| A D | vmstat.c | 274 struct pglist_data *pgdat; in refresh_zone_stat_thresholds() local 749 struct pglist_data *pgdat; in inc_node_page_state() local 751 pgdat = page_pgdat(page); in inc_node_page_state() 1474 pg_data_t *pgdat; in frag_start() local 1478 pgdat && node; in frag_start() 1479 pgdat = next_online_pgdat(pgdat)) in frag_start() 1482 return pgdat; in frag_start() 1515 print(m, pgdat, zone); in walk_zones_in_node() 1555 pgdat->node_id, in pagetypeinfo_showfree_print() 2236 pgdat->node_id, in unusable_show_print() [all …]
|
| A D | vmscan.c | 2634 lruvec->pgdat = pgdat; in get_lruvec() 3317 if (pfn < pgdat->node_start_pfn || pfn >= pgdat_end_pfn(pgdat)) in get_pte_pfn() 3342 if (pfn < pgdat->node_start_pfn || pfn >= pgdat_end_pfn(pgdat)) in get_pmd_pfn() 4205 WRITE_ONCE(pgdat->memcg_lru.seq, pgdat->memcg_lru.seq + 1); in lru_gen_rotate_memcg() 4268 WRITE_ONCE(pgdat->memcg_lru.seq, pgdat->memcg_lru.seq + 1); in lru_gen_release_memcg() 5861 .pgdat = pgdat, in shrink_node_memcgs() 6449 if (!pgdat) in throttle_direct_reclaim() 7253 pg_data_t *pgdat; in wakeup_kswapd() local 7343 pgdat->kswapd = kthread_run(kswapd, pgdat, "kswapd%d", nid); in kswapd_run() 7489 if (node_pagecache_reclaimable(pgdat) > pgdat->min_unmapped_pages || in __node_reclaim() [all …]
|
| A D | mmzone.c | 18 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) in next_online_pgdat() argument 20 int nid = next_online_node(pgdat->node_id); in next_online_pgdat() 32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() local 34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 37 pgdat = next_online_pgdat(pgdat); in next_zone() 38 if (pgdat) in next_zone() 39 zone = pgdat->node_zones; in next_zone()
|
| A D | compaction.c | 2189 pgdat_kswapd_lock(pgdat); in kswapd_is_running() 2190 running = pgdat->kswapd && task_is_running(pgdat->kswapd); in kswapd_is_running() 2191 pgdat_kswapd_unlock(pgdat); in kswapd_is_running() 2299 pg_data_t *pgdat; in __compact_finished() local 3121 pgdat->kcompactd_highest_zoneidx = pgdat->nr_zones - 1; in kcompactd_do_work() 3169 pgdat->kcompactd_highest_zoneidx = pgdat->nr_zones - 1; in kcompactd() 3186 kcompactd_do_work(pgdat); in kcompactd() 3233 if (pgdat->kcompactd) in kcompactd_run() 3236 pgdat->kcompactd = kthread_run(kcompactd, pgdat, "kcompactd%d", nid); in kcompactd_run() 3239 pgdat->kcompactd = NULL; in kcompactd_run() [all …]
|
| A D | workingset.c | 225 *pgdat = NODE_DATA(nid); in unpack_shadow() 244 struct pglist_data *pgdat = folio_pgdat(folio); in lru_gen_eviction() local 248 lruvec = mem_cgroup_lruvec(memcg, pgdat); in lru_gen_eviction() 269 struct pglist_data *pgdat; in lru_gen_test_recent() local 274 *lruvec = mem_cgroup_lruvec(memcg, pgdat); in lru_gen_test_recent() 384 struct pglist_data *pgdat = folio_pgdat(folio); in workingset_eviction() local 403 return pack_shadow(memcgid, pgdat, eviction, in workingset_eviction() 428 struct pglist_data *pgdat; in workingset_test_recent() local 541 struct pglist_data *pgdat; in workingset_refault() local 564 pgdat = folio_pgdat(folio); in workingset_refault() [all …]
|
| A D | numa.c | 39 pg_data_t *pgdat; in alloc_offline_node_data() local 41 pgdat = memblock_alloc(sizeof(*pgdat), SMP_CACHE_BYTES); in alloc_offline_node_data() 42 if (!pgdat) in alloc_offline_node_data() 44 sizeof(*pgdat), nid); in alloc_offline_node_data() 46 node_data[nid] = pgdat; in alloc_offline_node_data()
|
| A D | shuffle.h | 11 extern void __shuffle_free_memory(pg_data_t *pgdat); 13 static inline void __meminit shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument 17 __shuffle_free_memory(pgdat); in shuffle_free_memory() 40 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
|
| A D | memory-tiers.c | 262 pg_data_t *pgdat; in __node_get_memory_tier() local 264 pgdat = NODE_DATA(node); in __node_get_memory_tier() 265 if (!pgdat) in __node_get_memory_tier() 280 pg_data_t *pgdat; in node_is_toptier() local 283 pgdat = NODE_DATA(node); in node_is_toptier() 284 if (!pgdat) in node_is_toptier() 288 memtier = rcu_dereference(pgdat->memtier); in node_is_toptier() 537 pg_data_t *pgdat = NODE_DATA(node); in set_node_memory_tier() local 573 pg_data_t *pgdat; in clear_node_memory_tier() local 576 pgdat = NODE_DATA(node); in clear_node_memory_tier() [all …]
|
| A D | memory_hotplug.c | 504 for (zone = pgdat->node_zones; in update_pgdat_span() 557 update_pgdat_span(pgdat); in remove_pfn_range_from_zone() 715 if (!pgdat->node_spanned_pages || start_pfn < pgdat->node_start_pfn) in resize_pgdat_range() 718 pgdat->node_spanned_pages = max(start_pfn + nr_pages, old_end_pfn) - pgdat->node_start_pfn; in resize_pgdat_range() 749 int nid = pgdat->node_id; in move_pfn_range_to_zone() 1238 struct pglist_data *pgdat; in hotadd_init_pgdat() local 1246 pgdat = NODE_DATA(nid); in hotadd_init_pgdat() 1255 build_all_zonelists(pgdat); in hotadd_init_pgdat() 1257 return pgdat; in hotadd_init_pgdat() 1273 pg_data_t *pgdat; in __try_online_node() local [all …]
|
| A D | bootmem_info.c | 102 void __init register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument 105 int node = pgdat->node_id; in register_page_bootmem_info_node() 109 page = virt_to_page(pgdat); in register_page_bootmem_info_node() 114 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node() 115 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node()
|
| A D | shuffle.c | 153 void __meminit __shuffle_free_memory(pg_data_t *pgdat) in __shuffle_free_memory() argument 157 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
|
| A D | sparse.c | 318 static inline phys_addr_t pgdat_to_phys(struct pglist_data *pgdat) in pgdat_to_phys() argument 321 VM_BUG_ON(pgdat != &contig_page_data); in pgdat_to_phys() 324 return __pa(pgdat); in pgdat_to_phys() 329 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument 345 goal = pgdat_to_phys(pgdat) & (PAGE_SECTION_MASK << PAGE_SHIFT); in sparse_early_usemaps_alloc_pgdat_section() 363 struct pglist_data *pgdat = NODE_DATA(nid); in check_usemap_section_nr() local 373 pgdat_snr = pfn_to_section_nr(pgdat_to_phys(pgdat) >> PAGE_SHIFT); in check_usemap_section_nr() 401 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument 404 return memblock_alloc_node(size, SMP_CACHE_BYTES, pgdat->node_id); in sparse_early_usemaps_alloc_pgdat_section()
|
| /linux/include/linux/ |
| A D | memory_hotplug.h | 182 static inline void pgdat_kswapd_lock(pg_data_t *pgdat) in pgdat_kswapd_lock() argument 184 mutex_lock(&pgdat->kswapd_lock); in pgdat_kswapd_lock() 187 static inline void pgdat_kswapd_unlock(pg_data_t *pgdat) in pgdat_kswapd_unlock() argument 189 mutex_unlock(&pgdat->kswapd_lock); in pgdat_kswapd_unlock() 192 static inline void pgdat_kswapd_lock_init(pg_data_t *pgdat) in pgdat_kswapd_lock_init() argument 194 mutex_init(&pgdat->kswapd_lock); in pgdat_kswapd_lock_init() 239 static inline void pgdat_kswapd_lock(pg_data_t *pgdat) {} in pgdat_kswapd_lock() argument 258 spin_lock_irqsave(&pgdat->node_size_lock, *flags); in pgdat_resize_lock() 263 spin_unlock_irqrestore(&pgdat->node_size_lock, *flags); in pgdat_resize_unlock() 266 void pgdat_resize_init(struct pglist_data *pgdat) in pgdat_resize_init() argument [all …]
|
| A D | compaction.h | 91 extern void reset_isolation_suitable(pg_data_t *pgdat); 103 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx); 106 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument 123 static inline void wakeup_kcompactd(pg_data_t *pgdat, in wakeup_kcompactd() argument
|
| A D | vmstat.h | 176 static inline void node_page_state_add(long x, struct pglist_data *pgdat, in node_page_state_add() argument 179 atomic_long_add(x, &pgdat->vm_stat[item]); in node_page_state_add() 265 extern unsigned long node_page_state(struct pglist_data *pgdat, 267 extern unsigned long node_page_state_pages(struct pglist_data *pgdat, 315 void set_pgdat_percpu_threshold(pg_data_t *pgdat, 329 static inline void __mod_node_page_state(struct pglist_data *pgdat, in __mod_node_page_state() argument 343 node_page_state_add(delta, pgdat, item); in __mod_node_page_state() 352 static inline void __inc_node_state(struct pglist_data *pgdat, enum node_stat_item item) in __inc_node_state() argument 354 atomic_long_inc(&pgdat->vm_stat[item]); in __inc_node_state() 366 atomic_long_dec(&pgdat->vm_stat[item]); in __dec_node_state() [all …]
|
| A D | memcontrol.h | 59 pg_data_t *pgdat; member 742 struct pglist_data *pgdat) in mem_cgroup_lruvec() argument 748 lruvec = &pgdat->__lruvec; in mem_cgroup_lruvec() 755 mz = memcg->nodeinfo[pgdat->node_id]; in mem_cgroup_lruvec() 763 if (unlikely(lruvec->pgdat != pgdat)) in mem_cgroup_lruvec() 764 lruvec->pgdat = pgdat; in mem_cgroup_lruvec() 1239 return &pgdat->__lruvec; in mem_cgroup_lruvec() 1245 return &pgdat->__lruvec; in folio_lruvec() 1308 return &pgdat->__lruvec; in folio_lruvec_lock() 1316 return &pgdat->__lruvec; in folio_lruvec_lock_irq() [all …]
|
| A D | node.h | 130 struct pglist_data *pgdat = NODE_DATA(nid); in register_one_node() local 131 unsigned long start_pfn = pgdat->node_start_pfn; in register_one_node() 132 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node()
|
| A D | memory-tiers.h | 57 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets); 65 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() argument 109 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() argument
|
| A D | mmzone.h | 556 void lru_gen_init_pgdat(struct pglist_data *pgdat); 569 static inline void lru_gen_init_pgdat(struct pglist_data *pgdat) in lru_gen_init_pgdat() argument 634 struct pglist_data *pgdat; member 1446 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) in pgdat_end_pfn() argument 1448 return pgdat->node_start_pfn + pgdat->node_spanned_pages; in pgdat_end_pfn() 1453 void build_all_zonelists(pg_data_t *pgdat); 1481 return lruvec->pgdat; in lruvec_pgdat() 1602 #define for_each_online_pgdat(pgdat) \ argument 1603 for (pgdat = first_online_pgdat(); \ 1604 pgdat; \ [all …]
|
| A D | bootmem_info.h | 21 void __init register_page_bootmem_info_node(struct pglist_data *pgdat); 48 static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument
|
| /linux/tools/testing/vma/linux/ |
| A D | mmzone.h | 9 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat); 11 #define for_each_online_pgdat(pgdat) \ argument 12 for (pgdat = first_online_pgdat(); \ 13 pgdat; \ 14 pgdat = next_online_pgdat(pgdat))
|
| /linux/tools/testing/memblock/linux/ |
| A D | mmzone.h | 9 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat); 11 #define for_each_online_pgdat(pgdat) \ argument 12 for (pgdat = first_online_pgdat(); \ 13 pgdat; \ 14 pgdat = next_online_pgdat(pgdat))
|
| /linux/drivers/base/ |
| A D | node.c | 375 struct pglist_data *pgdat = NODE_DATA(nid); in node_read_meminfo() local 403 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON) + in node_read_meminfo() 404 node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo() 406 node_page_state(pgdat, NR_INACTIVE_FILE)), in node_read_meminfo() 407 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON)), in node_read_meminfo() 409 nid, K(node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo() 457 nid, K(node_page_state(pgdat, NR_WRITEBACK)), in node_read_meminfo() 466 nid, K(node_page_state(pgdat, NR_PAGETABLE)), in node_read_meminfo() 478 nid, K(node_page_state(pgdat, NR_ANON_THPS)), in node_read_meminfo() 481 nid, K(node_page_state(pgdat, NR_FILE_THPS)), in node_read_meminfo() [all …]
|