| /mm/ |
| A D | page_alloc.c | 2641 struct zone *zone; in drain_pages() local 2886 struct zone *zone; in __free_frozen_pages() local 3413 struct zone *zone; in unreserve_highatomic_pageblock() local 3723 struct zone *zone; in get_page_from_freelist() local 4190 struct zone *zone; in should_compact_retry() local 4360 struct zone *zone; in wake_all_kswapds() local 4490 struct zone *zone; in should_reclaim_retry() local 4964 struct zone *zone; in alloc_pages_bulk_noprof() local 5392 struct zone *zone; in nr_free_zone_pages() local 5437 struct zone *zone; in build_zonerefs_node() local [all …]
|
| A D | show_mem.c | 26 static inline void show_node(struct zone *zone) in show_node() argument 38 struct zone *zone; in si_mem_available() local 40 for_each_zone(zone) in si_mem_available() 98 struct zone *zone = &pgdat->node_zones[zone_type]; in si_meminfo_node() local 100 if (is_highmem(zone)) { in si_meminfo_node() 184 struct zone *zone; in show_free_areas() local 297 show_node(zone); in show_free_areas() 321 zone->name, in show_free_areas() 357 show_node(zone); in show_free_areas() 398 struct zone *zone; in __show_mem() local [all …]
|
| A D | compaction.c | 436 struct zone *zone = cc->zone; in update_cached_migrate() local 459 struct zone *zone = cc->zone; in update_pageblock_skip() local 1679 struct zone *zone = cc->zone; in isolate_freepages() local 2435 struct zone *zone; in compaction_zonelist_suitable() local 2759 .zone = zone, in compact_zone_order() 2819 struct zone *zone; in try_to_compact_pages() local 2896 struct zone *zone; in compact_node() local 2914 cc.zone = zone; in compact_node() 3032 struct zone *zone; in kcompactd_node_suitable() local 3062 struct zone *zone; in kcompactd_do_work() local [all …]
|
| A D | vmstat.c | 56 struct zone *zone; in zero_zones_numa_counters() local 192 struct zone *zone; in fold_vm_numa_events() local 278 struct zone *zone; in refresh_zone_stat_thresholds() local 323 struct zone *zone; in set_pgdat_percpu_threshold() local 709 struct zone *zone; in inc_zone_page_state() local 814 struct zone *zone; in refresh_cpu_vm_stats() local 902 struct zone *zone; in cpu_vm_stats_fold() local 1537 struct zone *zone; in walk_zones_in_node() local 1556 struct zone *zone) in frag_show_print() argument 1760 struct zone *zone) in zoneinfo_show_print() argument [all …]
|
| A D | mm_init.c | 69 struct zone *zone; in mminit_verify_zonelist() local 680 struct zone *zone = &pgdat->node_zones[zid]; in __init_page_from_nid() local 979 struct zone *zone = node->node_zones + j; in memmap_init() local 1348 struct zone *zone = pgdat->node_zones + i; in calculate_node_totalpages() local 1589 struct zone *zone = pgdat->node_zones + z; in free_area_init_core_hotplug() local 1606 struct zone *zone = pgdat->node_zones + j; in free_area_init_core() local 2142 struct zone *zone = arg; in deferred_init_memmap_chunk() local 2171 struct zone *zone; in deferred_init_memmap() local 2334 void set_zone_contiguous(struct zone *zone) in set_zone_contiguous() argument 2364 struct zone *zone, *izone = NULL; in pfn_range_intersects_zones() local [all …]
|
| A D | memory_hotplug.c | 495 zone->spanned_pages = zone_end_pfn(zone) - pfn; in shrink_zone_span() 511 zone->spanned_pages = pfn - zone->zone_start_pfn + 1; in shrink_zone_span() 522 struct zone *zone; in update_pgdat_span() local 525 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span() 547 void remove_pfn_range_from_zone(struct zone *zone, in remove_pfn_range_from_zone() argument 795 struct zone *zone) in auto_movable_stats_account_zone() argument 849 struct zone *zone; in auto_movable_can_online_movable() local 905 struct zone *zone = &pgdat->node_zones[zid]; in default_kernel_zone_for_pfn() local 1072 struct zone *zone = page_zone(page); in adjust_present_page_count() local 1091 struct zone *zone, bool mhp_off_inaccessible) in mhp_init_memmap_on_memory() argument [all …]
|
| A D | page_isolation.c | 37 struct zone *zone = page_zone(page); in has_unmovable_pages() local 72 if (zone_idx(zone) == ZONE_MOVABLE) in has_unmovable_pages() 157 struct zone *zone = page_zone(page); in set_migratetype_isolate() local 195 zone->nr_isolate_pageblock++; in set_migratetype_isolate() 214 struct zone *zone; in unset_migratetype_isolate() local 220 zone = page_zone(page); in unset_migratetype_isolate() 270 zone->nr_isolate_pageblock--; in unset_migratetype_isolate() 320 struct zone *zone; in isolate_single_pageblock() local 338 zone->zone_start_pfn); in isolate_single_pageblock() 602 struct zone *zone; in test_pages_isolated() local [all …]
|
| A D | page_reporting.c | 146 page_reporting_cycle(struct page_reporting_dev_info *prdev, struct zone *zone, in page_reporting_cycle() argument 164 spin_lock_irq(&zone->lock); in page_reporting_cycle() 222 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 234 spin_lock_irq(&zone->lock); in page_reporting_cycle() 254 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 261 struct scatterlist *sgl, struct zone *zone) in page_reporting_process_zone() argument 268 watermark = low_wmark_pages(zone) + in page_reporting_process_zone() 299 spin_lock_irq(&zone->lock); in page_reporting_process_zone() 301 spin_unlock_irq(&zone->lock); in page_reporting_process_zone() 314 struct zone *zone; in page_reporting_process() local [all …]
|
| A D | vmscan.c | 423 struct zone *zone; in lruvec_lru_size() local 523 struct zone *zone; in skip_throttle_noprogress() local 2424 struct zone *zone; in prepare_scan_control() local 5946 struct zone *zone; in should_continue_reclaim() local 6260 struct zone *zone; in shrink_zones() local 6386 struct zone *zone; in do_try_to_free_pages() local 6486 struct zone *zone; in allow_direct_reclaim() local 6533 struct zone *zone; in throttle_direct_reclaim() local 6774 struct zone *zone; in pgdat_watermark_boosted() local 6803 struct zone *zone; in pgdat_balanced() local [all …]
|
| A D | mmzone.c | 30 struct zone *next_zone(struct zone *zone) in next_zone() argument 32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() 34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 35 zone++; in next_zone() 39 zone = pgdat->node_zones; in next_zone() 41 zone = NULL; in next_zone() 43 return zone; in next_zone()
|
| A D | internal.h | 722 unsigned long end_pfn, struct zone *zone); 725 unsigned long end_pfn, struct zone *zone) in pageblock_pfn_to_page() argument 733 void set_zone_contiguous(struct zone *zone); 737 static inline void clear_zone_contiguous(struct zone *zone) in clear_zone_contiguous() argument 845 extern void zone_pcp_reset(struct zone *zone); 846 extern void zone_pcp_disable(struct zone *zone); 847 extern void zone_pcp_enable(struct zone *zone); 848 extern void zone_pcp_init(struct zone *zone); 884 struct zone *zone; member 1263 unsigned int reclaim_clean_pages_from_list(struct zone *zone, [all …]
|
| A D | page_owner.c | 414 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showmixedcount_print() argument 426 pfn = zone->zone_start_pfn; in pagetypeinfo_showmixedcount_print() 449 if (page_zone(page) != zone) in pagetypeinfo_showmixedcount_print() 760 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() argument 762 unsigned long pfn = zone->zone_start_pfn; in init_pages_in_zone() 786 if (page_zone(page) != zone) in init_pages_in_zone() 827 pgdat->node_id, zone->name, count); in init_pages_in_zone() 832 struct zone *zone; in init_zones_in_node() local 835 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in init_zones_in_node() 836 if (!populated_zone(zone)) in init_zones_in_node() [all …]
|
| A D | shuffle.c | 36 static struct page * __meminit shuffle_valid_page(struct zone *zone, in shuffle_valid_page() argument 51 if (page_zone(page) != zone) in shuffle_valid_page() 80 void __meminit __shuffle_zone(struct zone *z) in __shuffle_zone() 155 struct zone *z; in __shuffle_free_memory()
|
| A D | debug_page_alloc.c | 35 bool __set_page_guard(struct zone *zone, struct page *page, unsigned int order) in __set_page_guard() argument 47 void __clear_page_guard(struct zone *zone, struct page *page, unsigned int order) in __clear_page_guard() argument
|
| A D | highmem.c | 117 struct zone *zone; in __nr_free_highpages() local 119 for_each_populated_zone(zone) { in __nr_free_highpages() 120 if (is_highmem(zone)) in __nr_free_highpages() 121 pages += zone_page_state(zone, NR_FREE_PAGES); in __nr_free_highpages() 130 struct zone *zone; in __totalhigh_pages() local 132 for_each_populated_zone(zone) { in __totalhigh_pages() 133 if (is_highmem(zone)) in __totalhigh_pages() 134 pages += zone_managed_pages(zone); in __totalhigh_pages()
|
| A D | shuffle.h | 20 extern void __shuffle_zone(struct zone *z); 21 static inline void __meminit shuffle_zone(struct zone *z) in shuffle_zone() 44 static inline void shuffle_zone(struct zone *z) in shuffle_zone()
|
| A D | memremap.c | 226 struct zone *zone; in pagemap_range() local 228 zone = &NODE_DATA(nid)->node_zones[ZONE_DEVICE]; in pagemap_range() 229 move_pfn_range_to_zone(zone, PHYS_PFN(range->start), in pagemap_range()
|
| A D | oom_kill.c | 254 struct zone *zone; in constrained_alloc() local 295 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc() 297 if (!cpuset_zone_allowed(zone, oc->gfp_mask)) in constrained_alloc()
|
| A D | kmemleak.c | 1686 struct zone *zone; in kmemleak_scan() local 1741 for_each_populated_zone(zone) { in kmemleak_scan() 1742 unsigned long start_pfn = zone->zone_start_pfn; in kmemleak_scan() 1743 unsigned long end_pfn = zone_end_pfn(zone); in kmemleak_scan() 1756 if (page_zone(page) != zone) in kmemleak_scan()
|
| A D | memblock.c | 1458 __next_mem_pfn_range_in_zone(u64 *idx, struct zone *zone, in __next_mem_pfn_range_in_zone() argument 1461 int zone_nid = zone_to_nid(zone); in __next_mem_pfn_range_in_zone() 1476 if (zone->zone_start_pfn < epfn && spfn < epfn) { in __next_mem_pfn_range_in_zone() 1478 if (zone_end_pfn(zone) <= spfn) { in __next_mem_pfn_range_in_zone() 1484 *out_spfn = max(zone->zone_start_pfn, spfn); in __next_mem_pfn_range_in_zone() 1486 *out_epfn = min(zone_end_pfn(zone), epfn); in __next_mem_pfn_range_in_zone() 2373 struct zone *z; in reset_node_managed_pages()
|
| A D | swap.c | 755 void lru_add_drain_cpu_zone(struct zone *zone) in lru_add_drain_cpu_zone() argument 759 drain_local_pages(zone); in lru_add_drain_cpu_zone()
|
| A D | zpdesc.h | 165 static inline struct zone *zpdesc_zone(struct zpdesc *zpdesc) in zpdesc_zone()
|
| A D | migrate.c | 545 struct zone *oldzone, *newzone; in __folio_migrate_mapping() 2587 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat() local 2589 if (!managed_zone(zone)) in migrate_balanced_pgdat() 2593 if (!zone_watermark_ok(zone, 0, in migrate_balanced_pgdat() 2594 high_wmark_pages(zone) + in migrate_balanced_pgdat()
|
| A D | page-writeback.c | 258 struct zone *zone = pgdat->node_zones + z; in node_dirtyable_memory() local 260 if (!populated_zone(zone)) in node_dirtyable_memory() 263 nr_pages += zone_page_state(zone, NR_FREE_PAGES); in node_dirtyable_memory() 288 struct zone *z; in highmem_dirtyable_memory()
|
| A D | khugepaged.c | 2638 struct zone *zone; in set_recommended_min_free_kbytes() local 2647 for_each_populated_zone(zone) { in set_recommended_min_free_kbytes() 2652 if (zone_idx(zone) > gfp_zone(GFP_USER)) in set_recommended_min_free_kbytes()
|