Searched refs:node_zones (Results 1 – 24 of 24) sorted by relevance
34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone()39 zone = pgdat->node_zones; in next_zone()
157 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
437 for (zone = pgdat->node_zones; in update_pgdat_span()438 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span()797 zone = pgdat->node_zones + i; in auto_movable_can_online_movable()841 struct zone *zone = &pgdat->node_zones[zid]; in default_kernel_zone_for_pfn()847 return &pgdat->node_zones[ZONE_NORMAL]; in default_kernel_zone_for_pfn()956 return &NODE_DATA(nid)->node_zones[ZONE_MOVABLE]; in auto_movable_zone_for_pfn()966 struct zone *movable_zone = &NODE_DATA(nid)->node_zones[ZONE_MOVABLE]; in default_zone_for_pfn()993 return &NODE_DATA(nid)->node_zones[ZONE_MOVABLE]; in zone_for_pfn_range()1168 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_present_pages()1751 present_pages += pgdat->node_zones[zt].present_pages; in node_states_check_changes_offline()[all …]
693 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node() local695 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in init_zones_in_node()
326 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold()985 struct zone *zones = NODE_DATA(node)->node_zones; in sum_zone_node_page_state()999 struct zone *zones = NODE_DATA(node)->node_zones; in sum_zone_numa_event_state()1442 struct zone *node_zones = pgdat->node_zones; in walk_zones_in_node() local1445 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in walk_zones_in_node()1654 struct zone *compare = &pgdat->node_zones[zid]; in is_zone_first_populated()
44 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
253 zone = &NODE_DATA(nid)->node_zones[ZONE_DEVICE]; in pagemap_range()267 memmap_init_zone_device(&NODE_DATA(nid)->node_zones[ZONE_DEVICE], in pagemap_range()
381 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable()2035 zone = &pgdat->node_zones[zoneid]; in fragmentation_score_node()2679 zone = &pgdat->node_zones[zoneid]; in proactive_compact_node()2711 zone = &pgdat->node_zones[zoneid]; in compact_node()2821 zone = &pgdat->node_zones[zoneid]; in kcompactd_node_suitable()2857 zone = &pgdat->node_zones[zoneid]; in kcompactd_do_work()
1658 struct zone *zone = &pgdat->node_zones[zid]; in init_reserved_page()2144 zone = pgdat->node_zones + zid; in deferred_init_memmap()6044 struct zone *zone = &pgdat->node_zones[zone_type]; in si_meminfo_node()6361 zone = pgdat->node_zones + zone_type; in build_zonerefs_node()7054 struct zone *zone = node->node_zones + j; in memmap_init()7593 struct zone *zone = pgdat->node_zones + i; in calculate_node_totalpages()7836 struct zone *zone = pgdat->node_zones + j; in free_area_init_core()8294 struct zone *zone = &pgdat->node_zones[zone_type]; in check_for_memory()8696 struct zone *zone = pgdat->node_zones + i; in calculate_totalreserve_pages()8733 struct zone *zone = &pgdat->node_zones[i]; in setup_per_zone_lowmem_reserve()[all …]
1132 struct zone *zone = pgdat->node_zones + i; in skip_throttle_noprogress()2918 struct zone *zone = &pgdat->node_zones[z]; in prepare_scan_count()6367 struct zone *zone = &pgdat->node_zones[z]; in should_continue_reclaim()6856 zone = &pgdat->node_zones[i]; in allow_direct_reclaim()7137 zone = pgdat->node_zones + i; in pgdat_watermark_boosted()7163 zone = pgdat->node_zones + i; in pgdat_balanced()7251 zone = pgdat->node_zones + z; in kswapd_shrink_node()7285 zone = pgdat->node_zones + i; in update_reclaim_active()7351 zone = pgdat->node_zones + i; in balance_pgdat()7383 zone = pgdat->node_zones + i; in balance_pgdat()[all …]
2493 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat()2547 if (managed_zone(pgdat->node_zones + z)) in numamigrate_isolate_page()2550 wakeup_kswapd(pgdat->node_zones + z, 0, order, ZONE_MOVABLE); in numamigrate_isolate_page()
2120 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages()
278 struct zone *zone = pgdat->node_zones + z; in node_dirtyable_memory()314 z = &NODE_DATA(node)->node_zones[i]; in highmem_dirtyable_memory()
33 struct zone node_zones[MAX_NR_ZONES]; member
23 struct zone *zone = &pgdat->node_zones[zoneid]; in __show_mem()
1254 struct zone node_zones[MAX_NR_ZONES]; member1446 #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones)1585 for (zone = (first_online_pgdat())->node_zones; \1590 for (zone = (first_online_pgdat())->node_zones; \
47 __mod_zone_page_state(&pgdat->node_zones[zid], in __update_lru_size()
1664 return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)]; in page_zone()
461 VMCOREINFO_OFFSET(pglist_data, node_zones); in crash_save_vmcoreinfo_init()
670 zone = pgdat->node_zones + i; in early_node_zone_for_memory_block()
192 ``node_zones``
150 (pglist_data, node_zones|nr_zones|node_mem_map|node_start_pfn|node_spanned_pages|node_id)
1690 mem_in_bytes += zone_managed_pages(&pgdat->node_zones[zone_type]); in kfd_fill_mem_info_for_cpu()
1481 struct zone *zone = pgdat->node_zones + z; in pgdat_free_space_enough()
Completed in 105 milliseconds