Lines Matching refs:start_pfn
301 unsigned long start_pfn, end_pfn; in early_calculate_totalpages() local
304 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in early_calculate_totalpages()
305 unsigned long pages = end_pfn - start_pfn; in early_calculate_totalpages()
463 unsigned long start_pfn, end_pfn; in find_zone_movable_pfns_for_nodes() local
481 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in find_zone_movable_pfns_for_nodes()
484 start_pfn = max(start_pfn, zone_movable_pfn[nid]); in find_zone_movable_pfns_for_nodes()
485 if (start_pfn >= end_pfn) in find_zone_movable_pfns_for_nodes()
489 if (start_pfn < usable_startpfn) { in find_zone_movable_pfns_for_nodes()
492 - start_pfn; in find_zone_movable_pfns_for_nodes()
511 start_pfn = usable_startpfn; in find_zone_movable_pfns_for_nodes()
519 size_pages = end_pfn - start_pfn; in find_zone_movable_pfns_for_nodes()
522 zone_movable_pfn[nid] = start_pfn + size_pages; in find_zone_movable_pfns_for_nodes()
550 unsigned long start_pfn, end_pfn; in find_zone_movable_pfns_for_nodes() local
555 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in find_zone_movable_pfns_for_nodes()
603 unsigned long start_pfn, end_pfn; in __early_pfn_to_nid() local
609 nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn); in __early_pfn_to_nid()
611 state->last_start = start_pfn; in __early_pfn_to_nid()
754 unsigned long start_pfn = PFN_DOWN(start); in reserve_bootmem_region() local
757 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region()
758 if (pfn_valid(start_pfn)) { in reserve_bootmem_region()
759 struct page *page = pfn_to_page(start_pfn); in reserve_bootmem_region()
761 init_reserved_page(start_pfn, nid); in reserve_bootmem_region()
850 unsigned long start_pfn, unsigned long zone_end_pfn, in memmap_init_range() argument
854 unsigned long pfn, end_pfn = start_pfn + size; in memmap_init_range()
872 if (start_pfn == altmap->base_pfn) in memmap_init_range()
873 start_pfn += altmap->reserve; in memmap_init_range()
878 for (pfn = start_pfn; pfn < end_pfn; ) { in memmap_init_range()
917 unsigned long start_pfn, in memmap_init_zone_range() argument
925 start_pfn = clamp(start_pfn, zone_start_pfn, zone_end_pfn); in memmap_init_zone_range()
928 if (start_pfn >= end_pfn) in memmap_init_zone_range()
931 memmap_init_range(end_pfn - start_pfn, nid, zone_id, start_pfn, in memmap_init_zone_range()
934 if (*hole_pfn < start_pfn) in memmap_init_zone_range()
935 init_unavailable_range(*hole_pfn, start_pfn, zone_id, nid); in memmap_init_zone_range()
942 unsigned long start_pfn, end_pfn; in memmap_init() local
946 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in memmap_init()
955 memmap_init_zone_range(zone, start_pfn, end_pfn, in memmap_init()
1071 unsigned long start_pfn, in memmap_init_zone_device() argument
1075 unsigned long pfn, end_pfn = start_pfn + nr_pages; in memmap_init_zone_device()
1092 start_pfn = altmap->base_pfn + vmem_altmap_offset(altmap); in memmap_init_zone_device()
1093 nr_pages = end_pfn - start_pfn; in memmap_init_zone_device()
1096 for (pfn = start_pfn; pfn < end_pfn; pfn += pfns_per_compound) { in memmap_init_zone_device()
1158 unsigned long start_pfn, end_pfn; in __absent_pages_in_range() local
1161 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in __absent_pages_in_range()
1162 start_pfn = clamp(start_pfn, range_start_pfn, range_end_pfn); in __absent_pages_in_range()
1164 nr_absent -= end_pfn - start_pfn; in __absent_pages_in_range()
1176 unsigned long __init absent_pages_in_range(unsigned long start_pfn, in absent_pages_in_range() argument
1179 return __absent_pages_in_range(MAX_NUMNODES, start_pfn, end_pfn); in absent_pages_in_range()
1202 unsigned long start_pfn, end_pfn; in zone_absent_pages_in_node() local
1206 start_pfn = clamp(memblock_region_memory_base_pfn(r), in zone_absent_pages_in_node()
1213 nr_absent += end_pfn - start_pfn; in zone_absent_pages_in_node()
1217 nr_absent += end_pfn - start_pfn; in zone_absent_pages_in_node()
1276 unsigned long start_pfn, end_pfn; in calc_nr_kernel_pages() local
1284 start_pfn = PFN_UP(start_addr); in calc_nr_kernel_pages()
1287 if (start_pfn < end_pfn) { in calc_nr_kernel_pages()
1288 nr_all_pages += end_pfn - start_pfn; in calc_nr_kernel_pages()
1290 start_pfn = clamp(start_pfn, 0, high_zone_low); in calc_nr_kernel_pages()
1293 nr_kernel_pages += end_pfn - start_pfn; in calc_nr_kernel_pages()
1654 unsigned long *start_pfn, unsigned long *end_pfn) in get_pfn_range_for_nid() argument
1659 *start_pfn = -1UL; in get_pfn_range_for_nid()
1663 *start_pfn = min(*start_pfn, this_start_pfn); in get_pfn_range_for_nid()
1667 if (*start_pfn == -1UL) in get_pfn_range_for_nid()
1668 *start_pfn = 0; in get_pfn_range_for_nid()
1674 unsigned long start_pfn = 0; in free_area_init_node() local
1680 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in free_area_init_node()
1683 pgdat->node_start_pfn = start_pfn; in free_area_init_node()
1686 if (start_pfn != end_pfn) { in free_area_init_node()
1688 (u64)start_pfn << PAGE_SHIFT, in free_area_init_node()
1691 calculate_node_totalpages(pgdat, start_pfn, end_pfn); in free_area_init_node()
1759 unsigned long start_pfn, end_pfn; in free_area_init() local
1769 start_pfn = PHYS_PFN(memblock_start_of_DRAM()); in free_area_init()
1781 end_pfn = max(max_zone_pfn[zone], start_pfn); in free_area_init()
1782 arch_zone_lowest_possible_pfn[zone] = start_pfn; in free_area_init()
1785 start_pfn = end_pfn; in free_area_init()
1823 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in free_area_init()
1825 (u64)start_pfn << PAGE_SHIFT, in free_area_init()
1827 subsection_map_init(start_pfn, end_pfn - start_pfn); in free_area_init()
2019 deferred_init_maxorder(u64 *i, struct zone *zone, unsigned long *start_pfn, in deferred_init_maxorder() argument
2022 unsigned long mo_pfn = ALIGN(*start_pfn + 1, MAX_ORDER_NR_PAGES); in deferred_init_maxorder()
2023 unsigned long spfn = *start_pfn, epfn = *end_pfn; in deferred_init_maxorder()
2028 for_each_free_mem_pfn_range_in_zone_from(j, zone, start_pfn, end_pfn) { in deferred_init_maxorder()
2031 if (mo_pfn <= *start_pfn) in deferred_init_maxorder()
2035 nr_pages += deferred_init_pages(zone, *start_pfn, t); in deferred_init_maxorder()
2038 *start_pfn = mo_pfn; in deferred_init_maxorder()
2063 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn, in deferred_init_memmap_chunk() argument
2070 deferred_init_mem_pfn_range_in_zone(&i, zone, &spfn, &epfn, start_pfn); in deferred_init_memmap_chunk()