| /mm/ |
| A D | mm_init.c | 500 start_pfn = max(start_pfn, zone_movable_pfn[nid]); in find_zone_movable_pfns_for_nodes() 508 - start_pfn; in find_zone_movable_pfns_for_nodes() 953 start_pfn = clamp(start_pfn, zone_start_pfn, zone_end_pfn); in memmap_init_zone_range() 959 memmap_init_range(end_pfn - start_pfn, nid, zone_id, start_pfn, in memmap_init_zone_range() 1204 start_pfn = clamp(start_pfn, range_start_pfn, range_end_pfn); in __absent_pages_in_range() 1332 start_pfn = clamp(start_pfn, 0, high_zone_low); in calc_nr_kernel_pages() 1706 *start_pfn = -1UL; in get_pfn_range_for_nid() 1710 *start_pfn = min(*start_pfn, this_start_pfn); in get_pfn_range_for_nid() 1715 *start_pfn = 0; in get_pfn_range_for_nid() 1853 start_pfn = end_pfn; in free_area_init() [all …]
|
| A D | memory_hotplug.c | 435 unsigned long start_pfn, in find_smallest_section_pfn() argument 438 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 448 return start_pfn; in find_smallest_section_pfn() 456 unsigned long start_pfn, in find_biggest_section_pfn() argument 509 start_pfn); in shrink_zone_span() 548 unsigned long start_pfn, in remove_pfn_range_from_zone() argument 576 shrink_zone_span(zone, start_pfn, start_pfn + nr_pages); in remove_pfn_range_from_zone() 709 zone->zone_start_pfn = start_pfn; in resize_zone_range() 1157 .start_pfn = pfn, in online_pages() 1905 .start_pfn = start_pfn, in offline_pages() [all …]
|
| A D | page_isolation.c | 36 struct page *page = pfn_to_page(start_pfn); in has_unmovable_pages() 40 VM_BUG_ON(pageblock_start_pfn(start_pfn) != in has_unmovable_pages() 55 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in has_unmovable_pages() 155 unsigned long start_pfn, unsigned long end_pfn) in set_migratetype_isolate() argument 184 check_unmovable_start = max(page_to_pfn(page), start_pfn); in set_migratetype_isolate() 317 unsigned long start_pfn; in isolate_single_pageblock() local 371 for (pfn = start_pfn; pfn < boundary_pfn;) { in isolate_single_pageblock() 481 unsigned long isolate_start = pageblock_start_pfn(start_pfn); in start_isolate_page_range() 507 if (page && set_migratetype_isolate(page, mode, start_pfn, in start_isolate_page_range() 625 page = __first_valid_page(start_pfn, end_pfn - start_pfn); in test_pages_isolated() [all …]
|
| A D | page_ext.c | 374 int nid = pfn_to_nid(start_pfn); in online_page_ext() 378 start = SECTION_ALIGN_DOWN(start_pfn); in online_page_ext() 379 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in online_page_ext() 399 start = SECTION_ALIGN_DOWN(start_pfn); in offline_page_ext() 400 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in offline_page_ext() 428 ret = online_page_ext(mn->start_pfn, mn->nr_pages); in page_ext_callback() 431 offline_page_ext(mn->start_pfn, in page_ext_callback() 435 offline_page_ext(mn->start_pfn, in page_ext_callback() 457 unsigned long start_pfn, end_pfn; in page_ext_init() local 459 start_pfn = node_start_pfn(nid); in page_ext_init() [all …]
|
| A D | sparse.c | 138 if (*start_pfn > max_sparsemem_pfn) { in mminit_validate_memmodel_limits() 141 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 143 *start_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 148 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 631 void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in online_mem_sections() argument 635 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in online_mem_sections() 653 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in offline_mem_sections() 930 int __meminit sparse_add_section(int nid, unsigned long start_pfn, in sparse_add_section() argument 934 unsigned long section_nr = pfn_to_section_nr(start_pfn); in sparse_add_section() 943 memmap = section_activate(nid, start_pfn, nr_pages, altmap, pgmap); in sparse_add_section() [all …]
|
| A D | compaction.c | 557 unsigned long *start_pfn, in isolate_freepages_block() argument 567 unsigned long blockpfn = *start_pfn; in isolate_freepages_block() 664 *start_pfn = blockpfn; in isolate_freepages_block() 696 unsigned long start_pfn, unsigned long end_pfn) in isolate_freepages_range() argument 704 pfn = start_pfn; in isolate_freepages_range() 848 unsigned long start_pfn = low_pfn; in isolate_migratepages_block() local 1327 pfn = start_pfn; in isolate_migratepages_range() 1460 unsigned long start_pfn, end_pfn; in fast_isolate_around() local 1482 if (start_pfn == end_pfn && !cc->no_set_skip_hint) in fast_isolate_around() 2561 cc->migrate_pfn = start_pfn; in compact_zone() [all …]
|
| A D | shuffle.c | 83 unsigned long start_pfn = z->zone_start_pfn; in __shuffle_zone() local 89 start_pfn = ALIGN(start_pfn, order_pages); in __shuffle_zone() 90 for (i = start_pfn; i < end_pfn; i += order_pages) { in __shuffle_zone()
|
| A D | bootmem_info.c | 44 static void __init register_page_bootmem_info_section(unsigned long start_pfn) in register_page_bootmem_info_section() argument 51 section_nr = pfn_to_section_nr(start_pfn); in register_page_bootmem_info_section() 79 static void __init register_page_bootmem_info_section(unsigned long start_pfn) in register_page_bootmem_info_section() argument 86 section_nr = pfn_to_section_nr(start_pfn); in register_page_bootmem_info_section()
|
| A D | page_alloc.c | 583 unsigned long sp, start_pfn; in page_outside_zone_boundaries() local 595 start_pfn, start_pfn + sp); in page_outside_zone_boundaries() 1989 *start_pfn = start; in prep_move_freepages_block() 2020 unsigned long start_pfn; in move_freepages_block() local 2039 unsigned long pfn = start_pfn; in find_large_buddy() 2044 return start_pfn; in find_large_buddy() 2055 return start_pfn; in find_large_buddy() 2088 unsigned long start_pfn, pfn; in __move_freepages_block_isolate() local 2107 if (pfn != start_pfn) { in __move_freepages_block_isolate() 2293 unsigned long start_pfn; in try_to_claim_block() local [all …]
|
| A D | sparse-vmemmap.c | 479 static bool __meminit reuse_compound_section(unsigned long start_pfn, in reuse_compound_section() argument 483 unsigned long offset = start_pfn - in reuse_compound_section() 506 static int __meminit vmemmap_populate_compound_pages(unsigned long start_pfn, in vmemmap_populate_compound_pages() argument 515 if (reuse_compound_section(start_pfn, pgmap)) { in vmemmap_populate_compound_pages()
|
| A D | memblock.c | 773 unsigned long start_pfn, end_pfn, mem_size_mb; in memblock_validate_numa_coverage() local 777 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in memblock_validate_numa_coverage() 779 nr_pages += end_pfn - start_pfn; in memblock_validate_numa_coverage() 2030 unsigned long *start_pfn, unsigned long *end_pfn) in memblock_search_pfn_nid() argument 2038 *start_pfn = PFN_DOWN(type->regions[mid].base); in memblock_search_pfn_nid() 2175 static void __init free_memmap(unsigned long start_pfn, unsigned long end_pfn) in free_memmap() argument 2183 start_pg = pfn_to_page(start_pfn - 1) + 1; in free_memmap() 2284 unsigned long start_pfn = PFN_UP(start); in __free_memory_core() local 2290 if (start_pfn >= end_pfn) in __free_memory_core() 2293 __free_pages_memory(start_pfn, end_pfn); in __free_memory_core() [all …]
|
| A D | internal.h | 721 extern struct page *__pageblock_pfn_to_page(unsigned long start_pfn, 724 static inline struct page *pageblock_pfn_to_page(unsigned long start_pfn, in pageblock_pfn_to_page() argument 728 return pfn_to_page(start_pfn); in pageblock_pfn_to_page() 730 return __pageblock_pfn_to_page(start_pfn, end_pfn, zone); in pageblock_pfn_to_page() 734 bool pfn_range_intersects_zones(int nid, unsigned long start_pfn, 921 unsigned long start_pfn, unsigned long end_pfn);
|
| A D | numa_emulation.c | 37 unsigned long start_pfn = PFN_UP(start); in mem_hole_size() local 40 if (start_pfn < end_pfn) in mem_hole_size() 41 return PFN_PHYS(absent_pages_in_range(start_pfn, end_pfn)); in mem_hole_size()
|
| A D | ksm.c | 3167 unsigned long start_pfn, in stable_node_dup_remove_range() argument 3170 if (stable_node->kpfn >= start_pfn && in stable_node_dup_remove_range() 3183 unsigned long start_pfn, in stable_node_chain_remove_range() argument 3192 return stable_node_dup_remove_range(stable_node, start_pfn, in stable_node_chain_remove_range() 3199 stable_node_dup_remove_range(dup, start_pfn, end_pfn); in stable_node_chain_remove_range() 3208 static void ksm_check_stable_tree(unsigned long start_pfn, in ksm_check_stable_tree() argument 3220 start_pfn, end_pfn, in ksm_check_stable_tree() 3230 if (stable_node->kpfn >= start_pfn && in ksm_check_stable_tree() 3264 ksm_check_stable_tree(mn->start_pfn, in ksm_memory_callback() 3265 mn->start_pfn + mn->nr_pages); in ksm_memory_callback()
|
| A D | hugetlb.c | 2203 int dissolve_free_hugetlb_folios(unsigned long start_pfn, unsigned long end_pfn) in dissolve_free_hugetlb_folios() argument 2218 for (pfn = start_pfn; pfn < end_pfn; pfn += 1 << order) { in dissolve_free_hugetlb_folios() 2918 while (start_pfn < end_pfn) { in replace_free_hugepage_folios() 2919 folio = pfn_folio(start_pfn); in replace_free_hugepage_folios() 2929 start_pfn++; in replace_free_hugepage_folios() 3339 unsigned long start_pfn; in hugetlb_bootmem_page_zones_valid() local 3354 start_pfn = virt_to_phys(m) >> PAGE_SHIFT; in hugetlb_bootmem_page_zones_valid() 3356 valid = !pfn_range_intersects_zones(nid, start_pfn, in hugetlb_bootmem_page_zones_valid() 4999 unsigned long start_pfn, end_pfn; in hugetlb_bootmem_set_nodes() local 5004 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in hugetlb_bootmem_set_nodes() [all …]
|
| A D | gup.c | 2263 const unsigned long start_pfn = folio_pfn(folio); in pofs_next_folio() local 2264 const unsigned long end_pfn = start_pfn + folio_nr_pages(folio); in pofs_next_folio() 2270 if (pfn < start_pfn || pfn >= end_pfn) in pofs_next_folio()
|
| A D | kmemleak.c | 1742 unsigned long start_pfn = zone->zone_start_pfn; in kmemleak_scan() local 1746 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in kmemleak_scan()
|
| A D | vmstat.c | 1642 unsigned long start_pfn = zone->zone_start_pfn; in pagetypeinfo_showblockcount_print() local 1646 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in pagetypeinfo_showblockcount_print()
|
| /mm/kasan/ |
| A D | shadow.c | 221 start_kaddr = (unsigned long)pfn_to_kaddr(mem_data->start_pfn); in kasan_mem_notifier() 245 pfn_to_nid(mem_data->start_pfn), in kasan_mem_notifier()
|