| /linux/mm/ |
| A D | mm_init.c | 484 start_pfn = max(start_pfn, zone_movable_pfn[nid]); in find_zone_movable_pfns_for_nodes() 492 - start_pfn; in find_zone_movable_pfns_for_nodes() 757 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region() 925 start_pfn = clamp(start_pfn, zone_start_pfn, zone_end_pfn); in memmap_init_zone_range() 931 memmap_init_range(end_pfn - start_pfn, nid, zone_id, start_pfn, in memmap_init_zone_range() 1162 start_pfn = clamp(start_pfn, range_start_pfn, range_end_pfn); in __absent_pages_in_range() 1290 start_pfn = clamp(start_pfn, 0, high_zone_low); in calc_nr_kernel_pages() 1659 *start_pfn = -1UL; in get_pfn_range_for_nid() 1663 *start_pfn = min(*start_pfn, this_start_pfn); in get_pfn_range_for_nid() 1668 *start_pfn = 0; in get_pfn_range_for_nid() [all …]
|
| A D | page_isolation.c | 36 struct page *page = pfn_to_page(start_pfn); in has_unmovable_pages() 40 VM_BUG_ON(pageblock_start_pfn(start_pfn) != in has_unmovable_pages() 55 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in has_unmovable_pages() 148 unsigned long start_pfn, unsigned long end_pfn) in set_migratetype_isolate() argument 177 check_unmovable_start = max(page_to_pfn(page), start_pfn); in set_migratetype_isolate() 312 unsigned long start_pfn; in isolate_single_pageblock() local 367 for (pfn = start_pfn; pfn < boundary_pfn;) { in isolate_single_pageblock() 486 unsigned long isolate_start = pageblock_start_pfn(start_pfn); in start_isolate_page_range() 514 start_pfn, end_pfn)) { in start_isolate_page_range() 624 page = __first_valid_page(start_pfn, end_pfn - start_pfn); in test_pages_isolated() [all …]
|
| A D | memory_hotplug.c | 415 unsigned long start_pfn, in find_smallest_section_pfn() argument 418 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 428 return start_pfn; in find_smallest_section_pfn() 436 unsigned long start_pfn, in find_biggest_section_pfn() argument 489 start_pfn); in shrink_zone_span() 528 unsigned long start_pfn, in remove_pfn_range_from_zone() argument 556 shrink_zone_span(zone, start_pfn, start_pfn + nr_pages); in remove_pfn_range_from_zone() 705 zone->zone_start_pfn = start_pfn; in resize_zone_range() 1170 arg.start_pfn = pfn; in online_pages() 2002 arg.start_pfn = start_pfn; in offline_pages() [all …]
|
| A D | page_ext.c | 378 start = SECTION_ALIGN_DOWN(start_pfn); in online_page_ext() 379 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in online_page_ext() 387 nid = pfn_to_nid(start_pfn); in online_page_ext() 409 start = SECTION_ALIGN_DOWN(start_pfn); in offline_page_ext() 410 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in offline_page_ext() 438 ret = online_page_ext(mn->start_pfn, in page_ext_callback() 442 offline_page_ext(mn->start_pfn, in page_ext_callback() 446 offline_page_ext(mn->start_pfn, in page_ext_callback() 468 unsigned long start_pfn, end_pfn; in page_ext_init() local 470 start_pfn = node_start_pfn(nid); in page_ext_init() [all …]
|
| A D | sparse.c | 138 if (*start_pfn > max_sparsemem_pfn) { in mminit_validate_memmodel_limits() 141 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 143 *start_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 148 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 594 void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in online_mem_sections() argument 598 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in online_mem_sections() 616 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in offline_mem_sections() 893 int __meminit sparse_add_section(int nid, unsigned long start_pfn, in sparse_add_section() argument 897 unsigned long section_nr = pfn_to_section_nr(start_pfn); in sparse_add_section() 906 memmap = section_activate(nid, start_pfn, nr_pages, altmap, pgmap); in sparse_add_section() [all …]
|
| A D | shuffle.c | 83 unsigned long start_pfn = z->zone_start_pfn; in __shuffle_zone() local 89 start_pfn = ALIGN(start_pfn, order_pages); in __shuffle_zone() 90 for (i = start_pfn; i < end_pfn; i += order_pages) { in __shuffle_zone()
|
| /linux/arch/powerpc/platforms/powernv/ |
| A D | memtrace.c | 91 static void memtrace_clear_range(unsigned long start_pfn, in memtrace_clear_range() argument 97 for (pfn = start_pfn; pfn < start_pfn + nr_pages; pfn++) { in memtrace_clear_range() 114 unsigned long pfn, start_pfn; in memtrace_alloc_node() local 125 start_pfn = page_to_pfn(page); in memtrace_alloc_node() 132 memtrace_clear_range(start_pfn, nr_pages); in memtrace_alloc_node() 138 for (pfn = start_pfn; pfn < start_pfn + nr_pages; pfn++) in memtrace_alloc_node() 141 arch_remove_linear_mapping(PFN_PHYS(start_pfn), size); in memtrace_alloc_node() 143 return PFN_PHYS(start_pfn); in memtrace_alloc_node() 217 const unsigned long start_pfn = PHYS_PFN(start); in memtrace_free() local 225 for (pfn = start_pfn; pfn < start_pfn + nr_pages; pfn++) in memtrace_free() [all …]
|
| /linux/arch/sparc/mm/ |
| A D | init_32.c | 64 unsigned long start_pfn = sp_banks[i].base_addr >> PAGE_SHIFT; in calc_highpages() local 70 if (start_pfn < max_low_pfn) in calc_highpages() 71 start_pfn = max_low_pfn; in calc_highpages() 73 nr += end_pfn - start_pfn; in calc_highpages() 133 unsigned long start_pfn, bytes_avail, size; in bootmem_init() local 174 start_pfn >>= PAGE_SHIFT; in bootmem_init() 192 size = (start_pfn << PAGE_SHIFT) - phys_base; in bootmem_init() 243 for (tmp = start_pfn; tmp < end_pfn; tmp++) in map_high_region() 289 if (start_pfn < highstart_pfn) in mem_init() 290 start_pfn = highstart_pfn; in mem_init() [all …]
|
| /linux/include/trace/events/ |
| A D | page_isolation.h | 13 unsigned long start_pfn, 17 TP_ARGS(start_pfn, end_pfn, fin_pfn), 20 __field(unsigned long, start_pfn) 26 __entry->start_pfn = start_pfn; 32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn,
|
| A D | compaction.h | 17 unsigned long start_pfn, 22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 25 __field(unsigned long, start_pfn) 32 __entry->start_pfn = start_pfn; 39 __entry->start_pfn, 48 unsigned long start_pfn, 53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 59 unsigned long start_pfn, 64 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 70 unsigned long start_pfn, [all …]
|
| /linux/arch/x86/xen/ |
| A D | setup.c | 93 start_r = xen_extra_mem[i].start_pfn; in xen_del_extra_mem() 97 if (start_r == start_pfn) { in xen_del_extra_mem() 99 xen_extra_mem[i].start_pfn += n_pfns; in xen_del_extra_mem() 110 if (start_pfn > start_r && start_pfn < start_r + size_r) { in xen_del_extra_mem() 115 (start_pfn + n_pfns)); in xen_del_extra_mem() 132 if (pfn >= xen_extra_mem[i].start_pfn && in xen_chk_extra_mem() 151 pfn_s = xen_extra_mem[i].start_pfn; in xen_inv_extra_mem() 223 WARN_ON(start_pfn > end_pfn); in xen_set_identity_and_release_chunk() 351 unsigned long n = end_pfn - start_pfn; in xen_set_identity_and_remap_chunk() 405 if (start_pfn >= ini_nr_pages) in xen_count_remap_pages() [all …]
|
| A D | enlighten.c | 353 void __init xen_add_extra_mem(unsigned long start_pfn, unsigned long n_pfns) in xen_add_extra_mem() argument 364 xen_extra_mem[i].start_pfn = start_pfn; in xen_add_extra_mem() 369 if (xen_extra_mem[i].start_pfn + xen_extra_mem[i].n_pfns == in xen_add_extra_mem() 370 start_pfn) { in xen_add_extra_mem() 378 memblock_reserve(PFN_PHYS(start_pfn), PFN_PHYS(n_pfns)); in xen_add_extra_mem() 401 pfn_to_page(xen_extra_mem[i].start_pfn + j); in arch_xen_unpopulated_init()
|
| /linux/arch/x86/mm/ |
| A D | init.c | 328 if (start_pfn < end_pfn) { in save_mr() 331 mr[nr_range].start = start_pfn<<PAGE_SHIFT; in save_mr() 410 pfn = start_pfn = PFN_DOWN(start); in split_mem_range() 427 if (start_pfn < end_pfn) { in split_mem_range() 442 if (start_pfn < end_pfn) { in split_mem_range() 452 if (start_pfn < end_pfn) { in split_mem_range() 462 if (start_pfn < end_pfn) { in split_mem_range() 470 start_pfn = pfn; in split_mem_range() 510 if (start_pfn < (1UL<<(32-PAGE_SHIFT))) in add_pfn_range_mapped() 520 if ((start_pfn >= pfn_mapped[i].start) && in pfn_range_is_mapped() [all …]
|
| /linux/arch/sh/mm/ |
| A D | numa.c | 25 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 30 start_pfn = PFN_DOWN(start); in setup_bootmem_node() 38 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 48 NODE_DATA(nid)->node_start_pfn = start_pfn; in setup_bootmem_node() 49 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
|
| A D | init.c | 210 unsigned long start_pfn, end_pfn; in allocate_pgdat() local 212 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat() 218 NODE_DATA(nid)->node_start_pfn = start_pfn; in allocate_pgdat() 219 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat() 224 unsigned long start_pfn, end_pfn; in do_init_bootmem() local 228 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) in do_init_bootmem() 229 __add_active_range(0, start_pfn, end_pfn); in do_init_bootmem() 242 unsigned long start_pfn; in early_reserve_mem() local 250 start_pfn = PFN_UP(__pa(_end)); in early_reserve_mem() 258 memblock_reserve(start, (PFN_PHYS(start_pfn) + PAGE_SIZE - 1) - start); in early_reserve_mem()
|
| /linux/include/linux/ |
| A D | memory_hotplug.h | 130 extern unsigned long __offline_isolated_pages(unsigned long start_pfn, 157 extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages, 161 extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 165 static inline int add_pages(int nid, unsigned long start_pfn, in add_pages() argument 168 return __add_pages(nid, start_pfn, nr_pages, params); in add_pages() 171 int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 282 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages, 291 static inline int offline_pages(unsigned long start_pfn, unsigned long nr_pages, in offline_pages() argument 314 extern void move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, 318 unsigned long start_pfn, [all …]
|
| A D | node.h | 107 void register_memory_blocks_under_node(int nid, unsigned long start_pfn, 111 static inline void register_memory_blocks_under_node(int nid, unsigned long start_pfn, in register_memory_blocks_under_node() argument 131 unsigned long start_pfn = pgdat->node_start_pfn; in register_one_node() local 132 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node() 137 register_memory_blocks_under_node(nid, start_pfn, end_pfn, in register_one_node()
|
| A D | page-isolation.h | 41 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 44 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 47 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
|
| /linux/drivers/base/ |
| A D | memory.c | 199 start_pfn, nr_pages); in memory_block_online() 211 arg.altmap_start_pfn = start_pfn; in memory_block_online() 213 arg.start_pfn = start_pfn + nr_vmemmap_pages; in memory_block_online() 228 ret = online_pages(start_pfn + nr_vmemmap_pages, in memory_block_online() 280 ret = offline_pages(start_pfn + nr_vmemmap_pages, in memory_block_offline() 285 adjust_present_page_count(pfn_to_page(start_pfn), in memory_block_offline() 294 arg.altmap_start_pfn = start_pfn; in memory_block_offline() 296 arg.start_pfn = start_pfn + nr_vmemmap_pages; in memory_block_offline() 431 arch_get_memory_phys_device(start_pfn)); in phys_device_show() 477 start_pfn, nr_pages); in valid_zones_show() [all …]
|
| A D | arch_numa.c | 195 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) in setup_node_data() argument 197 if (start_pfn >= end_pfn) in setup_node_data() 203 NODE_DATA(nid)->node_start_pfn = start_pfn; in setup_node_data() 204 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_node_data() 213 unsigned long start_pfn, end_pfn; in numa_register_nodes() local 215 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in numa_register_nodes() 216 setup_node_data(nid, start_pfn, end_pfn); in numa_register_nodes()
|
| /linux/drivers/hv/ |
| A D | hv_balloon.c | 425 unsigned long start_pfn; member 438 unsigned long start_pfn; member 614 while ((pfn >= has->start_pfn) && in hv_page_offline_check() 712 unsigned long start_pfn; in hv_mem_hot_add() local 772 if (pfn < has->start_pfn || in hv_online_page() 794 if (start_pfn < has->start_pfn || start_pfn >= has->end_pfn) in pfn_covered() 810 gap->end_pfn = start_pfn; in pfn_covered() 813 has->covered_end_pfn = start_pfn; in pfn_covered() 853 if (start_pfn < has->start_pfn || start_pfn >= has->end_pfn) in handle_pg_range() 878 if (start_pfn > has->start_pfn && in handle_pg_range() [all …]
|
| /linux/arch/mips/loongson64/ |
| A D | numa.c | 85 unsigned long start_pfn, end_pfn; in node_mem_init() local 91 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 93 node, start_pfn, end_pfn); in node_mem_init() 97 NODE_DATA(node)->node_start_pfn = start_pfn; in node_mem_init() 98 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; in node_mem_init() 120 memblock_reserve(0, PAGE_SIZE * start_pfn); in node_mem_init()
|
| /linux/arch/parisc/mm/ |
| A D | init.c | 133 if (pmem_ranges[j-1].start_pfn < in setup_bootmem() 134 pmem_ranges[j].start_pfn) { in setup_bootmem() 149 if (pmem_ranges[i].start_pfn - in setup_bootmem() 150 (pmem_ranges[i-1].start_pfn + in setup_bootmem() 155 pmem_ranges[i].start_pfn - in setup_bootmem() 156 (pmem_ranges[i-1].start_pfn + in setup_bootmem() 249 unsigned long start_pfn; in setup_bootmem() local 254 start_pfn = pmem_ranges[i].start_pfn; in setup_bootmem() 257 start = start_pfn << PAGE_SHIFT; in setup_bootmem() 263 if ((start_pfn + npages) > max_pfn) in setup_bootmem() [all …]
|
| /linux/drivers/gpu/drm/imagination/ |
| A D | pvr_vm_mips.c | 156 s32 start_pfn; in pvr_vm_mips_map() local 172 start_pfn = (start & fw_dev->fw_heap_info.offset_mask) >> ROGUE_MIPSFW_LOG2_PAGE_SIZE_4K; in pvr_vm_mips_map() 182 for (pfn = start_pfn; pfn <= end_pfn; pfn++) { in pvr_vm_mips_map() 187 (pfn - start_pfn) << in pvr_vm_mips_map() 205 while (--pfn >= start_pfn) in pvr_vm_mips_map() 228 const u32 start_pfn = (start & fw_dev->fw_heap_info.offset_mask) >> in pvr_vm_mips_unmap() local 233 for (u32 pfn = start_pfn; pfn < end_pfn; pfn++) in pvr_vm_mips_unmap()
|
| /linux/arch/x86/virt/vmx/tdx/ |
| A D | tdx.c | 194 tmb->start_pfn = start_pfn; in add_tdx_memblock() 223 unsigned long start_pfn, end_pfn; in build_tdx_memlist() local 226 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in build_tdx_memlist() 233 start_pfn = max(start_pfn, PHYS_PFN(SZ_1M)); in build_tdx_memlist() 234 if (start_pfn >= end_pfn) in build_tdx_memlist() 242 ret = add_tdx_memblock(tmb_list, start_pfn, end_pfn, nid); in build_tdx_memlist() 425 start = TDMR_ALIGN_DOWN(PFN_PHYS(tmb->start_pfn)); in fill_out_tdmrs() 771 start = PFN_PHYS(tmb->start_pfn); in tdmr_populate_rsvd_holes() 1371 static bool is_tdx_memory(unsigned long start_pfn, unsigned long end_pfn) in is_tdx_memory() argument 1385 if (start_pfn >= tmb->start_pfn && end_pfn <= tmb->end_pfn) in is_tdx_memory() [all …]
|