| /arch/x86/mm/ |
| A D | init.c | 330 if (start_pfn < end_pfn) { in save_mr() 427 if (end_pfn > limit_pfn) in split_mem_range() 428 end_pfn = limit_pfn; in split_mem_range() 429 if (start_pfn < end_pfn) { in split_mem_range() 431 pfn = end_pfn; in split_mem_range() 444 if (start_pfn < end_pfn) { in split_mem_range() 447 pfn = end_pfn; in split_mem_range() 454 if (start_pfn < end_pfn) { in split_mem_range() 458 pfn = end_pfn; in split_mem_range() 467 pfn = end_pfn; in split_mem_range() [all …]
|
| A D | numa.c | 128 unsigned long start_pfn, end_pfn; in numa_register_nodes() local 134 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in numa_register_nodes() 135 if (start_pfn >= end_pfn) in numa_register_nodes()
|
| A D | init_32.c | 256 unsigned long start_pfn, end_pfn; in kernel_physical_mapping_init() local 267 end_pfn = end >> PAGE_SHIFT; in kernel_physical_mapping_init() 296 if (pfn >= end_pfn) in kernel_physical_mapping_init() 304 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init() 344 for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; in kernel_physical_mapping_init()
|
| /arch/sh/mm/ |
| A D | numa.c | 25 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 31 end_pfn = PFN_DOWN(end); in setup_bootmem_node() 38 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 49 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
|
| A D | init.c | 204 unsigned long start_pfn, end_pfn; in allocate_pgdat() local 206 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat() 213 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat() 218 unsigned long start_pfn, end_pfn; in do_init_bootmem() local 222 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) in do_init_bootmem() 223 __add_active_range(0, start_pfn, end_pfn); in do_init_bootmem()
|
| /arch/mips/loongson64/ |
| A D | numa.c | 85 unsigned long start_pfn, end_pfn; in node_mem_init() local 91 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 93 node, start_pfn, end_pfn); in node_mem_init() 98 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; in node_mem_init() 108 max_low_pfn = end_pfn; in node_mem_init()
|
| /arch/loongarch/kernel/ |
| A D | numa.c | 145 unsigned long start_pfn, end_pfn; in node_mem_init() local 152 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 154 node, start_pfn, end_pfn); in node_mem_init() 292 unsigned long start_pfn, end_pfn; in paging_init() local 294 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in paging_init() 296 if (end_pfn > max_low_pfn) in paging_init() 297 max_low_pfn = end_pfn; in paging_init()
|
| /arch/sh/kernel/ |
| A D | swsusp.c | 22 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local 24 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
| A D | setup.c | 194 unsigned long end_pfn) in __add_active_range() argument 202 end = end_pfn << PAGE_SHIFT; in __add_active_range() 211 start_pfn, end_pfn); in __add_active_range() 235 memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn), in __add_active_range()
|
| /arch/x86/platform/efi/ |
| A D | efi_32.c | 38 u64 start_pfn, end_pfn, end; in efi_map_region() local 45 end_pfn = PFN_UP(end); in efi_map_region() 47 if (pfn_range_is_mapped(start_pfn, end_pfn)) { in efi_map_region()
|
| /arch/x86/xen/ |
| A D | setup.c | 215 unsigned long end_pfn) in xen_set_identity_and_release_chunk() argument 220 WARN_ON(start_pfn > end_pfn); in xen_set_identity_and_release_chunk() 223 end = min(end_pfn, ini_nr_pages); in xen_set_identity_and_release_chunk() 242 set_phys_range_identity(start_pfn, end_pfn); in xen_set_identity_and_release_chunk() 348 unsigned long n = end_pfn - start_pfn; in xen_set_identity_and_remap_chunk() 399 unsigned long start_pfn, unsigned long end_pfn, in xen_count_remap_pages() argument 405 return remap_pages + min(end_pfn, ini_nr_pages) - start_pfn; in xen_count_remap_pages() 432 unsigned long end_pfn = PFN_UP(end); in xen_foreach_remap_area() local 435 end_pfn = PFN_UP(entry->addr); in xen_foreach_remap_area() 437 if (start_pfn < end_pfn) in xen_foreach_remap_area() [all …]
|
| /arch/s390/kernel/ |
| A D | early.c | 78 unsigned long end_pfn, init_pfn; in init_kernel_storage_key() local 80 end_pfn = PFN_UP(__pa(_end)); in init_kernel_storage_key() 82 for (init_pfn = 0 ; init_pfn < end_pfn; init_pfn++) in init_kernel_storage_key()
|
| /arch/powerpc/mm/ |
| A D | numa.c | 89 static int __init fake_numa_create_new_node(unsigned long end_pfn, in fake_numa_create_new_node() argument 120 if ((end_pfn << PAGE_SHIFT) > mem) { in fake_numa_create_new_node() 1041 unsigned long start_pfn, end_pfn; in setup_nonnuma() local 1048 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { in setup_nonnuma() 1049 fake_numa_create_new_node(end_pfn, &nid); in setup_nonnuma() 1051 PFN_PHYS(end_pfn - start_pfn), in setup_nonnuma() 1093 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) in setup_node_data() argument 1095 u64 spanned_pages = end_pfn - start_pfn; in setup_node_data() 1210 unsigned long start_pfn, end_pfn; in initmem_init() local 1212 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in initmem_init() [all …]
|
| A D | mem.c | 119 unsigned long end_pfn = PFN_UP(start + size); in update_end_of_memory_vars() local 121 if (end_pfn > max_pfn) { in update_end_of_memory_vars() 122 max_pfn = end_pfn; in update_end_of_memory_vars() 123 max_low_pfn = end_pfn; in update_end_of_memory_vars()
|
| /arch/x86/include/asm/ |
| A D | mtrr.h | 72 extern int mtrr_trim_uncached_memory(unsigned long end_pfn); 114 static inline int mtrr_trim_uncached_memory(unsigned long end_pfn) in mtrr_trim_uncached_memory() argument
|
| A D | page_types.h | 63 bool pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn);
|
| /arch/x86/virt/vmx/tdx/ |
| A D | tdx.c | 192 unsigned long end_pfn, int nid) in add_tdx_memblock() argument 202 tmb->end_pfn = end_pfn; in add_tdx_memblock() 230 unsigned long start_pfn, end_pfn; in build_tdx_memlist() local 233 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in build_tdx_memlist() 241 if (start_pfn >= end_pfn) in build_tdx_memlist() 249 ret = add_tdx_memblock(tmb_list, start_pfn, end_pfn, nid); in build_tdx_memlist() 391 end = TDMR_ALIGN_UP(PFN_PHYS(tmb->end_pfn)); in fill_out_tdmrs() 490 if (tmb->end_pfn > PHYS_PFN(tdmr->base)) in tdmr_get_nid() 737 end = PFN_PHYS(tmb->end_pfn); in tdmr_populate_rsvd_holes() 1344 static bool is_tdx_memory(unsigned long start_pfn, unsigned long end_pfn) in is_tdx_memory() argument [all …]
|
| A D | tdx.h | 105 unsigned long end_pfn; member
|
| /arch/sparc/mm/ |
| A D | init_32.c | 65 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in calc_highpages() local 67 if (end_pfn <= max_low_pfn) in calc_highpages() 73 nr += end_pfn - start_pfn; in calc_highpages()
|
| /arch/sh/include/asm/ |
| A D | mmzone.h | 38 unsigned long end_pfn);
|
| /arch/parisc/mm/ |
| A D | init.c | 222 unsigned long end_pfn; in setup_bootmem() local 226 end_pfn = pmem_ranges[0].start_pfn + pmem_ranges[0].pages; in setup_bootmem() 229 hole_pages = pmem_ranges[i].start_pfn - end_pfn; in setup_bootmem() 231 pmem_holes[npmem_holes].start_pfn = end_pfn; in setup_bootmem() 233 end_pfn += hole_pages; in setup_bootmem() 235 end_pfn += pmem_ranges[i].pages; in setup_bootmem() 238 pmem_ranges[0].pages = end_pfn - pmem_ranges[0].start_pfn; in setup_bootmem()
|
| /arch/mips/sgi-ip27/ |
| A D | ip27-memory.c | 354 unsigned long start_pfn, end_pfn; in node_mem_init() local 356 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 366 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; in node_mem_init()
|
| /arch/x86/kernel/ |
| A D | amd_gart_64.c | 712 unsigned long start_pfn, end_pfn; in gart_iommu_init() local 742 end_pfn = (aper_base>>PAGE_SHIFT) + (aper_size>>PAGE_SHIFT); in gart_iommu_init() 745 if (!pfn_range_is_mapped(start_pfn, end_pfn)) in gart_iommu_init() 746 init_memory_mapping(start_pfn<<PAGE_SHIFT, end_pfn<<PAGE_SHIFT, in gart_iommu_init()
|
| A D | e820.c | 838 unsigned long end_pfn; in e820__end_ram_pfn() local 845 end_pfn = (entry->addr + entry->size) >> PAGE_SHIFT; in e820__end_ram_pfn() 849 if (end_pfn > limit_pfn) { in e820__end_ram_pfn() 853 if (end_pfn > last_pfn) in e820__end_ram_pfn() 854 last_pfn = end_pfn; in e820__end_ram_pfn()
|
| /arch/x86/mm/pat/ |
| A D | memtype.c | 403 unsigned long end_pfn = (end + PAGE_SIZE - 1) >> PAGE_SHIFT; in pat_pagerange_is_ram() local 416 if (start_pfn < end_pfn) { in pat_pagerange_is_ram() 417 ret = walk_system_ram_range(start_pfn, end_pfn - start_pfn, in pat_pagerange_is_ram()
|