| /linux/arch/riscv/mm/ |
| A D | kasan_init.c | 36 set_pmd(pmd, pfn_pmd(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pte() 44 set_pte(ptep, pfn_pte(PFN_DOWN(phys_addr), PAGE_KERNEL)); in kasan_populate_pte() 58 set_pud(pud, pfn_pud(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pmd() 70 set_pmd(pmdp, pfn_pmd(PFN_DOWN(phys_addr), PAGE_KERNEL)); in kasan_populate_pmd() 89 set_p4d(p4d, pfn_p4d(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pud() 120 set_pgd(pgd, pfn_pgd(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_p4d() 335 pfn_pmd(PFN_DOWN in kasan_early_init() 342 pfn_pud(PFN_DOWN in kasan_early_init() 350 pfn_p4d(PFN_DOWN in kasan_early_init() 468 pfn_pgd(PFN_DOWN(__pa(tmp_p4d)), PAGE_TABLE)); in create_tmp_mapping() [all …]
|
| A D | init.c | 80 max_zone_pfns[ZONE_DMA32] = PFN_DOWN(dma32_phys_limit); in zone_sizes_init() 168 bool swiotlb = max_pfn > PFN_DOWN(dma32_phys_limit); in mem_init() 281 max_low_pfn = max_pfn = PFN_DOWN(phys_ram_end); in setup_bootmem() 408 ptep[pte_idx] = pfn_pte(PFN_DOWN(pa), prot); in create_pte_mapping() 490 pmdp[pmd_idx] = pfn_pmd(PFN_DOWN(pa), prot); in create_pmd_mapping() 496 pmdp[pmd_idx] = pfn_pmd(PFN_DOWN(pte_phys), PAGE_TABLE); in create_pmd_mapping() 592 pudp[pud_index] = pfn_pud(PFN_DOWN(pa), prot); in create_pud_mapping() 598 pudp[pud_index] = pfn_pud(PFN_DOWN(next_phys), PAGE_TABLE); in create_pud_mapping() 618 p4dp[p4d_index] = pfn_p4d(PFN_DOWN(pa), prot); in create_p4d_mapping() 675 pgdp[pgd_idx] = pfn_pgd(PFN_DOWN(pa), prot); in create_pgd_mapping() [all …]
|
| /linux/arch/x86/mm/ |
| A D | init.c | 407 limit_pfn = PFN_DOWN(end); in split_mem_range() 410 pfn = start_pfn = PFN_DOWN(start); in split_mem_range() 419 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range() 421 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 423 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 433 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 435 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 437 end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range() 450 start_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range() 451 end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range() [all …]
|
| /linux/arch/x86/xen/ |
| A D | p2m.c | 382 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL)); in xen_rebuild_p2m_list() 392 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list() 451 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) in get_phys_to_machine() 581 if (p2m_pfn == PFN_DOWN(__pa(p2m_identity)) || in xen_alloc_p2m_entry() 582 p2m_pfn == PFN_DOWN(__pa(p2m_missing))) { in xen_alloc_p2m_entry() 590 if (p2m_pfn == PFN_DOWN(__pa(p2m_missing))) in xen_alloc_p2m_entry() 601 pfn_pte(PFN_DOWN(__pa(p2m)), PAGE_KERNEL)); in xen_alloc_p2m_entry() 664 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_missing))) in __set_phys_to_machine() 720 mfn = PFN_DOWN(map_ops[i].dev_bus_addr); in set_foreign_p2m_mapping() 820 pfn = PFN_DOWN(remap->paddr); in xen_do_remap_nonram() [all …]
|
| A D | enlighten_pvh.c | 78 ram_pages += PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr); in pvh_reserve_extra_memory() 96 PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr)); in pvh_reserve_extra_memory() 98 if (pages != (PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr))) { in pvh_reserve_extra_memory()
|
| A D | mmu_pv.c | 759 PFN_DOWN(__pa(user_pgd))); in __xen_pgd_pin() 868 PFN_DOWN(__pa(user_pgd))); in __xen_pgd_unpin() 1325 mfn = pfn_to_mfn(PFN_DOWN(cr3)); in __xen_write_cr3() 1682 if (*pt_base == PFN_DOWN(__pa(addr))) { in check_pt_base() 1687 if (*pt_end == PFN_DOWN(__pa(addr))) { in check_pt_base() 1789 PFN_DOWN(__pa_symbol(init_top_pgt))); in xen_setup_kernel_pagetable() 1798 PFN_DOWN(__pa_symbol(level3_user_vsyscall))); in xen_setup_kernel_pagetable() 1925 p2m_pfn = PFN_DOWN(pt_phys) + n_pt; in xen_relocate_p2m() 1951 PFN_DOWN(pt_phys)); in xen_relocate_p2m() 1959 PFN_DOWN(pmd_phys)); in xen_relocate_p2m() [all …]
|
| /linux/arch/x86/kernel/cpu/sgx/ |
| A D | encl.c | 86 entry = xa_load(&encl->page_array, PFN_DOWN(addr)); in reclaimer_writing_to_pcmd() 149 page_index = PFN_DOWN(encl->size); in __sgx_encl_eldu() 287 entry = xa_load(&encl->page_array, PFN_DOWN(addr)); in sgx_encl_load_page_in_vma() 307 entry = xa_load(&encl->page_array, PFN_DOWN(addr)); in sgx_encl_load_page() 408 vmret = vmf_insert_pfn(vma, addr, PFN_DOWN(phys_addr)); in sgx_encl_eaug_page() 417 xa_erase(&encl->page_array, PFN_DOWN(encl_page->desc)); in sgx_encl_eaug_page() 457 (!xa_load(&encl->page_array, PFN_DOWN(addr)))) in sgx_vma_fault() 474 ret = vmf_insert_pfn(vma, addr, PFN_DOWN(phys_addr)); in sgx_vma_fault() 530 XA_STATE(xas, &encl->page_array, PFN_DOWN(start)); in sgx_encl_may_map() 546 xas_for_each(&xas, page, PFN_DOWN(end - 1)) { in sgx_encl_may_map() [all …]
|
| /linux/arch/csky/kernel/ |
| A D | setup.c | 17 unsigned long lowmem_size = PFN_DOWN(LOWMEM_LIMIT - PHYS_OFFSET_OFFSET); in csky_memblock_init() 18 unsigned long sseg_size = PFN_DOWN(SSEG_SIZE - PHYS_OFFSET_OFFSET); in csky_memblock_init() 30 max_low_pfn = max_pfn = PFN_DOWN(memblock_end_of_DRAM()); in csky_memblock_init()
|
| /linux/arch/csky/include/asm/ |
| A D | page.h | 47 #define phys_to_page(paddr) (pfn_to_page(PFN_DOWN(paddr))) 73 #define ARCH_PFN_OFFSET PFN_DOWN(va_pa_offset + PHYS_OFFSET_OFFSET) 85 #define MAP_NR(x) PFN_DOWN((unsigned long)(x) - PAGE_OFFSET - \
|
| /linux/arch/sparc/power/ |
| A D | hibernate.c | 25 unsigned long nosave_begin_pfn = PFN_DOWN((unsigned long)&__nosave_begin); in pfn_is_nosave() 26 unsigned long nosave_end_pfn = PFN_DOWN((unsigned long)&__nosave_end); in pfn_is_nosave()
|
| /linux/arch/arc/mm/ |
| A D | init.c | 94 max_low_pfn = max_pfn = PFN_DOWN(low_mem_start + low_mem_sz); in setup_arch_memory() 139 min_high_pfn = PFN_DOWN(high_mem_start); in setup_arch_memory() 140 max_high_pfn = PFN_DOWN(high_mem_start + high_mem_sz); in setup_arch_memory()
|
| /linux/arch/sh/mm/ |
| A D | numa.c | 30 start_pfn = PFN_DOWN(start); in setup_bootmem_node() 31 end_pfn = PFN_DOWN(end); in setup_bootmem_node()
|
| /linux/arch/m68k/mm/ |
| A D | mcfmmu.c | 82 max_zone_pfn[ZONE_DMA] = PFN_DOWN(_ramend); in paging_init() 171 num_pages = PFN_DOWN(_ramend - _rambase); in cf_bootmem_alloc() 175 min_low_pfn = PFN_DOWN(_rambase); in cf_bootmem_alloc() 176 max_pfn = max_low_pfn = PFN_DOWN(_ramend); in cf_bootmem_alloc()
|
| /linux/arch/hexagon/mm/ |
| A D | init.c | 131 bootmem_lastpg = PFN_DOWN(size); in early_mem() 156 bootmem_lastpg = PFN_DOWN((bootmem_lastpg << PAGE_SHIFT) & in setup_arch_memory() 169 max_low_pfn = bootmem_lastpg - PFN_DOWN(DMA_RESERVED_BYTES); in setup_arch_memory()
|
| /linux/arch/nios2/include/asm/ |
| A D | page.h | 89 # define virt_to_page(vaddr) pfn_to_page(PFN_DOWN(virt_to_phys(vaddr))) 90 # define virt_addr_valid(vaddr) pfn_valid(PFN_DOWN(virt_to_phys(vaddr)))
|
| /linux/arch/mips/kernel/ |
| A D | setup.c | 305 max_pfn = PFN_DOWN(ramend); in bootmem_init() 313 if (start >= PFN_DOWN(HIGHMEM_START)) in bootmem_init() 315 if (end > PFN_DOWN(HIGHMEM_START)) in bootmem_init() 316 end = PFN_DOWN(HIGHMEM_START); in bootmem_init() 324 if (max_pfn > PFN_DOWN(HIGHMEM_START)) { in bootmem_init() 325 max_low_pfn = PFN_DOWN(HIGHMEM_START); in bootmem_init()
|
| /linux/drivers/xen/ |
| A D | swiotlb-xen.c | 108 if (pfn_valid(PFN_DOWN(paddr))) in xen_swiotlb_find_pool() 245 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dev_addr)))) in xen_swiotlb_map_page() 270 if (pfn_valid(PFN_DOWN(dma_to_phys(hwdev, dev_addr)))) in xen_swiotlb_unmap_page() 291 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr)))) in xen_swiotlb_sync_single_for_cpu() 314 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr)))) in xen_swiotlb_sync_single_for_device()
|
| /linux/arch/mips/mm/ |
| A D | ioremap.c | 83 pfn = PFN_DOWN(phys_addr); in ioremap_prot() 84 last_pfn = PFN_DOWN(last_addr); in ioremap_prot()
|
| /linux/arch/nios2/kernel/ |
| A D | setup.c | 140 *max_low = PFN_DOWN(memblock_get_current_limit()); in find_limits() 142 *max_high = PFN_DOWN(memblock_end_of_DRAM()); in find_limits()
|
| /linux/arch/arm64/mm/ |
| A D | init.c | 144 max_zone_pfns[ZONE_DMA] = PFN_DOWN(arm64_dma_phys_limit); in zone_sizes_init() 147 max_zone_pfns[ZONE_DMA32] = PFN_DOWN(dma32_phys_limit); in zone_sizes_init() 321 max = PFN_DOWN(memblock_end_of_DRAM()); in bootmem_init() 369 bool swiotlb = max_pfn > PFN_DOWN(arm64_dma_phys_limit); in mem_init()
|
| /linux/arch/x86/include/asm/xen/ |
| A D | page.h | 233 return XMADDR(PFN_PHYS(pfn_to_mfn(PFN_DOWN(phys.paddr))) | offset); in phys_to_machine() 239 return XPADDR(PFN_PHYS(mfn_to_pfn(PFN_DOWN(machine.maddr))) | offset); in machine_to_phys() 300 return PFN_DOWN(__pa(v)); in virt_to_pfn()
|
| /linux/arch/m68k/kernel/ |
| A D | setup_no.c | 156 min_low_pfn = PFN_DOWN(memory_start); in setup_arch() 157 max_pfn = max_low_pfn = PFN_DOWN(memory_end); in setup_arch()
|
| /linux/kernel/dma/ |
| A D | direct.c | 601 for (m = dev->dma_range_map; PFN_DOWN(m->size); m++) { in check_ram_in_range_map() 602 unsigned long cpu_start_pfn = PFN_DOWN(m->cpu_start); in check_ram_in_range_map() 605 start_pfn - cpu_start_pfn < PFN_DOWN(m->size)) { in check_ram_in_range_map() 613 start_pfn = PFN_DOWN(bdr->cpu_start) + PFN_DOWN(bdr->size); in check_ram_in_range_map() 623 return !walk_system_ram_range(0, PFN_DOWN(ULONG_MAX) + 1, dev, in dma_direct_all_ram_mapped()
|
| /linux/arch/x86/kernel/ |
| A D | dumpstack_64.c | 72 [PFN_DOWN(CEA_ESTACK_OFFS(st)) ... \ 73 PFN_DOWN(CEA_ESTACK_OFFS(st) + CEA_ESTACK_SIZE(st) - 1)] = { \
|
| /linux/arch/powerpc/mm/ |
| A D | pgtable_32.c | 136 PFN_DOWN((unsigned long)_sinittext); in __mark_initmem_nx() 175 PFN_DOWN((unsigned long)_stext); in __mark_rodata_ro()
|