Home
last modified time | relevance | path

Searched refs:PFN_DOWN (Results 1 – 25 of 119) sorted by relevance

12345

/linux/arch/riscv/mm/
A Dkasan_init.c36 set_pmd(pmd, pfn_pmd(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pte()
44 set_pte(ptep, pfn_pte(PFN_DOWN(phys_addr), PAGE_KERNEL)); in kasan_populate_pte()
58 set_pud(pud, pfn_pud(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pmd()
70 set_pmd(pmdp, pfn_pmd(PFN_DOWN(phys_addr), PAGE_KERNEL)); in kasan_populate_pmd()
89 set_p4d(p4d, pfn_p4d(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pud()
120 set_pgd(pgd, pfn_pgd(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_p4d()
335 pfn_pmd(PFN_DOWN in kasan_early_init()
342 pfn_pud(PFN_DOWN in kasan_early_init()
350 pfn_p4d(PFN_DOWN in kasan_early_init()
468 pfn_pgd(PFN_DOWN(__pa(tmp_p4d)), PAGE_TABLE)); in create_tmp_mapping()
[all …]
A Dinit.c80 max_zone_pfns[ZONE_DMA32] = PFN_DOWN(dma32_phys_limit); in zone_sizes_init()
168 bool swiotlb = max_pfn > PFN_DOWN(dma32_phys_limit); in mem_init()
281 max_low_pfn = max_pfn = PFN_DOWN(phys_ram_end); in setup_bootmem()
408 ptep[pte_idx] = pfn_pte(PFN_DOWN(pa), prot); in create_pte_mapping()
490 pmdp[pmd_idx] = pfn_pmd(PFN_DOWN(pa), prot); in create_pmd_mapping()
496 pmdp[pmd_idx] = pfn_pmd(PFN_DOWN(pte_phys), PAGE_TABLE); in create_pmd_mapping()
592 pudp[pud_index] = pfn_pud(PFN_DOWN(pa), prot); in create_pud_mapping()
598 pudp[pud_index] = pfn_pud(PFN_DOWN(next_phys), PAGE_TABLE); in create_pud_mapping()
618 p4dp[p4d_index] = pfn_p4d(PFN_DOWN(pa), prot); in create_p4d_mapping()
675 pgdp[pgd_idx] = pfn_pgd(PFN_DOWN(pa), prot); in create_pgd_mapping()
[all …]
/linux/arch/x86/mm/
A Dinit.c407 limit_pfn = PFN_DOWN(end); in split_mem_range()
410 pfn = start_pfn = PFN_DOWN(start); in split_mem_range()
419 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range()
421 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
423 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
433 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
435 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
437 end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
450 start_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
451 end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
[all …]
/linux/arch/x86/xen/
A Dp2m.c382 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL)); in xen_rebuild_p2m_list()
392 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
451 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) in get_phys_to_machine()
581 if (p2m_pfn == PFN_DOWN(__pa(p2m_identity)) || in xen_alloc_p2m_entry()
582 p2m_pfn == PFN_DOWN(__pa(p2m_missing))) { in xen_alloc_p2m_entry()
590 if (p2m_pfn == PFN_DOWN(__pa(p2m_missing))) in xen_alloc_p2m_entry()
601 pfn_pte(PFN_DOWN(__pa(p2m)), PAGE_KERNEL)); in xen_alloc_p2m_entry()
664 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_missing))) in __set_phys_to_machine()
720 mfn = PFN_DOWN(map_ops[i].dev_bus_addr); in set_foreign_p2m_mapping()
820 pfn = PFN_DOWN(remap->paddr); in xen_do_remap_nonram()
[all …]
A Denlighten_pvh.c78 ram_pages += PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr); in pvh_reserve_extra_memory()
96 PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr)); in pvh_reserve_extra_memory()
98 if (pages != (PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr))) { in pvh_reserve_extra_memory()
A Dmmu_pv.c759 PFN_DOWN(__pa(user_pgd))); in __xen_pgd_pin()
868 PFN_DOWN(__pa(user_pgd))); in __xen_pgd_unpin()
1325 mfn = pfn_to_mfn(PFN_DOWN(cr3)); in __xen_write_cr3()
1682 if (*pt_base == PFN_DOWN(__pa(addr))) { in check_pt_base()
1687 if (*pt_end == PFN_DOWN(__pa(addr))) { in check_pt_base()
1789 PFN_DOWN(__pa_symbol(init_top_pgt))); in xen_setup_kernel_pagetable()
1798 PFN_DOWN(__pa_symbol(level3_user_vsyscall))); in xen_setup_kernel_pagetable()
1925 p2m_pfn = PFN_DOWN(pt_phys) + n_pt; in xen_relocate_p2m()
1951 PFN_DOWN(pt_phys)); in xen_relocate_p2m()
1959 PFN_DOWN(pmd_phys)); in xen_relocate_p2m()
[all …]
/linux/arch/x86/kernel/cpu/sgx/
A Dencl.c86 entry = xa_load(&encl->page_array, PFN_DOWN(addr)); in reclaimer_writing_to_pcmd()
149 page_index = PFN_DOWN(encl->size); in __sgx_encl_eldu()
287 entry = xa_load(&encl->page_array, PFN_DOWN(addr)); in sgx_encl_load_page_in_vma()
307 entry = xa_load(&encl->page_array, PFN_DOWN(addr)); in sgx_encl_load_page()
408 vmret = vmf_insert_pfn(vma, addr, PFN_DOWN(phys_addr)); in sgx_encl_eaug_page()
417 xa_erase(&encl->page_array, PFN_DOWN(encl_page->desc)); in sgx_encl_eaug_page()
457 (!xa_load(&encl->page_array, PFN_DOWN(addr)))) in sgx_vma_fault()
474 ret = vmf_insert_pfn(vma, addr, PFN_DOWN(phys_addr)); in sgx_vma_fault()
530 XA_STATE(xas, &encl->page_array, PFN_DOWN(start)); in sgx_encl_may_map()
546 xas_for_each(&xas, page, PFN_DOWN(end - 1)) { in sgx_encl_may_map()
[all …]
/linux/arch/csky/kernel/
A Dsetup.c17 unsigned long lowmem_size = PFN_DOWN(LOWMEM_LIMIT - PHYS_OFFSET_OFFSET); in csky_memblock_init()
18 unsigned long sseg_size = PFN_DOWN(SSEG_SIZE - PHYS_OFFSET_OFFSET); in csky_memblock_init()
30 max_low_pfn = max_pfn = PFN_DOWN(memblock_end_of_DRAM()); in csky_memblock_init()
/linux/arch/csky/include/asm/
A Dpage.h47 #define phys_to_page(paddr) (pfn_to_page(PFN_DOWN(paddr)))
73 #define ARCH_PFN_OFFSET PFN_DOWN(va_pa_offset + PHYS_OFFSET_OFFSET)
85 #define MAP_NR(x) PFN_DOWN((unsigned long)(x) - PAGE_OFFSET - \
/linux/arch/sparc/power/
A Dhibernate.c25 unsigned long nosave_begin_pfn = PFN_DOWN((unsigned long)&__nosave_begin); in pfn_is_nosave()
26 unsigned long nosave_end_pfn = PFN_DOWN((unsigned long)&__nosave_end); in pfn_is_nosave()
/linux/arch/arc/mm/
A Dinit.c94 max_low_pfn = max_pfn = PFN_DOWN(low_mem_start + low_mem_sz); in setup_arch_memory()
139 min_high_pfn = PFN_DOWN(high_mem_start); in setup_arch_memory()
140 max_high_pfn = PFN_DOWN(high_mem_start + high_mem_sz); in setup_arch_memory()
/linux/arch/sh/mm/
A Dnuma.c30 start_pfn = PFN_DOWN(start); in setup_bootmem_node()
31 end_pfn = PFN_DOWN(end); in setup_bootmem_node()
/linux/arch/m68k/mm/
A Dmcfmmu.c82 max_zone_pfn[ZONE_DMA] = PFN_DOWN(_ramend); in paging_init()
171 num_pages = PFN_DOWN(_ramend - _rambase); in cf_bootmem_alloc()
175 min_low_pfn = PFN_DOWN(_rambase); in cf_bootmem_alloc()
176 max_pfn = max_low_pfn = PFN_DOWN(_ramend); in cf_bootmem_alloc()
/linux/arch/hexagon/mm/
A Dinit.c131 bootmem_lastpg = PFN_DOWN(size); in early_mem()
156 bootmem_lastpg = PFN_DOWN((bootmem_lastpg << PAGE_SHIFT) & in setup_arch_memory()
169 max_low_pfn = bootmem_lastpg - PFN_DOWN(DMA_RESERVED_BYTES); in setup_arch_memory()
/linux/arch/nios2/include/asm/
A Dpage.h89 # define virt_to_page(vaddr) pfn_to_page(PFN_DOWN(virt_to_phys(vaddr)))
90 # define virt_addr_valid(vaddr) pfn_valid(PFN_DOWN(virt_to_phys(vaddr)))
/linux/arch/mips/kernel/
A Dsetup.c305 max_pfn = PFN_DOWN(ramend); in bootmem_init()
313 if (start >= PFN_DOWN(HIGHMEM_START)) in bootmem_init()
315 if (end > PFN_DOWN(HIGHMEM_START)) in bootmem_init()
316 end = PFN_DOWN(HIGHMEM_START); in bootmem_init()
324 if (max_pfn > PFN_DOWN(HIGHMEM_START)) { in bootmem_init()
325 max_low_pfn = PFN_DOWN(HIGHMEM_START); in bootmem_init()
/linux/drivers/xen/
A Dswiotlb-xen.c108 if (pfn_valid(PFN_DOWN(paddr))) in xen_swiotlb_find_pool()
245 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dev_addr)))) in xen_swiotlb_map_page()
270 if (pfn_valid(PFN_DOWN(dma_to_phys(hwdev, dev_addr)))) in xen_swiotlb_unmap_page()
291 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr)))) in xen_swiotlb_sync_single_for_cpu()
314 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr)))) in xen_swiotlb_sync_single_for_device()
/linux/arch/mips/mm/
A Dioremap.c83 pfn = PFN_DOWN(phys_addr); in ioremap_prot()
84 last_pfn = PFN_DOWN(last_addr); in ioremap_prot()
/linux/arch/nios2/kernel/
A Dsetup.c140 *max_low = PFN_DOWN(memblock_get_current_limit()); in find_limits()
142 *max_high = PFN_DOWN(memblock_end_of_DRAM()); in find_limits()
/linux/arch/arm64/mm/
A Dinit.c144 max_zone_pfns[ZONE_DMA] = PFN_DOWN(arm64_dma_phys_limit); in zone_sizes_init()
147 max_zone_pfns[ZONE_DMA32] = PFN_DOWN(dma32_phys_limit); in zone_sizes_init()
321 max = PFN_DOWN(memblock_end_of_DRAM()); in bootmem_init()
369 bool swiotlb = max_pfn > PFN_DOWN(arm64_dma_phys_limit); in mem_init()
/linux/arch/x86/include/asm/xen/
A Dpage.h233 return XMADDR(PFN_PHYS(pfn_to_mfn(PFN_DOWN(phys.paddr))) | offset); in phys_to_machine()
239 return XPADDR(PFN_PHYS(mfn_to_pfn(PFN_DOWN(machine.maddr))) | offset); in machine_to_phys()
300 return PFN_DOWN(__pa(v)); in virt_to_pfn()
/linux/arch/m68k/kernel/
A Dsetup_no.c156 min_low_pfn = PFN_DOWN(memory_start); in setup_arch()
157 max_pfn = max_low_pfn = PFN_DOWN(memory_end); in setup_arch()
/linux/kernel/dma/
A Ddirect.c601 for (m = dev->dma_range_map; PFN_DOWN(m->size); m++) { in check_ram_in_range_map()
602 unsigned long cpu_start_pfn = PFN_DOWN(m->cpu_start); in check_ram_in_range_map()
605 start_pfn - cpu_start_pfn < PFN_DOWN(m->size)) { in check_ram_in_range_map()
613 start_pfn = PFN_DOWN(bdr->cpu_start) + PFN_DOWN(bdr->size); in check_ram_in_range_map()
623 return !walk_system_ram_range(0, PFN_DOWN(ULONG_MAX) + 1, dev, in dma_direct_all_ram_mapped()
/linux/arch/x86/kernel/
A Ddumpstack_64.c72 [PFN_DOWN(CEA_ESTACK_OFFS(st)) ... \
73 PFN_DOWN(CEA_ESTACK_OFFS(st) + CEA_ESTACK_SIZE(st) - 1)] = { \
/linux/arch/powerpc/mm/
A Dpgtable_32.c136 PFN_DOWN((unsigned long)_sinittext); in __mark_initmem_nx()
175 PFN_DOWN((unsigned long)_stext); in __mark_rodata_ro()

Completed in 56 milliseconds

12345