/linux-6.3-rc2/arch/arm64/mm/ |
A D | hugetlbpage.c | 63 case PUD_SIZE: in __hugetlb_valid_size() 130 case PUD_SIZE: in num_contig_ptes() 302 if (sz == PUD_SIZE) { in huge_pte_alloc() 350 if (sz != PUD_SIZE && pud_none(pud)) in huge_pte_offset() 380 case PUD_SIZE: in hugetlb_mask_last_page() 381 return PGDIR_SIZE - PUD_SIZE; in hugetlb_mask_last_page() 384 return PUD_SIZE - CONT_PMD_SIZE; in hugetlb_mask_last_page() 386 return PUD_SIZE - PMD_SIZE; in hugetlb_mask_last_page() 405 } else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) { in arch_make_huge_pte()
|
/linux-6.3-rc2/include/asm-generic/ |
A D | pgtable-nopud.h | 20 #define PUD_SIZE (1UL << PUD_SHIFT) macro 21 #define PUD_MASK (~(PUD_SIZE-1))
|
/linux-6.3-rc2/arch/powerpc/include/asm/nohash/64/ |
A D | pgtable-4k.h | 36 #define PUD_SIZE (1UL << PUD_SHIFT) macro 37 #define PUD_MASK (~(PUD_SIZE-1))
|
/linux-6.3-rc2/arch/riscv/mm/ |
A D | kasan_init.c | 132 if (pud_none(*pudp) && IS_ALIGNED(vaddr, PUD_SIZE) && (next - vaddr) >= PUD_SIZE) { in kasan_populate_pud() 138 phys_addr = memblock_phys_alloc(PUD_SIZE, PUD_SIZE); in kasan_populate_pud() 363 if (IS_ALIGNED(vaddr, PUD_SIZE) && (next - vaddr) >= PUD_SIZE) in kasan_shallow_populate_pud()
|
A D | hugetlbpage.c | 19 else if (IS_ENABLED(CONFIG_64BIT) && size == PUD_SIZE) in arch_hugetlb_valid_size()
|
/linux-6.3-rc2/arch/x86/include/asm/ |
A D | pgtable_64_types.h | 101 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro 102 #define PUD_MASK (~(PUD_SIZE - 1))
|
/linux-6.3-rc2/arch/s390/mm/ |
A D | kasan_init.c | 146 IS_ALIGNED(address, PUD_SIZE) && in kasan_early_pgtable_populate() 147 end - address >= PUD_SIZE) { in kasan_early_pgtable_populate() 150 address = (address + PUD_SIZE) & PUD_MASK; in kasan_early_pgtable_populate()
|
A D | hugetlbpage.c | 133 size = PUD_SIZE; in clear_huge_pte_skeys() 198 if (sz == PUD_SIZE) in huge_pte_alloc() 244 else if (MACHINE_HAS_EDAT2 && size == PUD_SIZE) in arch_hugetlb_valid_size()
|
A D | vmem.c | 293 const unsigned long end = start + PUD_SIZE; in try_free_pmd_table() 327 if (IS_ALIGNED(addr, PUD_SIZE) && in modify_pud_table() 328 IS_ALIGNED(next, PUD_SIZE)) { in modify_pud_table() 335 if (IS_ALIGNED(addr, PUD_SIZE) && in modify_pud_table() 336 IS_ALIGNED(next, PUD_SIZE) && in modify_pud_table()
|
/linux-6.3-rc2/arch/powerpc/mm/book3s64/ |
A D | radix_pgtable.c | 91 if (map_page_size == PUD_SIZE) { in early_map_kernel_page() 157 if (map_page_size == PUD_SIZE) { in __map_kernel_page() 320 if (IS_ALIGNED(addr, PUD_SIZE) && gap >= PUD_SIZE && in create_physical_mapping() 322 mapping_size = PUD_SIZE; in create_physical_mapping() 820 if (!IS_ALIGNED(addr, PUD_SIZE) || in remove_pud_table() 821 !IS_ALIGNED(next, PUD_SIZE)) { in remove_pud_table() 1131 flush_tlb_kernel_range(addr, addr + PUD_SIZE); in pud_free_pmd_page()
|
A D | radix_hugetlbpage.c | 38 if (end - start >= PUD_SIZE) in radix__flush_hugetlb_tlb_range()
|
/linux-6.3-rc2/arch/x86/mm/ |
A D | kasan_init_64.c | 83 ((end - addr) == PUD_SIZE) && in kasan_populate_pud() 84 IS_ALIGNED(addr, PUD_SIZE)) { in kasan_populate_pud() 85 p = early_alloc(PUD_SIZE, nid, false); in kasan_populate_pud() 88 memblock_free(p, PUD_SIZE); in kasan_populate_pud()
|
A D | init.c | 353 unsigned long start = round_down(mr[i].start, PUD_SIZE); in adjust_range_page_size_mask() 354 unsigned long end = round_up(mr[i].end, PUD_SIZE); in adjust_range_page_size_mask() 425 end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range() 438 start_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range() 439 end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
|
/linux-6.3-rc2/mm/kasan/ |
A D | init.c | 150 if (IS_ALIGNED(addr, PUD_SIZE) && end - addr >= PUD_SIZE) { in zero_pud_populate() 401 if (IS_ALIGNED(addr, PUD_SIZE) && in kasan_remove_pud_table() 402 IS_ALIGNED(next, PUD_SIZE)) { in kasan_remove_pud_table()
|
/linux-6.3-rc2/arch/arc/include/asm/ |
A D | pgtable-levels.h | 75 #define PUD_SIZE BIT(PUD_SHIFT) macro 76 #define PUD_MASK (~(PUD_SIZE - 1))
|
/linux-6.3-rc2/drivers/dax/ |
A D | device.c | 188 unsigned int fault_size = PUD_SIZE; in __dev_dax_pud_fault() 194 if (dev_dax->align > PUD_SIZE) { in __dev_dax_pud_fault() 207 (pud_addr + PUD_SIZE) > vmf->vma->vm_end) in __dev_dax_pud_fault() 211 phys = dax_pgoff_to_phys(dev_dax, pgoff, PUD_SIZE); in __dev_dax_pud_fault()
|
A D | dax-private.h | 94 if (align == PUD_SIZE && IS_ENABLED(CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD)) in dax_align_valid()
|
/linux-6.3-rc2/arch/x86/mm/pat/ |
A D | set_memory.c | 1261 if (start & (PUD_SIZE - 1)) { in unmap_pud_range() 1262 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in unmap_pud_range() 1274 while (end - start >= PUD_SIZE) { in unmap_pud_range() 1279 unmap_pmd_range(pud, start, start + PUD_SIZE); in unmap_pud_range() 1281 start += PUD_SIZE; in unmap_pud_range() 1424 if (start & (PUD_SIZE - 1)) { in populate_pud() 1426 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in populate_pud() 1459 while (boot_cpu_has(X86_FEATURE_GBPAGES) && end - start >= PUD_SIZE) { in populate_pud() 1463 start += PUD_SIZE; in populate_pud() 1464 cpa->pfn += PUD_SIZE >> PAGE_SHIFT; in populate_pud() [all …]
|
/linux-6.3-rc2/mm/ |
A D | page_table_check.c | 182 PUD_SIZE >> PAGE_SHIFT); in __page_table_check_pud_clear() 226 PUD_SIZE >> PAGE_SHIFT, in __page_table_check_pud_set()
|
/linux-6.3-rc2/arch/mips/include/asm/ |
A D | pgtable-64.h | 60 #define PUD_SIZE (1UL << PUD_SHIFT) macro 61 #define PUD_MASK (~(PUD_SIZE-1))
|
/linux-6.3-rc2/arch/s390/boot/ |
A D | vmem.c | 117 IS_ALIGNED(addr, PUD_SIZE) && (end - addr) >= PUD_SIZE; in can_large_pud()
|
/linux-6.3-rc2/arch/arm64/include/asm/ |
A D | pgtable-hwdef.h | 60 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro 61 #define PUD_MASK (~(PUD_SIZE-1))
|
/linux-6.3-rc2/arch/sparc/mm/ |
A D | hugetlbpage.c | 294 if (sz >= PUD_SIZE) in huge_pte_alloc() 341 if (size >= PUD_SIZE) in set_huge_pte_at() 378 if (size >= PUD_SIZE) in huge_ptep_get_and_clear()
|
/linux-6.3-rc2/arch/riscv/include/asm/ |
A D | pgtable-64.h | 38 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro 39 #define PUD_MASK (~(PUD_SIZE - 1))
|
/linux-6.3-rc2/arch/riscv/kvm/ |
A D | mmu.c | 588 WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE); in kvm_age_gfn() 606 WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE); in kvm_test_age_gfn() 648 if (vma_pagesize == PMD_SIZE || vma_pagesize == PUD_SIZE) in kvm_riscv_gstage_map() 653 if (vma_pagesize != PUD_SIZE && in kvm_riscv_gstage_map()
|