Lines Matching refs:PMD_SIZE
102 if (map_page_size == PMD_SIZE) { in early_map_kernel_page()
165 if (map_page_size == PMD_SIZE) { in __map_kernel_page()
329 } else if (IS_ALIGNED(addr, PMD_SIZE) && gap >= PMD_SIZE && in create_physical_mapping()
331 mapping_size = PMD_SIZE; in create_physical_mapping()
740 unsigned long start = ALIGN_DOWN(addr, PMD_SIZE); in vmemmap_pmd_is_unused()
742 return !vmemmap_populated(start, PMD_SIZE); in vmemmap_pmd_is_unused()
835 if (IS_ALIGNED(addr, PMD_SIZE) && in remove_pmd_table()
836 IS_ALIGNED(next, PMD_SIZE)) { in remove_pmd_table()
838 free_vmemmap_pages(pmd_page(*pmd), altmap, get_order(PMD_SIZE)); in remove_pmd_table()
844 free_vmemmap_pages(pmd_page(*pmd), altmap, get_order(PMD_SIZE)); in remove_pmd_table()
1006 VM_BUG_ON(!IS_ALIGNED(addr, PMD_SIZE)); in vmemmap_set_pmd()
1142 start = ALIGN_DOWN(start, PMD_SIZE); in radix__vmemmap_populate()
1169 if (altmap && (!IS_ALIGNED(addr, PMD_SIZE) || in radix__vmemmap_populate()
1170 altmap_cross_boundary(altmap, addr, PMD_SIZE))) { in radix__vmemmap_populate()
1178 p = vmemmap_alloc_block_buf(PMD_SIZE, node, altmap); in radix__vmemmap_populate()
1337 addr_pfn += (PMD_SIZE >> PAGE_SHIFT); in vmemmap_populate_compound_pages()
1689 flush_tlb_kernel_range(addr, addr + PMD_SIZE); in pmd_free_pte_page()