/linux-6.3-rc2/mm/ |
A D | pgtable-generic.c | 114 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_set_access_flags() 128 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_clear_flush_young() 142 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_huge_clear_flush() 199 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_invalidate() 227 flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_collapse_flush()
|
A D | huge_memory.c | 389 return sysfs_emit(buf, "%lu\n", HPAGE_PMD_SIZE); in hpage_pmd_size_show() 1603 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in madvise_free_huge_pmd() 1634 if (next - addr != HPAGE_PMD_SIZE) { in madvise_free_huge_pmd() 1679 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in zap_huge_pmd() 1730 tlb_remove_page_size(tlb, page, HPAGE_PMD_SIZE); in zap_huge_pmd() 1828 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in change_huge_pmd() 1935 tlb_flush_pmd_range(tlb, addr, HPAGE_PMD_SIZE); in change_huge_pmd() 2083 VM_BUG_ON_VMA(vma->vm_end < haddr + HPAGE_PMD_SIZE, vma); in __split_huge_pmd_locked() 2280 (address & HPAGE_PMD_MASK) + HPAGE_PMD_SIZE); in __split_huge_pmd() 2337 if (!IS_ALIGNED(address, HPAGE_PMD_SIZE) && in split_huge_pmd_if_needed() [all …]
|
A D | khugepaged.c | 1055 address + HPAGE_PMD_SIZE); in collapse_huge_page() 1427 addr + HPAGE_PMD_SIZE); in collapse_and_free_pmd() 1469 !range_in_vma(vma, haddr, haddr + HPAGE_PMD_SIZE)) in collapse_pte_mapped_thp() 1670 vma->vm_end < addr + HPAGE_PMD_SIZE) { in retract_page_tables() 2315 hstart = round_up(vma->vm_start, HPAGE_PMD_SIZE); in khugepaged_scan_mm_slot() 2316 hend = round_down(vma->vm_end, HPAGE_PMD_SIZE); in khugepaged_scan_mm_slot() 2331 khugepaged_scan.address + HPAGE_PMD_SIZE > in khugepaged_scan_mm_slot() 2371 khugepaged_scan.address += HPAGE_PMD_SIZE; in khugepaged_scan_mm_slot() 2668 for (addr = hstart; addr < hend; addr += HPAGE_PMD_SIZE) { in madvise_collapse()
|
A D | shmem.c | 574 round_up(inode->i_size, HPAGE_PMD_SIZE)) { in shmem_unused_huge_shrink() 1071 stat->blksize = HPAGE_PMD_SIZE; in shmem_getattr() 2190 if (len < HPAGE_PMD_SIZE) in shmem_get_unmapped_area() 2222 offset = (pgoff << PAGE_SHIFT) & (HPAGE_PMD_SIZE-1); in shmem_get_unmapped_area() 2223 if (offset && offset + len < 2 * HPAGE_PMD_SIZE) in shmem_get_unmapped_area() 2225 if ((addr & (HPAGE_PMD_SIZE-1)) == offset) in shmem_get_unmapped_area() 2228 inflated_len = len + HPAGE_PMD_SIZE - PAGE_SIZE; in shmem_get_unmapped_area() 2240 inflated_offset = inflated_addr & (HPAGE_PMD_SIZE-1); in shmem_get_unmapped_area() 2243 inflated_addr += HPAGE_PMD_SIZE; in shmem_get_unmapped_area()
|
A D | madvise.c | 360 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in madvise_cold_or_pageout_pte_range() 384 if (next - addr != HPAGE_PMD_SIZE) { in madvise_cold_or_pageout_pte_range()
|
A D | mprotect.c | 406 if ((next - addr != HPAGE_PMD_SIZE) || in change_pmd_range()
|
A D | mremap.c | 542 if (extent == HPAGE_PMD_SIZE && in move_page_tables()
|
A D | memory.c | 1121 VM_BUG_ON_VMA(next-addr != HPAGE_PMD_SIZE, src_vma); in copy_pmd_range() 1510 if (next - addr != HPAGE_PMD_SIZE) in zap_pmd_range() 1517 next - addr == HPAGE_PMD_SIZE && pmd_none(*pmd)) { in zap_pmd_range()
|
A D | rmap.c | 978 address + HPAGE_PMD_SIZE); in page_vma_mkclean_one()
|
/linux-6.3-rc2/include/linux/ |
A D | huge_mm.h | 110 #define HPAGE_PMD_SIZE ((1UL) << HPAGE_PMD_SHIFT) macro 111 #define HPAGE_PMD_MASK (~(HPAGE_PMD_SIZE - 1)) 152 if (haddr < vma->vm_start || haddr + HPAGE_PMD_SIZE > vma->vm_end) in transhuge_vma_suitable() 299 #define HPAGE_PMD_SIZE ({ BUILD_BUG(); 0; }) macro
|
A D | mmu_notifier.h | 606 ___haddr + HPAGE_PMD_SIZE); \
|
/linux-6.3-rc2/mm/damon/ |
A D | ops-common.c | 80 if (mmu_notifier_clear_young(mm, addr, addr + HPAGE_PMD_SIZE)) in damon_pmdp_mkold()
|
A D | vaddr.c | 457 *priv->folio_sz = HPAGE_PMD_SIZE; in damon_young_pmd_entry()
|
/linux-6.3-rc2/arch/powerpc/mm/book3s64/ |
A D | pgtable.c | 128 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_invalidate() 146 flush_pmd_tlb_range(vma, addr, addr + HPAGE_PMD_SIZE); in pmdp_huge_get_and_clear_full()
|
A D | radix_tlb.c | 1417 end = addr + HPAGE_PMD_SIZE; in radix__flush_tlb_collapsed_pmd()
|
A D | hash_utils.c | 1973 max_hpte_count = HPAGE_PMD_SIZE >> shift; in flush_hash_hugepage()
|
/linux-6.3-rc2/arch/arc/mm/ |
A D | tlb.c | 675 mmu->s_pg_sz_m != TO_MB(HPAGE_PMD_SIZE)) in arc_mmu_init() 677 (unsigned long)TO_MB(HPAGE_PMD_SIZE)); in arc_mmu_init()
|
/linux-6.3-rc2/arch/sparc/mm/ |
A D | tlb.c | 250 flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_invalidate()
|
/linux-6.3-rc2/include/asm-generic/ |
A D | tlb.h | 625 tlb_flush_pmd_range(tlb, address, HPAGE_PMD_SIZE); \
|
/linux-6.3-rc2/drivers/nvdimm/ |
A D | pfn_devs.c | 102 alignments[1] = HPAGE_PMD_SIZE; in nd_pfn_supported_alignments() 117 return HPAGE_PMD_SIZE; in nd_pfn_default_alignment()
|
/linux-6.3-rc2/fs/proc/ |
A D | task_mmu.c | 602 mss->anonymous_thp += HPAGE_PMD_SIZE; in smaps_pmd_entry() 604 mss->shmem_thp += HPAGE_PMD_SIZE; in smaps_pmd_entry() 608 mss->file_thp += HPAGE_PMD_SIZE; in smaps_pmd_entry() 1882 HPAGE_PMD_SIZE/PAGE_SIZE); in gather_pte_stats()
|
/linux-6.3-rc2/arch/x86/mm/ |
A D | pgtable.c | 614 flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_clear_flush_young()
|