| /arch/powerpc/include/asm/nohash/32/ |
| A D | pte-8xx.h | 123 unsigned long clr, unsigned long set, int huge); 136 int huge = psize > mmu_virtual_psize ? 1 : 0; in __ptep_set_access_flags() local 138 pte_update(vma->vm_mm, address, ptep, clr, set, huge); in __ptep_set_access_flags() 177 if (!huge) in number_of_cells_per_pte() 188 unsigned long clr, unsigned long set, int huge) in __pte_update() argument 196 num = number_of_cells_per_pte(pmd, new, huge); in __pte_update() 211 unsigned long clr, unsigned long set, int huge) in pte_update() argument 215 if (huge && ptep_is_8m_pmdp(mm, addr, ptep)) { in pte_update() 218 old = __pte_update(mm, addr, pte_offset_kernel(pmdp, 0), clr, set, huge); in pte_update() 219 __pte_update(mm, addr, pte_offset_kernel(pmdp + 1, 0), clr, set, huge); in pte_update() [all …]
|
| /arch/powerpc/include/asm/book3s/64/ |
| A D | hash.h | 162 pte_t *ptep, unsigned long pte, int huge); 190 int huge) in hash__pte_update() argument 196 if (IS_ENABLED(CONFIG_PPC_4K_PAGES) && huge) { in hash__pte_update() 213 if (!huge) in hash__pte_update() 217 hpte_need_flush(mm, addr, ptep, old, huge); in hash__pte_update()
|
| A D | radix.h | 176 int huge) in radix__pte_update() argument 181 if (!huge) in radix__pte_update()
|
| A D | pgtable.h | 334 unsigned long set, int huge) in pte_update() argument 337 return radix__pte_update(mm, addr, ptep, clr, set, huge); in pte_update() 338 return hash__pte_update(mm, addr, ptep, clr, set, huge); in pte_update()
|
| /arch/powerpc/include/asm/nohash/ |
| A D | pgtable.h | 7 unsigned long clr, unsigned long set, int huge); 58 unsigned long clr, unsigned long set, int huge) in pte_update() argument 69 if (huge) in pte_update() 95 if (!huge) in pte_update() 143 int huge = psize > mmu_virtual_psize ? 1 : 0; in __ptep_set_access_flags() local 145 pte_update(vma->vm_mm, address, ptep, 0, set, huge); in __ptep_set_access_flags()
|
| /arch/loongarch/mm/ |
| A D | init.c | 125 int huge = pmd_val(pmdp_get(pmd)) & _PAGE_HUGE; in vmemmap_check_pmd() local 127 if (huge) in vmemmap_check_pmd() 130 return huge; in vmemmap_check_pmd()
|
| /arch/powerpc/mm/book3s64/ |
| A D | hash_tlb.c | 41 pte_t *ptep, unsigned long pte, int huge) in hpte_need_flush() argument 61 if (huge) { in hpte_need_flush()
|
| /arch/alpha/lib/ |
| A D | ev6-clear_user.S | 86 subq $1, 16, $4 # .. .. .. E : If < 16, we can not use the huge loop 87 and $16, 0x3f, $2 # .. .. E .. : Forward work for huge loop 88 subq $2, 0x40, $3 # .. E .. .. : bias counter (huge loop)
|
| /arch/parisc/mm/ |
| A D | init.c | 397 bool huge = false; in map_pages() local 407 huge = true; in map_pages() 412 huge = true; in map_pages() 418 if (huge) in map_pages()
|
| /arch/powerpc/include/asm/book3s/32/ |
| A D | pgtable.h | 260 unsigned long clr, unsigned long set, int huge) in pte_update() argument
|
| /arch/m68k/fpsp040/ |
| A D | kernel_ex.S | 27 huge: .long 0x7ffe0000,0xffffffff,0xffffffff label
|
| /arch/x86/kvm/svm/ |
| A D | sev.c | 3761 bool huge; in snp_begin_psc() local 3794 huge = entry_start.pagesize; in snp_begin_psc() 3795 npages = huge ? 512 : 1; in snp_begin_psc() 3812 huge = false; in snp_begin_psc() 3825 svm->sev_es.psc_2m = huge; in snp_begin_psc() 3839 entry.cur_page || !!entry.pagesize != huge) in snp_begin_psc() 3843 npages += huge ? 512 : 1; in snp_begin_psc()
|
| /arch/sparc/include/asm/ |
| A D | pgtable_64.h | 79 #error PMD_SHIFT must equal HPAGE_SHIFT for transparent huge pages.
|
| /arch/alpha/ |
| A D | Kconfig | 387 for systems that have huge holes in the physical address space.
|
| /arch/loongarch/ |
| A D | Kconfig | 713 or have huge holes in the physical address space for other reasons.
|
| /arch/x86/ |
| A D | Kconfig | 1474 helps to determine the effectiveness of preserving large and huge
|