/linux-6.3-rc2/arch/x86/include/asm/ |
A D | kfence.h | 33 if (level != PG_LEVEL_4K) in arch_kfence_init_pool() 46 if (WARN_ON(!pte || level != PG_LEVEL_4K)) in kfence_protect_page()
|
A D | pgtable_types.h | 521 PG_LEVEL_4K, enumerator
|
/linux-6.3-rc2/arch/x86/kvm/mmu/ |
A D | page_track.c | 93 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K); in update_gfn_track() 136 if (kvm_mmu_slot_gfn_write_protect(kvm, slot, gfn, PG_LEVEL_4K)) in kvm_slot_page_track_add_page() 194 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K); in kvm_slot_page_track_is_active()
|
A D | spte.c | 176 if (level > PG_LEVEL_4K && (pte_access & ACC_EXEC_MASK) && in make_spte() 189 if (level > PG_LEVEL_4K) in make_spte() 245 WARN_ON(level > PG_LEVEL_4K); in make_spte() 296 if (role.level == PG_LEVEL_4K) { in make_huge_page_split_spte()
|
A D | paging_tmpl.h | 74 #define gpte_to_gfn(pte) gpte_to_gfn_lvl((pte), PG_LEVEL_4K) 163 if (FNAME(is_rsvd_bits_set)(vcpu->arch.mmu, gpte, PG_LEVEL_4K)) in FNAME() 295 gpte |= level - PG_LEVEL_4K - 1; in FNAME() 429 if (walker->level > PG_LEVEL_4K && is_cpuid_PSE36()) in FNAME() 560 if (level == PG_LEVEL_4K) { in FNAME() 585 if (sp->role.level > PG_LEVEL_4K) in FNAME() 834 fault->max_level = PG_LEVEL_4K; in FNAME() 882 WARN_ON(sp->role.level != PG_LEVEL_4K); in FNAME()
|
A D | mmu.c | 821 if (sp->role.level > PG_LEVEL_4K) in account_shadowed() 868 if (sp->role.level > PG_LEVEL_4K) in unaccount_shadowed() 1315 PG_LEVEL_4K, slot); in kvm_mmu_write_protect_pt_masked() 1348 PG_LEVEL_4K, slot); in kvm_mmu_clear_dirty_pt_masked() 1973 if (level == PG_LEVEL_4K) in mmu_pages_next() 1995 WARN_ON(level == PG_LEVEL_4K); in mmu_pages_first() 2984 if (sp->role.level > PG_LEVEL_4K) in direct_pte_prefetch() 3025 int level = PG_LEVEL_4K; in host_pfn_mapping_level() 3100 if (max_level == PG_LEVEL_4K) in kvm_mmu_max_mapping_level() 3101 return PG_LEVEL_4K; in kvm_mmu_max_mapping_level() [all …]
|
A D | mmu_internal.h | 292 .req_level = PG_LEVEL_4K, in kvm_mmu_do_page_fault() 293 .goal_level = PG_LEVEL_4K, in kvm_mmu_do_page_fault()
|
A D | tdp_iter.h | 109 for_each_tdp_pte_min_level(iter, root, PG_LEVEL_4K, start, end)
|
A D | tdp_mmu.c | 354 if (level > PG_LEVEL_4K) in handle_changed_spte_dirty_log() 533 WARN_ON(level < PG_LEVEL_4K); in __handle_changed_spte() 950 for_each_tdp_pte_min_level(iter, root, PG_LEVEL_4K, start, end) { in tdp_mmu_zap_leafs() 1319 if (iter->level != PG_LEVEL_4K || in set_spte_gfn() 1688 if (iter.level > PG_LEVEL_4K || in clear_dirty_pt_masked()
|
A D | spte.h | 307 return (level == PG_LEVEL_4K) || is_large_pte(pte); in is_last_spte()
|
/linux-6.3-rc2/arch/x86/xen/ |
A D | p2m.c | 245 BUG_ON(!ptep || level != PG_LEVEL_4K); in xen_build_mfn_list_list() 443 BUG_ON(!ptep || level != PG_LEVEL_4K); in get_phys_to_machine() 537 BUG_ON(!ptep || level != PG_LEVEL_4K); in xen_alloc_p2m_entry() 662 BUG_ON(!ptep || level != PG_LEVEL_4K); in __set_phys_to_machine()
|
/linux-6.3-rc2/arch/x86/mm/ |
A D | mem_encrypt_amd.c | 258 case PG_LEVEL_4K: in pg_level_to_pfn() 423 if (level == PG_LEVEL_4K) { in early_set_memory_enc_dec()
|
A D | pti.c | 283 if (!pte || WARN_ON(level != PG_LEVEL_4K) || pte_none(*pte)) in pti_setup_vsyscall()
|
A D | kmmio.c | 174 case PG_LEVEL_4K: in clear_page_presence()
|
A D | init_32.c | 376 update_page_count(PG_LEVEL_4K, pages_4k); in kernel_physical_mapping_init()
|
A D | init_64.c | 497 update_page_count(PG_LEVEL_4K, pages); in phys_pte_init() 1099 update_page_count(PG_LEVEL_4K, -pages); in remove_pte_table()
|
/linux-6.3-rc2/tools/testing/selftests/kvm/lib/x86_64/ |
A D | vmx.c | 429 for (int level = PG_LEVEL_512G; level >= PG_LEVEL_4K; level--) { in __nested_pg_map() 453 __nested_pg_map(vmx, vm, nested_paddr, paddr, PG_LEVEL_4K); in nested_pg_map() 493 __nested_map(vmx, vm, nested_paddr, paddr, size, PG_LEVEL_4K); in nested_map()
|
A D | processor.c | 220 pte = virt_get_pte(vm, pde, vaddr, PG_LEVEL_4K); in __virt_pg_map() 228 __virt_pg_map(vm, vaddr, paddr, PG_LEVEL_4K); in virt_arch_pg_map() 295 return virt_get_pte(vm, pde, vaddr, PG_LEVEL_4K); in __vm_get_page_table_entry() 300 int level = PG_LEVEL_4K; in vm_get_page_table_entry()
|
/linux-6.3-rc2/tools/testing/selftests/kvm/x86_64/ |
A D | nx_huge_pages_test.c | 150 virt_map_level(vm, HPAGE_GVA, HPAGE_GPA, nr_bytes, PG_LEVEL_4K); in run_test()
|
A D | hyperv_tlb_flush.c | 630 __virt_pg_map(vm, gva + PAGE_SIZE * i, gpa & PAGE_MASK, PG_LEVEL_4K); in main()
|
/linux-6.3-rc2/arch/x86/kvm/vmx/ |
A D | capabilities.h | 330 return PG_LEVEL_4K; in ept_caps_to_lpage_level()
|
/linux-6.3-rc2/arch/x86/coco/tdx/ |
A D | tdx.c | 768 case PG_LEVEL_4K: in try_accept_one() 836 if (!try_accept_one(&start, len, PG_LEVEL_4K)) in tdx_enc_status_changed()
|
/linux-6.3-rc2/arch/x86/mm/pat/ |
A D | cpa-test.c | 206 if (level != PG_LEVEL_4K) { in pageattr_test()
|
A D | set_memory.c | 112 direct_pages_count[PG_LEVEL_4K] << 2); in arch_report_meminfo() 531 if (lookup_address(start, &level) && (level != PG_LEVEL_4K)) in protect_kernel_text_ro() 697 *level = PG_LEVEL_4K; in lookup_address_in_pgd() 1603 if (level == PG_LEVEL_4K) { in __change_page_attr()
|
/linux-6.3-rc2/tools/testing/selftests/kvm/include/x86_64/ |
A D | processor.h | 1074 PG_LEVEL_4K, enumerator 1084 #define PG_SIZE_4K PG_LEVEL_SIZE(PG_LEVEL_4K)
|