| /linux/virt/kvm/ |
| A D | pfncache.c | 99 static void *gpc_map(kvm_pfn_t pfn) in gpc_map() 111 static void gpc_unmap(kvm_pfn_t pfn, void *khva) in gpc_unmap() 155 static kvm_pfn_t hva_to_pfn_retry(struct gfn_to_pfn_cache *gpc) in hva_to_pfn_retry() 159 kvm_pfn_t new_pfn = KVM_PFN_ERR_FAULT; in hva_to_pfn_retry() 253 kvm_pfn_t old_pfn; in __kvm_gpc_refresh() 442 kvm_pfn_t old_pfn; in kvm_gpc_deactivate()
|
| A D | guest_memfd.c | 23 static inline kvm_pfn_t folio_file_pfn(struct folio *folio, pgoff_t index) in folio_file_pfn() 32 kvm_pfn_t pfn = folio_file_pfn(folio, index); in __kvm_gmem_prepare_folio() 359 kvm_pfn_t pfn = page_to_pfn(page); in kvm_gmem_free_folio() 556 gfn_t gfn, kvm_pfn_t *pfn, bool *is_prepared, in __kvm_gmem_get_pfn() 593 gfn_t gfn, kvm_pfn_t *pfn, int *max_order) in kvm_gmem_get_pfn() 652 kvm_pfn_t pfn; in kvm_gmem_populate()
|
| A D | kvm_mm.h | 23 kvm_pfn_t hva_to_pfn(unsigned long addr, bool atomic, bool interruptible,
|
| A D | kvm_main.c | 2763 bool *writable, kvm_pfn_t *pfn) in hva_to_pfn_fast() 2849 static int kvm_try_get_pfn(kvm_pfn_t pfn) in kvm_try_get_pfn() 2861 bool *writable, kvm_pfn_t *p_pfn) in hva_to_pfn_remapped() 2864 kvm_pfn_t pfn; in hva_to_pfn_remapped() 2941 kvm_pfn_t pfn; in hva_to_pfn() 3070 kvm_pfn_t pfn; in gfn_to_page() 3095 kvm_pfn_t pfn; in kvm_vcpu_map() 3182 void kvm_release_pfn_clean(kvm_pfn_t pfn) in kvm_release_pfn_clean() 3206 void kvm_release_pfn_dirty(kvm_pfn_t pfn) in kvm_release_pfn_dirty() 3226 void kvm_set_pfn_dirty(kvm_pfn_t pfn) in kvm_set_pfn_dirty() [all …]
|
| /linux/include/linux/ |
| A D | kvm_host.h | 105 static inline bool is_error_pfn(kvm_pfn_t pfn) in is_error_pfn() 130 static inline bool is_noslot_pfn(kvm_pfn_t pfn) in is_noslot_pfn() 295 kvm_pfn_t pfn; 296 kvm_pfn_t gfn; 1238 void kvm_release_pfn_clean(kvm_pfn_t pfn); 1239 void kvm_release_pfn_dirty(kvm_pfn_t pfn); 1240 void kvm_set_pfn_dirty(kvm_pfn_t pfn); 1241 void kvm_set_pfn_accessed(kvm_pfn_t pfn); 1831 static inline hpa_t pfn_to_hpa(kvm_pfn_t pfn) in pfn_to_hpa() 2468 kvm_pfn_t *pfn, int *max_order) in kvm_gmem_get_pfn() [all …]
|
| A D | kvm_types.h | 50 typedef hfn_t kvm_pfn_t; typedef 70 kvm_pfn_t pfn;
|
| /linux/arch/x86/kvm/svm/ |
| A D | svm.h | 763 int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order); 764 void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end); 765 int sev_private_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn); 787 static inline int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order) in sev_gmem_prepare() 791 static inline void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end) {} in sev_gmem_invalidate() 792 static inline int sev_private_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn) in sev_private_max_mapping_level()
|
| A D | sev.c | 3603 static int snp_rmptable_psmash(kvm_pfn_t pfn) in snp_rmptable_psmash() 3855 kvm_pfn_t pfn; in __sev_snp_update_protected_guest_state() 4695 kvm_pfn_t pfn; in sev_handle_rmp_fault() 4781 static bool is_pfn_range_shared(kvm_pfn_t start, kvm_pfn_t end) in is_pfn_range_shared() 4783 kvm_pfn_t pfn = start; in is_pfn_range_shared() 4816 static bool is_large_rmp_possible(struct kvm *kvm, kvm_pfn_t pfn, int order) in is_large_rmp_possible() 4818 kvm_pfn_t pfn_aligned = ALIGN_DOWN(pfn, PTRS_PER_PMD); in is_large_rmp_possible() 4835 kvm_pfn_t pfn_aligned; in sev_gmem_prepare() 4879 void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end) in sev_gmem_invalidate() 4881 kvm_pfn_t pfn; in sev_gmem_invalidate() [all …]
|
| /linux/arch/powerpc/kvm/ |
| A D | e500_mmu_host.c | 164 kvm_pfn_t pfn; in kvmppc_map_magic() 166 pfn = (kvm_pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT; in kvmppc_map_magic() 247 kvm_pfn_t pfn, unsigned int wimg) in kvmppc_e500_ref_setup() 310 kvm_pfn_t pfn = ref->pfn; in kvmppc_e500_setup_stlbe()
|
| A D | e500.h | 41 kvm_pfn_t pfn; /* valid only for TLB0, except briefly */
|
| A D | book3s.c | 424 kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing, in kvmppc_gpa_to_pfn() 437 kvm_pfn_t pfn; in kvmppc_gpa_to_pfn() 439 pfn = (kvm_pfn_t)virt_to_phys((void*)shared_page) >> PAGE_SHIFT; in kvmppc_gpa_to_pfn()
|
| A D | trace_pr.h | 32 TP_PROTO(int rflags, ulong hpteg, ulong va, kvm_pfn_t hpaddr,
|
| A D | book3s_32_mmu_host.c | 133 kvm_pfn_t hpaddr; in kvmppc_mmu_map_page()
|
| A D | book3s_64_mmu_host.c | 75 kvm_pfn_t hpaddr; in kvmppc_mmu_map_page()
|
| /linux/arch/x86/kvm/mmu/ |
| A D | spte.h | 353 static inline kvm_pfn_t spte_to_pfn(u64 pte) in spte_to_pfn() 501 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
|
| A D | spte.c | 108 static bool kvm_is_mmio_pfn(kvm_pfn_t pfn) in kvm_is_mmio_pfn() 159 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, in make_spte()
|
| A D | mmu_internal.h | 240 kvm_pfn_t pfn;
|
| A D | paging_tmpl.h | 539 kvm_pfn_t pfn; in FNAME()
|
| /linux/arch/mips/kvm/ |
| A D | mmu.c | 487 kvm_pfn_t pfn = 0; /* silence bogus GCC warning */ in _kvm_mips_map_page_fast() 564 kvm_pfn_t pfn; in kvm_mips_map_page()
|
| /linux/arch/loongarch/kvm/ |
| A D | mmu.c | 555 kvm_pfn_t pfn = 0; in kvm_map_page_fast() 793 kvm_pfn_t pfn; in kvm_map_page()
|
| /linux/arch/arm64/kvm/ |
| A D | mmu.c | 1330 unsigned long hva, kvm_pfn_t *pfnp, in transparent_hugepage_adjust() 1333 kvm_pfn_t pfn = *pfnp; in transparent_hugepage_adjust() 1400 static void sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn, in sanitise_mte_tags() 1438 kvm_pfn_t pfn; in user_mem_abort()
|
| /linux/arch/arm64/include/asm/ |
| A D | kvm_pgtable.h | 144 static inline kvm_pfn_t kvm_pte_to_pfn(kvm_pte_t pte) in kvm_pte_to_pfn()
|
| /linux/arch/x86/include/asm/ |
| A D | kvm_host.h | 1832 int (*gmem_prepare)(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order); 1833 void (*gmem_invalidate)(kvm_pfn_t start, kvm_pfn_t end); 1834 int (*private_max_mapping_level)(struct kvm *kvm, kvm_pfn_t pfn);
|
| /linux/arch/riscv/kvm/ |
| A D | mmu.c | 594 kvm_pfn_t hfn; in kvm_riscv_gstage_map()
|
| /linux/arch/powerpc/include/asm/ |
| A D | kvm_book3s.h | 237 extern kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa,
|