Lines Matching defs:gfn
287 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(sptep)); in kvm_flush_remote_tlbs_sptep() local
292 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, in mark_mmio_spte()
732 gfn_t gfn, unsigned int access) in kvm_mmu_page_set_translation()
753 gfn_t gfn = kvm_mmu_page_get_gfn(sp, index); in kvm_mmu_page_set_access() local
762 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot()
780 gfn_t gfn, int count) in update_gfn_disallow_lpage_count()
794 void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage()
799 void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage()
808 gfn_t gfn; in account_shadowed() local
865 gfn_t gfn; in unaccount_shadowed() local
894 gfn_t gfn, in gfn_to_memslot_dirty_bitmap()
1073 static struct kvm_rmap_head *gfn_to_rmap(gfn_t gfn, int level, in gfn_to_rmap()
1087 gfn_t gfn; in rmap_remove() local
1384 struct kvm_memory_slot *slot, u64 gfn, in kvm_mmu_slot_gfn_write_protect()
1405 static bool kvm_vcpu_write_protect_gfn(struct kvm_vcpu *vcpu, u64 gfn) in kvm_vcpu_write_protect_gfn()
1428 gfn_t gfn; member
1590 u64 *spte, gfn_t gfn, unsigned int access) in __rmap_add()
1612 u64 *spte, gfn_t gfn, unsigned int access) in rmap_add()
1734 static unsigned kvm_page_table_hashfn(gfn_t gfn) in kvm_page_table_hashfn()
2139 gfn_t gfn, in kvm_mmu_find_shadow_page()
2224 gfn_t gfn, in kvm_mmu_alloc_shadow_page()
2261 gfn_t gfn, in __kvm_mmu_get_shadow_page()
2281 gfn_t gfn, in kvm_mmu_get_shadow_page()
2340 u64 *sptep, gfn_t gfn, in kvm_mmu_get_child_sp()
2805 gfn_t gfn, bool can_unsync, bool prefetch) in mmu_try_to_unsync_pages()
2906 u64 *sptep, unsigned int pte_access, gfn_t gfn, in mmu_set_spte()
2982 gfn_t gfn; in direct_pte_prefetch_many() local
3079 static int host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn, in host_pfn_mapping_level()
3145 gfn_t gfn, int max_level, bool is_private) in __kvm_mmu_max_mapping_level()
3168 const struct kvm_memory_slot *slot, gfn_t gfn, in kvm_mmu_max_mapping_level()
3278 static void kvm_send_hwpoison_signal(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_send_hwpoison_signal()
3710 static hpa_t mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, int quadrant, in mmu_alloc_root()
4235 gfn_t gfn = get_mmio_spte_gfn(spte); in handle_mmio_page_fault() local
4962 static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in sync_mmio_spte()
5952 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mmu_track_write() local
6715 gfn_t gfn; in shadow_mmu_get_sp_for_split() local
6746 gfn_t gfn; in shadow_mmu_split_huge_page() local
6788 gfn_t gfn; in shadow_mmu_try_split_huge_page() local
7079 gfn_t gfn = slot->base_gfn + i; in kvm_mmu_zap_memslot_pages_and_flush() local
7612 static bool hugepage_test_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_test_mixed()
7618 static void hugepage_clear_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_clear_mixed()
7624 static void hugepage_set_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_set_mixed()
7631 gfn_t gfn, int level, unsigned long attrs) in hugepage_has_attrs()
7672 gfn_t gfn = gfn_round_for_level(range->start, level); in kvm_arch_post_set_memory_attributes() local
7731 gfn_t gfn; in kvm_mmu_init_memslot_memory_attributes() local