Lines Matching refs:gfn_t
274 gfn_t gfn, union kvm_mmu_page_role role) in tdp_mmu_init_sp()
333 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn,
348 static void handle_changed_spte_dirty_log(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte_dirty_log()
428 gfn_t base_gfn = sp->gfn; in handle_removed_pt()
437 gfn_t gfn = base_gfn + i * KVM_PAGES_PER_HPAGE(level); in handle_removed_pt()
522 static void __handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in __handle_changed_spte()
610 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte()
722 u64 old_spte, u64 new_spte, gfn_t gfn, int level, in __tdp_mmu_set_spte()
838 static inline gfn_t tdp_mmu_max_gfn_exclusive(void) in tdp_mmu_max_gfn_exclusive()
854 gfn_t end = tdp_mmu_max_gfn_exclusive(); in __tdp_mmu_zap_root()
855 gfn_t start = 0; in __tdp_mmu_zap_root()
940 gfn_t start, gfn_t end, bool can_yield, bool flush) in tdp_mmu_zap_leafs()
979 bool kvm_tdp_mmu_zap_leafs(struct kvm *kvm, int as_id, gfn_t start, gfn_t end, in kvm_tdp_mmu_zap_leafs()
1363 gfn_t start, gfn_t end, int min_level) in wrprot_gfn_range()
1514 gfn_t start, gfn_t end, in tdp_mmu_split_huge_pages_root()
1583 gfn_t start, gfn_t end, in kvm_tdp_mmu_try_split_huge_pages()
1608 gfn_t start, gfn_t end) in clear_dirty_gfn_range()
1676 gfn_t gfn, unsigned long mask, bool wrprot) in clear_dirty_pt_masked()
1721 gfn_t gfn, unsigned long mask, in kvm_tdp_mmu_clear_dirty_pt_masked()
1735 gfn_t start = slot->base_gfn; in zap_collapsible_spte_range()
1736 gfn_t end = start + slot->npages; in zap_collapsible_spte_range()
1803 gfn_t gfn, int min_level) in write_protect_gfn()
1839 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_tdp_mmu_write_protect_gfn()
1863 gfn_t gfn = addr >> PAGE_SHIFT; in kvm_tdp_mmu_get_walk()
1892 gfn_t gfn = addr >> PAGE_SHIFT; in kvm_tdp_mmu_fast_pf_get_last_sptep()