Lines Matching refs:gfn
274 gfn_t gfn, union kvm_mmu_page_role role) in tdp_mmu_init_sp() argument
281 sp->gfn = gfn; in tdp_mmu_init_sp()
299 tdp_mmu_init_sp(child_sp, iter->sptep, iter->gfn, role); in tdp_mmu_init_child_sp()
333 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn,
348 static void handle_changed_spte_dirty_log(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte_dirty_log() argument
361 slot = __gfn_to_memslot(__kvm_memslots(kvm, as_id), gfn); in handle_changed_spte_dirty_log()
362 mark_page_dirty_in_slot(kvm, slot, gfn); in handle_changed_spte_dirty_log()
428 gfn_t base_gfn = sp->gfn; in handle_removed_pt()
437 gfn_t gfn = base_gfn + i * KVM_PAGES_PER_HPAGE(level); in handle_removed_pt() local
500 handle_changed_spte(kvm, kvm_mmu_page_as_id(sp), gfn, in handle_removed_pt()
522 static void __handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in __handle_changed_spte() argument
534 WARN_ON(gfn & (KVM_PAGES_PER_HPAGE(level) - 1)); in __handle_changed_spte()
550 as_id, gfn, old_spte, new_spte, level); in __handle_changed_spte()
562 trace_kvm_tdp_mmu_spte_changed(as_id, gfn, level, old_spte, new_spte); in __handle_changed_spte()
588 as_id, gfn, old_spte, new_spte, level); in __handle_changed_spte()
610 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte() argument
614 __handle_changed_spte(kvm, as_id, gfn, old_spte, new_spte, level, in handle_changed_spte()
617 handle_changed_spte_dirty_log(kvm, as_id, gfn, old_spte, in handle_changed_spte()
661 __handle_changed_spte(kvm, iter->as_id, iter->gfn, iter->old_spte, in tdp_mmu_set_spte_atomic()
683 kvm_flush_remote_tlbs_gfn(kvm, iter->gfn, iter->level); in tdp_mmu_zap_spte_atomic()
722 u64 old_spte, u64 new_spte, gfn_t gfn, int level, in __tdp_mmu_set_spte() argument
738 __handle_changed_spte(kvm, as_id, gfn, old_spte, new_spte, level, false); in __tdp_mmu_set_spte()
743 handle_changed_spte_dirty_log(kvm, as_id, gfn, old_spte, in __tdp_mmu_set_spte()
756 iter->gfn, iter->level, in _tdp_mmu_set_spte()
830 WARN_ON(iter->gfn > iter->next_last_level_gfn); in tdp_mmu_iter_cond_resched()
927 sp->gfn, sp->role.level + 1, true, true); in kvm_tdp_mmu_zap_sp()
1070 new_spte = make_mmio_spte(vcpu, iter->gfn, ACC_ALL); in tdp_mmu_map_handle_target_level()
1072 wrprot = make_spte(vcpu, sp, fault->slot, ACC_ALL, iter->gfn, in tdp_mmu_map_handle_target_level()
1082 kvm_flush_remote_tlbs_gfn(vcpu->kvm, iter->gfn, iter->level); in tdp_mmu_map_handle_target_level()
1097 trace_mark_mmio_spte(rcu_dereference(iter->sptep), iter->gfn, in tdp_mmu_map_handle_target_level()
1101 trace_kvm_mmu_set_spte(iter->level, iter->gfn, in tdp_mmu_map_handle_target_level()
1160 tdp_mmu_for_each_pte(iter, mmu, fault->gfn, fault->gfn + 1) { in kvm_tdp_mmu_map()
1508 trace_kvm_mmu_split_huge_page(iter->gfn, huge_spte, level, ret); in tdp_mmu_split_huge_page()
1546 trace_kvm_mmu_split_huge_page(iter.gfn, in tdp_mmu_split_huge_pages_root()
1676 gfn_t gfn, unsigned long mask, bool wrprot) in clear_dirty_pt_masked() argument
1683 tdp_root_for_each_leaf_pte(iter, root, gfn + __ffs(mask), in clear_dirty_pt_masked()
1684 gfn + BITS_PER_LONG) { in clear_dirty_pt_masked()
1689 !(mask & (1UL << (iter.gfn - gfn)))) in clear_dirty_pt_masked()
1692 mask &= ~(1UL << (iter.gfn - gfn)); in clear_dirty_pt_masked()
1721 gfn_t gfn, unsigned long mask, in kvm_tdp_mmu_clear_dirty_pt_masked() argument
1728 clear_dirty_pt_masked(kvm, root, gfn, mask, wrprot); in kvm_tdp_mmu_clear_dirty_pt_masked()
1766 if (iter.gfn < start || iter.gfn >= end) in zap_collapsible_spte_range()
1770 iter.gfn, PG_LEVEL_NUM); in zap_collapsible_spte_range()
1803 gfn_t gfn, int min_level) in write_protect_gfn() argument
1813 for_each_tdp_pte_min_level(iter, root, min_level, gfn, gfn + 1) { in write_protect_gfn()
1839 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_tdp_mmu_write_protect_gfn() argument
1847 spte_set |= write_protect_gfn(kvm, root, gfn, min_level); in kvm_tdp_mmu_write_protect_gfn()
1863 gfn_t gfn = addr >> PAGE_SHIFT; in kvm_tdp_mmu_get_walk() local
1868 tdp_mmu_for_each_pte(iter, mmu, gfn, gfn + 1) { in kvm_tdp_mmu_get_walk()
1892 gfn_t gfn = addr >> PAGE_SHIFT; in kvm_tdp_mmu_fast_pf_get_last_sptep() local
1895 tdp_mmu_for_each_pte(iter, mmu, gfn, gfn + 1) { in kvm_tdp_mmu_fast_pf_get_last_sptep()