Lines Matching defs:spte
161 #define for_each_shadow_entry_lockless(_vcpu, _addr, _walker, spte) \ argument
287 u64 spte = make_mmio_spte(vcpu, gfn, access); in mark_mmio_spte() local
293 static gfn_t get_mmio_spte_gfn(u64 spte) in get_mmio_spte_gfn()
303 static unsigned get_mmio_spte_access(u64 spte) in get_mmio_spte_access()
308 static bool check_mmio_spte(struct kvm_vcpu *vcpu, u64 spte) in check_mmio_spte()
329 static void __set_spte(u64 *sptep, u64 spte) in __set_spte()
334 static void __update_clear_spte_fast(u64 *sptep, u64 spte) in __update_clear_spte_fast()
339 static u64 __update_clear_spte_slow(u64 *sptep, u64 spte) in __update_clear_spte_slow()
354 u64 spte; member
357 static void count_spte_clear(u64 *sptep, u64 spte) in count_spte_clear()
369 static void __set_spte(u64 *sptep, u64 spte) in __set_spte()
388 static void __update_clear_spte_fast(u64 *sptep, u64 spte) in __update_clear_spte_fast()
407 static u64 __update_clear_spte_slow(u64 *sptep, u64 spte) in __update_clear_spte_slow()
444 union split_spte spte, *orig = (union split_spte *)sptep; in __get_spte_lockless() local
609 u64 spte = mmu_spte_get_lockless(sptep); in mmu_spte_age() local
917 static int pte_list_add(struct kvm_mmu_memory_cache *cache, u64 *spte, in pte_list_add()
975 static void pte_list_remove(u64 *spte, struct kvm_rmap_head *rmap_head) in pte_list_remove()
1084 static void rmap_remove(struct kvm *kvm, u64 *spte) in rmap_remove()
1220 u64 spte = *sptep; in spte_write_protect() local
1250 u64 spte = *sptep; in spte_clear_dirty() local
1636 u64 *spte, gfn_t gfn, unsigned int access) in __rmap_add()
1658 u64 *spte, gfn_t gfn, unsigned int access) in rmap_add()
1780 static void mark_unsync(u64 *spte) in mark_unsync()
2070 static void clear_sp_write_flooding_count(u64 *spte) in clear_sp_write_flooding_count()
2343 u64 spte) in __shadow_walk_next()
2363 u64 spte; in __link_shadow_page() local
2424 u64 *spte, struct list_head *invalid_list) in mmu_page_zap_pte()
2849 u64 spte; in mmu_set_spte() local
2948 u64 *spte, *start = NULL; in __direct_pte_prefetch() local
3142 void disallowed_hugepage_adjust(struct kvm_page_fault *fault, u64 spte, int cur_level) in disallowed_hugepage_adjust()
3332 static bool is_access_allowed(struct kvm_page_fault *fault, u64 spte) in is_access_allowed()
3353 static u64 *fast_pf_get_last_sptep(struct kvm_vcpu *vcpu, gpa_t gpa, u64 *spte) in fast_pf_get_last_sptep()
3374 u64 spte = 0ull; in fast_page_fault() local
4036 u64 spte; in get_walk() local
4103 u64 spte; in handle_mmio_page_fault() local
4157 u64 spte; in shadow_page_table_clear_flood() local
5569 u64 *spte; in get_written_sptes() local
5604 u64 entry, gentry, *spte; in kvm_mmu_pte_write() local
6315 u64 *sptep, spte; in shadow_mmu_split_huge_page() local
6359 u64 spte; in shadow_mmu_try_split_huge_page() local