Home
last modified time | relevance | path

Searched refs:kvm_mmu_page (Results 1 – 12 of 12) sorted by relevance

/linux-6.3-rc2/arch/x86/kvm/mmu/
A Dtdp_mmu.c73 struct kvm_mmu_page *sp = container_of(head, struct kvm_mmu_page, in tdp_mmu_free_sp_rcu_callback()
84 struct kvm_mmu_page *root = container_of(work, struct kvm_mmu_page, in tdp_mmu_zap_root_work()
265 struct kvm_mmu_page *sp; in tdp_mmu_alloc_sp()
306 struct kvm_mmu_page *root; in kvm_tdp_mmu_get_vcpu_root_hpa()
982 struct kvm_mmu_page *root; in kvm_tdp_mmu_zap_leafs()
992 struct kvm_mmu_page *root; in kvm_tdp_mmu_zap_all()
1041 struct kvm_mmu_page *root; in kvm_tdp_mmu_invalidate_all_roots()
1151 struct kvm_mmu_page *sp; in kvm_tdp_mmu_map()
1242 struct kvm_mmu_page *root; in kvm_tdp_mmu_handle_gfn()
1417 struct kvm_mmu_page *sp; in __tdp_mmu_alloc_sp_for_split()
[all …]
A Dtdp_mmu.h15 __must_check static inline bool kvm_tdp_mmu_get_root(struct kvm_mmu_page *root) in kvm_tdp_mmu_get_root()
20 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root,
25 bool kvm_tdp_mmu_zap_sp(struct kvm *kvm, struct kvm_mmu_page *sp);
74 static inline bool is_tdp_mmu_page(struct kvm_mmu_page *sp) { return sp->tdp_mmu_page; } in is_tdp_mmu_page()
76 static inline bool is_tdp_mmu_page(struct kvm_mmu_page *sp) { return false; } in is_tdp_mmu_page()
A Dmmu.c1088 struct kvm_mmu_page *sp; in rmap_remove()
1194 struct kvm_mmu_page *sp; in drop_large_spte()
1638 struct kvm_mmu_page *sp; in __rmap_add()
1782 struct kvm_mmu_page *sp; in mark_unsync()
1985 struct kvm_mmu_page *sp; in mmu_pages_first()
2008 struct kvm_mmu_page *sp; in mmu_pages_clear_parents()
2027 struct kvm_mmu_page *sp; in mmu_sync_children()
2087 struct kvm_mmu_page *sp; in kvm_mmu_find_shadow_page()
2172 struct kvm_mmu_page *sp; in kvm_mmu_alloc_shadow_page()
2209 struct kvm_mmu_page *sp; in __kvm_mmu_get_shadow_page()
[all …]
A Dmmu_internal.h49 struct kvm_mmu_page { struct
141 static inline int kvm_mmu_page_as_id(struct kvm_mmu_page *sp) in kvm_mmu_page_as_id() argument
146 static inline bool kvm_mmu_page_ad_need_write_protect(struct kvm_mmu_page *sp) in kvm_mmu_page_ad_need_write_protect()
339 void track_possible_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp);
340 void untrack_possible_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp);
A Dspte.h222 static inline struct kvm_mmu_page *to_shadow_page(hpa_t shadow_page) in to_shadow_page()
226 return (struct kvm_mmu_page *)page_private(page); in to_shadow_page()
229 static inline struct kvm_mmu_page *spte_to_child_sp(u64 spte) in spte_to_child_sp()
234 static inline struct kvm_mmu_page *sptep_to_sp(u64 *sptep) in sptep_to_sp()
261 static inline bool sp_ad_disabled(struct kvm_mmu_page *sp) in sp_ad_disabled()
461 bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
A Dmmutrace.h159 TP_PROTO(struct kvm_mmu_page *sp, bool created),
178 TP_PROTO(struct kvm_mmu_page *sp),
193 TP_PROTO(struct kvm_mmu_page *sp),
199 TP_PROTO(struct kvm_mmu_page *sp),
205 TP_PROTO(struct kvm_mmu_page *sp),
A Dpaging_tmpl.h152 struct kvm_mmu_page *sp, u64 *spte, in FNAME()
521 FNAME(prefetch_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in FNAME()
578 struct kvm_mmu_page *sp; in FNAME()
621 struct kvm_mmu_page *sp = NULL; in FNAME()
878 static gpa_t FNAME(get_level1_sp_gpa)(struct kvm_mmu_page *sp) in FNAME()
893 struct kvm_mmu_page *sp; in FNAME()
984 static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) in FNAME()
A Dtdp_iter.h113 void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root,
A Dtdp_iter.c39 void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root, in tdp_iter_start()
A Dspte.c137 bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in make_spte()
/linux-6.3-rc2/arch/x86/include/asm/
A Dkvm_host.h425 struct kvm_mmu_page;
443 struct kvm_mmu_page *sp);
/linux-6.3-rc2/Documentation/virt/kvm/x86/
A Dmmu.rst121 The principal data structure is the shadow page, 'struct kvm_mmu_page'. A
198 shadow page; it is also used to go back from a struct kvm_mmu_page

Completed in 67 milliseconds