Lines Matching defs:kvm

250 static void kvm_flush_remote_tlbs_with_range(struct kvm *kvm,  in kvm_flush_remote_tlbs_with_range()
262 void kvm_flush_remote_tlbs_with_address(struct kvm *kvm, in kvm_flush_remote_tlbs_with_address()
276 static void kvm_flush_remote_tlbs_sptep(struct kvm *kvm, u64 *sptep) in kvm_flush_remote_tlbs_sptep()
553 static u64 mmu_spte_clear_track_bits(struct kvm *kvm, u64 *sptep) in mmu_spte_clear_track_bits()
809 static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in account_shadowed()
831 void track_possible_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp) in track_possible_nx_huge_page()
849 static void account_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp, in account_nx_huge_page()
858 static void unaccount_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_shadowed()
875 void untrack_possible_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp) in untrack_possible_nx_huge_page()
884 static void unaccount_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_nx_huge_page()
1011 static void kvm_zap_one_rmap_spte(struct kvm *kvm, in kvm_zap_one_rmap_spte()
1019 static bool kvm_zap_all_rmap_sptes(struct kvm *kvm, in kvm_zap_all_rmap_sptes()
1084 static void rmap_remove(struct kvm *kvm, u64 *spte) in rmap_remove()
1184 static void drop_spte(struct kvm *kvm, u64 *sptep) in drop_spte()
1192 static void drop_large_spte(struct kvm *kvm, u64 *sptep, bool flush) in drop_large_spte()
1275 static bool __rmap_clear_dirty(struct kvm *kvm, struct kvm_rmap_head *rmap_head, in __rmap_clear_dirty()
1300 static void kvm_mmu_write_protect_pt_masked(struct kvm *kvm, in kvm_mmu_write_protect_pt_masked()
1333 static void kvm_mmu_clear_dirty_pt_masked(struct kvm *kvm, in kvm_mmu_clear_dirty_pt_masked()
1366 void kvm_arch_mmu_enable_log_dirty_pt_masked(struct kvm *kvm, in kvm_arch_mmu_enable_log_dirty_pt_masked()
1407 bool kvm_mmu_slot_gfn_write_protect(struct kvm *kvm, in kvm_mmu_slot_gfn_write_protect()
1437 static bool __kvm_zap_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head, in __kvm_zap_rmap()
1443 static bool kvm_zap_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head, in kvm_zap_rmap()
1450 static bool kvm_set_pte_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head, in kvm_set_pte_rmap()
1563 static __always_inline bool kvm_handle_gfn_range(struct kvm *kvm, in kvm_handle_gfn_range()
1578 bool kvm_unmap_gfn_range(struct kvm *kvm, struct kvm_gfn_range *range) in kvm_unmap_gfn_range()
1591 bool kvm_set_spte_gfn(struct kvm *kvm, struct kvm_gfn_range *range) in kvm_set_spte_gfn()
1604 static bool kvm_age_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head, in kvm_age_rmap()
1618 static bool kvm_test_age_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head, in kvm_test_age_rmap()
1633 static void __rmap_add(struct kvm *kvm, in __rmap_add()
1665 bool kvm_age_gfn(struct kvm *kvm, struct kvm_gfn_range *range) in kvm_age_gfn()
1678 bool kvm_test_age_gfn(struct kvm *kvm, struct kvm_gfn_range *range) in kvm_test_age_gfn()
1713 static inline void kvm_mod_used_mmu_pages(struct kvm *kvm, long nr) in kvm_mod_used_mmu_pages()
1719 static void kvm_account_mmu_page(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_account_mmu_page()
1725 static void kvm_unaccount_mmu_page(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_unaccount_mmu_page()
1883 static void kvm_unlink_unsync_page(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_unlink_unsync_page()
1927 static bool kvm_mmu_remote_flush_or_zap(struct kvm *kvm, in kvm_mmu_remote_flush_or_zap()
1941 static bool is_obsolete_sp(struct kvm *kvm, struct kvm_mmu_page *sp) in is_obsolete_sp()
2081 static struct kvm_mmu_page *kvm_mmu_find_shadow_page(struct kvm *kvm, in kvm_mmu_find_shadow_page()
2166 static struct kvm_mmu_page *kvm_mmu_alloc_shadow_page(struct kvm *kvm, in kvm_mmu_alloc_shadow_page()
2202 static struct kvm_mmu_page *__kvm_mmu_get_shadow_page(struct kvm *kvm, in __kvm_mmu_get_shadow_page()
2359 static void __link_shadow_page(struct kvm *kvm, in __link_shadow_page()
2423 static int mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp, in mmu_page_zap_pte()
2453 static int kvm_mmu_page_unlink_children(struct kvm *kvm, in kvm_mmu_page_unlink_children()
2475 static int mmu_zap_unsync_children(struct kvm *kvm, in mmu_zap_unsync_children()
2499 static bool __kvm_mmu_prepare_zap_page(struct kvm *kvm, in __kvm_mmu_prepare_zap_page()
2564 static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp, in kvm_mmu_prepare_zap_page()
2573 static void kvm_mmu_commit_zap_page(struct kvm *kvm, in kvm_mmu_commit_zap_page()
2598 static unsigned long kvm_mmu_zap_oldest_mmu_pages(struct kvm *kvm, in kvm_mmu_zap_oldest_mmu_pages()
2635 static inline unsigned long kvm_mmu_available_pages(struct kvm *kvm) in kvm_mmu_available_pages()
2671 void kvm_mmu_change_mmu_pages(struct kvm *kvm, unsigned long goal_nr_mmu_pages) in kvm_mmu_change_mmu_pages()
2687 int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn) in kvm_mmu_unprotect_page()
2723 static void kvm_unsync_page(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_unsync_page()
2738 int mmu_try_to_unsync_pages(struct kvm *kvm, const struct kvm_memory_slot *slot, in mmu_try_to_unsync_pages()
3022 static int host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn, in host_pfn_mapping_level()
3086 int kvm_mmu_max_mapping_level(struct kvm *kvm, in kvm_mmu_max_mapping_level()
3484 static void mmu_free_root_page(struct kvm *kvm, hpa_t *root_hpa, in mmu_free_root_page()
3509 void kvm_mmu_free_roots(struct kvm *kvm, struct kvm_mmu *mmu, in kvm_mmu_free_roots()
3561 void kvm_mmu_free_guest_mode_roots(struct kvm *kvm, struct kvm_mmu *mmu) in kvm_mmu_free_guest_mode_roots()
3665 static int mmu_first_shadow_root_alloc(struct kvm *kvm) in mmu_first_shadow_root_alloc()
4492 static bool cached_root_find_and_keep_current(struct kvm *kvm, struct kvm_mmu *mmu, in cached_root_find_and_keep_current()
4526 static bool cached_root_find_without_current(struct kvm *kvm, struct kvm_mmu *mmu, in cached_root_find_without_current()
4547 static bool fast_pgd_switch(struct kvm *kvm, struct kvm_mmu *mmu, in fast_pgd_switch()
5441 struct kvm *kvm = vcpu->kvm; in kvm_mmu_unload() local
5450 static bool is_obsolete_root(struct kvm *kvm, hpa_t root_hpa) in is_obsolete_root()
5472 static void __kvm_mmu_free_obsolete_roots(struct kvm *kvm, struct kvm_mmu *mmu) in __kvm_mmu_free_obsolete_roots()
5822 slot_handle_level_range(struct kvm *kvm, const struct kvm_memory_slot *memslot, in slot_handle_level_range()
5849 slot_handle_level(struct kvm *kvm, const struct kvm_memory_slot *memslot, in slot_handle_level()
5860 slot_handle_level_4k(struct kvm *kvm, const struct kvm_memory_slot *memslot, in slot_handle_level_4k()
5959 static void kvm_zap_obsolete_pages(struct kvm *kvm) in kvm_zap_obsolete_pages()
6024 static void kvm_mmu_zap_all_fast(struct kvm *kvm) in kvm_mmu_zap_all_fast()
6075 static bool kvm_has_zapped_obsolete_pages(struct kvm *kvm) in kvm_has_zapped_obsolete_pages()
6080 static void kvm_mmu_invalidate_zap_pages_in_memslot(struct kvm *kvm, in kvm_mmu_invalidate_zap_pages_in_memslot()
6087 int kvm_mmu_init_vm(struct kvm *kvm) in kvm_mmu_init_vm()
6118 static void mmu_free_vm_memory_caches(struct kvm *kvm) in mmu_free_vm_memory_caches()
6125 void kvm_mmu_uninit_vm(struct kvm *kvm) in kvm_mmu_uninit_vm()
6137 static bool kvm_rmap_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end) in kvm_rmap_zap_gfn_range()
6172 void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end) in kvm_zap_gfn_range()
6201 static bool slot_rmap_write_protect(struct kvm *kvm, in slot_rmap_write_protect()
6208 void kvm_mmu_slot_remove_write_access(struct kvm *kvm, in kvm_mmu_slot_remove_write_access()
6231 static bool need_topup_split_caches_or_resched(struct kvm *kvm) in need_topup_split_caches_or_resched()
6246 static int topup_split_caches(struct kvm *kvm) in topup_split_caches()
6279 static struct kvm_mmu_page *shadow_mmu_get_sp_for_split(struct kvm *kvm, u64 *huge_sptep) in shadow_mmu_get_sp_for_split()
6306 static void shadow_mmu_split_huge_page(struct kvm *kvm, in shadow_mmu_split_huge_page()
6352 static int shadow_mmu_try_split_huge_page(struct kvm *kvm, in shadow_mmu_try_split_huge_page()
6391 static bool shadow_mmu_try_split_huge_pages(struct kvm *kvm, in shadow_mmu_try_split_huge_pages()
6437 static void kvm_shadow_mmu_try_split_huge_pages(struct kvm *kvm, in kvm_shadow_mmu_try_split_huge_pages()
6457 void kvm_mmu_try_split_huge_pages(struct kvm *kvm, in kvm_mmu_try_split_huge_pages()
6476 void kvm_mmu_slot_try_split_huge_pages(struct kvm *kvm, in kvm_mmu_slot_try_split_huge_pages()
6507 static bool kvm_mmu_zap_collapsible_spte(struct kvm *kvm, in kvm_mmu_zap_collapsible_spte()
6544 static void kvm_rmap_zap_collapsible_sptes(struct kvm *kvm, in kvm_rmap_zap_collapsible_sptes()
6556 void kvm_mmu_zap_collapsible_sptes(struct kvm *kvm, in kvm_mmu_zap_collapsible_sptes()
6572 void kvm_arch_flush_remote_tlbs_memslot(struct kvm *kvm, in kvm_arch_flush_remote_tlbs_memslot()
6587 void kvm_mmu_slot_leaf_clear_dirty(struct kvm *kvm, in kvm_mmu_slot_leaf_clear_dirty()
6616 void kvm_mmu_zap_all(struct kvm *kvm) in kvm_mmu_zap_all()
6641 void kvm_mmu_invalidate_mmio_sptes(struct kvm *kvm, u64 gen) in kvm_mmu_invalidate_mmio_sptes()
6669 struct kvm *kvm; in mmu_shrink_scan() local
6772 struct kvm *kvm; in set_nx_huge_pages() local
6915 struct kvm *kvm; in set_nx_huge_pages_recovery_param() local
6928 static void kvm_recover_nx_huge_pages(struct kvm *kvm) in kvm_recover_nx_huge_pages()
7031 static int kvm_nx_huge_page_recovery_worker(struct kvm *kvm, uintptr_t data) in kvm_nx_huge_page_recovery_worker()
7056 int kvm_mmu_post_init_vm(struct kvm *kvm) in kvm_mmu_post_init_vm()
7069 void kvm_mmu_pre_destroy_vm(struct kvm *kvm) in kvm_mmu_pre_destroy_vm()