Lines Matching refs:shared
31 bool shared) in kvm_lockdep_assert_mmu_lock_held() argument
33 if (shared) in kvm_lockdep_assert_mmu_lock_held()
80 bool shared);
130 bool shared) in kvm_tdp_mmu_put_root() argument
132 kvm_lockdep_assert_mmu_lock_held(kvm, shared); in kvm_tdp_mmu_put_root()
195 bool shared, bool only_valid) in tdp_mmu_next_root() argument
221 kvm_tdp_mmu_put_root(kvm, prev_root, shared); in tdp_mmu_next_root()
335 bool shared);
388 bool shared) in tdp_mmu_unlink_sp() argument
395 if (shared) in tdp_mmu_unlink_sp()
403 if (shared) in tdp_mmu_unlink_sp()
424 static void handle_removed_pt(struct kvm *kvm, tdp_ptep_t pt, bool shared) in handle_removed_pt() argument
433 tdp_mmu_unlink_sp(kvm, sp, shared); in handle_removed_pt()
440 if (shared) { in handle_removed_pt()
501 old_spte, REMOVED_SPTE, level, shared); in handle_removed_pt()
524 bool shared) in __handle_changed_spte() argument
607 handle_removed_pt(kvm, spte_to_child_pt(old_spte, level), shared); in __handle_changed_spte()
612 bool shared) in handle_changed_spte() argument
615 shared); in handle_changed_spte()
809 bool flush, bool shared) in tdp_mmu_iter_cond_resched() argument
823 if (shared) in tdp_mmu_iter_cond_resched()
850 bool shared, int zap_level) in __tdp_mmu_zap_root() argument
859 if (tdp_mmu_iter_cond_resched(kvm, &iter, false, shared)) in __tdp_mmu_zap_root()
868 if (!shared) in __tdp_mmu_zap_root()
876 bool shared) in tdp_mmu_zap_root() argument
891 kvm_lockdep_assert_mmu_lock_held(kvm, shared); in tdp_mmu_zap_root()
905 __tdp_mmu_zap_root(kvm, root, shared, PG_LEVEL_1G); in tdp_mmu_zap_root()
906 __tdp_mmu_zap_root(kvm, root, shared, root->role.level); in tdp_mmu_zap_root()
1121 struct kvm_mmu_page *sp, bool shared) in tdp_mmu_link_sp() argument
1126 if (shared) { in tdp_mmu_link_sp()
1140 struct kvm_mmu_page *sp, bool shared);
1436 bool shared) in tdp_mmu_alloc_sp_for_split() argument
1455 if (shared) in tdp_mmu_alloc_sp_for_split()
1463 if (shared) in tdp_mmu_alloc_sp_for_split()
1475 struct kvm_mmu_page *sp, bool shared) in tdp_mmu_split_huge_page() argument
1496 ret = tdp_mmu_link_sp(kvm, iter, sp, shared); in tdp_mmu_split_huge_page()
1515 int target_level, bool shared) in tdp_mmu_split_huge_pages_root() argument
1536 if (tdp_mmu_iter_cond_resched(kvm, &iter, false, shared)) in tdp_mmu_split_huge_pages_root()
1543 sp = tdp_mmu_alloc_sp_for_split(kvm, &iter, shared); in tdp_mmu_split_huge_pages_root()
1558 if (tdp_mmu_split_huge_page(kvm, &iter, sp, shared)) in tdp_mmu_split_huge_pages_root()
1584 int target_level, bool shared) in kvm_tdp_mmu_try_split_huge_pages() argument
1589 kvm_lockdep_assert_mmu_lock_held(kvm, shared); in kvm_tdp_mmu_try_split_huge_pages()
1591 for_each_valid_tdp_mmu_root_yield_safe(kvm, root, slot->as_id, shared) { in kvm_tdp_mmu_try_split_huge_pages()
1592 r = tdp_mmu_split_huge_pages_root(kvm, root, start, end, target_level, shared); in kvm_tdp_mmu_try_split_huge_pages()
1594 kvm_tdp_mmu_put_root(kvm, root, shared); in kvm_tdp_mmu_try_split_huge_pages()