Lines Matching refs:memslot

205 void kvmppc_map_vrma(struct kvm_vcpu *vcpu, struct kvm_memory_slot *memslot,  in kvmppc_map_vrma()  argument
219 npages = memslot->npages >> (porder - PAGE_SHIFT); in kvmppc_map_vrma()
501 struct kvm_memory_slot *memslot; in kvmppc_book3s_hv_page_fault() local
564 memslot = gfn_to_memslot(kvm, gfn); in kvmppc_book3s_hv_page_fault()
566 trace_kvm_page_fault_enter(vcpu, hpte, memslot, ea, dsisr); in kvmppc_book3s_hv_page_fault()
569 if (!memslot || (memslot->flags & KVM_MEMSLOT_INVALID)) in kvmppc_book3s_hv_page_fault()
577 if (gfn_base < memslot->base_gfn) in kvmppc_book3s_hv_page_fault()
589 hva = gfn_to_hva_memslot(memslot, gfn); in kvmppc_book3s_hv_page_fault()
601 pfn = __gfn_to_pfn_memslot(memslot, gfn, false, false, NULL, in kvmppc_book3s_hv_page_fault()
691 rmap = &memslot->arch.rmap[gfn_base - memslot->base_gfn]; in kvmppc_book3s_hv_page_fault()
744 struct kvm_memory_slot *memslot; in kvmppc_rmap_reset() local
749 kvm_for_each_memslot(memslot, bkt, slots) { in kvmppc_rmap_reset()
756 memset(memslot->arch.rmap, 0, in kvmppc_rmap_reset()
757 memslot->npages * sizeof(*memslot->arch.rmap)); in kvmppc_rmap_reset()
765 struct kvm_memory_slot *memslot, in kvmppc_unmap_hpte() argument
797 if ((rcbits & HPTE_R_C) && memslot->dirty_bitmap) in kvmppc_unmap_hpte()
798 kvmppc_update_dirty_map(memslot, gfn, psize); in kvmppc_unmap_hpte()
806 static void kvm_unmap_rmapp(struct kvm *kvm, struct kvm_memory_slot *memslot, in kvm_unmap_rmapp() argument
813 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_unmap_rmapp()
836 kvmppc_unmap_hpte(kvm, i, memslot, rmapp, gfn); in kvm_unmap_rmapp()
858 struct kvm_memory_slot *memslot) in kvmppc_core_flush_memslot_hv() argument
864 gfn = memslot->base_gfn; in kvmppc_core_flush_memslot_hv()
865 rmapp = memslot->arch.rmap; in kvmppc_core_flush_memslot_hv()
867 kvmppc_radix_flush_memslot(kvm, memslot); in kvmppc_core_flush_memslot_hv()
871 for (n = memslot->npages; n; --n, ++gfn) { in kvmppc_core_flush_memslot_hv()
879 kvm_unmap_rmapp(kvm, memslot, gfn); in kvmppc_core_flush_memslot_hv()
884 static bool kvm_age_rmapp(struct kvm *kvm, struct kvm_memory_slot *memslot, in kvm_age_rmapp() argument
893 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_age_rmapp()
955 static bool kvm_test_age_rmapp(struct kvm *kvm, struct kvm_memory_slot *memslot, in kvm_test_age_rmapp() argument
964 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_test_age_rmapp()
1101 struct kvm_memory_slot *memslot, in kvmppc_harvest_vpa_dirty() argument
1109 if (gfn < memslot->base_gfn || in kvmppc_harvest_vpa_dirty()
1110 gfn >= memslot->base_gfn + memslot->npages) in kvmppc_harvest_vpa_dirty()
1115 __set_bit_le(gfn - memslot->base_gfn, map); in kvmppc_harvest_vpa_dirty()
1119 struct kvm_memory_slot *memslot, unsigned long *map) in kvmppc_hv_get_dirty_log_hpt() argument
1125 rmapp = memslot->arch.rmap; in kvmppc_hv_get_dirty_log_hpt()
1126 for (i = 0; i < memslot->npages; ++i) { in kvmppc_hv_get_dirty_log_hpt()
1144 struct kvm_memory_slot *memslot; in kvmppc_pin_guest_page() local
1152 memslot = gfn_to_memslot(kvm, gfn); in kvmppc_pin_guest_page()
1153 if (!memslot || (memslot->flags & KVM_MEMSLOT_INVALID)) in kvmppc_pin_guest_page()
1155 hva = gfn_to_hva_memslot(memslot, gfn); in kvmppc_pin_guest_page()
1176 struct kvm_memory_slot *memslot; in kvmppc_unpin_guest_page() local
1188 memslot = gfn_to_memslot(kvm, gfn); in kvmppc_unpin_guest_page()
1189 if (memslot && memslot->dirty_bitmap) in kvmppc_unpin_guest_page()
1190 set_bit_le(gfn - memslot->base_gfn, memslot->dirty_bitmap); in kvmppc_unpin_guest_page()
1263 struct kvm_memory_slot *memslot = in resize_hpt_rehash_hpte() local
1266 if (memslot) { in resize_hpt_rehash_hpte()
1268 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in resize_hpt_rehash_hpte()
1271 kvmppc_unmap_hpte(kvm, idx, memslot, rmapp, gfn); in resize_hpt_rehash_hpte()