Lines Matching refs:pv

46 	kvm->arch.pv.handle = 0;  in kvm_s390_clear_pv_state()
47 kvm->arch.pv.guest_len = 0; in kvm_s390_clear_pv_state()
48 kvm->arch.pv.stor_base = 0; in kvm_s390_clear_pv_state()
49 kvm->arch.pv.stor_var = NULL; in kvm_s390_clear_pv_state()
67 free_pages(vcpu->arch.pv.stor_base, in kvm_s390_pv_destroy_cpu()
73 memset(&vcpu->arch.pv, 0, sizeof(vcpu->arch.pv)); in kvm_s390_pv_destroy_cpu()
98 vcpu->arch.pv.stor_base = __get_free_pages(GFP_KERNEL_ACCOUNT, in kvm_s390_pv_create_cpu()
100 if (!vcpu->arch.pv.stor_base) in kvm_s390_pv_create_cpu()
107 uvcb.stor_origin = virt_to_phys((void *)vcpu->arch.pv.stor_base); in kvm_s390_pv_create_cpu()
112 free_pages(vcpu->arch.pv.stor_base, in kvm_s390_pv_create_cpu()
134 vcpu->arch.pv.handle = uvcb.cpu_handle; in kvm_s390_pv_create_cpu()
145 vfree(kvm->arch.pv.stor_var); in kvm_s390_pv_dealloc_vm()
146 free_pages(kvm->arch.pv.stor_base, in kvm_s390_pv_dealloc_vm()
157 kvm->arch.pv.stor_var = NULL; in kvm_s390_pv_alloc_vm()
158 kvm->arch.pv.stor_base = __get_free_pages(GFP_KERNEL_ACCOUNT, get_order(base)); in kvm_s390_pv_alloc_vm()
159 if (!kvm->arch.pv.stor_base) in kvm_s390_pv_alloc_vm()
172 kvm->arch.pv.guest_len = npages * PAGE_SIZE; in kvm_s390_pv_alloc_vm()
177 kvm->arch.pv.stor_var = vzalloc(vlen); in kvm_s390_pv_alloc_vm()
178 if (!kvm->arch.pv.stor_var) in kvm_s390_pv_alloc_vm()
315 if (kvm->arch.pv.set_aside) in kvm_s390_pv_set_aside()
324 priv->stor_var = kvm->arch.pv.stor_var; in kvm_s390_pv_set_aside()
325 priv->stor_base = kvm->arch.pv.stor_base; in kvm_s390_pv_set_aside()
340 kvm->arch.pv.set_aside = priv; in kvm_s390_pv_set_aside()
420 if (kvm->arch.pv.set_aside) { in kvm_s390_pv_deinit_cleanup_all()
421 list_add(kvm->arch.pv.set_aside, &kvm->arch.pv.need_cleanup); in kvm_s390_pv_deinit_cleanup_all()
422 kvm->arch.pv.set_aside = NULL; in kvm_s390_pv_deinit_cleanup_all()
426 while (!list_empty(&kvm->arch.pv.need_cleanup)) { in kvm_s390_pv_deinit_cleanup_all()
427 cur = list_first_entry(&kvm->arch.pv.need_cleanup, typeof(*cur), list); in kvm_s390_pv_deinit_cleanup_all()
482 p = kvm->arch.pv.set_aside; in kvm_s390_pv_deinit_aside_vm()
483 kvm->arch.pv.set_aside = NULL; in kvm_s390_pv_deinit_aside_vm()
502 list_add(&p->list, &kvm->arch.pv.need_cleanup); in kvm_s390_pv_deinit_aside_vm()
514 struct kvm *kvm = container_of(subscription, struct kvm, arch.pv.mmu_notifier); in kvm_s390_pv_mmu_notifier_release()
549 uvcb.guest_stor_len = kvm->arch.pv.guest_len; in kvm_s390_pv_init_vm()
553 virt_to_phys((void *)kvm->arch.pv.stor_base); in kvm_s390_pv_init_vm()
554 uvcb.conf_virt_stor_origin = (u64)kvm->arch.pv.stor_var; in kvm_s390_pv_init_vm()
563 kvm->arch.pv.handle = uvcb.guest_handle; in kvm_s390_pv_init_vm()
577 if (kvm->arch.pv.mmu_notifier.ops != &kvm_s390_pv_mmu_notifier_ops) { in kvm_s390_pv_init_vm()
578 kvm->arch.pv.mmu_notifier.ops = &kvm_s390_pv_mmu_notifier_ops; in kvm_s390_pv_init_vm()
579 mmu_notifier_register(&kvm->arch.pv.mmu_notifier, kvm->mm); in kvm_s390_pv_init_vm()
678 .cpu_handle = vcpu->arch.pv.handle, in kvm_s390_pv_dump_cpu()
726 .config_handle = kvm->arch.pv.handle, in kvm_s390_pv_dump_stor_state()
857 kvm->arch.pv.dumping = false; in kvm_s390_pv_dump_complete()