Lines Matching refs:kvm_x86_ops
133 struct kvm_x86_ops kvm_x86_ops __read_mostly;
137 *(((struct kvm_x86_ops *)0)->func));
637 kvm_x86_ops.nested_ops->leave_nested(vcpu); in kvm_leave_nested()
657 kvm_x86_ops.nested_ops->is_exception_vmexit(vcpu, nr, error_code)) { in kvm_multiple_exception()
4511 r = kvm_x86_ops.nested_ops->get_state ? in kvm_vm_ioctl_check_extension()
4512 kvm_x86_ops.nested_ops->get_state(NULL, NULL, 0) : 0; in kvm_vm_ioctl_check_extension()
4515 r = kvm_x86_ops.enable_l2_tlb_flush != NULL; in kvm_vm_ioctl_check_extension()
4518 r = kvm_x86_ops.nested_ops->enable_evmcs != NULL; in kvm_vm_ioctl_check_extension()
5516 if (!kvm_x86_ops.nested_ops->enable_evmcs) in kvm_vcpu_ioctl_enable_cap()
5518 r = kvm_x86_ops.nested_ops->enable_evmcs(vcpu, &vmcs_version); in kvm_vcpu_ioctl_enable_cap()
5527 if (!kvm_x86_ops.enable_l2_tlb_flush) in kvm_vcpu_ioctl_enable_cap()
5869 if (!kvm_x86_ops.nested_ops->get_state) in kvm_arch_vcpu_ioctl()
5877 r = kvm_x86_ops.nested_ops->get_state(vcpu, user_kvm_nested_state, in kvm_arch_vcpu_ioctl()
5899 if (!kvm_x86_ops.nested_ops->set_state) in kvm_arch_vcpu_ioctl()
5922 r = kvm_x86_ops.nested_ops->set_state(vcpu, user_kvm_nested_state, &kvm_state); in kvm_arch_vcpu_ioctl()
6318 if (!kvm_x86_ops.vm_copy_enc_context_from) in kvm_vm_ioctl_enable_cap()
6325 if (!kvm_x86_ops.vm_move_enc_context_from) in kvm_vm_ioctl_enable_cap()
6958 if (!kvm_x86_ops.mem_enc_ioctl) in kvm_arch_vm_ioctl()
6972 if (!kvm_x86_ops.mem_enc_register_region) in kvm_arch_vm_ioctl()
6986 if (!kvm_x86_ops.mem_enc_unregister_region) in kvm_arch_vm_ioctl()
9317 memcpy(&kvm_x86_ops, ops->runtime_ops, sizeof(kvm_x86_ops)); in kvm_ops_update()
9320 static_call_update(kvm_x86_##func, kvm_x86_ops.func); in kvm_ops_update()
9322 WARN_ON(!kvm_x86_ops.func); __KVM_X86_OP(func) in kvm_ops_update()
9325 static_call_update(kvm_x86_##func, (void *)kvm_x86_ops.func ? : \ in kvm_ops_update()
9363 if (kvm_x86_ops.hardware_enable) { in __kvm_x86_vendor_init()
9364 pr_err("already loaded vendor module '%s'\n", kvm_x86_ops.name); in __kvm_x86_vendor_init()
9479 kvm_x86_ops.hardware_enable = NULL; in __kvm_x86_vendor_init()
9531 kvm_x86_ops.hardware_enable = NULL; in kvm_x86_vendor_exit()
9874 if (!kvm_x86_ops.update_cr8_intercept) in update_cr8_intercept()
9900 kvm_x86_ops.nested_ops->triple_fault(vcpu); in kvm_check_nested_events()
9904 return kvm_x86_ops.nested_ops->check_events(vcpu); in kvm_check_nested_events()
10131 kvm_x86_ops.nested_ops->has_events && in kvm_check_and_inject_events()
10132 kvm_x86_ops.nested_ops->has_events(vcpu)) in kvm_check_and_inject_events()
10241 kvm_x86_ops.allow_apicv_in_x2apic_without_x2apic_virtualization) in kvm_vcpu_update_apicv()
10254 if (!(kvm_x86_ops.required_apicv_inhibits & BIT(reason))) in __kvm_set_or_clear_apicv_inhibit()
10401 if (unlikely(!kvm_x86_ops.nested_ops->get_nested_state_pages(vcpu))) { in vcpu_enter_guest()
10451 kvm_x86_ops.nested_ops->triple_fault(vcpu); in vcpu_enter_guest()
11042 kvm_x86_ops.nested_ops->is_exception_vmexit(vcpu, ex->vector, in kvm_arch_vcpu_ioctl_run()
12529 if (!kvm_x86_ops.cpu_dirty_log_size) in kvm_mmu_update_cpu_dirty_logging()
12605 if (kvm_x86_ops.cpu_dirty_log_size) { in kvm_mmu_slot_apply_flags()
12732 kvm_x86_ops.nested_ops->has_events && in kvm_vcpu_has_events()
12733 kvm_x86_ops.nested_ops->has_events(vcpu)) in kvm_vcpu_has_events()