Searched refs:is_guest_mode (Results 1 – 19 of 19) sorted by relevance
182 bool is_guest_mode) in kvm_hv_get_tlb_flush_fifo() argument185 int i = is_guest_mode ? HV_L2_TLB_FLUSH_FIFO : in kvm_hv_get_tlb_flush_fifo()198 tlb_flush_fifo = kvm_hv_get_tlb_flush_fifo(vcpu, is_guest_mode(vcpu)); in kvm_hv_vcpu_purge_flush_tlb()
87 if (!is_guest_mode(v) && kvm_vcpu_apicv_active(v)) in kvm_cpu_has_injectable_intr()
210 static inline bool is_guest_mode(struct kvm_vcpu *vcpu) in is_guest_mode() function
1923 tlb_flush_fifo = kvm_hv_get_tlb_flush_fifo(vcpu, is_guest_mode(vcpu)); in kvm_hv_vcpu_flush_tlb()1986 if (!hc->fast && is_guest_mode(vcpu)) { in kvm_hv_flush_tlb()2008 is_guest_mode(vcpu)); in kvm_hv_flush_tlb()2039 flush_ex.flags, is_guest_mode(vcpu)); in kvm_hv_flush_tlb()2083 if (all_cpus && !is_guest_mode(vcpu)) { in kvm_hv_flush_tlb()2091 } else if (!is_guest_mode(vcpu)) { in kvm_hv_flush_tlb()2330 if (hv_result_success(result) && is_guest_mode(vcpu) && in kvm_hv_hypercall_complete()
224 bool (*is_guest_mode)(struct x86_emulate_ctxt *ctxt); member
692 if (!is_guest_mode(vcpu)) in kvm_multiple_exception()2578 if (is_guest_mode(vcpu)) in kvm_vcpu_write_tsc_offset()2594 if (is_guest_mode(vcpu)) in kvm_vcpu_write_tsc_multiplier()9969 if (is_guest_mode(vcpu)) in kvm_check_and_inject_events()10130 if (is_guest_mode(vcpu) && in kvm_check_and_inject_events()10315 if (is_guest_mode(vcpu)) in vcpu_scan_ioapic()10450 if (is_guest_mode(vcpu)) in vcpu_enter_guest()10803 if (is_guest_mode(vcpu)) { in vcpu_block()11897 if (is_guest_mode(vcpu)) in kvm_vcpu_reset()12731 if (is_guest_mode(vcpu) && in kvm_vcpu_has_events()[all …]
5136 bool is_guest_mode = ctxt->ops->is_guest_mode(ctxt); in x86_emulate_insn() local5184 if (unlikely(is_guest_mode) && ctxt->intercept) { in x86_emulate_insn()5213 if (unlikely(is_guest_mode) && (ctxt->d & Intercept)) { in x86_emulate_insn()5267 if (unlikely(is_guest_mode) && (ctxt->d & Intercept)) { in x86_emulate_insn()
3220 if (is_guest_mode(vcpu)) { in kvm_apic_accept_events()
875 if (is_guest_mode(vcpu)) in vmx_update_exception_bitmap()1720 if (!is_guest_mode(vcpu)) in vmx_update_emulated_instruction()2186 if (is_guest_mode(vcpu)) in vmx_set_msr()3184 if (is_guest_mode(vcpu)) in vmx_get_current_vpid()4194 if (is_guest_mode(vcpu) && in vmx_deliver_nested_posted_interrupt()4425 if (is_guest_mode(vcpu)) { in vmx_refresh_apicv_exec_ctrl()5382 if (is_guest_mode(vcpu)) { in handle_set_cr0()5415 if (is_guest_mode(vcpu)) { in handle_set_cr4()6676 if (is_guest_mode(vcpu)) in vmx_update_cr8_intercept()6808 if (!is_guest_mode(vcpu)) in vmx_hwapic_irr_update()[all …]
3306 if (WARN_ON_ONCE(!is_guest_mode(vcpu))) in nested_vmx_write_pml_buffer()5359 (is_guest_mode(vcpu) && in handle_vmread()5367 if (!is_guest_mode(vcpu) && is_vmcs12_ext_field(field)) in handle_vmread()5383 if (WARN_ON_ONCE(is_guest_mode(vcpu))) in handle_vmread()5469 (is_guest_mode(vcpu) && in handle_vmwrite()5503 if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) in handle_vmwrite()5870 if (WARN_ON_ONCE(!is_guest_mode(vcpu))) { in handle_vmfunc()6431 if (is_guest_mode(vcpu) && in vmx_get_nested_state()6443 if (is_guest_mode(vcpu)) { in vmx_get_nested_state()6478 if (is_guest_mode(vcpu)) { in vmx_get_nested_state()[all …]
493 if (!vmcs12 && is_guest_mode(vcpu)) in vmx_write_encls_bitmap()
720 return enable_unrestricted_guest && (!is_guest_mode(vcpu) || in is_unrestricted_guest()
112 if (is_guest_mode(&svm->vcpu) && in avic_deactivate_vmcb()522 if (is_guest_mode(vcpu)) in avic_vcpu_get_apicv_inhibit_reasons()906 pi.is_guest_mode = true; in avic_pi_update_irte()917 if (!ret && pi.is_guest_mode) in avic_pi_update_irte()929 pi.is_guest_mode = false; in avic_pi_update_irte()
934 if (is_guest_mode(vcpu)) in svm_enable_lbrv()952 if (is_guest_mode(vcpu)) in svm_disable_lbrv()2277 if (is_guest_mode(vcpu)) { in emulate_svm_instr()2319 if (!is_guest_mode(vcpu)) in gp_interception()2913 if (is_guest_mode(vcpu)) in svm_set_msr()3413 if (is_guest_mode(vcpu)) { in svm_handle_exit()3642 if (is_guest_mode(vcpu)) { in svm_interrupt_blocked()3703 if (!is_guest_mode(vcpu)) in svm_enable_irq_window()4029 if (is_guest_mode(vcpu)) { in svm_vcpu_run()4060 if (is_guest_mode(vcpu)) in svm_vcpu_run()[all …]
130 if (!is_guest_mode(&svm->vcpu)) in recalc_intercepts()1178 if (is_guest_mode(vcpu)) { in svm_leave_nested()1544 if (is_guest_mode(vcpu)) { in svm_get_nested_state()1559 if (!is_guest_mode(vcpu)) in svm_get_nested_state()1681 if (is_guest_mode(vcpu)) in svm_set_nested_state()1726 if (WARN_ON(!is_guest_mode(vcpu))) in svm_get_nested_state_pages()
501 if (is_guest_mode(&svm->vcpu) && !nested_vgif_enabled(svm)) in get_vgif_vmcb()590 return is_guest_mode(vcpu) && (svm->nested.ctl.int_ctl & V_INTR_MASKING_MASK); in nested_svm_virtualize_tpr()
24 bool is_guest_mode; member
3569 pi_data->is_guest_mode = false; in amd_ir_set_vcpu_affinity()3573 if (pi_data->is_guest_mode) { in amd_ir_set_vcpu_affinity()
4224 if (is_guest_mode(vcpu)) { in __kvm_faultin_pfn()5082 role.base.guest_mode = is_guest_mode(vcpu); in kvm_calc_cpu_role()5701 if (!mmio_info_in_cache(vcpu, cr2_or_gpa, direct) && !is_guest_mode(vcpu)) in kvm_mmu_page_fault()
Completed in 117 milliseconds