Home
last modified time | relevance | path

Searched refs:vcpu (Results 1 – 25 of 444) sorted by relevance

12345678910>>...18

/linux-6.3-rc2/arch/arm64/include/asm/
A Dkvm_emulate.h65 struct kvm *kvm = vcpu->kvm; in vcpu_el1_is_32bit()
98 if (vcpu_el1_is_32bit(vcpu)) in vcpu_reset_hcr()
101 if (kvm_has_mte(vcpu->kvm)) in vcpu_reset_hcr()
122 vcpu->arch.hcr_el2 |= HCR_TWE; in vcpu_set_wfx_traps()
138 return vcpu->arch.vsesr_el2; in vcpu_get_vsesr()
143 vcpu->arch.vsesr_el2 = vsesr; in vcpu_set_vsesr()
163 if (vcpu_mode_is_32bit(vcpu)) in kvm_condition_valid()
393 return kvm_vcpu_trap_is_iabt(vcpu) && !kvm_vcpu_abt_iss1tw(vcpu); in kvm_vcpu_trap_is_exec_fault()
486 if (vcpu_mode_priv(vcpu)) in kvm_vcpu_is_be()
496 if (kvm_vcpu_is_be(vcpu)) { in vcpu_data_guest_to_host()
[all …]
/linux-6.3-rc2/arch/s390/kvm/
A Dpriv.c35 vcpu->stat.instruction_ri++; in handle_ri()
40 kvm_s390_retry_instr(vcpu); in handle_ri()
49 return handle_ri(vcpu); in kvm_s390_handle_aa()
67 vcpu->arch.gs_enabled = 1; in handle_gs()
79 return handle_gs(vcpu); in kvm_s390_handle_e3()
445 wait_event(vcpu->kvm->arch.ipte_wq, !ipte_lock_held(vcpu->kvm)); in handle_ipte_interlock()
465 return kvm_s390_inject_prog_irq(vcpu, &vcpu->arch.pgm); in handle_test_block()
496 inti = kvm_s390_get_io_int(vcpu->kvm, vcpu->arch.sie_block->gcr[6], 0); in handle_tpi()
575 vcpu->run->s390_tsch.ipb = vcpu->arch.sie_block->ipb; in handle_tsch()
1020 vcpu->run->s.regs.gprs[reg1] |= vcpu->arch.sie_block->gpsw.mask >> 32; in handle_epsw()
[all …]
A Dintercept.c87 kvm_s390_vcpu_stop(vcpu); in handle_stop()
216 if (!IS_TE_ENABLED(vcpu) || !IS_ITDB_VALID(vcpu)) in handle_itdb()
229 #define per_event(vcpu) (vcpu->arch.sie_block->iprcc & PGM_PER) argument
245 if (guestdbg_enabled(vcpu) && per_event(vcpu)) { in handle_prog()
254 trace_kvm_s390_intercept_prog(vcpu, vcpu->arch.sie_block->iprcc); in handle_prog()
263 rc = handle_itdb(vcpu); in handle_prog()
335 rc = guest_translate_address_with_key(vcpu, vcpu->run->s.regs.gprs[reg2], in handle_mvpg_pei()
344 rc = guest_translate_address_with_key(vcpu, vcpu->run->s.regs.gprs[reg1], in handle_mvpg_pei()
359 vcpu->stat.exit_pei++; in handle_partial_execution()
434 trace_kvm_s390_handle_operexc(vcpu, vcpu->arch.sie_block->ipa, in handle_operexc()
[all …]
A Ddiag.c25 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages()
26 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; in diag_release_pages()
76 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service()
80 rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); in __diag_page_ref_service()
149 kvm_vcpu_on_spin(vcpu, true); in __diag_time_slice_end()
171 tid = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in __diag_time_slice_end_directed()
175 if (tid == vcpu->vcpu_id) in __diag_time_slice_end_directed()
192 VCPU_EVENT(vcpu, 5, in __diag_time_slice_end_directed()
195 vcpu->stat.diag_9c_forward++; in __diag_time_slice_end_directed()
206 vcpu->stat.diag_9c_ignored++; in __diag_time_slice_end_directed()
[all …]
A Dguestdbg.c132 vcpu->arch.guestdbg.cr0 = vcpu->arch.sie_block->gcr[0]; in kvm_s390_backup_guest_per_regs()
133 vcpu->arch.guestdbg.cr9 = vcpu->arch.sie_block->gcr[9]; in kvm_s390_backup_guest_per_regs()
134 vcpu->arch.guestdbg.cr10 = vcpu->arch.sie_block->gcr[10]; in kvm_s390_backup_guest_per_regs()
135 vcpu->arch.guestdbg.cr11 = vcpu->arch.sie_block->gcr[11]; in kvm_s390_backup_guest_per_regs()
140 vcpu->arch.sie_block->gcr[0] = vcpu->arch.guestdbg.cr0; in kvm_s390_restore_guest_per_regs()
141 vcpu->arch.sie_block->gcr[9] = vcpu->arch.guestdbg.cr9; in kvm_s390_restore_guest_per_regs()
142 vcpu->arch.sie_block->gcr[10] = vcpu->arch.guestdbg.cr10; in kvm_s390_restore_guest_per_regs()
143 vcpu->arch.sie_block->gcr[11] = vcpu->arch.guestdbg.cr11; in kvm_s390_restore_guest_per_regs()
590 if (debug_exit_required(vcpu, vcpu->arch.sie_block->perc, in kvm_s390_handle_per_event()
614 (pssec(vcpu) || hssec(vcpu))) in kvm_s390_handle_per_event()
[all …]
A Dkvm-s390.h24 #define IS_TE_ENABLED(vcpu) ((vcpu->arch.sie_block->ecb & ECB_TE)) argument
26 #define IS_ITDB_VALID(vcpu) \ argument
83 return test_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in is_vcpu_idle()
105 VCPU_EVENT(vcpu, 3, "set prefix of cpu %03u to 0x%x", vcpu->vcpu_id, in kvm_s390_set_prefix()
270 return vcpu->arch.pv.handle; in kvm_s390_pv_cpu_get_handle()
338 kvm_s390_rewind_psw(vcpu, kvm_s390_get_ilen(vcpu)); in kvm_s390_retry_instr()
393 struct kvm_vcpu *vcpu; in kvm_s390_vcpu_block_all() local
397 kvm_s390_vcpu_block(vcpu); in kvm_s390_vcpu_block_all()
403 struct kvm_vcpu *vcpu; in kvm_s390_vcpu_unblock_all() local
406 kvm_s390_vcpu_unblock(vcpu); in kvm_s390_vcpu_unblock_all()
[all …]
/linux-6.3-rc2/arch/arm64/kvm/
A Ddebug.c111 if (vcpu->guest_debug) in kvm_arm_setup_mdcr_el2()
140 kvm_arm_setup_mdcr_el2(vcpu); in kvm_arm_vcpu_init_debug()
150 vcpu->arch.debug_ptr = &vcpu->arch.vcpu_debug_state; in kvm_arm_reset_debug_ptr()
173 trace_kvm_arm_setup_debug(vcpu, vcpu->guest_debug); in kvm_arm_setup_debug()
175 kvm_arm_setup_mdcr_el2(vcpu); in kvm_arm_setup_debug()
178 if (vcpu->guest_debug || kvm_vcpu_os_lock_enabled(vcpu)) { in kvm_arm_setup_debug()
180 save_guest_debug_regs(vcpu); in kvm_arm_setup_debug()
240 vcpu->arch.debug_ptr = &vcpu->arch.external_debug_state; in kvm_arm_setup_debug()
266 BUG_ON(!vcpu->guest_debug && in kvm_arm_setup_debug()
267 vcpu->arch.debug_ptr != &vcpu->arch.vcpu_debug_state); in kvm_arm_setup_debug()
[all …]
A Dinject_fault.c21 if (likely(!vcpu_has_nv(vcpu))) { in pend_sync_exception()
42 if (vcpu_el2_tge_is_set(vcpu)) in pend_sync_exception()
63 pend_sync_exception(vcpu); in inject_abt64()
99 pend_sync_exception(vcpu); in inject_undef64()
168 if (vcpu_el1_is_32bit(vcpu)) in kvm_inject_dabt()
184 if (vcpu_el1_is_32bit(vcpu)) in kvm_inject_pabt()
198 kvm_inject_pabt(vcpu, addr); in kvm_inject_size_fault()
227 if (vcpu_el1_is_32bit(vcpu)) in kvm_inject_undefined()
228 inject_undef32(vcpu); in kvm_inject_undefined()
230 inject_undef64(vcpu); in kvm_inject_undefined()
[all …]
A Dreset.c135 if (!vcpu_has_sve(vcpu)) in kvm_arm_vcpu_finalize()
149 if (vcpu_has_sve(vcpu) && !kvm_arm_vcpu_sve_finalized(vcpu)) in kvm_arm_vcpu_is_finalized()
160 kvm_unshare_hyp(vcpu, vcpu + 1); in kvm_arm_vcpu_destroy()
164 kfree(vcpu->arch.ccsidr); in kvm_arm_vcpu_destroy()
169 if (vcpu_has_sve(vcpu)) in kvm_vcpu_reset_sve()
170 memset(vcpu->arch.sve_state, 0, vcpu_sve_state_size(vcpu)); in kvm_vcpu_reset_sve()
285 if (vcpu_has_nv(vcpu) && vcpu_has_feature(vcpu, KVM_ARM_VCPU_SVE)) { in kvm_reset_vcpu()
326 memset(vcpu_gp_regs(vcpu), 0, sizeof(*vcpu_gp_regs(vcpu))); in kvm_reset_vcpu()
327 memset(&vcpu->arch.ctxt.fp_regs, 0, sizeof(vcpu->arch.ctxt.fp_regs)); in kvm_reset_vcpu()
347 vcpu_set_thumb(vcpu); in kvm_reset_vcpu()
[all …]
A Dpsci.c49 kvm_vcpu_wfi(vcpu); in kvm_psci_vcpu_suspend()
64 struct kvm_vcpu *vcpu = NULL; in kvm_psci_vcpu_on() local
77 if (!vcpu) in kvm_psci_vcpu_on()
109 kvm_vcpu_wake_up(vcpu); in kvm_psci_vcpu_on()
121 struct kvm *kvm = vcpu->kvm; in kvm_psci_vcpu_affinity_info()
175 memset(&vcpu->run->system_event, 0, sizeof(vcpu->run->system_event)); in kvm_prepare_system_event()
216 vcpu_set_reg(vcpu, i, lower_32_bits(vcpu_get_reg(vcpu, i))); in kvm_psci_narrow_to_32bit()
232 struct kvm *kvm = vcpu->kvm; in kvm_psci_0_2_call()
276 kvm_psci_system_off(vcpu); in kvm_psci_0_2_call()
312 struct kvm *kvm = vcpu->kvm; in kvm_psci_1_x_call()
[all …]
/linux-6.3-rc2/arch/powerpc/kvm/
A Dbook3s_emulate.c92 vcpu->arch.ppr_tm = vcpu->arch.ppr; in kvmppc_copyto_vcpu_tm()
94 vcpu->arch.amr_tm = vcpu->arch.amr; in kvmppc_copyto_vcpu_tm()
96 vcpu->arch.tar_tm = vcpu->arch.tar; in kvmppc_copyto_vcpu_tm()
111 vcpu->arch.ppr = vcpu->arch.ppr_tm; in kvmppc_copyfrom_vcpu_tm()
113 vcpu->arch.amr = vcpu->arch.amr_tm; in kvmppc_copyfrom_vcpu_tm()
115 vcpu->arch.tar = vcpu->arch.tar_tm; in kvmppc_copyfrom_vcpu_tm()
337 vcpu->arch.mmu.mtsrin(vcpu, in kvmppc_core_emulate_op_pr()
342 vcpu->arch.mmu.mtsrin(vcpu, in kvmppc_core_emulate_op_pr()
388 vcpu->arch.mmu.slbmte(vcpu, in kvmppc_core_emulate_op_pr()
396 vcpu->arch.mmu.slbie(vcpu, in kvmppc_core_emulate_op_pr()
[all …]
A Dbooke.c510 set_guest_srr(vcpu, vcpu->arch.regs.nip, in kvmppc_booke_irqprio_deliver()
514 set_guest_csrr(vcpu, vcpu->arch.regs.nip, in kvmppc_booke_irqprio_deliver()
518 set_guest_dsrr(vcpu, vcpu->arch.regs.nip, in kvmppc_booke_irqprio_deliver()
522 set_guest_mcsrr(vcpu, vcpu->arch.regs.nip, in kvmppc_booke_irqprio_deliver()
527 vcpu->arch.regs.nip = vcpu->arch.ivpr | in kvmppc_booke_irqprio_deliver()
802 vcpu->arch.pgdir = vcpu->kvm->mm->pgd; in kvmppc_vcpu_run()
1887 vcpu->arch.dec = vcpu->arch.decar; in kvmppc_decrementer_func()
2132 vcpu->arch.shared->pir = vcpu->vcpu_id; in kvmppc_core_vcpu_create()
2152 vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu); in kvmppc_core_vcpu_create()
2159 vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu); in kvmppc_core_vcpu_free()
[all …]
A Dbooke_emulate.c26 vcpu->arch.regs.nip = vcpu->arch.shared->srr0; in kvmppc_emul_rfi()
27 kvmppc_set_msr(vcpu, vcpu->arch.shared->srr1); in kvmppc_emul_rfi()
32 vcpu->arch.regs.nip = vcpu->arch.dsrr0; in kvmppc_emul_rfdi()
33 kvmppc_set_msr(vcpu, vcpu->arch.dsrr1); in kvmppc_emul_rfdi()
38 vcpu->arch.regs.nip = vcpu->arch.csrr0; in kvmppc_emul_rfci()
39 kvmppc_set_msr(vcpu, vcpu->arch.csrr1); in kvmppc_emul_rfci()
53 kvmppc_emul_rfi(vcpu); in kvmppc_booke_emulate_op()
80 kvmppc_set_gpr(vcpu, rt, vcpu->arch.shared->msr); in kvmppc_booke_emulate_op()
86 kvmppc_set_msr(vcpu, kvmppc_get_gpr(vcpu, rs)); in kvmppc_booke_emulate_op()
90 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) in kvmppc_booke_emulate_op()
[all …]
A Dbook3s_pr.c307 vcpu->arch.purr += get_tb() - vcpu->arch.entry_tb; in kvmppc_copy_from_svcpu()
308 vcpu->arch.spurr += get_tb() - vcpu->arch.entry_tb; in kvmppc_copy_from_svcpu()
411 _kvmppc_restore_tm_pr(vcpu, kvmppc_get_msr(vcpu)); in kvmppc_restore_tm_pr()
742 if (vcpu->arch.mmu.is_dcbz32(vcpu) && in kvmppc_handle_pagefault()
787 else if (vcpu->arch.mmu.is_dcbz32(vcpu) && in kvmppc_handle_pagefault()
1189 } else if (vcpu->arch.mmu.is_dcbz32(vcpu) && in kvmppc_handle_exit_pr()
1335 kvmppc_set_gpr(vcpu, 3, kvmppc_kvm_pv(vcpu)); in kvmppc_handle_exit_pr()
1499 vcpu->arch.mmu.slbmte(vcpu, 0, 0); in kvm_arch_vcpu_ioctl_set_sregs_pr()
1500 vcpu->arch.mmu.slbia(vcpu); in kvm_arch_vcpu_ioctl_set_sregs_pr()
1507 vcpu->arch.mmu.slbmte(vcpu, rs, rb); in kvm_arch_vcpu_ioctl_set_sregs_pr()
[all …]
A Demulate_loadstore.c86 vcpu->arch.mmio_vsx_offset = 0; in kvmppc_emulate_loadstore()
91 vcpu->arch.mmio_vmx_offset = 0; in kvmppc_emulate_loadstore()
95 vcpu->arch.regs.msr = vcpu->arch.shared->msr; in kvmppc_emulate_loadstore()
148 vcpu->arch.mmio_copy_type = in kvmppc_emulate_loadstore()
151 vcpu->arch.mmio_copy_type = in kvmppc_emulate_loadstore()
154 vcpu->arch.mmio_copy_type = in kvmppc_emulate_loadstore()
157 vcpu->arch.mmio_copy_type = in kvmppc_emulate_loadstore()
246 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, in kvmppc_emulate_loadstore()
270 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, in kvmppc_emulate_loadstore()
315 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, in kvmppc_emulate_loadstore()
[all …]
A Dbook3s_hv_tm.c29 vcpu->arch.tfiar = tfiar; in emulate_tx_failure()
31 vcpu->arch.texasr = (vcpu->arch.texasr & 0x3ffffff) | texasr; in emulate_tx_failure()
56 vcpu->arch.regs.nip -= 4; in kvmhv_p9_tm_emulation()
79 vcpu->arch.cfar = vcpu->arch.regs.nip; in kvmhv_p9_tm_emulation()
80 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; in kvmhv_p9_tm_emulation()
103 bescr = vcpu->arch.bescr; in kvmhv_p9_tm_emulation()
113 vcpu->arch.cfar = vcpu->arch.regs.nip; in kvmhv_p9_tm_emulation()
114 vcpu->arch.regs.nip = vcpu->arch.ebbrr; in kvmhv_p9_tm_emulation()
156 vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) | in kvmhv_p9_tm_emulation()
203 vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) | in kvmhv_p9_tm_emulation()
[all …]
A Dbook3s.c516 ret = vcpu->kvm->arch.kvm_ops->get_sregs(vcpu, sregs); in kvm_arch_vcpu_ioctl_get_sregs()
528 ret = vcpu->kvm->arch.kvm_ops->set_sregs(vcpu, sregs); in kvm_arch_vcpu_ioctl_set_sregs()
638 if (!vcpu->arch.icp && !vcpu->arch.xive_vcpu) { in kvmppc_get_one_reg()
723 if (!vcpu->arch.icp && !vcpu->arch.xive_vcpu) { in kvmppc_set_one_reg()
774 vcpu->kvm->arch.kvm_ops->vcpu_load(vcpu, cpu); in kvmppc_core_vcpu_load()
779 vcpu->kvm->arch.kvm_ops->vcpu_put(vcpu); in kvmppc_core_vcpu_put()
784 vcpu->kvm->arch.kvm_ops->set_msr(vcpu, msr); in kvmppc_set_msr()
790 return vcpu->kvm->arch.kvm_ops->vcpu_run(vcpu); in kvmppc_vcpu_run()
816 return vcpu->kvm->arch.kvm_ops->vcpu_create(vcpu); in kvmppc_core_vcpu_create()
821 vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu); in kvmppc_core_vcpu_free()
[all …]
A Dbook3s_paired_singles.c152 kvm_cvt_df(&VCPU_FPR(vcpu, rt), &vcpu->arch.qpr[rt]); in kvmppc_sync_qpr()
757 VCPU_FPR(vcpu, ax_rd) = VCPU_FPR(vcpu, ax_rb); in kvmppc_emulate_paired_single()
759 vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rb]; in kvmppc_emulate_paired_single()
768 VCPU_FPR(vcpu, ax_rd) = VCPU_FPR(vcpu, ax_rb); in kvmppc_emulate_paired_single()
769 vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rb]; in kvmppc_emulate_paired_single()
777 VCPU_FPR(vcpu, ax_rd) = VCPU_FPR(vcpu, ax_rb); in kvmppc_emulate_paired_single()
779 vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rb]; in kvmppc_emulate_paired_single()
784 VCPU_FPR(vcpu, ax_rd) = VCPU_FPR(vcpu, ax_rb); in kvmppc_emulate_paired_single()
786 vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rb]; in kvmppc_emulate_paired_single()
791 VCPU_FPR(vcpu, ax_rd) = VCPU_FPR(vcpu, ax_ra); in kvmppc_emulate_paired_single()
[all …]
/linux-6.3-rc2/arch/mips/kvm/
A Demulate.c248 err = kvm_compute_return_epc(vcpu, vcpu->arch.pc, in update_pc()
1003 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1013 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1022 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1031 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1061 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1091 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1139 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1186 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1290 vcpu->arch.io_pc = vcpu->arch.pc; in kvm_mips_emulate_load()
[all …]
/linux-6.3-rc2/arch/x86/kvm/
A Dx86.c964 !load_pdptrs(vcpu, kvm_read_cr3(vcpu))) in kvm_set_cr0()
1020 if (vcpu->arch.pkru != vcpu->arch.host_pkru) in kvm_load_host_xsave_state()
1094 __kvm_set_xcr(vcpu, kvm_rcx_read(vcpu), kvm_read_edx_eax(vcpu))) { in kvm_emulate_xsetbv()
1318 vcpu->arch.eff_db[i] = vcpu->arch.db[i]; in kvm_update_dr0123()
2603 vcpu, vcpu->arch.tsc_scaling_ratio); in kvm_vcpu_write_tsc_multiplier()
8310 ctxt->vcpu = vcpu; in alloc_emulate_ctxt()
10688 vcpu->arch.last_vmentry_cpu = vcpu->cpu; in vcpu_enter_guest()
11412 load_pdptrs(vcpu, kvm_read_cr3(vcpu)); in __set_sregs_common()
11671 vcpu, &vcpu->run->s.regs.events); in store_regs()
11687 vcpu, &vcpu->run->s.regs.events)) in sync_regs()
[all …]
A Dkvm_cache_regs.h103 return vcpu->arch.regs[reg]; in kvm_register_read_raw()
112 vcpu->arch.regs[reg] = val; in kvm_register_write_raw()
113 kvm_register_mark_dirty(vcpu, reg); in kvm_register_write_raw()
157 return vcpu->arch.cr0 & mask; in kvm_read_cr0_bits()
171 return vcpu->arch.cr4 & mask; in kvm_read_cr4_bits()
178 return vcpu->arch.cr3; in kvm_read_cr3()
188 return (kvm_rax_read(vcpu) & -1u) in kvm_read_edx_eax()
194 vcpu->arch.hflags |= HF_GUEST_MASK; in enter_guest_mode()
195 vcpu->stat.guest_mode = 1; in enter_guest_mode()
200 vcpu->arch.hflags &= ~HF_GUEST_MASK; in leave_guest_mode()
[all …]
A Dx86.h115 return vcpu->arch.exception.injected || vcpu->arch.interrupt.injected || in kvm_event_needs_reinjection()
116 vcpu->arch.nmi_injected; in kvm_event_needs_reinjection()
144 if (!is_long_mode(vcpu)) in is_64_bit_mode()
157 return vcpu->arch.guest_state_protected || is_64_bit_mode(vcpu); in is_64_bit_hypercall()
171 return vcpu->arch.walk_mmu == &vcpu->arch.nested_mmu; in mmu_is_nested()
191 return !is_long_mode(vcpu) && is_pae(vcpu) && is_paging(vcpu); in is_pae_paging()
216 vcpu->arch.mmio_gva = mmu_is_nested(vcpu) ? 0 : gva & PAGE_MASK; in vcpu_cache_mmio_info()
224 return vcpu->arch.mmio_gen == kvm_memslots(vcpu->kvm)->generation; in vcpu_match_mmio_gen()
238 vcpu->arch.mmio_gva = 0; in vcpu_clear_mmio_info()
243 if (vcpu_match_mmio_gen(vcpu) && vcpu->arch.mmio_gva && in vcpu_match_mmio_gva()
[all …]
/linux-6.3-rc2/arch/riscv/kvm/
A Dvcpu.c145 memset(vcpu->arch.hfence_queue, 0, sizeof(vcpu->arch.hfence_queue)); in kvm_riscv_reset_vcpu()
246 !vcpu->arch.power_off && !vcpu->arch.pause); in kvm_arch_vcpu_runnable()
898 vcpu->cpu = cpu; in kvm_arch_vcpu_load()
905 vcpu->cpu = -1; in kvm_arch_vcpu_put()
932 (!vcpu->arch.power_off) && (!vcpu->arch.pause), in kvm_riscv_check_vcpu_requests()
936 if (vcpu->arch.power_off || vcpu->arch.pause) { in kvm_riscv_check_vcpu_requests()
987 vcpu->arch.last_exit_cpu = vcpu->cpu; in kvm_riscv_vcpu_enter_exit()
1005 ret = kvm_riscv_vcpu_mmio_return(vcpu, vcpu->run); in kvm_arch_vcpu_ioctl_run()
1009 ret = kvm_riscv_vcpu_sbi_return(vcpu, vcpu->run); in kvm_arch_vcpu_ioctl_run()
1013 ret = kvm_riscv_vcpu_csr_return(vcpu, vcpu->run); in kvm_arch_vcpu_ioctl_run()
[all …]
/linux-6.3-rc2/arch/x86/kvm/vmx/
A Dsgx.c32 if (!is_long_mode(vcpu)) { in sgx_get_encls_gva()
50 kvm_inject_gp(vcpu, 0); in sgx_get_encls_gva()
135 kvm_inject_gp(vcpu, 0); in sgx_inject_fault()
169 kvm_inject_gp(vcpu, 0); in __handle_encls_ecreate()
179 kvm_inject_gp(vcpu, 0); in __handle_encls_ecreate()
187 kvm_inject_gp(vcpu, 0); in __handle_encls_ecreate()
220 if (sgx_get_encls_gva(vcpu, kvm_rbx_read(vcpu), 32, 32, &pageinfo_gva) || in handle_encls_ecreate()
221 sgx_get_encls_gva(vcpu, kvm_rcx_read(vcpu), 4096, 4096, &secs_gva)) in handle_encls_ecreate()
297 if (sgx_get_encls_gva(vcpu, kvm_rbx_read(vcpu), 1808, 4096, &sig_gva) || in handle_encls_einit()
298 sgx_get_encls_gva(vcpu, kvm_rcx_read(vcpu), 4096, 4096, &secs_gva) || in handle_encls_einit()
[all …]
/linux-6.3-rc2/include/kvm/
A Darm_pmu.h53 void kvm_pmu_vcpu_init(struct kvm_vcpu *vcpu);
54 void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu);
55 void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu);
59 void kvm_pmu_sync_hwstate(struct kvm_vcpu *vcpu);
61 void kvm_pmu_update_run(struct kvm_vcpu *vcpu);
78 #define kvm_vcpu_has_pmu(vcpu) \ argument
86 #define kvm_pmu_update_vcpu_events(vcpu) \ argument
88 if (!has_vhe() && kvm_vcpu_has_pmu(vcpu)) \
95 #define kvm_pmu_is_3p5(vcpu) \ argument
161 #define kvm_vcpu_has_pmu(vcpu) ({ false; }) argument
[all …]

Completed in 134 milliseconds

12345678910>>...18