Home
last modified time | relevance | path

Searched refs:vcpu (Results 1 – 25 of 507) sorted by relevance

12345678910>>...21

/linux/arch/s390/kvm/
A Dpriv.c34 vcpu->stat.instruction_ri++; in handle_ri()
39 kvm_s390_retry_instr(vcpu); in handle_ri()
48 return handle_ri(vcpu); in kvm_s390_handle_aa()
66 vcpu->arch.gs_enabled = 1; in handle_gs()
78 return handle_gs(vcpu); in kvm_s390_handle_e3()
444 wait_event(vcpu->kvm->arch.ipte_wq, !ipte_lock_held(vcpu->kvm)); in handle_ipte_interlock()
464 return kvm_s390_inject_prog_irq(vcpu, &vcpu->arch.pgm); in handle_test_block()
495 inti = kvm_s390_get_io_int(vcpu->kvm, vcpu->arch.sie_block->gcr[6], 0); in handle_tpi()
574 vcpu->run->s390_tsch.ipb = vcpu->arch.sie_block->ipb; in handle_tsch()
1053 vcpu->run->s.regs.gprs[reg1] |= vcpu->arch.sie_block->gpsw.mask >> 32; in handle_epsw()
[all …]
A Dintercept.c216 if (!IS_TE_ENABLED(vcpu) || !IS_ITDB_VALID(vcpu)) in handle_itdb()
229 #define per_event(vcpu) (vcpu->arch.sie_block->iprcc & PGM_PER) argument
233 if (!guestdbg_enabled(vcpu) || !per_event(vcpu)) in should_handle_per_event()
269 trace_kvm_s390_intercept_prog(vcpu, vcpu->arch.sie_block->iprcc); in handle_prog()
278 rc = handle_itdb(vcpu); in handle_prog()
366 rc = guest_translate_address_with_key(vcpu, vcpu->run->s.regs.gprs[reg2], in handle_mvpg_pei()
375 rc = guest_translate_address_with_key(vcpu, vcpu->run->s.regs.gprs[reg1], in handle_mvpg_pei()
390 vcpu->stat.exit_pei++; in handle_partial_execution()
468 trace_kvm_s390_handle_operexc(vcpu, vcpu->arch.sie_block->ipa, in handle_operexc()
474 if (vcpu->arch.sie_block->ipa == 0 && vcpu->kvm->arch.user_instr0) in handle_operexc()
[all …]
A Ddiag.c25 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages()
26 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; in diag_release_pages()
76 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service()
80 rc = read_guest_real(vcpu, vcpu->run->s.regs.gprs[rx], &parm, sizeof(parm)); in __diag_page_ref_service()
149 kvm_vcpu_on_spin(vcpu, true); in __diag_time_slice_end()
172 tid = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in __diag_time_slice_end_directed()
176 if (tid == vcpu->vcpu_id) in __diag_time_slice_end_directed()
194 VCPU_EVENT(vcpu, 5, in __diag_time_slice_end_directed()
197 vcpu->stat.diag_9c_forward++; in __diag_time_slice_end_directed()
208 vcpu->stat.diag_9c_ignored++; in __diag_time_slice_end_directed()
[all …]
A Dguestdbg.c132 vcpu->arch.guestdbg.cr0 = vcpu->arch.sie_block->gcr[0]; in kvm_s390_backup_guest_per_regs()
133 vcpu->arch.guestdbg.cr9 = vcpu->arch.sie_block->gcr[9]; in kvm_s390_backup_guest_per_regs()
134 vcpu->arch.guestdbg.cr10 = vcpu->arch.sie_block->gcr[10]; in kvm_s390_backup_guest_per_regs()
135 vcpu->arch.guestdbg.cr11 = vcpu->arch.sie_block->gcr[11]; in kvm_s390_backup_guest_per_regs()
140 vcpu->arch.sie_block->gcr[0] = vcpu->arch.guestdbg.cr0; in kvm_s390_restore_guest_per_regs()
141 vcpu->arch.sie_block->gcr[9] = vcpu->arch.guestdbg.cr9; in kvm_s390_restore_guest_per_regs()
142 vcpu->arch.sie_block->gcr[10] = vcpu->arch.guestdbg.cr10; in kvm_s390_restore_guest_per_regs()
143 vcpu->arch.sie_block->gcr[11] = vcpu->arch.guestdbg.cr11; in kvm_s390_restore_guest_per_regs()
590 if (debug_exit_required(vcpu, vcpu->arch.sie_block->perc, in kvm_s390_handle_per_event()
614 (pssec(vcpu) || hssec(vcpu))) in kvm_s390_handle_per_event()
[all …]
/linux/arch/riscv/kvm/
A Dvcpu.c90 memset(vcpu->arch.hfence_queue, 0, sizeof(vcpu->arch.hfence_queue)); in kvm_riscv_reset_vcpu()
216 !kvm_riscv_vcpu_stopped(vcpu) && !vcpu->arch.pause); in kvm_arch_vcpu_runnable()
601 vcpu->cpu = cpu; in kvm_arch_vcpu_load()
608 vcpu->cpu = -1; in kvm_arch_vcpu_put()
640 (!kvm_riscv_vcpu_stopped(vcpu)) && (!vcpu->arch.pause), in kvm_riscv_check_vcpu_requests()
644 if (kvm_riscv_vcpu_stopped(vcpu) || vcpu->arch.pause) { in kvm_riscv_check_vcpu_requests()
726 vcpu->arch.last_exit_cpu = vcpu->cpu; in kvm_riscv_vcpu_enter_exit()
748 ret = kvm_riscv_vcpu_mmio_return(vcpu, vcpu->run); in kvm_arch_vcpu_ioctl_run()
752 ret = kvm_riscv_vcpu_sbi_return(vcpu, vcpu->run); in kvm_arch_vcpu_ioctl_run()
756 ret = kvm_riscv_vcpu_csr_return(vcpu, vcpu->run); in kvm_arch_vcpu_ioctl_run()
[all …]
/linux/arch/arm64/kvm/
A Ddebug.c111 if (vcpu->guest_debug) in kvm_arm_setup_mdcr_el2()
140 kvm_arm_setup_mdcr_el2(vcpu); in kvm_arm_vcpu_init_debug()
151 vcpu->arch.debug_ptr = &vcpu->arch.vcpu_debug_state; in kvm_arm_reset_debug_ptr()
174 trace_kvm_arm_setup_debug(vcpu, vcpu->guest_debug); in kvm_arm_setup_debug()
176 kvm_arm_setup_mdcr_el2(vcpu); in kvm_arm_setup_debug()
179 if (vcpu->guest_debug || kvm_vcpu_os_lock_enabled(vcpu)) { in kvm_arm_setup_debug()
181 save_guest_debug_regs(vcpu); in kvm_arm_setup_debug()
241 vcpu->arch.debug_ptr = &vcpu->arch.external_debug_state; in kvm_arm_setup_debug()
267 BUG_ON(!vcpu->guest_debug && in kvm_arm_setup_debug()
268 vcpu->arch.debug_ptr != &vcpu->arch.vcpu_debug_state); in kvm_arm_setup_debug()
[all …]
A Dinject_fault.c21 if (likely(!vcpu_has_nv(vcpu))) { in pend_sync_exception()
42 if (vcpu_el2_tge_is_set(vcpu)) in pend_sync_exception()
63 pend_sync_exception(vcpu); in inject_abt64()
99 pend_sync_exception(vcpu); in inject_undef64()
168 if (vcpu_el1_is_32bit(vcpu)) in kvm_inject_dabt()
184 if (vcpu_el1_is_32bit(vcpu)) in kvm_inject_pabt()
198 kvm_inject_pabt(vcpu, addr); in kvm_inject_size_fault()
227 if (vcpu_el1_is_32bit(vcpu)) in kvm_inject_undefined()
228 inject_undef32(vcpu); in kvm_inject_undefined()
230 inject_undef64(vcpu); in kvm_inject_undefined()
[all …]
A Dhandle_exit.c39 trace_kvm_hvc_arm64(*vcpu_pc(vcpu), vcpu_get_reg(vcpu, 0), in handle_hvc()
48 kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in handle_hvc()
107 return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in kvm_handle_fpasimd()
146 val = vcpu_get_reg(vcpu, kvm_vcpu_sys_get_rt(vcpu)); in kvm_handle_wfx()
157 kvm_vcpu_on_spin(vcpu, vcpu_mode_priv(vcpu)); in kvm_handle_wfx()
221 return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in handle_sve()
247 if (vcpu_has_nv(vcpu) && !is_hyp_ctxt(vcpu)) { in kvm_handle_ptrauth()
248 kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in kvm_handle_ptrauth()
278 kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in kvm_handle_eret()
290 kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in handle_svc()
[all …]
/linux/arch/powerpc/kvm/
A Dbook3s_emulate.c92 vcpu->arch.ppr_tm = vcpu->arch.ppr; in kvmppc_copyto_vcpu_tm()
94 vcpu->arch.amr_tm = vcpu->arch.amr; in kvmppc_copyto_vcpu_tm()
96 vcpu->arch.tar_tm = vcpu->arch.tar; in kvmppc_copyto_vcpu_tm()
111 vcpu->arch.ppr = vcpu->arch.ppr_tm; in kvmppc_copyfrom_vcpu_tm()
113 vcpu->arch.amr = vcpu->arch.amr_tm; in kvmppc_copyfrom_vcpu_tm()
115 vcpu->arch.tar = vcpu->arch.tar_tm; in kvmppc_copyfrom_vcpu_tm()
337 vcpu->arch.mmu.mtsrin(vcpu, in kvmppc_core_emulate_op_pr()
342 vcpu->arch.mmu.mtsrin(vcpu, in kvmppc_core_emulate_op_pr()
388 vcpu->arch.mmu.slbmte(vcpu, in kvmppc_core_emulate_op_pr()
396 vcpu->arch.mmu.slbie(vcpu, in kvmppc_core_emulate_op_pr()
[all …]
A Dbooke.c513 set_guest_srr(vcpu, vcpu->arch.regs.nip, in kvmppc_booke_irqprio_deliver()
517 set_guest_csrr(vcpu, vcpu->arch.regs.nip, in kvmppc_booke_irqprio_deliver()
521 set_guest_dsrr(vcpu, vcpu->arch.regs.nip, in kvmppc_booke_irqprio_deliver()
525 set_guest_mcsrr(vcpu, vcpu->arch.regs.nip, in kvmppc_booke_irqprio_deliver()
530 vcpu->arch.regs.nip = vcpu->arch.ivpr | in kvmppc_booke_irqprio_deliver()
805 vcpu->arch.pgdir = vcpu->kvm->mm->pgd; in kvmppc_vcpu_run()
1894 vcpu->arch.dec = vcpu->arch.decar; in kvmppc_decrementer_func()
2140 vcpu->arch.shared->pir = vcpu->vcpu_id; in kvmppc_core_vcpu_create()
2160 vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu); in kvmppc_core_vcpu_create()
2167 vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu); in kvmppc_core_vcpu_free()
[all …]
A Dbooke_emulate.c26 vcpu->arch.regs.nip = vcpu->arch.shared->srr0; in kvmppc_emul_rfi()
27 kvmppc_set_msr(vcpu, vcpu->arch.shared->srr1); in kvmppc_emul_rfi()
32 vcpu->arch.regs.nip = vcpu->arch.dsrr0; in kvmppc_emul_rfdi()
33 kvmppc_set_msr(vcpu, vcpu->arch.dsrr1); in kvmppc_emul_rfdi()
38 vcpu->arch.regs.nip = vcpu->arch.csrr0; in kvmppc_emul_rfci()
39 kvmppc_set_msr(vcpu, vcpu->arch.csrr1); in kvmppc_emul_rfci()
53 kvmppc_emul_rfi(vcpu); in kvmppc_booke_emulate_op()
80 kvmppc_set_gpr(vcpu, rt, vcpu->arch.shared->msr); in kvmppc_booke_emulate_op()
86 kvmppc_set_msr(vcpu, kvmppc_get_gpr(vcpu, rs)); in kvmppc_booke_emulate_op()
90 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) in kvmppc_booke_emulate_op()
[all …]
A Dbook3s_pr.c307 vcpu->arch.purr += get_tb() - vcpu->arch.entry_tb; in kvmppc_copy_from_svcpu()
308 vcpu->arch.spurr += get_tb() - vcpu->arch.entry_tb; in kvmppc_copy_from_svcpu()
411 _kvmppc_restore_tm_pr(vcpu, kvmppc_get_msr(vcpu)); in kvmppc_restore_tm_pr()
737 if (vcpu->arch.mmu.is_dcbz32(vcpu) && in kvmppc_handle_pagefault()
782 else if (vcpu->arch.mmu.is_dcbz32(vcpu) && in kvmppc_handle_pagefault()
1186 } else if (vcpu->arch.mmu.is_dcbz32(vcpu) && in kvmppc_handle_exit_pr()
1332 kvmppc_set_gpr(vcpu, 3, kvmppc_kvm_pv(vcpu)); in kvmppc_handle_exit_pr()
1496 vcpu->arch.mmu.slbmte(vcpu, 0, 0); in kvm_arch_vcpu_ioctl_set_sregs_pr()
1497 vcpu->arch.mmu.slbia(vcpu); in kvm_arch_vcpu_ioctl_set_sregs_pr()
1504 vcpu->arch.mmu.slbmte(vcpu, rs, rb); in kvm_arch_vcpu_ioctl_set_sregs_pr()
[all …]
A Demulate_loadstore.c31 kvmppc_core_queue_fpunavail(vcpu, kvmppc_get_msr(vcpu) & SRR1_PREFIXED); in kvmppc_check_fp_disabled()
43 kvmppc_core_queue_vsx_unavail(vcpu, kvmppc_get_msr(vcpu) & SRR1_PREFIXED); in kvmppc_check_vsx_disabled()
95 vcpu->arch.regs.msr = kvmppc_get_msr(vcpu); in kvmppc_emulate_loadstore()
114 kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed); in kvmppc_emulate_loadstore()
134 kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed); in kvmppc_emulate_loadstore()
233 kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed); in kvmppc_emulate_loadstore()
247 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, in kvmppc_emulate_loadstore()
257 kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed); in kvmppc_emulate_loadstore()
271 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, in kvmppc_emulate_loadstore()
316 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, in kvmppc_emulate_loadstore()
[all …]
A Dbook3s_hv_tm.c29 vcpu->arch.tfiar = tfiar; in emulate_tx_failure()
31 vcpu->arch.texasr = (vcpu->arch.texasr & 0x3ffffff) | texasr; in emulate_tx_failure()
56 vcpu->arch.regs.nip -= 4; in kvmhv_p9_tm_emulation()
79 vcpu->arch.cfar = vcpu->arch.regs.nip; in kvmhv_p9_tm_emulation()
80 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; in kvmhv_p9_tm_emulation()
103 bescr = vcpu->arch.bescr; in kvmhv_p9_tm_emulation()
113 vcpu->arch.cfar = vcpu->arch.regs.nip; in kvmhv_p9_tm_emulation()
114 vcpu->arch.regs.nip = vcpu->arch.ebbrr; in kvmhv_p9_tm_emulation()
156 vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) | in kvmhv_p9_tm_emulation()
203 vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) | in kvmhv_p9_tm_emulation()
[all …]
A Dbook3s.c534 ret = vcpu->kvm->arch.kvm_ops->get_sregs(vcpu, sregs); in kvm_arch_vcpu_ioctl_get_sregs()
546 ret = vcpu->kvm->arch.kvm_ops->set_sregs(vcpu, sregs); in kvm_arch_vcpu_ioctl_set_sregs()
656 if (!vcpu->arch.icp && !vcpu->arch.xive_vcpu) { in kvmppc_get_one_reg()
741 if (!vcpu->arch.icp && !vcpu->arch.xive_vcpu) { in kvmppc_set_one_reg()
792 vcpu->kvm->arch.kvm_ops->vcpu_load(vcpu, cpu); in kvmppc_core_vcpu_load()
797 vcpu->kvm->arch.kvm_ops->vcpu_put(vcpu); in kvmppc_core_vcpu_put()
802 vcpu->kvm->arch.kvm_ops->set_msr(vcpu, msr); in kvmppc_set_msr()
808 return vcpu->kvm->arch.kvm_ops->vcpu_run(vcpu); in kvmppc_vcpu_run()
834 return vcpu->kvm->arch.kvm_ops->vcpu_create(vcpu); in kvmppc_core_vcpu_create()
839 vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu); in kvmppc_core_vcpu_free()
[all …]
/linux/arch/arm64/include/asm/
A Dkvm_emulate.h81 if (!vcpu_has_run_once(vcpu)) in vcpu_reset_hcr()
111 vcpu->arch.hcr_el2 |= HCR_TWE; in vcpu_set_wfx_traps()
117 return vcpu->arch.vsesr_el2; in vcpu_get_vsesr()
122 vcpu->arch.vsesr_el2 = vsesr; in vcpu_set_vsesr()
142 if (vcpu_mode_is_32bit(vcpu)) in kvm_condition_valid()
225 return vcpu_has_nv(vcpu) && __is_hyp_ctxt(&vcpu->arch.ctxt); in is_hyp_ctxt()
373 return kvm_vcpu_trap_is_iabt(vcpu) && !kvm_vcpu_abt_iss1tw(vcpu); in kvm_vcpu_trap_is_exec_fault()
466 if (vcpu_mode_priv(vcpu)) in kvm_vcpu_is_be()
476 if (kvm_vcpu_is_be(vcpu)) { in vcpu_data_guest_to_host()
507 if (kvm_vcpu_is_be(vcpu)) { in vcpu_data_host_to_guest()
[all …]
/linux/arch/mips/kvm/
A Demulate.c248 err = kvm_compute_return_epc(vcpu, vcpu->arch.pc, in update_pc()
1003 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1013 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1022 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1031 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1061 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1091 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1139 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1186 vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr, in kvm_mips_emulate_store()
1290 vcpu->arch.io_pc = vcpu->arch.pc; in kvm_mips_emulate_load()
[all …]
/linux/arch/x86/kvm/vmx/
A Dx86_ops.h22 int vmx_vcpu_create(struct kvm_vcpu *vcpu);
23 int vmx_vcpu_pre_run(struct kvm_vcpu *vcpu);
25 void vmx_vcpu_free(struct kvm_vcpu *vcpu);
28 void vmx_vcpu_put(struct kvm_vcpu *vcpu);
42 int vmx_check_intercept(struct kvm_vcpu *vcpu,
47 void vmx_migrate_timers(struct kvm_vcpu *vcpu);
65 int vmx_get_cpl(struct kvm_vcpu *vcpu);
82 bool vmx_get_if_flag(struct kvm_vcpu *vcpu);
91 void vmx_inject_nmi(struct kvm_vcpu *vcpu);
96 bool vmx_get_nmi_mask(struct kvm_vcpu *vcpu);
[all …]
/linux/arch/arm64/kvm/hyp/vhe/
A Dswitch.c53 if (!vcpu_has_nv(vcpu)) in __compute_hcr()
99 if (vcpu_el2_e2h_is_set(vcpu) && is_hyp_ctxt(vcpu)) in __activate_cptr_traps()
106 if (is_hyp_ctxt(vcpu)) in __activate_cptr_traps()
143 ___activate_traps(vcpu, __compute_hcr(vcpu)); in __activate_traps()
248 __load_stage2(vcpu->arch.hw_mmu, vcpu->arch.hw_mmu->arch); in kvm_vcpu_load_vhe()
283 if (!(vcpu_el2_e2h_is_set(vcpu) && vcpu_el2_tge_is_set(vcpu))) in kvm_hyp_handle_eret()
298 if (!(vcpu_has_ptrauth(vcpu) && kvm_auth_eretax(vcpu, &elr))) in kvm_hyp_handle_eret()
346 val = vcpu_get_reg(vcpu, kvm_vcpu_sys_get_rt(vcpu)); in kvm_hyp_handle_tlbi_el2()
349 vcpu_el2_e2h_is_set(vcpu) && vcpu_el2_tge_is_set(vcpu)) || in kvm_hyp_handle_tlbi_el2()
372 vcpu_set_reg(vcpu, rt, __vcpu_sys_reg(vcpu, CPTR_EL2)); in kvm_hyp_handle_cpacr_el1()
[all …]
/linux/arch/x86/kvm/
A Dhyperv.h66 return vcpu->arch.hyperv; in to_hv_vcpu()
80 return hv_vcpu->vcpu; in hv_synic_to_vcpu()
100 return vcpu->arch.hyperv_enabled && to_kvm_hv(vcpu->kvm)->hv_guest_os_id; in kvm_hv_hypercall_enabled()
112 return to_hv_vcpu(vcpu) && test_bit(vector, to_hv_synic(vcpu)->vec_bitmap); in kvm_hv_synic_has_vector()
117 return to_hv_vcpu(vcpu) && in kvm_hv_synic_auto_eoi_set()
138 return hv_vcpu->vcpu; in hv_stimer_to_vcpu()
210 if (!to_hv_vcpu(vcpu) || !kvm_check_request(KVM_REQ_HV_TLB_FLUSH, vcpu)) in kvm_hv_vcpu_purge_flush_tlb()
213 tlb_flush_fifo = kvm_hv_get_tlb_flush_fifo(vcpu, is_guest_mode(vcpu)); in kvm_hv_vcpu_purge_flush_tlb()
234 code = is_64_bit_hypercall(vcpu) ? kvm_rcx_read(vcpu) : in kvm_hv_is_tlb_flush_hcall()
245 if (!to_hv_vcpu(vcpu)) in kvm_hv_verify_vp_assist()
[all …]
A Dx86.c1182 !load_pdptrs(vcpu, kvm_read_cr3(vcpu))) in kvm_set_cr0()
1219 vcpu->arch.pkru != vcpu->arch.host_pkru && in kvm_load_guest_xsave_state()
1235 if (vcpu->arch.pkru != vcpu->arch.host_pkru) in kvm_load_host_xsave_state()
1308 __kvm_set_xcr(vcpu, kvm_rcx_read(vcpu), kvm_read_edx_eax(vcpu))) { in kvm_emulate_xsetbv()
1391 && !load_pdptrs(vcpu, kvm_read_cr3(vcpu))) in kvm_set_cr4()
1527 vcpu->arch.eff_db[i] = vcpu->arch.db[i]; in kvm_update_dr0123()
8705 ctxt->vcpu = vcpu; in alloc_emulate_ctxt()
11027 vcpu->arch.last_vmentry_cpu = vcpu->cpu; in vcpu_enter_guest()
11933 load_pdptrs(vcpu, kvm_read_cr3(vcpu)); in __set_sregs_common()
12093 vcpu->arch.eff_db[i] = vcpu->arch.db[i]; in kvm_arch_vcpu_ioctl_set_guest_debug()
[all …]
A Dkvm_cache_regs.h101 kvm_x86_call(cache_reg)(vcpu, reg); in kvm_register_read_raw()
103 return vcpu->arch.regs[reg]; in kvm_register_read_raw()
112 vcpu->arch.regs[reg] = val; in kvm_register_write_raw()
113 kvm_register_mark_dirty(vcpu, reg); in kvm_register_write_raw()
157 return vcpu->arch.cr0 & mask; in kvm_read_cr0_bits()
179 return vcpu->arch.cr4 & mask; in kvm_read_cr4_bits()
194 return vcpu->arch.cr3; in kvm_read_cr3()
204 return (kvm_rax_read(vcpu) & -1u) in kvm_read_edx_eax()
210 vcpu->arch.hflags |= HF_GUEST_MASK; in enter_guest_mode()
211 vcpu->stat.guest_mode = 1; in enter_guest_mode()
[all …]
/linux/arch/loongarch/kvm/
A Dvcpu.c227 kvm_flush_tlb_gpa(vcpu, vcpu->arch.flush_gpa); in kvm_late_check_requests()
399 vcpu->arch.gprs[i], vcpu->arch.gprs[i + 1], in kvm_arch_vcpu_dump_regs()
400 vcpu->arch.gprs[i + 2], vcpu->arch.gprs[i + 3]); in kvm_arch_vcpu_dump_regs()
480 if (vcpu == map->phys_map[val].vcpu) { in kvm_set_cpuid()
495 map->phys_map[val].vcpu = vcpu; in kvm_set_cpuid()
553 vcpu_put(vcpu); in _kvm_getcsr()
1294 kvm_check_fcsr(vcpu, vcpu->arch.fpu.fcsr); in kvm_own_fpu()
1314 kvm_check_fcsr(vcpu, vcpu->arch.fpu.fcsr); in kvm_own_lsx()
1344 …if (!kvm_guest_has_fpu(&vcpu->arch) || !kvm_guest_has_lsx(&vcpu->arch) || !kvm_guest_has_lasx(&vcp… in kvm_own_lasx()
1349 kvm_check_fcsr(vcpu, vcpu->arch.fpu.fcsr); in kvm_own_lasx()
[all …]
/linux/arch/arm64/kvm/hyp/include/hyp/
A Dswitch.h113 if (vcpu_has_nv(vcpu) && !is_hyp_ctxt(vcpu)) \
180 if ((vcpu_has_nv(vcpu) && !is_hyp_ctxt(vcpu)) || \
234 if (vcpu_has_nv(vcpu) && !is_hyp_ctxt(vcpu)) { in __activate_traps_common()
295 return __get_fault_info(vcpu->arch.fault.esr_el2, &vcpu->arch.fault); in __populate_fault_info()
301 arm64_mops_reset_regs(vcpu_gp_regs(vcpu), vcpu->arch.fault.esr_el2); in kvm_hyp_handle_mops()
330 if (vcpu_has_nv(vcpu) && !is_hyp_ctxt(vcpu)) in __hyp_sve_restore_guest()
333 write_sysreg_el1(__vcpu_sys_reg(vcpu, vcpu_sve_zcr_elx(vcpu)), SYS_ZCR); in __hyp_sve_restore_guest()
469 __kvm_skip_instr(vcpu); in handle_tx2_tvm()
517 vcpu_set_reg(vcpu, kvm_vcpu_sys_get_rt(vcpu), val); in kvm_hyp_handle_cntpct()
518 __kvm_skip_instr(vcpu); in kvm_hyp_handle_cntpct()
[all …]
/linux/arch/powerpc/include/asm/
A Dkvm_book3s.h389 return vcpu->arch.book3s; in to_book3s()
415 vcpu->arch.regs.ccr = val; in kvmppc_set_cr()
422 return vcpu->arch.regs.ccr; in kvmppc_get_cr()
427 vcpu->arch.regs.xer = val; in kvmppc_set_xer()
434 return vcpu->arch.regs.xer; in kvmppc_get_xer()
439 vcpu->arch.regs.ctr = val; in kvmppc_set_ctr()
446 return vcpu->arch.regs.ctr; in kvmppc_get_ctr()
463 vcpu->arch.regs.nip = val; in kvmppc_set_pc()
504 vcpu->arch.fp.fpscr = val; in kvmppc_set_fpscr()
526 *v = vcpu->arch.vr.vr[i]; in kvmppc_get_vsx_vr()
[all …]

Completed in 140 milliseconds

12345678910>>...21