Lines Matching refs:arch
43 &vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx; in vcpu_get_gpreg()
51 &vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx; in vcpu_set_gpreg()
59 &vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx; in vcpu_get_rip()
70 vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx.rip = val; in vcpu_set_rip()
77 &vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx; in vcpu_get_rsp()
85 &vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx; in vcpu_set_rsp()
94 &vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx; in vcpu_get_efer()
101 vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx.ia32_efer in vcpu_set_efer()
119 &vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx; in vcpu_get_rflags()
130 vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx.rflags = in vcpu_set_rflags()
141 val = vcpu->arch.guest_msrs[index]; in vcpu_get_guest_msr()
152 vcpu->arch.guest_msrs[index] = val; in vcpu_set_guest_msr()
164 exec_vmwrite64(VMX_EOI_EXIT0_FULL, vcpu->arch.eoi_exit_bitmap[0]); in vcpu_set_vmcs_eoi_exit()
165 exec_vmwrite64(VMX_EOI_EXIT1_FULL, vcpu->arch.eoi_exit_bitmap[1]); in vcpu_set_vmcs_eoi_exit()
166 exec_vmwrite64(VMX_EOI_EXIT2_FULL, vcpu->arch.eoi_exit_bitmap[2]); in vcpu_set_vmcs_eoi_exit()
167 exec_vmwrite64(VMX_EOI_EXIT3_FULL, vcpu->arch.eoi_exit_bitmap[3]); in vcpu_set_vmcs_eoi_exit()
180 &(vcpu->arch.eoi_exit_bitmap[(vector & 0xffU) >> 6U]))) { in vcpu_set_eoi_exit_bitmap()
190 &(vcpu->arch.eoi_exit_bitmap[(vector & 0xffU) >> 6U]))) { in vcpu_clear_eoi_exit_bitmap()
202 (void)memset((void *)(vcpu->arch.eoi_exit_bitmap), 0U, sizeof(vcpu->arch.eoi_exit_bitmap)); in vcpu_reset_eoi_exit_bitmaps()
210 vcpu->arch.IWKey.integrity_key[0] = get_random_value(); in init_iwkey()
211 vcpu->arch.IWKey.integrity_key[1] = get_random_value(); in init_iwkey()
212 vcpu->arch.IWKey.encryption_key[0] = get_random_value(); in init_iwkey()
213 vcpu->arch.IWKey.encryption_key[1] = get_random_value(); in init_iwkey()
214 vcpu->arch.IWKey.encryption_key[2] = get_random_value(); in init_iwkey()
215 vcpu->arch.IWKey.encryption_key[3] = get_random_value(); in init_iwkey()
226 if (pcpu_has_cap(X86_FEATURE_KEYLOCKER) && vcpu->arch.cr4_kl_enabled && in load_iwkey()
230 write_xmm_0_2(&vcpu->arch.IWKey.integrity_key[0], &vcpu->arch.IWKey.encryption_key[0], in load_iwkey()
231 &vcpu->arch.IWKey.encryption_key[2]); in load_iwkey()
245 vcpu->arch.nr_sipi = 0U; in vcpu_reset_internal()
247 vcpu->arch.exception_info.exception = VECTOR_INVALID; in vcpu_reset_internal()
248 vcpu->arch.cur_context = NORMAL_WORLD; in vcpu_reset_internal()
249 vcpu->arch.lapic_pt_enabled = false; in vcpu_reset_internal()
250 vcpu->arch.irq_window_enabled = false; in vcpu_reset_internal()
251 vcpu->arch.emulating_lock = false; in vcpu_reset_internal()
252 (void)memset((void *)vcpu->arch.vmcs, 0U, PAGE_SIZE); in vcpu_reset_internal()
255 (void)memset((void *)(&vcpu->arch.contexts[i]), 0U, in vcpu_reset_internal()
269 vcpu->arch.iwkey_copy_status = 0UL; in vcpu_reset_internal()
295 vcpu->arch.cpu_mode = CPU_MODE_64BIT; in set_vcpu_mode()
297 vcpu->arch.cpu_mode = CPU_MODE_COMPATIBILITY; in set_vcpu_mode()
300 vcpu->arch.cpu_mode = CPU_MODE_PROTECTED; in set_vcpu_mode()
302 vcpu->arch.cpu_mode = CPU_MODE_REAL; in set_vcpu_mode()
308 struct ext_context *ectx = &(vcpu->arch.contexts[vcpu->arch.cur_context].ext_ctx); in init_xsave()
333 ectx = &(vcpu->arch.contexts[vcpu->arch.cur_context].ext_ctx); in set_vcpu_regs()
334 ctx = &(vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx); in set_vcpu_regs()
473 struct ext_context *ectx = &(vcpu->arch.contexts[vcpu->arch.cur_context].ext_ctx); in reset_vcpu_regs()
498 ectx = &(vcpu->arch.contexts[vcpu->arch.cur_context].ext_ctx); in set_vcpu_startup_entry()
573 vcpu->arch.vpid = ALLOCATED_MIN_L1_VPID + (vm->vm_id * MAX_VCPUS_PER_VM) + vcpu->vcpu_id; in create_vcpu()
609 vcpu->arch.pid.control.bits.nv = POSTED_INTR_VECTOR + vm->vm_id; in create_vcpu()
614 vcpu->arch.pid.control.bits.ndst = per_cpu(lapic_id, pcpu_id); in create_vcpu()
661 &vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx; in write_cached_registers()
698 &vcpu->arch.contexts[vcpu->arch.cur_context].run_ctx; in run_vcpu()
725 if (vcpu->arch.vpid != 0U) { in run_vcpu()
726 exec_vmwrite16(VMX_VPID, vcpu->arch.vpid); in run_vcpu()
767 if (vcpu->arch.inst_len != 0U) { in run_vcpu()
768 exec_vmwrite(VMX_GUEST_RIP, vcpu_get_rip(vcpu) + vcpu->arch.inst_len); in run_vcpu()
784 vcpu->arch.inst_len = exec_vmread32(VMX_EXIT_INSTR_LEN); in run_vcpu()
787 vcpu->arch.exit_reason = exec_vmread32(VMX_EXIT_REASON); in run_vcpu()
791 if ((vcpu->arch.exit_reason & VMX_VMENTRY_FAIL) != 0U) { in run_vcpu()
792 pr_fatal("vmentry fail reason=%lx", vcpu->arch.exit_reason); in run_vcpu()
822 if ((get_pcpu_id() != pcpu_id) && (per_cpu(vmcs_run, pcpu_id) == vcpu->arch.vmcs)) { in kick_vcpu()
927 struct ext_context *ectx = &(vcpu->arch.contexts[vcpu->arch.cur_context].ext_ctx); in context_switch_out()
943 struct ext_context *ectx = &(vcpu->arch.contexts[vcpu->arch.cur_context].ext_ctx); in context_switch_in()