| /qemu/target/i386/hvf/ |
| A D | vmx.h | 95 efer |= MSR_EFER_LMA; in enter_long_mode() 96 wvmcs(vcpu, VMCS_GUEST_IA32_EFER, efer); in enter_long_mode() 101 if ((efer & MSR_EFER_LME) && in enter_long_mode() 108 static void exit_long_mode(hv_vcpuid_t vcpu, uint64_t cr0, uint64_t efer) in exit_long_mode() argument 115 efer &= ~MSR_EFER_LMA; in exit_long_mode() 116 wvmcs(vcpu, VMCS_GUEST_IA32_EFER, efer); in exit_long_mode() 123 uint64_t efer = rvmcs(vcpu, VMCS_GUEST_IA32_EFER); in macvm_set_cr0() local 131 !(efer & MSR_EFER_LME)) { in macvm_set_cr0() 144 if (efer & MSR_EFER_LME) { in macvm_set_cr0() 147 enter_long_mode(vcpu, cr0, efer); in macvm_set_cr0() [all …]
|
| A D | x86hvf.c | 99 wvmcs(cs->accel->fd, VMCS_GUEST_IA32_EFER, env->efer); in hvf_put_segments() 208 env->efer = rvmcs(cs->accel->fd, VMCS_GUEST_IA32_EFER); in hvf_get_segments()
|
| /qemu/target/i386/tcg/sysemu/ |
| A D | svm_helper.c | 89 if (!(env->efer & MSR_EFER_SVME)) { in is_efer_invalid_state() 93 if (env->efer & MSR_EFER_RESERVED) { in is_efer_invalid_state() 97 if ((env->efer & (MSR_EFER_LMA | MSR_EFER_LME)) && in is_efer_invalid_state() 102 if ((env->efer & MSR_EFER_LME) && (env->cr[0] & CR0_PG_MASK) in is_efer_invalid_state() 107 if ((env->efer & MSR_EFER_LME) && (env->cr[0] & CR0_PG_MASK) in is_efer_invalid_state() 112 if ((env->efer & MSR_EFER_LME) && (env->cr[0] & CR0_PG_MASK) in is_efer_invalid_state() 211 env->vm_hsave + offsetof(struct vmcb, save.efer), env->efer); in helper_vmrun() 317 if ((env->efer & MSR_EFER_LMA) && in helper_vmrun() 346 env->vm_vmcb + offsetof(struct vmcb, save.efer))); in helper_vmrun() 793 env->vm_vmcb + offsetof(struct vmcb, save.efer), env->efer); in do_vmexit() [all …]
|
| A D | misc_helper.c | 91 if ((env->efer & MSR_EFER_LMA) && in helper_write_crN() 95 if (!(env->efer & MSR_EFER_LMA)) { in helper_write_crN() 187 cpu_load_efer(env, (env->efer & ~update_mask) | in helper_wrmsr() 349 val = env->efer; in helper_rdmsr()
|
| A D | seg_helper.c | 34 if (!(env->efer & MSR_EFER_SCE)) { in helper_syscall()
|
| A D | smm_helper.c | 86 x86_stq_phys(cs, sm_state + 0x7ed0, env->efer); in do_smm_enter()
|
| /qemu/target/i386/ |
| A D | helper.c | 143 (env->efer & MSR_EFER_LME)) { in cpu_x86_update_cr0() 148 env->efer |= MSR_EFER_LMA; in cpu_x86_update_cr0() 151 (env->efer & MSR_EFER_LMA)) { in cpu_x86_update_cr0() 153 env->efer &= ~MSR_EFER_LMA; in cpu_x86_update_cr0() 624 env->efer = val; in cpu_load_efer() 626 if (env->efer & MSR_EFER_LMA) { in cpu_load_efer() 629 if (env->efer & MSR_EFER_SVME) { in cpu_load_efer()
|
| A D | svm.h | 205 uint64_t efer; member
|
| A D | machine.c | 1445 return env->efer != 0; in intel_efer32_needed() 1454 VMSTATE_UINT64(env.efer, X86CPU), 1695 VMSTATE_UINT64(env.efer, X86CPU),
|
| A D | gdbstub.c | 225 return gdb_read_reg_cs64(env->hflags, mem_buf, env->efer); in x86_cpu_gdb_read_register()
|
| A D | cpu-dump.c | 480 qemu_fprintf(f, "EFER=%016" PRIx64 "\n", env->efer); in x86_cpu_dump_state()
|
| A D | cpu.h | 1770 uint64_t efer; member
|
| A D | cpu.c | 8388 if (env->efer & MSR_EFER_LMA) { in x86_update_hflags()
|
| /qemu/bsd-user/x86_64/ |
| A D | target_arch_cpu.h | 45 env->efer |= MSR_EFER_LMA | MSR_EFER_LME; in target_cpu_init()
|
| /qemu/target/i386/tcg/ |
| A D | seg_helper.c | 106 if (env->efer & MSR_EFER_NXE) { in get_pg_mode() 1063 if (!(env->efer & MSR_EFER_SCE)) { in helper_sysret() 1506 if (env->efer & MSR_EFER_LMA) { in helper_ljmp_protected() 1536 if (env->efer & MSR_EFER_LMA) { in helper_ljmp_protected() 1565 if (env->efer & MSR_EFER_LMA) { in helper_ljmp_protected() 1717 if (env->efer & MSR_EFER_LMA) { in helper_lcall_protected() 1753 if (env->efer & MSR_EFER_LMA) { in helper_lcall_protected() 1782 if (env->efer & MSR_EFER_LMA) { in helper_lcall_protected()
|
| A D | fpu_helper.c | 2676 if (!(env->efer & MSR_EFER_FFXSR) in do_fxsave() 2925 if (!(env->efer & MSR_EFER_FFXSR) in do_fxrstor()
|
| /qemu/linux-headers/asm-x86/ |
| A D | kvm.h | 151 __u64 efer; member 162 __u64 efer; member
|
| /qemu/target/i386/kvm/ |
| A D | kvm.c | 3512 sregs.efer = env->efer; in kvm_put_sregs() 3559 sregs.efer = env->efer; in kvm_put_sregs2() 4287 env->efer = sregs.efer; in kvm_get_sregs() 4288 if (sev_es_enabled() && env->efer & MSR_EFER_LME && in kvm_get_sregs() 4290 env->efer |= MSR_EFER_LMA; in kvm_get_sregs() 4330 env->efer = sregs.efer; in kvm_get_sregs2() 4331 if (sev_es_enabled() && env->efer & MSR_EFER_LME && in kvm_get_sregs2() 4333 env->efer |= MSR_EFER_LMA; in kvm_get_sregs2()
|
| /qemu/linux-user/i386/ |
| A D | cpu_loop.c | 351 env->efer |= MSR_EFER_LMA | MSR_EFER_LME; in target_cpu_copy_regs()
|
| /qemu/target/i386/nvmm/ |
| A D | nvmm-all.c | 169 state->msrs[NVMM_X64_MSR_EFER] = env->efer; in nvmm_set_registers() 324 env->efer = state->msrs[NVMM_X64_MSR_EFER]; in nvmm_get_registers()
|
| /qemu/target/i386/whpx/ |
| A D | whpx-all.c | 510 vcxt.values[idx++].Reg64 = env->efer; in whpx_set_registers() 733 env->efer = vcxt.values[idx++].Reg64; in whpx_get_registers()
|