/linux-6.3-rc2/arch/x86/include/asm/ |
A D | virtext.h | 125 uint64_t efer; in cpu_svm_disable() local 128 rdmsrl(MSR_EFER, efer); in cpu_svm_disable() 129 if (efer & EFER_SVME) { in cpu_svm_disable() 142 wrmsrl(MSR_EFER, efer & ~EFER_SVME); in cpu_svm_disable()
|
A D | suspend_64.h | 44 unsigned long efer; member
|
A D | realmode.h | 52 u64 efer;
|
A D | svm.h | 306 u64 efer; member 362 u64 efer; member
|
A D | kvm_host.h | 738 u64 efer; member 1591 int (*set_efer)(struct kvm_vcpu *vcpu, u64 efer); 1928 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer);
|
/linux-6.3-rc2/arch/x86/realmode/ |
A D | init.c | 100 u64 efer; in setup_real_mode() local 148 rdmsrl(MSR_EFER, efer); in setup_real_mode() 149 trampoline_header->efer = efer & ~EFER_LMA; in setup_real_mode()
|
/linux-6.3-rc2/tools/testing/selftests/kvm/lib/x86_64/ |
A D | svm.c | 73 uint64_t efer; in generic_svm_setup() local 75 efer = rdmsr(MSR_EFER); in generic_svm_setup() 76 wrmsr(MSR_EFER, efer | EFER_SVME); in generic_svm_setup() 90 save->efer = rdmsr(MSR_EFER); in generic_svm_setup()
|
/linux-6.3-rc2/include/xen/interface/hvm/ |
A D | hvm_vcpu.h | 35 uint64_t efer; member 89 uint64_t efer; member
|
/linux-6.3-rc2/arch/x86/kvm/svm/ |
A D | nested.c | 94 svm->vmcb01.ptr->save.efer, in nested_svm_init_mmu_context() 286 if (CC(!(save->efer & EFER_SVME))) in __nested_vmcb_check_save() 301 if ((save->efer & EFER_LME) && (save->cr0 & X86_CR0_PG)) { in __nested_vmcb_check_save() 312 if (CC(!kvm_valid_efer(vcpu, save->efer))) in __nested_vmcb_check_save() 392 to->efer = from->efer; in __nested_copy_vmcb_save_to_cache() 570 svm_set_efer(&svm->vcpu, svm->nested.save.efer); in nested_vmcb02_prepare_save() 877 vmcb01->save.efer = vcpu->arch.efer; in nested_svm_vmrun() 923 to_save->efer = from_save->efer; in svm_copy_vmrun_state() 985 vmcb12->save.efer = svm->vcpu.arch.efer; in nested_svm_vmexit() 1057 svm_set_efer(vcpu, vmcb01->save.efer); in nested_svm_vmexit() [all …]
|
A D | svm.c | 301 u64 old_efer = vcpu->arch.efer; in svm_set_efer() 302 vcpu->arch.efer = efer; in svm_set_efer() 306 efer |= EFER_NX; in svm_set_efer() 308 if (!(efer & EFER_LMA)) in svm_set_efer() 309 efer &= ~EFER_LME; in svm_set_efer() 313 if (!(efer & EFER_SVME)) { in svm_set_efer() 332 vcpu->arch.efer = old_efer; in svm_set_efer() 345 svm->vmcb->save.efer = efer | EFER_SVME; in svm_set_efer() 583 uint64_t efer; in svm_hardware_enable() local 587 rdmsrl(MSR_EFER, efer); in svm_hardware_enable() [all …]
|
A D | svm.h | 114 u64 efer; member 565 int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer);
|
/linux-6.3-rc2/arch/x86/kvm/ |
A D | smm.c | 82 CHECK_SMRAM64_OFFSET(efer, 0xFED0); in check_smram_offsets() 256 smram->efer = vcpu->arch.efer; in enter_smm_save_state_64() 539 if (kvm_set_msr(vcpu, MSR_EFER, smstate->efer & ~EFER_LMA)) in rsm_load_state_64() 621 unsigned long cr4, efer; in emulator_leave_smm() local 629 efer = 0; in emulator_leave_smm() 630 kvm_set_msr(vcpu, MSR_EFER, efer); in emulator_leave_smm()
|
A D | emulate.c | 782 u64 efer; in emulator_recalc_and_set_mode() local 791 if (efer & EFER_LMA) in emulator_recalc_and_set_mode() 799 if (efer & EFER_LMA) in emulator_recalc_and_set_mode() 808 if (efer & EFER_LMA) { in emulator_recalc_and_set_mode() 1511 u64 efer = 0; in get_descriptor_ptr() local 1677 u64 efer = 0; in __load_segment_descriptor() local 1680 if (efer & EFER_LMA) in __load_segment_descriptor() 2398 u64 efer = 0; in em_syscall() local 2457 u64 efer = 0; in em_sysenter() local 2498 if (efer & EFER_LMA) in em_sysenter() [all …]
|
A D | smm.h | 104 u64 efer; member
|
A D | x86.c | 1706 if (efer & (EFER_LME | EFER_LMA) && in __kvm_valid_efer() 1718 if (efer & efer_reserved_bits) in kvm_valid_efer() 1727 u64 old_efer = vcpu->arch.efer; in set_efer() 1728 u64 efer = msr_info->data; in set_efer() local 1731 if (efer & efer_reserved_bits) in set_efer() 1739 (vcpu->arch.efer & EFER_LME) != (efer & EFER_LME)) in set_efer() 1743 efer &= ~EFER_LMA; in set_efer() 1744 efer |= vcpu->arch.efer & EFER_LMA; in set_efer() 11202 sregs->efer = vcpu->arch.efer; in __get_sregs_common() 11399 *mmu_reset_needed |= vcpu->arch.efer != sregs->efer; in __set_sregs_common() [all …]
|
A D | mmu.h | 109 unsigned long cr4, u64 efer, gpa_t nested_cr3);
|
A D | x86.h | 132 return vcpu->arch.efer & EFER_LMA; in is_long_mode()
|
/linux-6.3-rc2/arch/x86/power/ |
A D | cpu.c | 117 rdmsrl(MSR_EFER, ctxt->efer); in __save_processor_state() 210 wrmsrl(MSR_EFER, ctxt->efer); in __restore_processor_state()
|
/linux-6.3-rc2/arch/x86/include/uapi/asm/ |
A D | kvm.h | 150 __u64 efer; member 161 __u64 efer; member
|
/linux-6.3-rc2/tools/arch/x86/include/uapi/asm/ |
A D | kvm.h | 150 __u64 efer; member 161 __u64 efer; member
|
/linux-6.3-rc2/tools/testing/selftests/kvm/include/x86_64/ |
A D | svm.h | 209 u64 efer; member
|
/linux-6.3-rc2/drivers/parport/ |
A D | parport_pc.c | 1102 static void decode_winbond(int efer, int key, int devid, int devrev, int oldid) in decode_winbond() argument 1147 efer, key, devid, devrev, oldid, type); in decode_winbond() 1150 show_parconfig_winbond(efer, key); in decode_winbond() 1153 static void decode_smsc(int efer, int key, int devid, int devrev) in decode_smsc() argument 1179 efer, key, devid, devrev, type); in decode_smsc() 1182 func(efer, key); in decode_smsc()
|
/linux-6.3-rc2/arch/x86/kvm/vmx/ |
A D | vmx.c | 1073 u64 guest_efer = vmx->vcpu.arch.efer; in update_transition_efer() 1838 (vmx->vcpu.arch.efer & EFER_SCE); in vmx_setup_uret_msrs() 3110 int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer) in vmx_set_efer() argument 3118 vcpu->arch.efer = efer; in vmx_set_efer() 3120 if (efer & EFER_LMA) in vmx_set_efer() 3125 if (KVM_BUG_ON(efer & EFER_LMA, vcpu->kvm)) in vmx_set_efer() 3149 vmx_set_efer(vcpu, vcpu->arch.efer | EFER_LMA); in enter_lmode() 3154 vmx_set_efer(vcpu, vcpu->arch.efer & ~EFER_LMA); in exit_lmode() 3288 if (vcpu->arch.efer & EFER_LME) { in vmx_set_cr0() 6290 vcpu->arch.efer | (EFER_LMA | EFER_LME)); in dump_vmcs() [all …]
|
A D | nested.c | 2152 return vmx->vcpu.arch.efer | (EFER_LMA | EFER_LME); in nested_vmx_calc_efer() 2154 return vmx->vcpu.arch.efer & ~(EFER_LMA | EFER_LME); in nested_vmx_calc_efer() 2602 vcpu->arch.efer = nested_vmx_calc_efer(vmx, vmcs12); in prepare_vmcs02() 2604 vmx_set_efer(vcpu, vcpu->arch.efer); in prepare_vmcs02() 2897 !!(vcpu->arch.efer & EFER_LMA))) in nested_vmx_check_address_space_size() 4385 vmcs12->guest_ia32_efer = vcpu->arch.efer; in sync_vmcs02_to_vmcs12() 4462 vcpu->arch.efer = vmcs12->host_ia32_efer; in load_vmcs12_host_state() 4464 vcpu->arch.efer |= (EFER_LMA | EFER_LME); in load_vmcs12_host_state() 4466 vcpu->arch.efer &= ~(EFER_LMA | EFER_LME); in load_vmcs12_host_state() 4467 vmx_set_efer(vcpu, vcpu->arch.efer); in load_vmcs12_host_state()
|
/linux-6.3-rc2/arch/x86/kvm/mmu/ |
A D | mmu.c | 176 const u64 efer; member 201 BUILD_MMU_ROLE_REGS_ACCESSOR(efer, nx, EFER_NX); 202 BUILD_MMU_ROLE_REGS_ACCESSOR(efer, lma, EFER_LMA); 221 BUILD_MMU_ROLE_ACCESSOR(base, efer, nx); 222 BUILD_MMU_ROLE_ACCESSOR(ext, efer, lma); 239 .efer = vcpu->arch.efer, in vcpu_to_role_regs() 5225 unsigned long cr4, u64 efer, gpa_t nested_cr3) in kvm_init_shadow_npt_mmu() argument 5231 .efer = efer, in kvm_init_shadow_npt_mmu()
|