Lines Matching refs:efer
813 u64 efer = 0; in assign_eip_far() local
815 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in assign_eip_far()
816 if (efer & EFER_LMA) in assign_eip_far()
1497 u64 efer = 0; in get_descriptor_ptr() local
1499 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in get_descriptor_ptr()
1500 if (!(efer & EFER_LMA)) in get_descriptor_ptr()
1648 u64 efer = 0; in __load_segment_descriptor() local
1650 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in __load_segment_descriptor()
1651 if (efer & EFER_LMA) in __load_segment_descriptor()
2520 unsigned long cr0, cr4, efer; in em_rsm() local
2571 efer = 0; in em_rsm()
2572 ctxt->ops->set_msr(ctxt, MSR_EFER, efer); in em_rsm()
2685 u64 efer = 0; in em_syscall() local
2695 ops->get_msr(ctxt, MSR_EFER, &efer); in em_syscall()
2696 if (!(efer & EFER_SCE)) in em_syscall()
2705 if (efer & EFER_LMA) { in em_syscall()
2713 if (efer & EFER_LMA) { in em_syscall()
2744 u64 efer = 0; in em_sysenter() local
2746 ops->get_msr(ctxt, MSR_EFER, &efer); in em_sysenter()
2755 if ((ctxt->mode != X86EMUL_MODE_PROT64) && (efer & EFER_LMA) in em_sysenter()
2771 if (efer & EFER_LMA) { in em_sysenter()
2780 ctxt->_eip = (efer & EFER_LMA) ? msr_data : (u32)msr_data; in em_sysenter()
2783 *reg_write(ctxt, VCPU_REGS_RSP) = (efer & EFER_LMA) ? msr_data : in em_sysenter()
2785 if (efer & EFER_LMA) in em_sysenter()
4182 u64 efer = 0; in check_svme() local
4184 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_svme()
4186 if (!(efer & EFER_SVME)) in check_svme()