Lines Matching refs:guest_cr
675 if ( unlikely(!(v->arch.hvm_vcpu.guest_cr[0] & X86_CR0_PE)) ) in vmx_guest_x86_mode()
739 c->cr0 = v->arch.hvm_vcpu.guest_cr[0]; in vmx_vmcs_save()
740 c->cr2 = v->arch.hvm_vcpu.guest_cr[2]; in vmx_vmcs_save()
741 c->cr3 = v->arch.hvm_vcpu.guest_cr[3]; in vmx_vmcs_save()
742 c->cr4 = v->arch.hvm_vcpu.guest_cr[4]; in vmx_vmcs_save()
790 v->arch.hvm_vcpu.guest_cr[0] = cr0 | X86_CR0_ET; in vmx_restore_cr0_cr3()
791 v->arch.hvm_vcpu.guest_cr[3] = cr3; in vmx_restore_cr0_cr3()
825 v->arch.hvm_vcpu.guest_cr[2] = c->cr2; in vmx_vmcs_restore()
826 v->arch.hvm_vcpu.guest_cr[4] = c->cr4; in vmx_vmcs_restore()
991 if ( !(v->arch.hvm_vcpu.guest_cr[0] & X86_CR0_TS) ) in vmx_fpu_leave()
1468 unsigned long cr3 = v->arch.hvm_vcpu.guest_cr[3]; in vmx_load_pdptrs()
1577 __vmwrite(CR0_READ_SHADOW, v->arch.hvm_vcpu.guest_cr[0]); in vmx_update_guest_cr()
1581 if ( !(v->arch.hvm_vcpu.guest_cr[0] & X86_CR0_TS) ) in vmx_update_guest_cr()
1589 realmode = !(v->arch.hvm_vcpu.guest_cr[0] & X86_CR0_PE); in vmx_update_guest_cr()
1623 v->arch.hvm_vcpu.guest_cr[0] | hw_cr0_mask; in vmx_update_guest_cr()
1633 __vmwrite(CR4_READ_SHADOW, v->arch.hvm_vcpu.guest_cr[4]); in vmx_update_guest_cr()
1637 v->arch.hvm_vcpu.hw_cr[4] |= v->arch.hvm_vcpu.guest_cr[4]; in vmx_update_guest_cr()
1864 curr->arch.hvm_vcpu.guest_cr[2] = _event.cr2; in vmx_inject_event()
1902 TRC_PAR_LONG(curr->arch.hvm_vcpu.guest_cr[2])); in vmx_inject_event()
2562 if ( !(curr->arch.hvm_vcpu.guest_cr[0] & X86_CR0_TS) ) in vmx_fpu_dirty_intercept()
2649 unsigned long old = curr->arch.hvm_vcpu.guest_cr[0]; in vmx_cr_access()
2658 curr->arch.hvm_vcpu.guest_cr[0] = value; in vmx_cr_access()
2664 unsigned long value = curr->arch.hvm_vcpu.guest_cr[0]; in vmx_cr_access()
3523 v->arch.hvm_vcpu.guest_cr[3] = v->arch.hvm_vcpu.hw_cr[3]; in vmx_vmexit_handler()