Searched refs:cr4 (Results 1 – 14 of 14) sorted by relevance
| /hypervisor/arch/x86/guest/ |
| A D | virtual_cr.c | 328 static inline bool is_valid_cr4(uint64_t cr4) in is_valid_cr4() argument 336 bool is_valid_cr0_cr4(uint64_t cr0, uint64_t cr4) in is_valid_cr0_cr4() argument 338 return is_valid_cr4(cr4) & is_valid_cr0(cr0); in is_valid_cr0_cr4() 345 if (!is_valid_cr4(cr4) || (is_long_mode(vcpu) && ((cr4 & CR4_PAE) == 0UL))) { in is_cr4_write_valid() 364 if (!is_cr4_write_valid(vcpu, cr4)) { in vmx_write_cr4() 369 uint64_t cr4_changed_bits = vcpu_get_cr4(vcpu) ^ cr4; in vmx_write_cr4() 386 if ((cr4 & CR4_PCIDE) != 0UL) { in vmx_write_cr4() 400 if ((cr4 & CR4_KL) != 0UL) { in vmx_write_cr4() 413 tmp = (initial_guest_cr4 & ~mask) | (cr4 & mask); in vmx_write_cr4() 420 exec_vmwrite(VMX_CR4_READ_SHADOW, cr4); in vmx_write_cr4() [all …]
|
| A D | vmcs.c | 24 uint64_t cr4) in init_guest_vmx() argument 29 pr_dbg("%s,cr0:0x%lx, cr4:0x%lx.", __func__, cr0, cr4); in init_guest_vmx() 31 vcpu_set_cr4(vcpu, cr4); in init_guest_vmx() 80 ctx->run_ctx.cr0, ctx->run_ctx.cr4); in init_guest_state() 83 ctx->run_ctx.cr4 & ~(CR4_VMXE | CR4_SMXE | CR4_MCE)); in init_guest_state() 191 CPU_CR_READ(cr4, &value); in init_host_state()
|
| A D | vcpu.c | 402 ctx->cr4 = vcpu_regs->cr4; in set_vcpu_regs() 424 .cr4 = 0UL, 448 if (is_valid_cr0_cr4(realmode_init_vregs.cr0, realmode_init_vregs.cr4) && in sanitize_cr0_cr4_pattern() 449 is_valid_cr0_cr4(protect_mode_init_vregs.cr0, protect_mode_init_vregs.cr4)) { in sanitize_cr0_cr4_pattern() 453 realmode_init_vregs.cr4, protect_mode_init_vregs.cr0, protect_mode_init_vregs.cr4); in sanitize_cr0_cr4_pattern() 686 vcpu_set_cr4(vcpu, ctx->cr4); in write_cached_registers()
|
| A D | vcpuid.c | 770 uint64_t cr4; in guest_cpuid_01h() local 772 cr4 = vcpu_get_cr4(vcpu); in guest_cpuid_01h() 773 if ((cr4 & CR4_OSXSAVE) != 0UL) { in guest_cpuid_01h()
|
| /hypervisor/arch/x86/ |
| A D | vmx.c | 53 CPU_CR_READ(cr4, &tmp64); in vmx_on() 54 CPU_CR_WRITE(cr4, tmp64 | CR4_VMXE); in vmx_on()
|
| A D | mmu.c | 183 CPU_CR_READ(cr4, &val64); in enable_smep() 184 CPU_CR_WRITE(cr4, val64 | CR4_SMEP); in enable_smep() 192 CPU_CR_READ(cr4, &val64); in enable_smap() 193 CPU_CR_WRITE(cr4, val64 | CR4_SMAP); in enable_smap()
|
| A D | wakeup.S | 84 mov %cr4, %rax 107 mov %rax, %cr4
|
| A D | cpu.c | 574 CPU_CR_READ(cr4, &val64); in init_pcpu_xsave() 576 CPU_CR_WRITE(cr4, val64); in init_pcpu_xsave() 616 CPU_CR_READ(cr4, &val64); in init_keylocker() 618 CPU_CR_WRITE(cr4, val64); in init_keylocker()
|
| /hypervisor/arch/x86/boot/ |
| A D | trampoline.S | 91 mov %eax, %cr4
|
| A D | cpu_primary.S | 172 mov %eax, %cr4
|
| /hypervisor/include/arch/x86/asm/ |
| A D | vmx.h | 465 bool is_valid_cr0_cr4(uint64_t cr0, uint64_t cr4);
|
| A D | cpu.h | 355 uint64_t cr4; member
|
| /hypervisor/include/public/ |
| A D | acrn_common.h | 433 uint64_t cr4; member
|
| /hypervisor/common/ |
| A D | hypercall.c | 383 if (is_valid_cr0_cr4(vcpu_regs.vcpu_regs.cr0, vcpu_regs.vcpu_regs.cr4)) { in hcall_set_vcpu_regs()
|
Completed in 33 milliseconds