Lines Matching refs:vmcs01
283 if (WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01)) in free_nested()
284 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in free_nested()
299 vmcs_clear(vmx->vmcs01.shadow_vmcs); in free_nested()
300 free_vmcs(vmx->vmcs01.shadow_vmcs); in free_nested()
301 vmx->vmcs01.shadow_vmcs = NULL; in free_nested()
561 if (vmx_test_msr_bitmap_##rw(vmx->vmcs01.msr_bitmap, msr) || \
1507 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs; in copy_shadow_to_vmcs12()
1542 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs; in copy_vmcs12_to_shadow()
2178 static void prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01, in prepare_vmcs02_early() argument
2190 exec_control = __pin_controls_get(vmcs01); in prepare_vmcs02_early()
2205 exec_control = __exec_controls_get(vmcs01); /* L0's desires */ in prepare_vmcs02_early()
2242 exec_control = __secondary_exec_controls_get(vmcs01); in prepare_vmcs02_early()
2296 exec_control = __vm_entry_controls_get(vmcs01); in prepare_vmcs02_early()
2314 exec_control = __vm_exit_controls_get(vmcs01); in prepare_vmcs02_early()
3380 prepare_vmcs02_early(vmx, &vmx->vmcs01, vmcs12); in nested_vmx_enter_non_root_mode()
3384 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
3389 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
3476 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
4065 WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01); in copy_vmcs02_to_vmcs12_rare()
4069 vmx_vcpu_load_vmcs(vcpu, cpu, &vmx->vmcs01); in copy_vmcs02_to_vmcs12_rare()
4073 vmx->loaded_vmcs = &vmx->vmcs01; in copy_vmcs02_to_vmcs12_rare()
4554 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_vmexit()
4834 WARN_ON(loaded_vmcs == &vmx->vmcs01 && loaded_vmcs->shadow_vmcs); in alloc_shadow_vmcs()
5237 vmcs_load(vmx->vmcs01.shadow_vmcs); in handle_vmwrite()
5241 vmcs_clear(vmx->vmcs01.shadow_vmcs); in handle_vmwrite()
5257 __pa(vmx->vmcs01.shadow_vmcs)); in set_current_vmptr()