Lines Matching refs:vmcs01
290 if (WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01)) in free_nested()
291 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in free_nested()
306 vmcs_clear(vmx->vmcs01.shadow_vmcs); in free_nested()
307 free_vmcs(vmx->vmcs01.shadow_vmcs); in free_nested()
308 vmx->vmcs01.shadow_vmcs = NULL; in free_nested()
531 if (vmx_test_msr_bitmap_##rw(vmx->vmcs01.msr_bitmap, msr) || \
1497 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs; in copy_shadow_to_vmcs12()
1532 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs; in copy_vmcs12_to_shadow()
2207 static void prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01, in prepare_vmcs02_early() argument
2219 exec_control = __pin_controls_get(vmcs01); in prepare_vmcs02_early()
2234 exec_control = __exec_controls_get(vmcs01); /* L0's desires */ in prepare_vmcs02_early()
2271 exec_control = __secondary_exec_controls_get(vmcs01); in prepare_vmcs02_early()
2328 exec_control = __vm_entry_controls_get(vmcs01); in prepare_vmcs02_early()
2347 exec_control = __vm_exit_controls_get(vmcs01); in prepare_vmcs02_early()
3412 prepare_vmcs02_early(vmx, &vmx->vmcs01, vmcs12); in nested_vmx_enter_non_root_mode()
3416 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
3421 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
3500 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
4270 WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01); in copy_vmcs02_to_vmcs12_rare()
4274 vmx_vcpu_load_vmcs(vcpu, cpu, &vmx->vmcs01); in copy_vmcs02_to_vmcs12_rare()
4278 vmx->loaded_vmcs = &vmx->vmcs01; in copy_vmcs02_to_vmcs12_rare()
4768 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_vmexit()
5032 if (WARN_ON(loaded_vmcs != &vmx->vmcs01 || loaded_vmcs->shadow_vmcs)) in alloc_shadow_vmcs()
5462 vmcs_load(vmx->vmcs01.shadow_vmcs); in handle_vmwrite()
5466 vmcs_clear(vmx->vmcs01.shadow_vmcs); in handle_vmwrite()
5482 __pa(vmx->vmcs01.shadow_vmcs)); in set_current_vmptr()