Lines Matching refs:vmcs01
283 if (WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01)) in free_nested()
284 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in free_nested()
298 vmcs_clear(vmx->vmcs01.shadow_vmcs); in free_nested()
299 free_vmcs(vmx->vmcs01.shadow_vmcs); in free_nested()
300 vmx->vmcs01.shadow_vmcs = NULL; in free_nested()
537 msr_bitmap = to_vmx(vcpu)->vmcs01.msr_bitmap; in msr_write_intercepted_l01()
1539 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs; in copy_shadow_to_vmcs12()
1574 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs; in copy_vmcs12_to_shadow()
2210 static void prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01, in prepare_vmcs02_early() argument
2222 exec_control = __pin_controls_get(vmcs01); in prepare_vmcs02_early()
2237 exec_control = __exec_controls_get(vmcs01); /* L0's desires */ in prepare_vmcs02_early()
2274 exec_control = __secondary_exec_controls_get(vmcs01); in prepare_vmcs02_early()
2328 exec_control = __vm_entry_controls_get(vmcs01); in prepare_vmcs02_early()
2346 exec_control = __vm_exit_controls_get(vmcs01); in prepare_vmcs02_early()
3398 prepare_vmcs02_early(vmx, &vmx->vmcs01, vmcs12); in nested_vmx_enter_non_root_mode()
3402 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
3407 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
3494 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
4080 WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01); in copy_vmcs02_to_vmcs12_rare()
4084 vmx_vcpu_load_vmcs(vcpu, cpu, &vmx->vmcs01); in copy_vmcs02_to_vmcs12_rare()
4088 vmx->loaded_vmcs = &vmx->vmcs01; in copy_vmcs02_to_vmcs12_rare()
4570 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_vmexit()
4850 WARN_ON(loaded_vmcs == &vmx->vmcs01 && loaded_vmcs->shadow_vmcs); in alloc_shadow_vmcs()
5252 vmcs_load(vmx->vmcs01.shadow_vmcs); in handle_vmwrite()
5256 vmcs_clear(vmx->vmcs01.shadow_vmcs); in handle_vmwrite()
5272 __pa(vmx->vmcs01.shadow_vmcs)); in set_current_vmptr()