Lines Matching refs:vmcs01
298 if (WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01)) in free_nested()
299 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in free_nested()
314 vmcs_clear(vmx->vmcs01.shadow_vmcs); in free_nested()
315 free_vmcs(vmx->vmcs01.shadow_vmcs); in free_nested()
316 vmx->vmcs01.shadow_vmcs = NULL; in free_nested()
542 if (vmx_test_msr_bitmap_##rw(vmx->vmcs01.msr_bitmap, msr) || \
1520 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs; in copy_shadow_to_vmcs12()
1555 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs; in copy_vmcs12_to_shadow()
2239 static void prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01, in prepare_vmcs02_early() argument
2251 exec_control = __pin_controls_get(vmcs01); in prepare_vmcs02_early()
2266 exec_control = __exec_controls_get(vmcs01); /* L0's desires */ in prepare_vmcs02_early()
2303 exec_control = __secondary_exec_controls_get(vmcs01); in prepare_vmcs02_early()
2359 exec_control = __vm_entry_controls_get(vmcs01); in prepare_vmcs02_early()
2378 exec_control = __vm_exit_controls_get(vmcs01); in prepare_vmcs02_early()
3446 prepare_vmcs02_early(vmx, &vmx->vmcs01, vmcs12); in nested_vmx_enter_non_root_mode()
3450 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
3455 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
3534 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
4309 WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01); in copy_vmcs02_to_vmcs12_rare()
4313 vmx_vcpu_load_vmcs(vcpu, cpu, &vmx->vmcs01); in copy_vmcs02_to_vmcs12_rare()
4317 vmx->loaded_vmcs = &vmx->vmcs01; in copy_vmcs02_to_vmcs12_rare()
4807 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_vmexit()
5082 if (WARN_ON(loaded_vmcs != &vmx->vmcs01 || loaded_vmcs->shadow_vmcs)) in alloc_shadow_vmcs()
5542 vmcs_load(vmx->vmcs01.shadow_vmcs); in handle_vmwrite()
5546 vmcs_clear(vmx->vmcs01.shadow_vmcs); in handle_vmwrite()
5562 __pa(vmx->vmcs01.shadow_vmcs)); in set_current_vmptr()