Lines Matching refs:to_vmx
173 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_failValid()
220 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_release_evmcs()
236 struct vcpu_vmx *vmx = to_vmx(vcpu); in free_nested()
295 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_switch_vmcs()
320 vmx_switch_vmcs(vcpu, &to_vmx(vcpu)->vmcs01); in nested_vmx_free_vcpu()
329 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_ept_inject_page_fault()
352 to_vmx(vcpu)->nested.msrs.ept_caps & in nested_ept_init_mmu_context()
429 !to_vmx(vcpu)->nested.nested_run_pending) { in vmx_inject_page_fault_nested()
493 msr_bitmap = to_vmx(vcpu)->vmcs01.msr_bitmap; in msr_write_intercepted_l01()
566 unsigned long *msr_bitmap_l0 = to_vmx(vcpu)->nested.vmcs02.msr_bitmap; in nested_vmx_prepare_msr_bitmap()
567 struct kvm_host_map *map = &to_vmx(vcpu)->nested.msr_bitmap_map; in nested_vmx_prepare_msr_bitmap()
653 kvm_vcpu_unmap(vcpu, &to_vmx(vcpu)->nested.msr_bitmap_map, false); in nested_vmx_prepare_msr_bitmap()
680 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_flush_cached_shadow_vmcs12()
881 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_max_atomic_switch_msrs()
1038 (nested_cpu_has_vpid(vmcs12) && to_vmx(vcpu)->nested.vpid02); in nested_has_guest_tlb_tag()
1043 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_vpid02()
1218 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_vmx_msr()
1821 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_handle_enlightened_vmptrld()
1902 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_sync_vmcs12_to_shadow()
1939 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_start_preemption_timer()
2308 struct vcpu_vmx *vmx = to_vmx(vcpu); in prepare_vmcs02()
2453 struct vcpu_vmx *vmx = to_vmx(vcpu); in valid_ept_address()
2493 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_check_vm_execution_controls()
2552 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_check_vm_exit_controls()
2569 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_check_vm_entry_controls()
2791 if (to_vmx(vcpu)->nested.nested_run_pending && in nested_vmx_check_guest_state()
2814 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_check_vmentry_hw()
2923 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_vmcs12_pages()
3008 if (!to_vmx(vcpu)->nested.vmxon) { in nested_vmx_check_permission()
3045 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_enter_non_root_mode()
3189 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_run()
3397 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_complete_nested_posted_interrupt()
3454 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_check_nested_events()
3514 hrtimer_get_remaining(&to_vmx(vcpu)->nested.preemption_timer); in vmx_get_preemption_timer_value()
3575 struct vcpu_vmx *vmx = to_vmx(vcpu); in sync_vmcs02_to_vmcs12_rare()
3622 struct vcpu_vmx *vmx = to_vmx(vcpu); in copy_vmcs02_to_vmcs12_rare()
3650 struct vcpu_vmx *vmx = to_vmx(vcpu); in sync_vmcs02_to_vmcs12()
3709 (vm_entry_controls_get(to_vmx(vcpu)) & VM_ENTRY_IA32E_MODE); in sync_vmcs02_to_vmcs12()
3953 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_restore_host_state()
4066 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_vmexit()
4075 hrtimer_cancel(&to_vmx(vcpu)->nested.preemption_timer); in nested_vmx_vmexit()
4355 struct vcpu_vmx *vmx = to_vmx(vcpu); in alloc_shadow_vmcs()
4376 struct vcpu_vmx *vmx = to_vmx(vcpu); in enter_vmx_operation()
4436 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmon()
4498 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_release_vmcs12()
4540 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmclear()
4613 if (to_vmx(vcpu)->nested.current_vmptr == -1ull) in handle_vmread()
4692 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmwrite()
4817 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmptrld()
4883 gpa_t current_vmptr = to_vmx(vcpu)->nested.current_vmptr; in handle_vmptrst()
4890 if (unlikely(to_vmx(vcpu)->nested.hv_evmcs)) in handle_vmptrst()
4908 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_invept()
4965 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_invvpid()
5088 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmfunc()
5321 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_exit_reflected()
5519 vmx = to_vmx(vcpu); in vmx_get_nested_state()
5606 to_vmx(vcpu)->nested.nested_run_pending = 0; in vmx_leave_nested()
5616 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_nested_state()