Searched refs:is_guest_mode (Results 1 – 15 of 15) sorted by relevance
746 if (is_guest_mode(vcpu)) in vmx_update_exception_bitmap()1563 if (!is_guest_mode(vcpu)) in vmx_update_emulated_instruction()1968 if (is_guest_mode(vcpu)) in vmx_set_msr()1973 if (is_guest_mode(vcpu)) { in vmx_set_msr()1980 if (is_guest_mode(vcpu)) { in vmx_set_msr()1999 if (is_guest_mode(vcpu) && get_vmcs12(vcpu)->vm_exit_controls & in vmx_set_msr()2095 if (is_guest_mode(vcpu) && in vmx_set_msr()2942 else if (!is_guest_mode(vcpu)) in vmx_flush_tlb_current()3066 } else if (!is_guest_mode(vcpu)) { in vmx_set_cr0()3098 (!is_guest_mode(vcpu) || nested_ept_ad_enabled(vcpu))) in construct_eptp()[all …]
475 WARN_ON(!is_guest_mode(vcpu)); in vmx_inject_page_fault_nested()3264 if (is_guest_mode(vcpu) && !nested_get_vmcs12_pages(vcpu)) in vmx_get_nested_state_pages()3276 if (WARN_ON_ONCE(!is_guest_mode(vcpu))) in nested_vmx_write_pml_buffer()5077 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) in handle_vmread()5097 (is_guest_mode(vcpu) && in handle_vmread()5108 if (!is_guest_mode(vcpu) && is_vmcs12_ext_field(field)) in handle_vmread()5161 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) in handle_vmwrite()5189 (is_guest_mode(vcpu) && in handle_vmwrite()5223 if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) in handle_vmwrite()5245 if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) { in handle_vmwrite()[all …]
512 return enable_unrestricted_guest && (!is_guest_mode(vcpu) || in is_unrestricted_guest()
496 if (!vmcs12 && is_guest_mode(vcpu)) in vmx_write_encls_bitmap()
83 if (!is_guest_mode(v) && kvm_vcpu_apicv_active(v)) in kvm_cpu_has_injectable_intr()
192 static inline bool is_guest_mode(struct kvm_vcpu *vcpu) in is_guest_mode() function
641 if (!is_guest_mode(vcpu)) in kvm_multiple_exception()717 is_guest_mode(vcpu) && fault->async_page_fault; in kvm_inject_page_fault()2416 if (is_guest_mode(vcpu)) in kvm_vcpu_write_tsc_offset()2432 if (is_guest_mode(vcpu)) in kvm_vcpu_write_tsc_multiplier()7538 if (!is_guest_mode(vcpu) && static_call(kvm_x86_get_cpl)(vcpu) == 0) { in handle_emulation_failure()7558 if (WARN_ON_ONCE(is_guest_mode(vcpu)) || in reexecute_instruction()7650 if (WARN_ON_ONCE(is_guest_mode(vcpu)) || in retry_instruction()8907 if (is_guest_mode(vcpu)) { in inject_pending_event()8994 if (is_guest_mode(vcpu) && in inject_pending_event()9368 if (is_guest_mode(vcpu)) in vcpu_scan_ioapic()[all …]
2898 if (is_guest_mode(vcpu)) { in kvm_apic_accept_events()
24 bool is_guest_mode; member
59 WARN_ON(!is_guest_mode(vcpu)); in svm_inject_page_fault_nested()129 if (!is_guest_mode(&svm->vcpu)) in recalc_intercepts()949 if (is_guest_mode(vcpu)) { in svm_leave_nested()1244 if (is_guest_mode(vcpu)) { in svm_get_nested_state()1259 if (!is_guest_mode(vcpu)) in svm_get_nested_state()1380 if (is_guest_mode(vcpu)) in svm_set_nested_state()1410 if (WARN_ON(!is_guest_mode(vcpu))) in svm_get_nested_state_pages()
641 msrpm = is_guest_mode(vcpu) ? to_svm(vcpu)->nested.msrpm: in msr_write_intercepted()1561 if (is_guest_mode(&svm->vcpu)) { in svm_clear_vintr()2199 if (is_guest_mode(vcpu)) { in emulate_svm_instr()2245 if (!is_guest_mode(vcpu)) in gp_interception()2442 if (!is_guest_mode(vcpu) || in check_selective_cr0_intercepted()2814 if (is_guest_mode(vcpu)) in svm_set_msr()3312 if (is_guest_mode(vcpu)) { in handle_exit()3438 if (is_guest_mode(vcpu) && nested_exit_on_nmi(svm)) in svm_nmi_blocked()3454 if (for_injection && is_guest_mode(vcpu) && nested_exit_on_nmi(svm)) in svm_nmi_allowed()3495 } else if (is_guest_mode(vcpu)) { in svm_interrupt_blocked()[all …]
840 pi.is_guest_mode = true; in svm_update_pi_irte()851 if (!ret && pi.is_guest_mode) in svm_update_pi_irte()863 pi.is_guest_mode = false; in svm_update_pi_irte()
444 return is_guest_mode(vcpu) && (svm->nested.ctl.int_ctl & V_INTR_MASKING_MASK); in nested_svm_virtualize_tpr()
3395 pi_data->is_guest_mode = false; in amd_ir_set_vcpu_affinity()3403 if (pi_data->is_guest_mode) { in amd_ir_set_vcpu_affinity()
3923 if (is_guest_mode(vcpu)) { in kvm_faultin_pfn()4701 role.base.guest_mode = is_guest_mode(vcpu); in kvm_calc_mmu_role_common()5322 if (!mmio_info_in_cache(vcpu, cr2_or_gpa, direct) && !is_guest_mode(vcpu)) in kvm_mmu_page_fault()