Searched refs:is_guest_mode (Results 1 – 15 of 15) sorted by relevance
807 if (is_guest_mode(vcpu)) in vmx_update_exception_bitmap()1650 if (!is_guest_mode(vcpu)) in vmx_update_emulated_instruction()2089 if (is_guest_mode(vcpu)) in vmx_set_msr()2094 if (is_guest_mode(vcpu)) { in vmx_set_msr()2101 if (is_guest_mode(vcpu)) { in vmx_set_msr()2122 if (is_guest_mode(vcpu) && get_vmcs12(vcpu)->vm_exit_controls & in vmx_set_msr()2141 if (is_guest_mode(vcpu) && in vmx_set_msr()2224 if (is_guest_mode(vcpu) && in vmx_set_msr()3057 if (is_guest_mode(vcpu)) in vmx_get_current_vpid()3196 } else if (!is_guest_mode(vcpu)) { in vmx_set_cr0()[all …]
3265 if (is_guest_mode(vcpu) && !nested_get_vmcs12_pages(vcpu)) in vmx_get_nested_state_pages()3277 if (WARN_ON_ONCE(!is_guest_mode(vcpu))) in nested_vmx_write_pml_buffer()5265 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) in handle_vmread()5289 (is_guest_mode(vcpu) && in handle_vmread()5297 if (!is_guest_mode(vcpu) && is_vmcs12_ext_field(field)) in handle_vmread()5313 if (WARN_ON_ONCE(is_guest_mode(vcpu))) in handle_vmread()5371 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) in handle_vmwrite()5399 (is_guest_mode(vcpu) && in handle_vmwrite()5433 if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) in handle_vmwrite()5455 if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) { in handle_vmwrite()[all …]
490 if (!vmcs12 && is_guest_mode(vcpu)) in vmx_write_encls_bitmap()
722 return enable_unrestricted_guest && (!is_guest_mode(vcpu) || in is_unrestricted_guest()
107 if (is_guest_mode(&svm->vcpu) && in avic_deactivate_vmcb()536 if (is_guest_mode(vcpu)) in avic_vcpu_get_apicv_inhibit_reasons()938 pi.is_guest_mode = true; in avic_pi_update_irte()949 if (!ret && pi.is_guest_mode) in avic_pi_update_irte()961 pi.is_guest_mode = false; in avic_pi_update_irte()
726 msrpm = is_guest_mode(vcpu) ? to_svm(vcpu)->nested.msrpm: in msr_write_intercepted()921 if (is_guest_mode(vcpu)) in svm_enable_lbrv()939 if (is_guest_mode(vcpu)) in svm_disable_lbrv()984 if (unlikely(is_guest_mode(vcpu) && svm->lbrv_enabled)) in svm_update_lbrv()1588 if (is_guest_mode(&svm->vcpu)) { in svm_clear_vintr()2261 if (is_guest_mode(vcpu)) { in emulate_svm_instr()2303 if (!is_guest_mode(vcpu)) in gp_interception()2507 if (!is_guest_mode(vcpu) || in check_selective_cr0_intercepted()2890 if (svm->tsc_scaling_enabled && is_guest_mode(vcpu)) in svm_set_msr()2899 if (is_guest_mode(vcpu)) in svm_set_msr()[all …]
129 if (!is_guest_mode(&svm->vcpu)) in recalc_intercepts()1159 if (is_guest_mode(vcpu)) { in svm_leave_nested()1515 if (is_guest_mode(vcpu)) { in svm_get_nested_state()1530 if (!is_guest_mode(vcpu)) in svm_get_nested_state()1652 if (is_guest_mode(vcpu)) in svm_set_nested_state()1697 if (WARN_ON(!is_guest_mode(vcpu))) in svm_get_nested_state_pages()
484 if (is_guest_mode(&svm->vcpu) && !nested_vgif_enabled(svm)) in get_vgif_vmcb()573 return is_guest_mode(vcpu) && (svm->nested.ctl.int_ctl & V_INTR_MASKING_MASK); in nested_svm_virtualize_tpr()
24 bool is_guest_mode; member
87 if (!is_guest_mode(v) && kvm_vcpu_apicv_active(v)) in kvm_cpu_has_injectable_intr()
198 static inline bool is_guest_mode(struct kvm_vcpu *vcpu) in is_guest_mode() function
653 if (!reinject && is_guest_mode(vcpu) && in kvm_multiple_exception()689 if (!is_guest_mode(vcpu)) in kvm_multiple_exception()778 if (is_guest_mode(vcpu) && fault->async_page_fault) in kvm_inject_page_fault()2572 if (is_guest_mode(vcpu)) in kvm_vcpu_write_tsc_offset()2588 if (is_guest_mode(vcpu)) in kvm_vcpu_write_tsc_multiplier()8456 if (!is_guest_mode(vcpu) && static_call(kvm_x86_get_cpl)(vcpu) == 0) { in handle_emulation_failure()8474 if (WARN_ON_ONCE(is_guest_mode(vcpu)) || in reexecute_instruction()8566 if (WARN_ON_ONCE(is_guest_mode(vcpu)) || in retry_instruction()9884 if (is_guest_mode(vcpu)) in kvm_check_and_inject_events()10043 if (is_guest_mode(vcpu) && in kvm_check_and_inject_events()[all …]
3032 if (is_guest_mode(vcpu)) { in kvm_apic_accept_events()
3575 pi_data->is_guest_mode = false; in amd_ir_set_vcpu_affinity()3579 if (pi_data->is_guest_mode) { in amd_ir_set_vcpu_affinity()
4155 if (is_guest_mode(vcpu)) { in kvm_faultin_pfn()4966 role.base.guest_mode = is_guest_mode(vcpu); in kvm_calc_cpu_role()5585 if (!mmio_info_in_cache(vcpu, cr2_or_gpa, direct) && !is_guest_mode(vcpu)) in kvm_mmu_page_fault()