Lines Matching refs:efer
772 if ((vcpu->arch.efer & EFER_LME)) { in kvm_set_cr0()
1368 static bool __kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer) in __kvm_valid_efer() argument
1370 if (efer & EFER_FFXSR && !guest_cpuid_has(vcpu, X86_FEATURE_FXSR_OPT)) in __kvm_valid_efer()
1373 if (efer & EFER_SVME && !guest_cpuid_has(vcpu, X86_FEATURE_SVM)) in __kvm_valid_efer()
1376 if (efer & (EFER_LME | EFER_LMA) && in __kvm_valid_efer()
1380 if (efer & EFER_NX && !guest_cpuid_has(vcpu, X86_FEATURE_NX)) in __kvm_valid_efer()
1386 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer) in kvm_valid_efer() argument
1388 if (efer & efer_reserved_bits) in kvm_valid_efer()
1391 return __kvm_valid_efer(vcpu, efer); in kvm_valid_efer()
1397 u64 old_efer = vcpu->arch.efer; in set_efer()
1398 u64 efer = msr_info->data; in set_efer() local
1400 if (efer & efer_reserved_bits) in set_efer()
1404 if (!__kvm_valid_efer(vcpu, efer)) in set_efer()
1408 (vcpu->arch.efer & EFER_LME) != (efer & EFER_LME)) in set_efer()
1412 efer &= ~EFER_LMA; in set_efer()
1413 efer |= vcpu->arch.efer & EFER_LMA; in set_efer()
1415 kvm_x86_ops->set_efer(vcpu, efer); in set_efer()
1418 if ((efer ^ old_efer) & EFER_NX) in set_efer()
3021 msr_info->data = vcpu->arch.efer; in kvm_get_msr_common()
7756 put_smstate(u64, buf, 0x7ed0, vcpu->arch.efer); in enter_smm_save_state_64()
8659 sregs->efer = vcpu->arch.efer; in __get_sregs()
8750 if ((sregs->efer & EFER_LME) && (sregs->cr0 & X86_CR0_PG)) { in kvm_valid_sregs()
8757 || !(sregs->efer & EFER_LMA)) in kvm_valid_sregs()
8764 if (sregs->efer & EFER_LMA || sregs->cs.l) in kvm_valid_sregs()
8802 mmu_reset_needed |= vcpu->arch.efer != sregs->efer; in __set_sregs()
8803 kvm_x86_ops->set_efer(vcpu, sregs->efer); in __set_sregs()