Lines Matching refs:efer
2782 u64 guest_efer = vmx->vcpu.arch.efer; in update_transition_efer()
2816 (enable_ept && ((vmx->vcpu.arch.efer ^ host_efer) & EFER_NX))) { in update_transition_efer()
3408 if ((index >= 0) && (vmx->vcpu.arch.efer & EFER_SCE)) in setup_msrs()
5064 static void vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer) in vmx_set_efer() argument
5072 vcpu->arch.efer = efer; in vmx_set_efer()
5073 if (efer & EFER_LMA) { in vmx_set_efer()
5075 msr->data = efer; in vmx_set_efer()
5079 msr->data = efer & ~EFER_LME; in vmx_set_efer()
5100 vmx_set_efer(vcpu, vcpu->arch.efer | EFER_LMA); in enter_lmode()
5106 vmx_set_efer(vcpu, vcpu->arch.efer & ~EFER_LMA); in exit_lmode()
5281 if (vcpu->arch.efer & EFER_LME) { in vmx_set_cr0()
9903 u64 efer = vmcs_read64(GUEST_IA32_EFER); in dump_vmcs() local
9917 (cr4 & X86_CR4_PAE) && !(efer & EFER_LMA)) in dump_vmcs()
9944 efer, vmcs_read64(GUEST_IA32_PAT)); in dump_vmcs()
12339 vcpu->arch.efer = vmcs12->guest_ia32_efer; in prepare_vmcs02()
12341 vcpu->arch.efer |= (EFER_LMA | EFER_LME); in prepare_vmcs02()
12343 vcpu->arch.efer &= ~(EFER_LMA | EFER_LME); in prepare_vmcs02()
12345 vmx_set_efer(vcpu, vcpu->arch.efer); in prepare_vmcs02()
13066 vmcs12->guest_ia32_efer = vcpu->arch.efer; in sync_vmcs12()
13158 vcpu->arch.efer = vmcs12->host_ia32_efer; in load_vmcs12_host_state()
13160 vcpu->arch.efer |= (EFER_LMA | EFER_LME); in load_vmcs12_host_state()
13162 vcpu->arch.efer &= ~(EFER_LMA | EFER_LME); in load_vmcs12_host_state()
13163 vmx_set_efer(vcpu, vcpu->arch.efer); in load_vmcs12_host_state()