Lines Matching refs:efer
815 u64 efer = 0; in assign_eip_far() local
817 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in assign_eip_far()
818 if (efer & EFER_LMA) in assign_eip_far()
1601 u64 efer = 0; in get_descriptor_ptr() local
1603 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in get_descriptor_ptr()
1604 if (!(efer & EFER_LMA)) in get_descriptor_ptr()
1752 u64 efer = 0; in __load_segment_descriptor() local
1754 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in __load_segment_descriptor()
1755 if (efer & EFER_LMA) in __load_segment_descriptor()
2624 unsigned long cr0, cr4, efer; in em_rsm() local
2676 efer = 0; in em_rsm()
2677 ctxt->ops->set_msr(ctxt, MSR_EFER, efer); in em_rsm()
2782 u64 efer = 0; in em_syscall() local
2792 ops->get_msr(ctxt, MSR_EFER, &efer); in em_syscall()
2793 if (!(efer & EFER_SCE)) in em_syscall()
2802 if (efer & EFER_LMA) { in em_syscall()
2810 if (efer & EFER_LMA) { in em_syscall()
2841 u64 efer = 0; in em_sysenter() local
2843 ops->get_msr(ctxt, MSR_EFER, &efer); in em_sysenter()
2852 if ((ctxt->mode != X86EMUL_MODE_PROT64) && (efer & EFER_LMA) in em_sysenter()
2868 if (efer & EFER_LMA) { in em_sysenter()
2877 ctxt->_eip = (efer & EFER_LMA) ? msr_data : (u32)msr_data; in em_sysenter()
2880 *reg_write(ctxt, VCPU_REGS_RSP) = (efer & EFER_LMA) ? msr_data : in em_sysenter()
4233 u64 efer = 0; in check_cr_write() local
4257 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_cr_write()
4259 if ((new_val & X86_CR0_PG) && (efer & EFER_LME) && in check_cr_write()
4268 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_cr_write()
4269 if (efer & EFER_LMA) { in check_cr_write()
4291 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_cr_write()
4293 if ((efer & EFER_LMA) && !(new_val & X86_CR4_PAE)) in check_cr_write()
4351 u64 efer = 0; in check_svme() local
4353 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_svme()
4355 if (!(efer & EFER_SVME)) in check_svme()