Lines Matching refs:efer
809 u64 efer = 0; in assign_eip_far() local
811 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in assign_eip_far()
812 if (efer & EFER_LMA) in assign_eip_far()
1568 u64 efer = 0; in get_descriptor_ptr() local
1570 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in get_descriptor_ptr()
1571 if (!(efer & EFER_LMA)) in get_descriptor_ptr()
1719 u64 efer = 0; in __load_segment_descriptor() local
1721 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in __load_segment_descriptor()
1722 if (efer & EFER_LMA) in __load_segment_descriptor()
2584 unsigned long cr0, cr4, efer; in em_rsm() local
2636 efer = 0; in em_rsm()
2637 ctxt->ops->set_msr(ctxt, MSR_EFER, efer); in em_rsm()
2762 u64 efer = 0; in em_syscall() local
2772 ops->get_msr(ctxt, MSR_EFER, &efer); in em_syscall()
2775 if (!(efer & EFER_SCE)) in em_syscall()
2783 if (efer & EFER_LMA) { in em_syscall()
2791 if (efer & EFER_LMA) { in em_syscall()
2822 u64 efer = 0; in em_sysenter() local
2824 ops->get_msr(ctxt, MSR_EFER, &efer); in em_sysenter()
2833 if ((ctxt->mode != X86EMUL_MODE_PROT64) && (efer & EFER_LMA) in em_sysenter()
2850 if (efer & EFER_LMA) { in em_sysenter()
2859 ctxt->_eip = (efer & EFER_LMA) ? msr_data : (u32)msr_data; in em_sysenter()
2862 *reg_write(ctxt, VCPU_REGS_RSP) = (efer & EFER_LMA) ? msr_data : in em_sysenter()
4197 u64 efer = 0; in check_cr_write() local
4221 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_cr_write()
4223 if ((new_val & X86_CR0_PG) && (efer & EFER_LME) && in check_cr_write()
4232 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_cr_write()
4233 if (efer & EFER_LMA) { in check_cr_write()
4255 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_cr_write()
4257 if ((efer & EFER_LMA) && !(new_val & X86_CR4_PAE)) in check_cr_write()
4315 u64 efer = 0; in check_svme() local
4317 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_svme()
4319 if (!(efer & EFER_SVME)) in check_svme()