Home
last modified time | relevance | path

Searched refs:efer (Results 1 – 25 of 27) sorted by relevance

12

/Linux-v5.10/arch/x86/include/asm/
Dvirtext.h112 uint64_t efer; in cpu_svm_disable() local
115 rdmsrl(MSR_EFER, efer); in cpu_svm_disable()
116 wrmsrl(MSR_EFER, efer & ~EFER_SVME); in cpu_svm_disable()
Dsuspend_64.h41 unsigned long efer; member
Drealmode.h51 u64 efer;
Dsvm.h222 u64 efer; member
Dkvm_host.h543 u64 efer; member
1119 int (*set_efer)(struct kvm_vcpu *vcpu, u64 efer);
1453 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer);
/Linux-v5.10/arch/x86/realmode/
Dinit.c72 u64 efer; in setup_real_mode() local
119 rdmsrl(MSR_EFER, efer); in setup_real_mode()
120 trampoline_header->efer = efer & ~EFER_LMA; in setup_real_mode()
/Linux-v5.10/tools/testing/selftests/kvm/lib/x86_64/
Dsvm.c70 uint64_t efer; in generic_svm_setup() local
72 efer = rdmsr(MSR_EFER); in generic_svm_setup()
73 wrmsr(MSR_EFER, efer | EFER_SVME); in generic_svm_setup()
87 save->efer = rdmsr(MSR_EFER); in generic_svm_setup()
Dprocessor.c203 sregs->cr8, sregs->efer, sregs->apic_base); in sregs_dump()
566 sregs.efer |= (EFER_LME | EFER_LMA | EFER_NX); in vcpu_setup()
/Linux-v5.10/include/xen/interface/hvm/
Dhvm_vcpu.h52 uint64_t efer; member
106 uint64_t efer; member
/Linux-v5.10/arch/x86/kvm/svm/
Dnested.c83 kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, hsave->save.cr4, hsave->save.efer, in nested_svm_init_mmu_context()
232 if ((vmcb12->save.efer & EFER_SVME) == 0) in nested_vmcb_checks()
241 vmcb12_lma = (vmcb12->save.efer & EFER_LME) && (vmcb12->save.cr0 & X86_CR0_PG); in nested_vmcb_checks()
385 svm_set_efer(&svm->vcpu, vmcb12->save.efer); in nested_prepare_vmcb_save()
522 hsave->save.efer = svm->vcpu.arch.efer; in nested_svm_vmrun()
609 vmcb12->save.efer = svm->vcpu.arch.efer; in nested_svm_vmexit()
663 svm_set_efer(&svm->vcpu, hsave->save.efer); in nested_svm_vmexit()
869 if (!(svm->vcpu.arch.efer & EFER_SVME) || in nested_svm_check_permissions()
1137 if (!(vcpu->arch.efer & EFER_SVME)) { in svm_set_nested_state()
Dsvm.c266 int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer) in svm_set_efer() argument
269 u64 old_efer = vcpu->arch.efer; in svm_set_efer()
270 vcpu->arch.efer = efer; in svm_set_efer()
274 efer |= EFER_NX; in svm_set_efer()
276 if (!(efer & EFER_LMA)) in svm_set_efer()
277 efer &= ~EFER_LME; in svm_set_efer()
280 if ((old_efer & EFER_SVME) != (efer & EFER_SVME)) { in svm_set_efer()
281 if (!(efer & EFER_SVME)) { in svm_set_efer()
297 vcpu->arch.efer = old_efer; in svm_set_efer()
303 svm->vmcb->save.efer = efer | EFER_SVME; in svm_set_efer()
[all …]
Dsvm.h359 int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer);
/Linux-v5.10/arch/x86/power/
Dcpu.c116 rdmsrl(MSR_EFER, ctxt->efer); in __save_processor_state()
209 wrmsrl(MSR_EFER, ctxt->efer); in __restore_processor_state()
/Linux-v5.10/arch/x86/kvm/
Demulate.c815 u64 efer = 0; in assign_eip_far() local
817 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in assign_eip_far()
818 if (efer & EFER_LMA) in assign_eip_far()
1601 u64 efer = 0; in get_descriptor_ptr() local
1603 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in get_descriptor_ptr()
1604 if (!(efer & EFER_LMA)) in get_descriptor_ptr()
1752 u64 efer = 0; in __load_segment_descriptor() local
1754 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in __load_segment_descriptor()
1755 if (efer & EFER_LMA) in __load_segment_descriptor()
2624 unsigned long cr0, cr4, efer; in em_rsm() local
[all …]
Dx86.h88 return vcpu->arch.efer & EFER_LMA; in is_long_mode()
107 return (vcpu->arch.efer & EFER_LMA) && in is_la57_mode()
Dmmu.h61 void kvm_init_shadow_npt_mmu(struct kvm_vcpu *vcpu, u32 cr0, u32 cr4, u32 efer,
Dx86.c829 if ((vcpu->arch.efer & EFER_LME) && !is_paging(vcpu) && in kvm_set_cr0()
840 if (!(vcpu->arch.efer & EFER_LME) && (cr0 & X86_CR0_PG) && in kvm_set_cr0()
1430 static bool __kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer) in __kvm_valid_efer() argument
1432 if (efer & EFER_FFXSR && !guest_cpuid_has(vcpu, X86_FEATURE_FXSR_OPT)) in __kvm_valid_efer()
1435 if (efer & EFER_SVME && !guest_cpuid_has(vcpu, X86_FEATURE_SVM)) in __kvm_valid_efer()
1438 if (efer & (EFER_LME | EFER_LMA) && in __kvm_valid_efer()
1442 if (efer & EFER_NX && !guest_cpuid_has(vcpu, X86_FEATURE_NX)) in __kvm_valid_efer()
1448 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer) in kvm_valid_efer() argument
1450 if (efer & efer_reserved_bits) in kvm_valid_efer()
1453 return __kvm_valid_efer(vcpu, efer); in kvm_valid_efer()
[all …]
/Linux-v5.10/tools/testing/selftests/kvm/include/x86_64/
Dsvm.h178 u64 efer; member
/Linux-v5.10/arch/x86/kvm/vmx/
Dvmx.c1017 u64 guest_efer = vmx->vcpu.arch.efer; in update_transition_efer()
1042 (enable_ept && ((vmx->vcpu.arch.efer ^ host_efer) & EFER_NX))) { in update_transition_efer()
1717 if (is_long_mode(&vmx->vcpu) && (vmx->vcpu.arch.efer & EFER_SCE)) { in setup_msrs()
2835 int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer) in vmx_set_efer() argument
2844 vcpu->arch.efer = efer; in vmx_set_efer()
2845 if (efer & EFER_LMA) { in vmx_set_efer()
2847 msr->data = efer; in vmx_set_efer()
2851 msr->data = efer & ~EFER_LME; in vmx_set_efer()
2873 vmx_set_efer(vcpu, vcpu->arch.efer | EFER_LMA); in enter_lmode()
2879 vmx_set_efer(vcpu, vcpu->arch.efer & ~EFER_LMA); in exit_lmode()
[all …]
Dvmx.h322 int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer);
Dnested.c2143 return vmx->vcpu.arch.efer | (EFER_LMA | EFER_LME); in nested_vmx_calc_efer()
2145 return vmx->vcpu.arch.efer & ~(EFER_LMA | EFER_LME); in nested_vmx_calc_efer()
2571 vcpu->arch.efer = nested_vmx_calc_efer(vmx, vmcs12); in prepare_vmcs02()
2573 vmx_set_efer(vcpu, vcpu->arch.efer); in prepare_vmcs02()
2870 ia32e = !!(vcpu->arch.efer & EFER_LMA); in nested_vmx_check_host_state()
4079 vmcs12->guest_ia32_efer = vcpu->arch.efer; in sync_vmcs02_to_vmcs12()
4157 vcpu->arch.efer = vmcs12->host_ia32_efer; in load_vmcs12_host_state()
4159 vcpu->arch.efer |= (EFER_LMA | EFER_LME); in load_vmcs12_host_state()
4161 vcpu->arch.efer &= ~(EFER_LMA | EFER_LME); in load_vmcs12_host_state()
4162 vmx_set_efer(vcpu, vcpu->arch.efer); in load_vmcs12_host_state()
/Linux-v5.10/drivers/parport/
Dparport_pc.c1095 static void decode_winbond(int efer, int key, int devid, int devrev, int oldid) in decode_winbond() argument
1140 efer, key, devid, devrev, oldid, type); in decode_winbond()
1143 show_parconfig_winbond(efer, key); in decode_winbond()
1146 static void decode_smsc(int efer, int key, int devid, int devrev) in decode_smsc() argument
1172 efer, key, devid, devrev, type); in decode_smsc()
1175 func(efer, key); in decode_smsc()
/Linux-v5.10/arch/x86/include/uapi/asm/
Dkvm.h155 __u64 efer; member
/Linux-v5.10/tools/arch/x86/include/uapi/asm/
Dkvm.h155 __u64 efer; member
/Linux-v5.10/arch/x86/kvm/mmu/
Dmmu.c284 return vcpu->arch.efer & EFER_NX; in is_nx()
4562 u32 cr0, u32 cr4, u32 efer, in shadow_mmu_init_context() argument
4567 else if (efer & EFER_LMA) in shadow_mmu_init_context()
4578 static void kvm_init_shadow_mmu(struct kvm_vcpu *vcpu, u32 cr0, u32 cr4, u32 efer) in kvm_init_shadow_mmu() argument
4585 shadow_mmu_init_context(vcpu, context, cr0, cr4, efer, new_role); in kvm_init_shadow_mmu()
4600 void kvm_init_shadow_npt_mmu(struct kvm_vcpu *vcpu, u32 cr0, u32 cr4, u32 efer, in kvm_init_shadow_npt_mmu() argument
4611 shadow_mmu_init_context(vcpu, context, cr0, cr4, efer, new_role); in kvm_init_shadow_npt_mmu()
4686 vcpu->arch.efer); in init_kvm_softmmu()

12