Lines Matching full:save
92 kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, svm->vmcb01.ptr->save.cr4, in nested_svm_init_mmu_context()
93 svm->vmcb01.ptr->save.efer, in nested_svm_init_mmu_context()
282 struct vmcb_save_area_cached *save) in __nested_vmcb_check_save() argument
284 if (CC(!(save->efer & EFER_SVME))) in __nested_vmcb_check_save()
287 if (CC((save->cr0 & X86_CR0_CD) == 0 && (save->cr0 & X86_CR0_NW)) || in __nested_vmcb_check_save()
288 CC(save->cr0 & ~0xffffffffULL)) in __nested_vmcb_check_save()
291 if (CC(!kvm_dr6_valid(save->dr6)) || CC(!kvm_dr7_valid(save->dr7))) in __nested_vmcb_check_save()
299 if ((save->efer & EFER_LME) && (save->cr0 & X86_CR0_PG)) { in __nested_vmcb_check_save()
300 if (CC(!(save->cr4 & X86_CR4_PAE)) || in __nested_vmcb_check_save()
301 CC(!(save->cr0 & X86_CR0_PE)) || in __nested_vmcb_check_save()
302 CC(kvm_vcpu_is_illegal_gpa(vcpu, save->cr3))) in __nested_vmcb_check_save()
307 if (CC(!__kvm_is_valid_cr4(vcpu, save->cr4))) in __nested_vmcb_check_save()
310 if (CC(!kvm_valid_efer(vcpu, save->efer))) in __nested_vmcb_check_save()
319 struct vmcb_save_area_cached *save = &svm->nested.save; in nested_vmcb_check_save() local
321 return __nested_vmcb_check_save(vcpu, save); in nested_vmcb_check_save()
400 struct vmcb_save_area *save) in nested_copy_vmcb_save_to_cache() argument
402 __nested_copy_vmcb_save_to_cache(&svm->nested.save, save); in nested_copy_vmcb_save_to_cache()
524 svm->nested.vmcb02.ptr->save.g_pat = svm->vmcb01.ptr->save.g_pat; in nested_vmcb02_compute_g_pat()
543 vmcb02->save.es = vmcb12->save.es; in nested_vmcb02_prepare_save()
544 vmcb02->save.cs = vmcb12->save.cs; in nested_vmcb02_prepare_save()
545 vmcb02->save.ss = vmcb12->save.ss; in nested_vmcb02_prepare_save()
546 vmcb02->save.ds = vmcb12->save.ds; in nested_vmcb02_prepare_save()
547 vmcb02->save.cpl = vmcb12->save.cpl; in nested_vmcb02_prepare_save()
552 vmcb02->save.gdtr = vmcb12->save.gdtr; in nested_vmcb02_prepare_save()
553 vmcb02->save.idtr = vmcb12->save.idtr; in nested_vmcb02_prepare_save()
557 kvm_set_rflags(&svm->vcpu, vmcb12->save.rflags | X86_EFLAGS_FIXED); in nested_vmcb02_prepare_save()
559 svm_set_efer(&svm->vcpu, svm->nested.save.efer); in nested_vmcb02_prepare_save()
561 svm_set_cr0(&svm->vcpu, svm->nested.save.cr0); in nested_vmcb02_prepare_save()
562 svm_set_cr4(&svm->vcpu, svm->nested.save.cr4); in nested_vmcb02_prepare_save()
564 svm->vcpu.arch.cr2 = vmcb12->save.cr2; in nested_vmcb02_prepare_save()
566 kvm_rax_write(&svm->vcpu, vmcb12->save.rax); in nested_vmcb02_prepare_save()
567 kvm_rsp_write(&svm->vcpu, vmcb12->save.rsp); in nested_vmcb02_prepare_save()
568 kvm_rip_write(&svm->vcpu, vmcb12->save.rip); in nested_vmcb02_prepare_save()
571 vmcb02->save.rax = vmcb12->save.rax; in nested_vmcb02_prepare_save()
572 vmcb02->save.rsp = vmcb12->save.rsp; in nested_vmcb02_prepare_save()
573 vmcb02->save.rip = vmcb12->save.rip; in nested_vmcb02_prepare_save()
577 vmcb02->save.dr7 = svm->nested.save.dr7 | DR7_FIXED_1; in nested_vmcb02_prepare_save()
578 svm->vcpu.arch.dr6 = svm->nested.save.dr6 | DR6_ACTIVE_LOW; in nested_vmcb02_prepare_save()
588 vmcb02->save.dbgctl &= ~DEBUGCTL_RESERVED_BITS; in nested_vmcb02_prepare_save()
753 to_vmcb->save.spec_ctrl = from_vmcb->save.spec_ctrl; in nested_svm_copy_common_state()
762 trace_kvm_nested_vmenter(svm->vmcb->save.rip, in enter_svm_guest_mode()
764 vmcb12->save.rip, in enter_svm_guest_mode()
769 vmcb12->save.cr3, in enter_svm_guest_mode()
787 nested_vmcb02_prepare_control(svm, vmcb12->save.rip, vmcb12->save.cs.base); in enter_svm_guest_mode()
790 ret = nested_svm_load_cr3(&svm->vcpu, svm->nested.save.cr3, in enter_svm_guest_mode()
825 vmcb12_gpa = svm->vmcb->save.rax; in nested_svm_vmrun()
842 nested_copy_vmcb_save_to_cache(svm, &vmcb12->save); in nested_svm_vmrun()
857 vmcb01->save.efer = vcpu->arch.efer; in nested_svm_vmrun()
858 vmcb01->save.cr0 = kvm_read_cr0(vcpu); in nested_svm_vmrun()
859 vmcb01->save.cr4 = vcpu->arch.cr4; in nested_svm_vmrun()
860 vmcb01->save.rflags = kvm_get_rflags(vcpu); in nested_svm_vmrun()
861 vmcb01->save.rip = kvm_rip_read(vcpu); in nested_svm_vmrun()
864 vmcb01->save.cr3 = kvm_read_cr3(vcpu); in nested_svm_vmrun()
892 /* Copy state save area fields which are handled by VMRUN */
915 to_vmcb->save.fs = from_vmcb->save.fs; in svm_copy_vmloadsave_state()
916 to_vmcb->save.gs = from_vmcb->save.gs; in svm_copy_vmloadsave_state()
917 to_vmcb->save.tr = from_vmcb->save.tr; in svm_copy_vmloadsave_state()
918 to_vmcb->save.ldtr = from_vmcb->save.ldtr; in svm_copy_vmloadsave_state()
919 to_vmcb->save.kernel_gs_base = from_vmcb->save.kernel_gs_base; in svm_copy_vmloadsave_state()
920 to_vmcb->save.star = from_vmcb->save.star; in svm_copy_vmloadsave_state()
921 to_vmcb->save.lstar = from_vmcb->save.lstar; in svm_copy_vmloadsave_state()
922 to_vmcb->save.cstar = from_vmcb->save.cstar; in svm_copy_vmloadsave_state()
923 to_vmcb->save.sfmask = from_vmcb->save.sfmask; in svm_copy_vmloadsave_state()
924 to_vmcb->save.sysenter_cs = from_vmcb->save.sysenter_cs; in svm_copy_vmloadsave_state()
925 to_vmcb->save.sysenter_esp = from_vmcb->save.sysenter_esp; in svm_copy_vmloadsave_state()
926 to_vmcb->save.sysenter_eip = from_vmcb->save.sysenter_eip; in svm_copy_vmloadsave_state()
959 vmcb12->save.es = vmcb02->save.es; in nested_svm_vmexit()
960 vmcb12->save.cs = vmcb02->save.cs; in nested_svm_vmexit()
961 vmcb12->save.ss = vmcb02->save.ss; in nested_svm_vmexit()
962 vmcb12->save.ds = vmcb02->save.ds; in nested_svm_vmexit()
963 vmcb12->save.gdtr = vmcb02->save.gdtr; in nested_svm_vmexit()
964 vmcb12->save.idtr = vmcb02->save.idtr; in nested_svm_vmexit()
965 vmcb12->save.efer = svm->vcpu.arch.efer; in nested_svm_vmexit()
966 vmcb12->save.cr0 = kvm_read_cr0(vcpu); in nested_svm_vmexit()
967 vmcb12->save.cr3 = kvm_read_cr3(vcpu); in nested_svm_vmexit()
968 vmcb12->save.cr2 = vmcb02->save.cr2; in nested_svm_vmexit()
969 vmcb12->save.cr4 = svm->vcpu.arch.cr4; in nested_svm_vmexit()
970 vmcb12->save.rflags = kvm_get_rflags(vcpu); in nested_svm_vmexit()
971 vmcb12->save.rip = kvm_rip_read(vcpu); in nested_svm_vmexit()
972 vmcb12->save.rsp = kvm_rsp_read(vcpu); in nested_svm_vmexit()
973 vmcb12->save.rax = kvm_rax_read(vcpu); in nested_svm_vmexit()
974 vmcb12->save.dr7 = vmcb02->save.dr7; in nested_svm_vmexit()
975 vmcb12->save.dr6 = svm->vcpu.arch.dr6; in nested_svm_vmexit()
976 vmcb12->save.cpl = vmcb02->save.cpl; in nested_svm_vmexit()
1037 kvm_set_rflags(vcpu, vmcb01->save.rflags); in nested_svm_vmexit()
1038 svm_set_efer(vcpu, vmcb01->save.efer); in nested_svm_vmexit()
1039 svm_set_cr0(vcpu, vmcb01->save.cr0 | X86_CR0_PE); in nested_svm_vmexit()
1040 svm_set_cr4(vcpu, vmcb01->save.cr4); in nested_svm_vmexit()
1041 kvm_rax_write(vcpu, vmcb01->save.rax); in nested_svm_vmexit()
1042 kvm_rsp_write(vcpu, vmcb01->save.rsp); in nested_svm_vmexit()
1043 kvm_rip_write(vcpu, vmcb01->save.rip); in nested_svm_vmexit()
1061 rc = nested_svm_load_cr3(vcpu, vmcb01->save.cr3, false, true); in nested_svm_vmexit()
1079 if (unlikely(vmcb01->save.rflags & X86_EFLAGS_TF)) in nested_svm_vmexit()
1287 if (to_svm(vcpu)->vmcb->save.cpl) { in nested_svm_check_permissions()
1409 trace_kvm_nested_intr_vmexit(svm->vmcb->save.rip); in svm_check_nested_events()
1551 if (copy_to_user(&user_vmcb->save, &svm->vmcb01.ptr->save, in svm_get_nested_state()
1552 sizeof(user_vmcb->save))) in svm_get_nested_state()
1566 struct vmcb_save_area *save; in svm_set_nested_state() local
1610 save = kzalloc(sizeof(*save), GFP_KERNEL_ACCOUNT); in svm_set_nested_state()
1611 if (!ctl || !save) in svm_set_nested_state()
1617 if (copy_from_user(save, &user_vmcb->save, sizeof(*save))) in svm_set_nested_state()
1637 __nested_copy_vmcb_save_to_cache(&save_cached, save); in svm_set_nested_state()
1638 if (!(save->cr0 & X86_CR0_PG) || in svm_set_nested_state()
1639 !(save->cr0 & X86_CR0_PE) || in svm_set_nested_state()
1640 (save->rflags & X86_EFLAGS_VM) || in svm_set_nested_state()
1648 * vmcb02, and the L1 save state which we store in vmcb01. in svm_set_nested_state()
1655 svm->nested.vmcb02.ptr->save = svm->vmcb01.ptr->save; in svm_set_nested_state()
1664 svm_copy_vmrun_state(&svm->vmcb01.ptr->save, save); in svm_set_nested_state()
1668 nested_vmcb02_prepare_control(svm, svm->vmcb->save.rip, svm->vmcb->save.cs.base); in svm_set_nested_state()
1687 kfree(save); in svm_set_nested_state()