Lines Matching refs:vmcb01
107 kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, svm->vmcb01.ptr->save.cr4, in nested_svm_init_mmu_context()
108 svm->vmcb01.ptr->save.efer, in nested_svm_init_mmu_context()
133 h = &svm->vmcb01.ptr->control; in recalc_intercepts()
446 svm->nested.vmcb02.ptr->save.g_pat = svm->vmcb01.ptr->save.g_pat; in nested_vmcb02_compute_g_pat()
528 svm->vmcb->control.nested_ctl = svm->vmcb01.ptr->control.nested_ctl; in nested_vmcb02_prepare_control()
529 svm->vmcb->control.iopm_base_pa = svm->vmcb01.ptr->control.iopm_base_pa; in nested_vmcb02_prepare_control()
530 svm->vmcb->control.msrpm_base_pa = svm->vmcb01.ptr->control.msrpm_base_pa; in nested_vmcb02_prepare_control()
546 (svm->vmcb01.ptr->control.int_ctl & int_ctl_vmcb01_bits); in nested_vmcb02_prepare_control()
604 nested_svm_copy_common_state(svm->vmcb01.ptr, svm->nested.vmcb02.ptr); in enter_svm_guest_mode()
675 svm->vmcb01.ptr->save.efer = vcpu->arch.efer; in nested_svm_vmrun()
676 svm->vmcb01.ptr->save.cr0 = kvm_read_cr0(vcpu); in nested_svm_vmrun()
677 svm->vmcb01.ptr->save.cr4 = vcpu->arch.cr4; in nested_svm_vmrun()
678 svm->vmcb01.ptr->save.rflags = kvm_get_rflags(vcpu); in nested_svm_vmrun()
679 svm->vmcb01.ptr->save.rip = kvm_rip_read(vcpu); in nested_svm_vmrun()
682 svm->vmcb01.ptr->save.cr3 = kvm_read_cr3(vcpu); in nested_svm_vmrun()
818 nested_svm_copy_common_state(svm->nested.vmcb02.ptr, svm->vmcb01.ptr); in nested_svm_vmexit()
820 svm_switch_vmcb(svm, &svm->vmcb01); in nested_svm_vmexit()
955 svm_switch_vmcb(svm, &svm->vmcb01); in svm_leave_nested()
1206 if (svm->vmcb01.ptr->control.intercepts[INTERCEPT_EXCEPTION] & in nested_svm_exit_special()
1271 if (copy_to_user(&user_vmcb->save, &svm->vmcb01.ptr->save, in svm_get_nested_state()
1383 svm->nested.vmcb02.ptr->save = svm->vmcb01.ptr->save; in svm_set_nested_state()
1392 svm_copy_vmrun_state(&svm->vmcb01.ptr->save, save); in svm_set_nested_state()