Lines Matching refs:vmcb01
92 kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, svm->vmcb01.ptr->save.cr4, in nested_svm_init_mmu_context()
93 svm->vmcb01.ptr->save.efer, in nested_svm_init_mmu_context()
133 h = &svm->vmcb01.ptr->control; in recalc_intercepts()
524 svm->nested.vmcb02.ptr->save.g_pat = svm->vmcb01.ptr->save.g_pat; in nested_vmcb02_compute_g_pat()
530 struct vmcb *vmcb01 = svm->vmcb01.ptr; in nested_vmcb02_prepare_save() local
591 } else if (unlikely(vmcb01->control.virt_ext & LBR_CTL_ENABLE_MASK)) { in nested_vmcb02_prepare_save()
592 svm_copy_lbrs(vmcb02, vmcb01); in nested_vmcb02_prepare_save()
628 struct vmcb *vmcb01 = svm->vmcb01.ptr; in nested_vmcb02_prepare_control() local
644 vmcb02->control.nested_ctl = vmcb01->control.nested_ctl; in nested_vmcb02_prepare_control()
645 vmcb02->control.iopm_base_pa = vmcb01->control.iopm_base_pa; in nested_vmcb02_prepare_control()
646 vmcb02->control.msrpm_base_pa = vmcb01->control.msrpm_base_pa; in nested_vmcb02_prepare_control()
671 (vmcb01->control.int_ctl & int_ctl_vmcb01_bits); in nested_vmcb02_prepare_control()
702 vmcb02->control.virt_ext = vmcb01->control.virt_ext & in nested_vmcb02_prepare_control()
720 vmcb02->control.pause_filter_count = vmcb01->control.pause_filter_count; in nested_vmcb02_prepare_control()
721 vmcb02->control.pause_filter_thresh = vmcb01->control.pause_filter_thresh; in nested_vmcb02_prepare_control()
784 nested_svm_copy_common_state(svm->vmcb01.ptr, svm->nested.vmcb02.ptr); in enter_svm_guest_mode()
813 struct vmcb *vmcb01 = svm->vmcb01.ptr; in nested_svm_vmrun() local
857 vmcb01->save.efer = vcpu->arch.efer; in nested_svm_vmrun()
858 vmcb01->save.cr0 = kvm_read_cr0(vcpu); in nested_svm_vmrun()
859 vmcb01->save.cr4 = vcpu->arch.cr4; in nested_svm_vmrun()
860 vmcb01->save.rflags = kvm_get_rflags(vcpu); in nested_svm_vmrun()
861 vmcb01->save.rip = kvm_rip_read(vcpu); in nested_svm_vmrun()
864 vmcb01->save.cr3 = kvm_read_cr3(vcpu); in nested_svm_vmrun()
932 struct vmcb *vmcb01 = svm->vmcb01.ptr; in nested_svm_vmexit() local
996 vmcb01->control.pause_filter_count = vmcb02->control.pause_filter_count; in nested_svm_vmexit()
997 vmcb_mark_dirty(vmcb01, VMCB_INTERCEPTS); in nested_svm_vmexit()
1001 nested_svm_copy_common_state(svm->nested.vmcb02.ptr, svm->vmcb01.ptr); in nested_svm_vmexit()
1003 svm_switch_vmcb(svm, &svm->vmcb01); in nested_svm_vmexit()
1008 } else if (unlikely(vmcb01->control.virt_ext & LBR_CTL_ENABLE_MASK)) { in nested_svm_vmexit()
1009 svm_copy_lbrs(vmcb01, vmcb02); in nested_svm_vmexit()
1018 vmcb01->control.exit_int_info = 0; in nested_svm_vmexit()
1021 if (vmcb01->control.tsc_offset != svm->vcpu.arch.tsc_offset) { in nested_svm_vmexit()
1022 vmcb01->control.tsc_offset = svm->vcpu.arch.tsc_offset; in nested_svm_vmexit()
1023 vmcb_mark_dirty(vmcb01, VMCB_INTERCEPTS); in nested_svm_vmexit()
1037 kvm_set_rflags(vcpu, vmcb01->save.rflags); in nested_svm_vmexit()
1038 svm_set_efer(vcpu, vmcb01->save.efer); in nested_svm_vmexit()
1039 svm_set_cr0(vcpu, vmcb01->save.cr0 | X86_CR0_PE); in nested_svm_vmexit()
1040 svm_set_cr4(vcpu, vmcb01->save.cr4); in nested_svm_vmexit()
1041 kvm_rax_write(vcpu, vmcb01->save.rax); in nested_svm_vmexit()
1042 kvm_rsp_write(vcpu, vmcb01->save.rsp); in nested_svm_vmexit()
1043 kvm_rip_write(vcpu, vmcb01->save.rip); in nested_svm_vmexit()
1061 rc = nested_svm_load_cr3(vcpu, vmcb01->save.cr3, false, true); in nested_svm_vmexit()
1079 if (unlikely(vmcb01->save.rflags & X86_EFLAGS_TF)) in nested_svm_vmexit()
1134 if (WARN_ON_ONCE(svm->vmcb != svm->vmcb01.ptr)) in svm_free_nested()
1135 svm_switch_vmcb(svm, &svm->vmcb01); in svm_free_nested()
1165 svm_switch_vmcb(svm, &svm->vmcb01); in svm_leave_nested()
1429 if (svm->vmcb01.ptr->control.intercepts[INTERCEPT_EXCEPTION] & in nested_svm_exit_special()
1551 if (copy_to_user(&user_vmcb->save, &svm->vmcb01.ptr->save, in svm_get_nested_state()
1655 svm->nested.vmcb02.ptr->save = svm->vmcb01.ptr->save; in svm_set_nested_state()
1664 svm_copy_vmrun_state(&svm->vmcb01.ptr->save, save); in svm_set_nested_state()