| /Linux-v5.15/arch/x86/kvm/ |
| D | mmu.h | 86 if (likely(vcpu->arch.mmu->root_hpa != INVALID_PAGE)) in kvm_mmu_reload() 108 u64 root_hpa = vcpu->arch.mmu->root_hpa; in kvm_mmu_load_pgd() local 110 if (!VALID_PAGE(root_hpa)) in kvm_mmu_load_pgd() 113 static_call(kvm_x86_load_mmu_pgd)(vcpu, root_hpa, in kvm_mmu_load_pgd()
|
| D | x86.c | 744 fault_mmu->root_hpa); in kvm_inject_emulated_page_fault()
|
| /Linux-v5.15/arch/x86/kvm/mmu/ |
| D | mmu.c | 2201 BUG_ON(root != vcpu->arch.mmu->root_hpa); in shadow_walk_init_using_root() 2215 shadow_walk_init_using_root(iterator, vcpu, vcpu->arch.mmu->root_hpa, in shadow_walk_init() 3308 static void mmu_free_root_page(struct kvm *kvm, hpa_t *root_hpa, in mmu_free_root_page() argument 3313 if (!VALID_PAGE(*root_hpa)) in mmu_free_root_page() 3316 sp = to_shadow_page(*root_hpa & PT64_BASE_ADDR_MASK); in mmu_free_root_page() 3323 *root_hpa = INVALID_PAGE; in mmu_free_root_page() 3338 if (!(free_active_root && VALID_PAGE(mmu->root_hpa))) { in kvm_mmu_free_roots() 3358 mmu_free_root_page(kvm, &mmu->root_hpa, &invalid_list); in kvm_mmu_free_roots() 3369 mmu->root_hpa = INVALID_PAGE; in kvm_mmu_free_roots() 3381 hpa_t root_hpa; in kvm_mmu_free_guest_mode_roots() local [all …]
|
| D | mmu_audit.c | 59 if (!VALID_PAGE(vcpu->arch.mmu->root_hpa)) in mmu_spte_walk() 63 hpa_t root = vcpu->arch.mmu->root_hpa; in mmu_spte_walk()
|
| D | tdp_mmu.h | 101 hpa_t hpa = mmu->root_hpa; in is_tdp_mmu()
|
| D | paging_tmpl.h | 695 if (WARN_ON(!VALID_PAGE(vcpu->arch.mmu->root_hpa))) in FNAME() 960 static void FNAME(invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa) in FNAME() 976 if (!VALID_PAGE(root_hpa)) { in FNAME() 982 for_each_shadow_entry_using_root(vcpu, root_hpa, gva, iterator) { in FNAME()
|
| D | tdp_mmu.c | 671 for_each_tdp_pte(_iter, __va(_mmu->root_hpa), \
|
| /Linux-v5.15/arch/x86/include/asm/ |
| D | kvm_host.h | 429 void (*invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa); 430 hpa_t root_hpa; member 1398 void (*load_mmu_pgd)(struct kvm_vcpu *vcpu, hpa_t root_hpa, 1778 gva_t gva, hpa_t root_hpa);
|
| /Linux-v5.15/arch/x86/kvm/vmx/ |
| D | vmx.h | 370 u64 construct_eptp(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level);
|
| D | vmx.c | 2933 u64 root_hpa = mmu->root_hpa; in vmx_flush_tlb_current() local 2936 if (!VALID_PAGE(root_hpa)) in vmx_flush_tlb_current() 2940 ept_sync_context(construct_eptp(vcpu, root_hpa, in vmx_flush_tlb_current() 3091 u64 construct_eptp(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level) in construct_eptp() argument 3100 eptp |= root_hpa; in construct_eptp() 3105 static void vmx_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, in vmx_load_mmu_pgd() argument 3114 eptp = construct_eptp(vcpu, root_hpa, root_level); in vmx_load_mmu_pgd() 3117 hv_track_root_tdp(vcpu, root_hpa); in vmx_load_mmu_pgd() 3127 guest_cr3 = root_hpa | kvm_get_active_pcid(vcpu); in vmx_load_mmu_pgd()
|
| D | nested.c | 335 static bool nested_ept_root_matches(hpa_t root_hpa, u64 root_eptp, u64 eptp) in nested_ept_root_matches() argument 337 return VALID_PAGE(root_hpa) && in nested_ept_root_matches() 5424 if (nested_ept_root_matches(mmu->root_hpa, mmu->root_pgd, in handle_invept()
|
| /Linux-v5.15/arch/x86/kvm/svm/ |
| D | svm.c | 3904 static void svm_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, in svm_load_mmu_pgd() argument 3911 svm->vmcb->control.nested_cr3 = __sme_set(root_hpa); in svm_load_mmu_pgd() 3914 hv_track_root_tdp(vcpu, root_hpa); in svm_load_mmu_pgd() 3921 cr3 = __sme_set(root_hpa) | kvm_get_active_pcid(vcpu); in svm_load_mmu_pgd() 3925 cr3 = root_hpa; in svm_load_mmu_pgd()
|