Lines Matching refs:kvm_mmu_page
357 static inline bool sp_ad_disabled(struct kvm_mmu_page *sp) in sp_ad_disabled()
671 struct kvm_mmu_page *sp = page_header(__pa(sptep)); in count_spte_clear()
755 struct kvm_mmu_page *sp = page_header(__pa(sptep)); in __get_spte_lockless()
1150 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn()
1158 static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) in kvm_mmu_page_set_gfn()
1209 static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in account_shadowed()
1228 static void account_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) in account_huge_nx_page()
1239 static void unaccount_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_shadowed()
1256 static void unaccount_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_huge_nx_page()
1474 struct kvm_mmu_page *sp) in gfn_to_rmap()
1494 struct kvm_mmu_page *sp; in rmap_add()
1505 struct kvm_mmu_page *sp; in rmap_remove()
1614 struct kvm_mmu_page *sp = page_header(__pa(sptep)); in drop_large_spte()
2086 struct kvm_mmu_page *sp; in rmap_recycle()
2135 static void kvm_mmu_free_page(struct kvm_mmu_page *sp) in kvm_mmu_free_page()
2152 struct kvm_mmu_page *sp, u64 *parent_pte) in mmu_page_add_parent_pte()
2160 static void mmu_page_remove_parent_pte(struct kvm_mmu_page *sp, in mmu_page_remove_parent_pte()
2166 static void drop_parent_pte(struct kvm_mmu_page *sp, in drop_parent_pte()
2173 static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, int direct) in kvm_mmu_alloc_page()
2175 struct kvm_mmu_page *sp; in kvm_mmu_alloc_page()
2195 static void kvm_mmu_mark_parents_unsync(struct kvm_mmu_page *sp) in kvm_mmu_mark_parents_unsync()
2207 struct kvm_mmu_page *sp; in mark_unsync()
2220 struct kvm_mmu_page *sp) in nonpaging_sync_page()
2230 struct kvm_mmu_page *sp, u64 *spte, in nonpaging_update_pte()
2240 struct kvm_mmu_page *sp;
2246 static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp, in mmu_pages_add()
2262 static inline void clear_unsync_child_bit(struct kvm_mmu_page *sp, int idx) in clear_unsync_child_bit()
2269 static int __mmu_unsync_walk(struct kvm_mmu_page *sp, in __mmu_unsync_walk()
2275 struct kvm_mmu_page *child; in __mmu_unsync_walk()
2310 static int mmu_unsync_walk(struct kvm_mmu_page *sp, in mmu_unsync_walk()
2321 static void kvm_unlink_unsync_page(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_unlink_unsync_page()
2329 static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp,
2345 static inline bool is_ept_sp(struct kvm_mmu_page *sp) in is_ept_sp()
2351 static bool __kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in __kvm_sync_page()
2395 static bool is_obsolete_sp(struct kvm *kvm, struct kvm_mmu_page *sp) in is_obsolete_sp()
2401 static bool kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in kvm_sync_page()
2412 struct kvm_mmu_page *s; in kvm_sync_pages()
2427 struct kvm_mmu_page *parent[PT64_ROOT_MAX_LEVEL];
2443 struct kvm_mmu_page *sp = pvec->page[n].sp; in mmu_pages_next()
2460 struct kvm_mmu_page *sp; in mmu_pages_first()
2483 struct kvm_mmu_page *sp; in mmu_pages_clear_parents()
2499 struct kvm_mmu_page *parent) in mmu_sync_children()
2502 struct kvm_mmu_page *sp; in mmu_sync_children()
2533 static void __clear_sp_write_flooding_count(struct kvm_mmu_page *sp) in __clear_sp_write_flooding_count()
2540 struct kvm_mmu_page *sp = page_header(__pa(spte)); in clear_sp_write_flooding_count()
2545 static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu, in kvm_mmu_get_page()
2554 struct kvm_mmu_page *sp; in kvm_mmu_get_page()
2699 struct kvm_mmu_page *sp) in link_shadow_page()
2725 struct kvm_mmu_page *child; in validate_direct_spte()
2743 static bool mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp, in mmu_page_zap_pte()
2747 struct kvm_mmu_page *child; in mmu_page_zap_pte()
2769 struct kvm_mmu_page *sp) in kvm_mmu_page_unlink_children()
2777 static void kvm_mmu_unlink_parents(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_mmu_unlink_parents()
2787 struct kvm_mmu_page *parent, in mmu_zap_unsync_children()
2798 struct kvm_mmu_page *sp; in mmu_zap_unsync_children()
2811 struct kvm_mmu_page *sp, in __kvm_mmu_prepare_zap_page()
2855 static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp, in kvm_mmu_prepare_zap_page()
2867 struct kvm_mmu_page *sp, *nsp; in kvm_mmu_commit_zap_page()
2892 struct kvm_mmu_page *sp; in prepare_zap_oldest_mmu_page()
2898 struct kvm_mmu_page, link); in prepare_zap_oldest_mmu_page()
2929 struct kvm_mmu_page *sp; in kvm_mmu_unprotect_page()
2949 static void kvm_unsync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) in kvm_unsync_page()
2961 struct kvm_mmu_page *sp; in mmu_need_write_protect()
3051 struct kvm_mmu_page *sp; in set_spte()
3168 struct kvm_mmu_page *child; in mmu_set_spte()
3227 struct kvm_mmu_page *sp, in direct_pte_prefetch_many()
3255 struct kvm_mmu_page *sp, u64 *sptep) in __direct_pte_prefetch()
3279 struct kvm_mmu_page *sp; in direct_pte_prefetch()
3325 struct kvm_mmu_page *sp; in __direct_map()
3480 fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in fast_pf_fix_direct_spte()
3535 struct kvm_mmu_page *sp; in fast_page_fault()
3697 struct kvm_mmu_page *sp; in mmu_free_root_page()
3773 struct kvm_mmu_page *sp; in mmu_alloc_direct_roots()
3814 struct kvm_mmu_page *sp; in mmu_alloc_shadow_roots()
3928 struct kvm_mmu_page *sp; in kvm_mmu_sync_roots()
5288 struct kvm_mmu_page *sp, u64 *spte, in mmu_pte_write_new_pte()
5343 static bool detect_write_flooding(struct kvm_mmu_page *sp) in detect_write_flooding()
5360 static bool detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa, in detect_write_misaligned()
5384 static u64 *get_written_sptes(struct kvm_mmu_page *sp, gpa_t gpa, int *nspte) in get_written_sptes()
5420 struct kvm_mmu_page *sp; in kvm_mmu_pte_write()
5815 struct kvm_mmu_page *sp, *node; in kvm_zap_obsolete_pages()
6003 struct kvm_mmu_page *sp; in kvm_mmu_zap_collapsible_spte()
6107 struct kvm_mmu_page *sp, *node; in kvm_mmu_zap_all()
6332 sizeof(struct kvm_mmu_page), in kvm_mmu_module_init()
6419 struct kvm_mmu_page *sp; in kvm_recover_nx_lpages()
6436 struct kvm_mmu_page, in kvm_recover_nx_lpages()