Lines Matching refs:kvm_mmu_page
383 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in count_spte_clear()
467 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in __get_spte_lockless()
770 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn()
778 static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) in kvm_mmu_page_set_gfn()
828 static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in account_shadowed()
847 void account_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) in account_huge_nx_page()
858 static void unaccount_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_shadowed()
875 void unaccount_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_huge_nx_page()
1077 struct kvm_mmu_page *sp; in rmap_add()
1092 struct kvm_mmu_page *sp; in rmap_remove()
1211 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in drop_large_spte()
1646 struct kvm_mmu_page *sp; in rmap_recycle()
1711 static void kvm_mmu_free_page(struct kvm_mmu_page *sp) in kvm_mmu_free_page()
1728 struct kvm_mmu_page *sp, u64 *parent_pte) in mmu_page_add_parent_pte()
1736 static void mmu_page_remove_parent_pte(struct kvm_mmu_page *sp, in mmu_page_remove_parent_pte()
1742 static void drop_parent_pte(struct kvm_mmu_page *sp, in drop_parent_pte()
1749 static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, int direct) in kvm_mmu_alloc_page()
1751 struct kvm_mmu_page *sp; in kvm_mmu_alloc_page()
1771 static void kvm_mmu_mark_parents_unsync(struct kvm_mmu_page *sp) in kvm_mmu_mark_parents_unsync()
1783 struct kvm_mmu_page *sp; in mark_unsync()
1796 struct kvm_mmu_page *sp) in nonpaging_sync_page()
1805 struct kvm_mmu_page *sp;
1811 static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp, in mmu_pages_add()
1827 static inline void clear_unsync_child_bit(struct kvm_mmu_page *sp, int idx) in clear_unsync_child_bit()
1834 static int __mmu_unsync_walk(struct kvm_mmu_page *sp, in __mmu_unsync_walk()
1840 struct kvm_mmu_page *child; in __mmu_unsync_walk()
1875 static int mmu_unsync_walk(struct kvm_mmu_page *sp, in mmu_unsync_walk()
1886 static void kvm_unlink_unsync_page(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_unlink_unsync_page()
1894 static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp,
1909 static bool kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in kvm_sync_page()
1952 static bool is_obsolete_sp(struct kvm *kvm, struct kvm_mmu_page *sp) in is_obsolete_sp()
1959 struct kvm_mmu_page *parent[PT64_ROOT_MAX_LEVEL];
1975 struct kvm_mmu_page *sp = pvec->page[n].sp; in mmu_pages_next()
1992 struct kvm_mmu_page *sp; in mmu_pages_first()
2015 struct kvm_mmu_page *sp; in mmu_pages_clear_parents()
2031 struct kvm_mmu_page *parent, bool can_yield) in mmu_sync_children()
2034 struct kvm_mmu_page *sp; in mmu_sync_children()
2072 static void __clear_sp_write_flooding_count(struct kvm_mmu_page *sp) in __clear_sp_write_flooding_count()
2082 static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu, in kvm_mmu_get_page()
2093 struct kvm_mmu_page *sp; in kvm_mmu_get_page()
2247 struct kvm_mmu_page *sp) in link_shadow_page()
2267 struct kvm_mmu_page *child; in validate_direct_spte()
2286 static int mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp, in mmu_page_zap_pte()
2290 struct kvm_mmu_page *child; in mmu_page_zap_pte()
2317 struct kvm_mmu_page *sp, in kvm_mmu_page_unlink_children()
2329 static void kvm_mmu_unlink_parents(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_mmu_unlink_parents()
2339 struct kvm_mmu_page *parent, in mmu_zap_unsync_children()
2350 struct kvm_mmu_page *sp; in mmu_zap_unsync_children()
2363 struct kvm_mmu_page *sp, in __kvm_mmu_prepare_zap_page()
2420 static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp, in kvm_mmu_prepare_zap_page()
2432 struct kvm_mmu_page *sp, *nsp; in kvm_mmu_commit_zap_page()
2458 struct kvm_mmu_page *sp, *tmp; in kvm_mmu_zap_oldest_mmu_pages()
2545 struct kvm_mmu_page *sp; in kvm_mmu_unprotect_page()
2579 static void kvm_unsync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) in kvm_unsync_page()
2596 struct kvm_mmu_page *sp; in mmu_try_to_unsync_pages()
2697 struct kvm_mmu_page *sp; in set_spte()
2740 struct kvm_mmu_page *child; in mmu_set_spte()
2802 struct kvm_mmu_page *sp, in direct_pte_prefetch_many()
2830 struct kvm_mmu_page *sp, u64 *sptep) in __direct_pte_prefetch()
2854 struct kvm_mmu_page *sp; in direct_pte_prefetch()
2998 struct kvm_mmu_page *sp; in __direct_map()
3132 fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in fast_pf_fix_direct_spte()
3209 struct kvm_mmu_page *sp; in fast_page_fault()
3311 struct kvm_mmu_page *sp; in mmu_free_root_page()
3420 struct kvm_mmu_page *sp; in mmu_alloc_root()
3659 struct kvm_mmu_page *sp; in kvm_mmu_sync_roots()
5134 static bool detect_write_flooding(struct kvm_mmu_page *sp) in detect_write_flooding()
5151 static bool detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa, in detect_write_misaligned()
5175 static u64 *get_written_sptes(struct kvm_mmu_page *sp, gpa_t gpa, int *nspte) in get_written_sptes()
5211 struct kvm_mmu_page *sp; in kvm_mmu_pte_write()
5576 struct kvm_mmu_page *sp, *node; in kvm_zap_obsolete_pages()
5824 struct kvm_mmu_page *sp; in kvm_mmu_zap_collapsible_spte()
5923 struct kvm_mmu_page *sp, *node; in kvm_mmu_zap_all()
6120 sizeof(struct kvm_mmu_page), in kvm_mmu_module_init()
6208 struct kvm_mmu_page *sp; in kvm_recover_nx_lpages()
6229 struct kvm_mmu_page, in kvm_recover_nx_lpages()