Home
last modified time | relevance | path

Searched refs:kvm_mmu_page (Results 1 – 12 of 12) sorted by relevance

/Linux-v6.1/arch/x86/kvm/mmu/
Dtdp_mmu.c68 static void tdp_mmu_free_sp(struct kvm_mmu_page *sp) in tdp_mmu_free_sp()
84 struct kvm_mmu_page *sp = container_of(head, struct kvm_mmu_page, in tdp_mmu_free_sp_rcu_callback()
90 static void tdp_mmu_zap_root(struct kvm *kvm, struct kvm_mmu_page *root,
95 struct kvm_mmu_page *root = container_of(work, struct kvm_mmu_page, in tdp_mmu_zap_root_work()
123 static void tdp_mmu_schedule_zap_root(struct kvm *kvm, struct kvm_mmu_page *root) in tdp_mmu_schedule_zap_root()
130 static inline bool kvm_tdp_root_mark_invalid(struct kvm_mmu_page *page) in kvm_tdp_root_mark_invalid()
140 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root, in kvm_tdp_mmu_put_root()
204 static struct kvm_mmu_page *tdp_mmu_next_root(struct kvm *kvm, in tdp_mmu_next_root()
205 struct kvm_mmu_page *prev_root, in tdp_mmu_next_root()
208 struct kvm_mmu_page *next_root; in tdp_mmu_next_root()
[all …]
Dtdp_mmu.h10 __must_check static inline bool kvm_tdp_mmu_get_root(struct kvm_mmu_page *root) in kvm_tdp_mmu_get_root()
15 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root,
20 bool kvm_tdp_mmu_zap_sp(struct kvm *kvm, struct kvm_mmu_page *sp);
71 static inline bool is_tdp_mmu_page(struct kvm_mmu_page *sp) { return sp->tdp_mmu_page; } in is_tdp_mmu_page()
75 struct kvm_mmu_page *sp; in is_tdp_mmu()
92 static inline bool is_tdp_mmu_page(struct kvm_mmu_page *sp) { return false; } in is_tdp_mmu_page()
Dmmu_internal.h49 struct kvm_mmu_page { struct
123 static inline struct kvm_mmu_page *to_shadow_page(hpa_t shadow_page) in to_shadow_page() argument
127 return (struct kvm_mmu_page *)page_private(page); in to_shadow_page()
130 static inline struct kvm_mmu_page *sptep_to_sp(u64 *sptep) in sptep_to_sp()
140 static inline int kvm_mmu_page_as_id(struct kvm_mmu_page *sp) in kvm_mmu_page_as_id()
145 static inline bool kvm_mmu_page_ad_need_write_protect(struct kvm_mmu_page *sp) in kvm_mmu_page_ad_need_write_protect()
318 void account_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp);
319 void unaccount_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp);
Dmmu.c338 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in count_spte_clear()
422 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in __get_spte_lockless()
681 static bool sp_has_gptes(struct kvm_mmu_page *sp);
683 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn()
700 static u32 kvm_mmu_page_get_access(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_access()
720 static void kvm_mmu_page_set_translation(struct kvm_mmu_page *sp, int index, in kvm_mmu_page_set_translation()
739 static void kvm_mmu_page_set_access(struct kvm_mmu_page *sp, int index, in kvm_mmu_page_set_access()
783 static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in account_shadowed()
805 void account_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) in account_huge_nx_page()
816 static void unaccount_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_shadowed()
[all …]
Dmmutrace.h159 TP_PROTO(struct kvm_mmu_page *sp, bool created),
178 TP_PROTO(struct kvm_mmu_page *sp),
193 TP_PROTO(struct kvm_mmu_page *sp),
199 TP_PROTO(struct kvm_mmu_page *sp),
205 TP_PROTO(struct kvm_mmu_page *sp),
Dpaging_tmpl.h152 struct kvm_mmu_page *sp, u64 *spte, in FNAME()
521 FNAME(prefetch_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in FNAME()
578 struct kvm_mmu_page *sp; in FNAME()
621 struct kvm_mmu_page *sp = NULL; in FNAME()
888 static gpa_t FNAME(get_level1_sp_gpa)(struct kvm_mmu_page *sp) in FNAME()
903 struct kvm_mmu_page *sp; in FNAME()
995 static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) in FNAME()
Dtdp_iter.h113 void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root,
Dspte.h244 static inline bool sp_ad_disabled(struct kvm_mmu_page *sp) in sp_ad_disabled()
444 bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
Dtdp_iter.c43 void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root, in tdp_iter_start()
Dspte.c137 bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in make_spte()
/Linux-v6.1/arch/x86/include/asm/
Dkvm_host.h421 struct kvm_mmu_page;
439 struct kvm_mmu_page *sp);
/Linux-v6.1/Documentation/virt/kvm/x86/
Dmmu.rst121 The principal data structure is the shadow page, 'struct kvm_mmu_page'. A
198 shadow page; it is also used to go back from a struct kvm_mmu_page