/Linux-v4.19/arch/x86/kvm/ |
D | mmu_audit.c | 35 typedef void (*inspect_spte_fn) (struct kvm_vcpu *vcpu, u64 *sptep, int level); 96 static void audit_mappings(struct kvm_vcpu *vcpu, u64 *sptep, int level) in audit_mappings() argument 103 sp = page_header(__pa(sptep)); in audit_mappings() 113 if (!is_shadow_present_pte(*sptep) || !is_last_spte(*sptep, level)) in audit_mappings() 116 gfn = kvm_mmu_page_get_gfn(sp, sptep - sp->spt); in audit_mappings() 123 if ((*sptep & PT64_BASE_ADDR_MASK) != hpa) in audit_mappings() 126 hpa, *sptep); in audit_mappings() 129 static void inspect_spte_has_rmap(struct kvm *kvm, u64 *sptep) in inspect_spte_has_rmap() argument 138 rev_sp = page_header(__pa(sptep)); in inspect_spte_has_rmap() 139 gfn = kvm_mmu_page_get_gfn(rev_sp, sptep - rev_sp->spt); in inspect_spte_has_rmap() [all …]
|
D | mmu.c | 176 u64 *sptep; member 206 ({ spte = mmu_spte_get_lockless(_walker.sptep); 1; }); \ 263 static void mmu_spte_set(u64 *sptep, u64 spte); 347 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, in mark_mmio_spte() argument 360 trace_mark_mmio_spte(sptep, gfn, access, gen); in mark_mmio_spte() 361 mmu_spte_set(sptep, mask); in mark_mmio_spte() 385 static bool set_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in set_mmio_spte() argument 389 mark_mmio_spte(vcpu, sptep, gfn, access); in set_mmio_spte() 511 static void __set_spte(u64 *sptep, u64 spte) in __set_spte() argument 513 WRITE_ONCE(*sptep, spte); in __set_spte() [all …]
|
D | mmutrace.h | 205 TP_PROTO(u64 *sptep, gfn_t gfn, unsigned access, unsigned int gen), 206 TP_ARGS(sptep, gfn, access, gen), 209 __field(void *, sptep) 216 __entry->sptep = sptep; 222 TP_printk("sptep:%p gfn %llx access %x gen %x", __entry->sptep, 253 u64 *sptep, u64 old_spte, bool retry), 254 TP_ARGS(vcpu, gva, error_code, sptep, old_spte, retry), 260 __field(u64 *, sptep) 270 __entry->sptep = sptep; 272 __entry->new_spte = *sptep; [all …]
|
D | paging_tmpl.h | 560 u64 *sptep) in FNAME() 567 sp = page_header(__pa(sptep)); in FNAME() 573 return __direct_pte_prefetch(vcpu, sp, sptep); in FNAME() 575 i = (sptep - sp->spt) & ~(PTE_PREFETCH_NUM - 1); in FNAME() 579 if (spte == sptep) in FNAME() 627 clear_sp_write_flooding_count(it.sptep); in FNAME() 628 drop_large_spte(vcpu, it.sptep); in FNAME() 631 if (!is_shadow_present_pte(*it.sptep)) { in FNAME() 645 link_shadow_page(vcpu, it.sptep, sp); in FNAME() 653 clear_sp_write_flooding_count(it.sptep); in FNAME() [all …]
|
/Linux-v4.19/arch/s390/mm/ |
D | pgtable.c | 629 pte_t *sptep, pte_t *tptep, pte_t pte) in ptep_shadow_pte() argument 637 spgste = pgste_get_lock(sptep); in ptep_shadow_pte() 638 spte = *sptep; in ptep_shadow_pte() 651 pgste_set_unlock(sptep, spgste); in ptep_shadow_pte()
|
D | gmap.c | 2100 pte_t *sptep, *tptep; in gmap_shadow_page() local 2124 sptep = gmap_pte_op_walk(parent, paddr, &ptl); in gmap_shadow_page() 2125 if (sptep) { in gmap_shadow_page() 2135 rc = ptep_shadow_pte(sg->mm, saddr, sptep, tptep, pte); in gmap_shadow_page()
|
/Linux-v4.19/arch/s390/include/asm/ |
D | pgtable.h | 1103 pte_t *sptep, pte_t *tptep, pte_t pte);
|