| /Linux-v5.4/arch/x86/kvm/ |
| D | mmu_audit.c | 32 typedef void (*inspect_spte_fn) (struct kvm_vcpu *vcpu, u64 *sptep, int level); 93 static void audit_mappings(struct kvm_vcpu *vcpu, u64 *sptep, int level) in audit_mappings() argument 100 sp = page_header(__pa(sptep)); in audit_mappings() 110 if (!is_shadow_present_pte(*sptep) || !is_last_spte(*sptep, level)) in audit_mappings() 113 gfn = kvm_mmu_page_get_gfn(sp, sptep - sp->spt); in audit_mappings() 120 if ((*sptep & PT64_BASE_ADDR_MASK) != hpa) in audit_mappings() 123 hpa, *sptep); in audit_mappings() 126 static void inspect_spte_has_rmap(struct kvm *kvm, u64 *sptep) in inspect_spte_has_rmap() argument 135 rev_sp = page_header(__pa(sptep)); in inspect_spte_has_rmap() 136 gfn = kvm_mmu_page_get_gfn(rev_sp, sptep - rev_sp->spt); in inspect_spte_has_rmap() [all …]
|
| D | mmutrace.h | 205 TP_PROTO(u64 *sptep, gfn_t gfn, unsigned access, unsigned int gen), 206 TP_ARGS(sptep, gfn, access, gen), 209 __field(void *, sptep) 216 __entry->sptep = sptep; 222 TP_printk("sptep:%p gfn %llx access %x gen %x", __entry->sptep, 253 u64 *sptep, u64 old_spte, bool retry), 254 TP_ARGS(vcpu, gva, error_code, sptep, old_spte, retry), 260 __field(u64 *, sptep) 270 __entry->sptep = sptep; 272 __entry->new_spte = *sptep; [all …]
|
| D | mmu.c | 213 u64 *sptep; member 243 ({ spte = mmu_spte_get_lockless(_walker.sptep); 1; }); \ 305 static void mmu_spte_set(u64 *sptep, u64 spte); 454 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, in mark_mmio_spte() argument 467 trace_mark_mmio_spte(sptep, gfn, access, gen); in mark_mmio_spte() 468 mmu_spte_set(sptep, mask); in mark_mmio_spte() 486 static bool set_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in set_mmio_spte() argument 490 mark_mmio_spte(vcpu, sptep, gfn, access); in set_mmio_spte() 641 static void __set_spte(u64 *sptep, u64 spte) in __set_spte() argument 643 WRITE_ONCE(*sptep, spte); in __set_spte() [all …]
|
| D | paging_tmpl.h | 579 u64 *sptep) in FNAME() 586 sp = page_header(__pa(sptep)); in FNAME() 592 return __direct_pte_prefetch(vcpu, sp, sptep); in FNAME() 594 i = (sptep - sp->spt) & ~(PTE_PREFETCH_NUM - 1); in FNAME() 598 if (spte == sptep) in FNAME() 648 clear_sp_write_flooding_count(it.sptep); in FNAME() 649 drop_large_spte(vcpu, it.sptep); in FNAME() 652 if (!is_shadow_present_pte(*it.sptep)) { in FNAME() 666 link_shadow_page(vcpu, it.sptep, sp); in FNAME() 679 clear_sp_write_flooding_count(it.sptep); in FNAME() [all …]
|
| /Linux-v5.4/arch/s390/mm/ |
| D | pgtable.c | 631 pte_t *sptep, pte_t *tptep, pte_t pte) in ptep_shadow_pte() argument 639 spgste = pgste_get_lock(sptep); in ptep_shadow_pte() 640 spte = *sptep; in ptep_shadow_pte() 653 pgste_set_unlock(sptep, spgste); in ptep_shadow_pte()
|
| D | gmap.c | 2106 pte_t *sptep, *tptep; in gmap_shadow_page() local 2130 sptep = gmap_pte_op_walk(parent, paddr, &ptl); in gmap_shadow_page() 2131 if (sptep) { in gmap_shadow_page() 2141 rc = ptep_shadow_pte(sg->mm, saddr, sptep, tptep, pte); in gmap_shadow_page()
|
| /Linux-v5.4/arch/s390/include/asm/ |
| D | pgtable.h | 1144 pte_t *sptep, pte_t *tptep, pte_t pte);
|