Lines Matching refs:old_spte
462 u64 old_spte = *sptep; in mmu_spte_update_no_track() local
467 if (!is_shadow_present_pte(old_spte)) { in mmu_spte_update_no_track()
469 return old_spte; in mmu_spte_update_no_track()
472 if (!spte_has_volatile_bits(old_spte)) in mmu_spte_update_no_track()
475 old_spte = __update_clear_spte_slow(sptep, new_spte); in mmu_spte_update_no_track()
477 WARN_ON(spte_to_pfn(old_spte) != spte_to_pfn(new_spte)); in mmu_spte_update_no_track()
479 return old_spte; in mmu_spte_update_no_track()
494 u64 old_spte = mmu_spte_update_no_track(sptep, new_spte); in mmu_spte_update() local
496 if (!is_shadow_present_pte(old_spte)) in mmu_spte_update()
504 if (is_mmu_writable_spte(old_spte) && in mmu_spte_update()
513 if (is_accessed_spte(old_spte) && !is_accessed_spte(new_spte)) { in mmu_spte_update()
515 kvm_set_pfn_accessed(spte_to_pfn(old_spte)); in mmu_spte_update()
518 if (is_dirty_spte(old_spte) && !is_dirty_spte(new_spte)) { in mmu_spte_update()
520 kvm_set_pfn_dirty(spte_to_pfn(old_spte)); in mmu_spte_update()
535 u64 old_spte = *sptep; in mmu_spte_clear_track_bits() local
539 if (!is_shadow_present_pte(old_spte) || in mmu_spte_clear_track_bits()
540 !spte_has_volatile_bits(old_spte)) in mmu_spte_clear_track_bits()
543 old_spte = __update_clear_spte_slow(sptep, 0ull); in mmu_spte_clear_track_bits()
545 if (!is_shadow_present_pte(old_spte)) in mmu_spte_clear_track_bits()
546 return old_spte; in mmu_spte_clear_track_bits()
550 pfn = spte_to_pfn(old_spte); in mmu_spte_clear_track_bits()
561 if (is_accessed_spte(old_spte)) in mmu_spte_clear_track_bits()
564 if (is_dirty_spte(old_spte)) in mmu_spte_clear_track_bits()
567 return old_spte; in mmu_spte_clear_track_bits()
1135 u64 old_spte = mmu_spte_clear_track_bits(kvm, sptep); in drop_spte() local
1137 if (is_shadow_present_pte(old_spte)) in drop_spte()
3242 u64 *sptep, u64 old_spte, u64 new_spte) in fast_pf_fix_direct_spte() argument
3256 if (!try_cmpxchg64(sptep, &old_spte, new_spte)) in fast_pf_fix_direct_spte()
3259 if (is_writable_pte(new_spte) && !is_writable_pte(old_spte)) in fast_pf_fix_direct_spte()
3289 u64 old_spte; in fast_pf_get_last_sptep() local
3292 for_each_shadow_entry_lockless(vcpu, gpa, iterator, old_spte) { in fast_pf_get_last_sptep()
3294 *spte = old_spte; in fast_pf_get_last_sptep()