Lines Matching refs:gfn_t

57 			  gfn_t start, gfn_t end, bool can_yield, bool flush,
174 static struct kvm_mmu_page *alloc_tdp_mmu_page(struct kvm_vcpu *vcpu, gfn_t gfn, in alloc_tdp_mmu_page()
220 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn,
235 static void handle_changed_spte_dirty_log(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte_dirty_log()
318 gfn_t base_gfn = sp->gfn; in handle_removed_tdp_mmu_page()
321 gfn_t gfn; in handle_removed_tdp_mmu_page()
397 static void __handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in __handle_changed_spte()
480 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte()
738 gfn_t start, gfn_t end, bool can_yield, bool flush, in zap_gfn_range()
741 gfn_t max_gfn_host = 1ULL << (shadow_phys_bits - PAGE_SHIFT); in zap_gfn_range()
808 bool __kvm_tdp_mmu_zap_gfn_range(struct kvm *kvm, int as_id, gfn_t start, in __kvm_tdp_mmu_zap_gfn_range()
809 gfn_t end, bool can_yield, bool flush) in __kvm_tdp_mmu_zap_gfn_range()
1003 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_tdp_mmu_map()
1221 gfn_t start, gfn_t end, int min_level) in wrprot_gfn_range()
1288 gfn_t start, gfn_t end) in clear_dirty_gfn_range()
1359 gfn_t gfn, unsigned long mask, bool wrprot) in clear_dirty_pt_masked()
1404 gfn_t gfn, unsigned long mask, in kvm_tdp_mmu_clear_dirty_pt_masked()
1423 gfn_t start = slot->base_gfn; in zap_collapsible_spte_range()
1424 gfn_t end = start + slot->npages; in zap_collapsible_spte_range()
1487 gfn_t gfn, int min_level) in write_protect_gfn()
1524 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_tdp_mmu_write_protect_gfn()
1548 gfn_t gfn = addr >> PAGE_SHIFT; in kvm_tdp_mmu_get_walk()
1577 gfn_t gfn = addr >> PAGE_SHIFT; in kvm_tdp_mmu_fast_pf_get_last_sptep()