Lines Matching refs:gfn
295 kvm_pfn_t gfn; member
1168 int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn,
1171 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn);
1172 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn);
1173 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable);
1174 unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
1175 unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn,
1180 kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn);
1181 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,
1183 kvm_pfn_t gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn);
1184 kvm_pfn_t gfn_to_pfn_memslot_atomic(const struct kvm_memory_slot *slot, gfn_t gfn);
1185 kvm_pfn_t __gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn,
1195 int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset,
1203 int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data,
1215 #define __kvm_get_guest(kvm, gfn, offset, v) \ argument
1217 unsigned long __addr = gfn_to_hva(kvm, gfn); \
1235 #define __kvm_put_guest(kvm, gfn, offset, v) \ argument
1237 unsigned long __addr = gfn_to_hva(kvm, gfn); \
1244 mark_page_dirty(kvm, gfn); \
1258 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn);
1259 bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn);
1260 bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
1261 unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn);
1262 void mark_page_dirty_in_slot(struct kvm *kvm, const struct kvm_memory_slot *memslot, gfn_t gfn);
1263 void mark_page_dirty(struct kvm *kvm, gfn_t gfn);
1266 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn);
1267 kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn);
1268 kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn);
1271 unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn);
1272 unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable);
1273 int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset,
1279 int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data,
1283 void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn);
1383 void kvm_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages);
1514 gfn_t gfn, u64 nr_pages) in kvm_arch_flush_remote_tlbs_range() argument
1519 int kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages);
1648 try_get_memslot(struct kvm_memory_slot *slot, gfn_t gfn) in try_get_memslot() argument
1653 if (gfn >= slot->base_gfn && gfn < slot->base_gfn + slot->npages) in try_get_memslot()
1667 search_memslots(struct kvm_memslots *slots, gfn_t gfn, bool approx) in search_memslots() argument
1676 if (gfn >= slot->base_gfn) { in search_memslots()
1677 if (gfn < slot->base_gfn + slot->npages) in search_memslots()
1688 ____gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn, bool approx) in ____gfn_to_memslot() argument
1693 slot = try_get_memslot(slot, gfn); in ____gfn_to_memslot()
1697 slot = search_memslots(slots, gfn, approx); in ____gfn_to_memslot()
1712 __gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn) in __gfn_to_memslot() argument
1714 return ____gfn_to_memslot(slots, gfn, false); in __gfn_to_memslot()
1718 __gfn_to_hva_memslot(const struct kvm_memory_slot *slot, gfn_t gfn) in __gfn_to_hva_memslot() argument
1726 unsigned long offset = gfn - slot->base_gfn; in __gfn_to_hva_memslot()
1731 static inline int memslot_id(struct kvm *kvm, gfn_t gfn) in memslot_id() argument
1733 return gfn_to_memslot(kvm, gfn)->id; in memslot_id()
1744 static inline gpa_t gfn_to_gpa(gfn_t gfn) in gfn_to_gpa() argument
1746 return (gpa_t)gfn << PAGE_SHIFT; in gfn_to_gpa()