Lines Matching refs:gfn
282 kvm_pfn_t gfn; member
1126 int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn,
1129 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn);
1130 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn);
1131 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable);
1132 unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
1133 unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn,
1138 kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn);
1139 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,
1141 kvm_pfn_t gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn);
1142 kvm_pfn_t gfn_to_pfn_memslot_atomic(const struct kvm_memory_slot *slot, gfn_t gfn);
1143 kvm_pfn_t __gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn,
1153 int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset,
1161 int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data,
1173 #define __kvm_get_guest(kvm, gfn, offset, v) \ argument
1175 unsigned long __addr = gfn_to_hva(kvm, gfn); \
1193 #define __kvm_put_guest(kvm, gfn, offset, v) \ argument
1195 unsigned long __addr = gfn_to_hva(kvm, gfn); \
1202 mark_page_dirty(kvm, gfn); \
1216 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn);
1217 bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn);
1218 bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
1219 unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn);
1220 void mark_page_dirty_in_slot(struct kvm *kvm, const struct kvm_memory_slot *memslot, gfn_t gfn);
1221 void mark_page_dirty(struct kvm *kvm, gfn_t gfn);
1224 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn);
1225 kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn);
1226 kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn);
1229 unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn);
1230 unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable);
1231 int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset,
1237 int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data,
1241 void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn);
1619 try_get_memslot(struct kvm_memory_slot *slot, gfn_t gfn) in try_get_memslot() argument
1624 if (gfn >= slot->base_gfn && gfn < slot->base_gfn + slot->npages) in try_get_memslot()
1638 search_memslots(struct kvm_memslots *slots, gfn_t gfn, bool approx) in search_memslots() argument
1647 if (gfn >= slot->base_gfn) { in search_memslots()
1648 if (gfn < slot->base_gfn + slot->npages) in search_memslots()
1659 ____gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn, bool approx) in ____gfn_to_memslot() argument
1664 slot = try_get_memslot(slot, gfn); in ____gfn_to_memslot()
1668 slot = search_memslots(slots, gfn, approx); in ____gfn_to_memslot()
1683 __gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn) in __gfn_to_memslot() argument
1685 return ____gfn_to_memslot(slots, gfn, false); in __gfn_to_memslot()
1689 __gfn_to_hva_memslot(const struct kvm_memory_slot *slot, gfn_t gfn) in __gfn_to_hva_memslot() argument
1697 unsigned long offset = gfn - slot->base_gfn; in __gfn_to_hva_memslot()
1702 static inline int memslot_id(struct kvm *kvm, gfn_t gfn) in memslot_id() argument
1704 return gfn_to_memslot(kvm, gfn)->id; in memslot_id()
1715 static inline gpa_t gfn_to_gpa(gfn_t gfn) in gfn_to_gpa() argument
1717 return (gpa_t)gfn << PAGE_SHIFT; in gfn_to_gpa()