Lines Matching refs:gfn

240 	kvm_pfn_t gfn;  member
704 int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn,
707 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn);
708 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn);
709 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable);
710 unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
711 unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn,
717 kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn);
718 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,
720 kvm_pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
721 kvm_pfn_t gfn_to_pfn_memslot_atomic(struct kvm_memory_slot *slot, gfn_t gfn);
722 kvm_pfn_t __gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn,
733 int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset,
741 int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data,
753 #define __kvm_get_guest(kvm, gfn, offset, v) \ argument
755 unsigned long __addr = gfn_to_hva(kvm, gfn); \
773 #define __kvm_put_guest(kvm, gfn, offset, v) \ argument
775 unsigned long __addr = gfn_to_hva(kvm, gfn); \
782 mark_page_dirty(kvm, gfn); \
795 int kvm_clear_guest_page(struct kvm *kvm, gfn_t gfn, int offset, int len);
797 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn);
798 bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn);
799 bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
800 unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn);
801 void mark_page_dirty_in_slot(struct kvm_memory_slot *memslot, gfn_t gfn);
802 void mark_page_dirty(struct kvm *kvm, gfn_t gfn);
805 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn);
806 kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn);
807 kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn);
809 int kvm_map_gfn(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map,
811 struct page *kvm_vcpu_gfn_to_page(struct kvm_vcpu *vcpu, gfn_t gfn);
815 unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn);
816 unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable);
817 int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset,
823 int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data,
827 void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn);
1067 search_memslots(struct kvm_memslots *slots, gfn_t gfn) in search_memslots() argument
1076 if (gfn >= memslots[slot].base_gfn && in search_memslots()
1077 gfn < memslots[slot].base_gfn + memslots[slot].npages) in search_memslots()
1083 if (gfn >= memslots[slot].base_gfn) in search_memslots()
1089 if (start < slots->used_slots && gfn >= memslots[start].base_gfn && in search_memslots()
1090 gfn < memslots[start].base_gfn + memslots[start].npages) { in search_memslots()
1099 __gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn) in __gfn_to_memslot() argument
1101 return search_memslots(slots, gfn); in __gfn_to_memslot()
1105 __gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn) in __gfn_to_hva_memslot() argument
1107 return slot->userspace_addr + (gfn - slot->base_gfn) * PAGE_SIZE; in __gfn_to_hva_memslot()
1110 static inline int memslot_id(struct kvm *kvm, gfn_t gfn) in memslot_id() argument
1112 return gfn_to_memslot(kvm, gfn)->id; in memslot_id()
1123 static inline gpa_t gfn_to_gpa(gfn_t gfn) in gfn_to_gpa() argument
1125 return (gpa_t)gfn << PAGE_SHIFT; in gfn_to_gpa()