Lines Matching refs:gfn
241 kvm_pfn_t gfn; member
702 int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn,
705 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn);
706 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn);
707 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable);
708 unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
709 unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn,
715 kvm_pfn_t gfn_to_pfn_atomic(struct kvm *kvm, gfn_t gfn);
716 kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn);
717 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,
719 kvm_pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
720 kvm_pfn_t gfn_to_pfn_memslot_atomic(struct kvm_memory_slot *slot, gfn_t gfn);
721 kvm_pfn_t __gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn,
731 int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset,
738 int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data,
749 int kvm_clear_guest_page(struct kvm *kvm, gfn_t gfn, int offset, int len);
751 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn);
752 bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn);
753 unsigned long kvm_host_page_size(struct kvm *kvm, gfn_t gfn);
754 void mark_page_dirty(struct kvm *kvm, gfn_t gfn);
757 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn);
758 kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn);
759 kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn);
761 struct page *kvm_vcpu_gfn_to_page(struct kvm_vcpu *vcpu, gfn_t gfn);
763 unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn);
764 unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable);
765 int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset,
771 int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data,
775 void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn);
1006 search_memslots(struct kvm_memslots *slots, gfn_t gfn) in search_memslots() argument
1012 if (gfn >= memslots[slot].base_gfn && in search_memslots()
1013 gfn < memslots[slot].base_gfn + memslots[slot].npages) in search_memslots()
1019 if (gfn >= memslots[slot].base_gfn) in search_memslots()
1025 if (gfn >= memslots[start].base_gfn && in search_memslots()
1026 gfn < memslots[start].base_gfn + memslots[start].npages) { in search_memslots()
1035 __gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn) in __gfn_to_memslot() argument
1037 return search_memslots(slots, gfn); in __gfn_to_memslot()
1041 __gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn) in __gfn_to_hva_memslot() argument
1043 return slot->userspace_addr + (gfn - slot->base_gfn) * PAGE_SIZE; in __gfn_to_hva_memslot()
1046 static inline int memslot_id(struct kvm *kvm, gfn_t gfn) in memslot_id() argument
1048 return gfn_to_memslot(kvm, gfn)->id; in memslot_id()
1059 static inline gpa_t gfn_to_gpa(gfn_t gfn) in gfn_to_gpa() argument
1061 return (gpa_t)gfn << PAGE_SHIFT; in gfn_to_gpa()