Lines Matching refs:ghc

3207 				       struct gfn_to_hva_cache *ghc,  in __kvm_gfn_to_hva_cache_init()  argument
3217 ghc->generation = slots->generation; in __kvm_gfn_to_hva_cache_init()
3220 ghc->hva = KVM_HVA_ERR_BAD; in __kvm_gfn_to_hva_cache_init()
3229 ghc->memslot = __gfn_to_memslot(slots, start_gfn); in __kvm_gfn_to_hva_cache_init()
3230 ghc->hva = gfn_to_hva_many(ghc->memslot, start_gfn, in __kvm_gfn_to_hva_cache_init()
3232 if (kvm_is_error_hva(ghc->hva)) in __kvm_gfn_to_hva_cache_init()
3238 ghc->hva += offset; in __kvm_gfn_to_hva_cache_init()
3240 ghc->memslot = NULL; in __kvm_gfn_to_hva_cache_init()
3242 ghc->gpa = gpa; in __kvm_gfn_to_hva_cache_init()
3243 ghc->len = len; in __kvm_gfn_to_hva_cache_init()
3247 int kvm_gfn_to_hva_cache_init(struct kvm *kvm, struct gfn_to_hva_cache *ghc, in kvm_gfn_to_hva_cache_init() argument
3251 return __kvm_gfn_to_hva_cache_init(slots, ghc, gpa, len); in kvm_gfn_to_hva_cache_init()
3255 int kvm_write_guest_offset_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, in kvm_write_guest_offset_cached() argument
3261 gpa_t gpa = ghc->gpa + offset; in kvm_write_guest_offset_cached()
3263 if (WARN_ON_ONCE(len + offset > ghc->len)) in kvm_write_guest_offset_cached()
3266 if (slots->generation != ghc->generation) { in kvm_write_guest_offset_cached()
3267 if (__kvm_gfn_to_hva_cache_init(slots, ghc, ghc->gpa, ghc->len)) in kvm_write_guest_offset_cached()
3271 if (kvm_is_error_hva(ghc->hva)) in kvm_write_guest_offset_cached()
3274 if (unlikely(!ghc->memslot)) in kvm_write_guest_offset_cached()
3277 r = __copy_to_user((void __user *)ghc->hva + offset, data, len); in kvm_write_guest_offset_cached()
3280 mark_page_dirty_in_slot(kvm, ghc->memslot, gpa >> PAGE_SHIFT); in kvm_write_guest_offset_cached()
3286 int kvm_write_guest_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, in kvm_write_guest_cached() argument
3289 return kvm_write_guest_offset_cached(kvm, ghc, data, 0, len); in kvm_write_guest_cached()
3293 int kvm_read_guest_offset_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, in kvm_read_guest_offset_cached() argument
3299 gpa_t gpa = ghc->gpa + offset; in kvm_read_guest_offset_cached()
3301 if (WARN_ON_ONCE(len + offset > ghc->len)) in kvm_read_guest_offset_cached()
3304 if (slots->generation != ghc->generation) { in kvm_read_guest_offset_cached()
3305 if (__kvm_gfn_to_hva_cache_init(slots, ghc, ghc->gpa, ghc->len)) in kvm_read_guest_offset_cached()
3309 if (kvm_is_error_hva(ghc->hva)) in kvm_read_guest_offset_cached()
3312 if (unlikely(!ghc->memslot)) in kvm_read_guest_offset_cached()
3315 r = __copy_from_user(data, (void __user *)ghc->hva + offset, len); in kvm_read_guest_offset_cached()
3323 int kvm_read_guest_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, in kvm_read_guest_cached() argument
3326 return kvm_read_guest_offset_cached(kvm, ghc, data, 0, len); in kvm_read_guest_cached()