/Linux-v4.19/include/linux/ |
D | kvm_host.h | 298 gfn_t base_gfn; 654 int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn, 657 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn); 658 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn); 659 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable); 660 unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn); 661 unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn, 667 kvm_pfn_t gfn_to_pfn_atomic(struct kvm *kvm, gfn_t gfn); 668 kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn); 669 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, [all …]
|
D | kvm_types.h | 50 typedef u64 gfn_t; typedef
|
/Linux-v4.19/arch/mips/kvm/ |
D | mmu.c | 299 bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn) in kvm_mips_flush_gpa_pt() 423 int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn) in BUILD_PTE_RANGE_OP() 443 gfn_t gfn_offset, unsigned long mask) in kvm_arch_mmu_enable_log_dirty_pt_masked() 445 gfn_t base_gfn = slot->base_gfn + gfn_offset; in kvm_arch_mmu_enable_log_dirty_pt_masked() 446 gfn_t start = base_gfn + __ffs(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked() 447 gfn_t end = base_gfn + __fls(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked() 460 static int kvm_mips_mkold_gpa_pt(struct kvm *kvm, gfn_t start_gfn, in BUILD_PTE_RANGE_OP() 461 gfn_t end_gfn) in BUILD_PTE_RANGE_OP() 471 int (*handler)(struct kvm *kvm, gfn_t gfn, in handle_hva_to_gpa() 486 gfn_t gfn, gfn_end; in handle_hva_to_gpa() [all …]
|
/Linux-v4.19/arch/x86/include/asm/ |
D | kvm_page_track.h | 58 struct kvm_memory_slot *slot, gfn_t gfn, 61 struct kvm_memory_slot *slot, gfn_t gfn, 63 bool kvm_page_track_is_active(struct kvm_vcpu *vcpu, gfn_t gfn,
|
D | kvm_host.h | 112 static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level) in gfn_to_index() 288 gfn_t gfn; 293 gfn_t *gfns; 667 gfn_t mmio_gfn; 684 gfn_t gfns[roundup_pow_of_two(ASYNC_PF_PER_VCPU)]; 1036 u64 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio); 1088 gfn_t offset, unsigned long mask); 1140 gfn_t gfn; 1193 gfn_t gfn_offset, unsigned long mask); 1295 gfn_t gfn, void *data, int offset, int len, [all …]
|
/Linux-v4.19/virt/kvm/ |
D | kvm_main.c | 130 static void mark_page_dirty_in_slot(struct kvm_memory_slot *memslot, gfn_t gfn); 917 gfn_t base_gfn; in __kvm_set_memory_region() 1185 gfn_t offset; in kvm_get_dirty_log_protect() 1221 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) in gfn_to_memslot() 1227 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_vcpu_gfn_to_memslot() 1232 bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn) in kvm_is_visible_gfn() 1244 unsigned long kvm_host_page_size(struct kvm *kvm, gfn_t gfn) in kvm_host_page_size() 1273 static unsigned long __gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, in __gfn_to_hva_many() 1274 gfn_t *nr_pages, bool write) in __gfn_to_hva_many() 1288 static unsigned long gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, in gfn_to_hva_many() [all …]
|
/Linux-v4.19/arch/x86/kvm/ |
D | mmu.h | 212 void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end); 214 void kvm_mmu_gfn_disallow_lpage(struct kvm_memory_slot *slot, gfn_t gfn); 215 void kvm_mmu_gfn_allow_lpage(struct kvm_memory_slot *slot, gfn_t gfn);
|
D | mmu.c | 369 static gfn_t get_mmio_spte_gfn(u64 spte) in get_mmio_spte_gfn() 385 static bool set_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in set_mmio_spte() 503 static gfn_t pse36_gfn_delta(u32 gpte) in pse36_gfn_delta() 1020 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn() 1028 static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) in kvm_mmu_page_set_gfn() 1040 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot() 1051 gfn_t gfn, int count) in update_gfn_disallow_lpage_count() 1063 void kvm_mmu_gfn_disallow_lpage(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage() 1068 void kvm_mmu_gfn_allow_lpage(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage() 1077 gfn_t gfn; in account_shadowed() [all …]
|
D | page_track.c | 65 static void update_gfn_track(struct kvm_memory_slot *slot, gfn_t gfn, in update_gfn_track() 93 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_slot_page_track_add_page() 128 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_slot_page_track_remove_page() 147 bool kvm_page_track_is_active(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_page_track_is_active()
|
D | paging_tmpl.h | 90 gfn_t table_gfn[PT_MAX_FULL_LEVELS]; 98 gfn_t gfn; 102 static gfn_t gpte_to_gfn_lvl(pt_element_t gpte, int lvl) in gpte_to_gfn_lvl() 210 gfn_t table_gfn; in FNAME() 285 gfn_t table_gfn; in FNAME() 298 gfn_t gfn; in FNAME() 330 gfn_t real_gfn; in FNAME() 502 gfn_t gfn; in FNAME() 625 gfn_t table_gfn; in FNAME() 651 gfn_t direct_gfn; in FNAME() [all …]
|
D | mmutrace.h | 205 TP_PROTO(u64 *sptep, gfn_t gfn, unsigned access, unsigned int gen), 210 __field(gfn_t, gfn) 228 TP_PROTO(u64 addr, gfn_t gfn, unsigned access), 233 __field(gfn_t, gfn)
|
D | x86.h | 182 gva_t gva, gfn_t gfn, unsigned access) in vcpu_cache_mmio_info() 270 u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn); 274 bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn,
|
D | mmu_audit.c | 99 gfn_t gfn; in audit_mappings() 136 gfn_t gfn; in inspect_spte_has_rmap()
|
D | mtrr.c | 319 gfn_t start, end; in update_mtrr() 625 u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_mtrr_get_guest_memory_type() 701 bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_mtrr_check_gfn_range_consistency()
|
D | x86.c | 559 gfn_t ngfn, void *data, int offset, int len, in kvm_read_guest_page_mmu() 563 gfn_t real_gfn; in kvm_read_guest_page_mmu() 577 static int kvm_read_nested_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_read_nested_guest_page() 589 gfn_t pdpt_gfn = cr3 >> PAGE_SHIFT; in load_pdptrs() 628 gfn_t gfn; in pdptrs_changed() 9345 static inline u32 kvm_async_pf_hash_fn(gfn_t gfn) in kvm_async_pf_hash_fn() 9355 static void kvm_add_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_add_async_pf_gfn() 9365 static u32 kvm_async_pf_gfn_slot(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_async_pf_gfn_slot() 9378 bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_find_async_pf_gfn() 9383 static void kvm_del_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_del_async_pf_gfn()
|
/Linux-v4.19/drivers/gpu/drm/i915/gvt/ |
D | kvmgt.c | 80 gfn_t gfn; 98 gfn_t gfn; 231 static struct gvt_dma *__gvt_cache_find_gfn(struct intel_vgpu *vgpu, gfn_t gfn) in __gvt_cache_find_gfn() 249 static int __gvt_cache_add(struct intel_vgpu *vgpu, gfn_t gfn, in __gvt_cache_add() 352 __kvmgt_protect_table_find(struct kvmgt_guest_info *info, gfn_t gfn) in __kvmgt_protect_table_find() 367 gfn_t gfn) in kvmgt_gfn_is_write_protected() 375 static void kvmgt_protect_table_add(struct kvmgt_guest_info *info, gfn_t gfn) in kvmgt_protect_table_add() 391 gfn_t gfn) in kvmgt_protect_table_del() 1560 gfn_t gfn; in kvmgt_page_track_flush_slot()
|
/Linux-v4.19/arch/powerpc/kvm/ |
D | e500_mmu_host.c | 326 u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, in kvmppc_e500_shadow_map() 563 u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, in kvmppc_e500_tlb1_map() 615 gfn_t gfn = gpaddr >> PAGE_SHIFT; in kvmppc_mmu_map()
|
D | book3s.c | 397 gfn_t gfn = gpa >> PAGE_SHIFT; in kvmppc_gpa_to_pfn()
|
D | booke.c | 1250 gfn_t gfn; in kvmppc_handle_exit() 1307 gfn_t gfn; in kvmppc_handle_exit()
|
D | book3s_64_mmu_hv.c | 773 gfn_t gfn, gfn_end; in kvm_handle_hva_range()
|
D | book3s_pr.c | 406 gfn_t gfn, gfn_end; in do_kvm_unmap_hva()
|
/Linux-v4.19/arch/mips/include/asm/ |
D | kvm_host.h | 912 bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn); 913 int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
|
/Linux-v4.19/virt/kvm/arm/ |
D | mmu.c | 1232 gfn_t gfn = *ipap >> PAGE_SHIFT; in transparent_hugepage_adjust() 1417 gfn_t gfn_offset, unsigned long mask) in kvm_mmu_write_protect_pt_masked() 1435 gfn_t gfn_offset, unsigned long mask) in kvm_arch_mmu_enable_log_dirty_pt_masked() 1476 gfn_t gfn = fault_ipa >> PAGE_SHIFT; in user_mem_abort() 1688 gfn_t gfn; in kvm_handle_guest_abort() 1799 gfn_t gpa; in handle_hva_to_gpa()
|
/Linux-v4.19/arch/s390/kvm/ |
D | kvm-s390.c | 526 gfn_t cur_gfn, last_gfn; in kvm_s390_sync_dirty_log() 1636 static int gfn_to_memslot_approx(struct kvm_memslots *slots, gfn_t gfn) in gfn_to_memslot_approx()
|
/Linux-v4.19/virt/kvm/arm/vgic/ |
D | vgic-its.c | 764 gfn_t gfn; in vgic_its_check_id()
|