/Linux-v6.1/arch/x86/kvm/mmu/ |
D | tdp_mmu.h | 18 bool kvm_tdp_mmu_zap_leafs(struct kvm *kvm, int as_id, gfn_t start, 19 gfn_t end, bool can_yield, bool flush); 39 gfn_t gfn, unsigned long mask, 45 struct kvm_memory_slot *slot, gfn_t gfn, 50 gfn_t start, gfn_t end,
|
D | tdp_mmu.c | 285 gfn_t gfn, union kvm_mmu_page_role role) in tdp_mmu_init_sp() 342 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, 357 static void handle_changed_spte_dirty_log(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte_dirty_log() 432 gfn_t base_gfn = sp->gfn; in handle_removed_pt() 441 gfn_t gfn = base_gfn + i * KVM_PAGES_PER_HPAGE(level); in handle_removed_pt() 526 static void __handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in __handle_changed_spte() 614 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte() 727 u64 old_spte, u64 new_spte, gfn_t gfn, int level, in __tdp_mmu_set_spte() 843 static inline gfn_t tdp_mmu_max_gfn_exclusive(void) in tdp_mmu_max_gfn_exclusive() 859 gfn_t end = tdp_mmu_max_gfn_exclusive(); in __tdp_mmu_zap_root() [all …]
|
D | tdp_iter.h | 63 gfn_t next_last_level_gfn; 69 gfn_t yielded_gfn; 75 gfn_t gfn; 114 int min_level, gfn_t next_last_level_gfn);
|
D | mmu_internal.h | 67 gfn_t gfn; 159 gfn_t gfn, bool can_unsync, bool prefetch); 161 void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn); 162 void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn); 218 gfn_t gfn; 311 const struct kvm_memory_slot *slot, gfn_t gfn,
|
D | tdp_iter.c | 18 static gfn_t round_gfn_for_level(gfn_t gfn, int level) in round_gfn_for_level() 44 int min_level, gfn_t next_last_level_gfn) in tdp_iter_start()
|
D | paging_tmpl.h | 83 gfn_t table_gfn[PT_MAX_FULL_LEVELS]; 91 gfn_t gfn; 96 static inline gfn_t pse36_gfn_delta(u32 gpte) in pse36_gfn_delta() 104 static gfn_t gpte_to_gfn_lvl(pt_element_t gpte, int lvl) in gpte_to_gfn_lvl() 205 gfn_t table_gfn; in FNAME() 309 gfn_t table_gfn; in FNAME() 322 gfn_t gfn; in FNAME() 526 gfn_t gfn; in FNAME() 625 gfn_t base_gfn = fault->gfn; in FNAME() 648 gfn_t table_gfn; in FNAME() [all …]
|
D | mmu.c | 272 static gfn_t get_mmio_spte_gfn(u64 spte) in get_mmio_spte_gfn() 683 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn() 721 gfn_t gfn, unsigned int access) in kvm_mmu_page_set_translation() 742 gfn_t gfn = kvm_mmu_page_get_gfn(sp, index); in kvm_mmu_page_set_access() 751 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot() 761 gfn_t gfn, int count) in update_gfn_disallow_lpage_count() 773 void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage() 778 void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage() 787 gfn_t gfn; in account_shadowed() 820 gfn_t gfn; in unaccount_shadowed() [all …]
|
D | page_track.c | 87 static void update_gfn_track(struct kvm_memory_slot *slot, gfn_t gfn, in update_gfn_track() 115 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_slot_page_track_add_page() 154 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_slot_page_track_remove_page() 179 gfn_t gfn, enum kvm_page_track_mode mode) in kvm_slot_page_track_is_active()
|
D | mmutrace.h | 212 TP_PROTO(u64 *sptep, gfn_t gfn, u64 spte), 217 __field(gfn_t, gfn) 235 TP_PROTO(u64 addr, gfn_t gfn, unsigned access), 240 __field(gfn_t, gfn) 335 TP_PROTO(int level, gfn_t gfn, u64 *sptep), 393 TP_PROTO(int as_id, gfn_t gfn, int level, u64 old_spte, u64 new_spte),
|
D | spte.h | 446 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
|
/Linux-v6.1/include/linux/ |
D | kvm_host.h | 250 gfn_t start; 251 gfn_t end; 570 gfn_t base_gfn; 1012 gfn_t start) in kvm_memslot_iter_start() 1070 static inline bool kvm_memslot_iter_is_valid(struct kvm_memslot_iter *iter, gfn_t end) in kvm_memslot_iter_is_valid() 1126 int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn, 1129 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn); 1130 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn); 1131 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable); 1132 unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn); [all …]
|
D | kvm_types.h | 41 typedef u64 gfn_t; typedef
|
/Linux-v6.1/arch/mips/kvm/ |
D | mmu.c | 271 bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn) in kvm_mips_flush_gpa_pt() 397 int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn) in BUILD_PTE_RANGE_OP() 417 gfn_t gfn_offset, unsigned long mask) in kvm_arch_mmu_enable_log_dirty_pt_masked() 419 gfn_t base_gfn = slot->base_gfn + gfn_offset; in kvm_arch_mmu_enable_log_dirty_pt_masked() 420 gfn_t start = base_gfn + __ffs(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked() 421 gfn_t end = base_gfn + __fls(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked() 434 static int kvm_mips_mkold_gpa_pt(struct kvm *kvm, gfn_t start_gfn, in BUILD_PTE_RANGE_OP() 435 gfn_t end_gfn) in BUILD_PTE_RANGE_OP() 516 gfn_t gfn = gpa >> PAGE_SHIFT; in _kvm_mips_map_page_fast() 593 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mips_map_page()
|
/Linux-v6.1/arch/x86/kvm/ |
D | mmu.h | 66 static inline gfn_t kvm_mmu_max_gfn(void) in kvm_mmu_max_gfn() 214 void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end); 243 static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level) in gfn_to_index()
|
D | x86.h | 204 gva_t gva, gfn_t gfn, unsigned access) in vcpu_cache_mmio_info() 296 u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn); 300 bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn,
|
D | mtrr.c | 309 gfn_t start, end; in update_mtrr() 615 u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_mtrr_get_guest_memory_type() 691 bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_mtrr_check_gfn_range_consistency()
|
/Linux-v6.1/arch/x86/include/asm/ |
D | kvm_page_track.h | 61 struct kvm_memory_slot *slot, gfn_t gfn, 64 struct kvm_memory_slot *slot, gfn_t gfn, 68 gfn_t gfn, enum kvm_page_track_mode mode);
|
D | kvm_host.h | 847 gfn_t mmio_gfn; 866 gfn_t gfns[ASYNC_PF_PER_VCPU]; 1561 u8 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio); 1681 gfn_t gfn; 1906 int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn); 2048 extern bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
|
/Linux-v6.1/virt/kvm/ |
D | kvm_main.c | 1900 gfn_t start, gfn_t end) in kvm_check_memslot_overlap() 1927 gfn_t base_gfn; in __kvm_set_memory_region() 2163 gfn_t offset; in kvm_get_dirty_log_protect() 2232 gfn_t offset; in kvm_clear_dirty_log_protect() 2316 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) in gfn_to_memslot() 2322 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_vcpu_gfn_to_memslot() 2355 bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn) in kvm_is_visible_gfn() 2363 bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_vcpu_is_visible_gfn() 2371 unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_host_page_size() 2400 static unsigned long __gfn_to_hva_many(const struct kvm_memory_slot *slot, gfn_t gfn, in __gfn_to_hva_many() [all …]
|
/Linux-v6.1/drivers/gpu/drm/i915/gvt/ |
D | kvmgt.c | 92 gfn_t gfn; 100 gfn_t gfn; 230 static struct gvt_dma *__gvt_cache_find_gfn(struct intel_vgpu *vgpu, gfn_t gfn) in __gvt_cache_find_gfn() 248 static int __gvt_cache_add(struct intel_vgpu *vgpu, gfn_t gfn, in __gvt_cache_add() 351 __kvmgt_protect_table_find(struct intel_vgpu *info, gfn_t gfn) in __kvmgt_protect_table_find() 365 static bool kvmgt_gfn_is_write_protected(struct intel_vgpu *info, gfn_t gfn) in kvmgt_gfn_is_write_protected() 373 static void kvmgt_protect_table_add(struct intel_vgpu *info, gfn_t gfn) in kvmgt_protect_table_add() 388 static void kvmgt_protect_table_del(struct intel_vgpu *info, gfn_t gfn) in kvmgt_protect_table_del() 1625 gfn_t gfn; in kvmgt_page_track_flush_slot()
|
/Linux-v6.1/arch/mips/include/asm/ |
D | kvm_host.h | 808 bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn); 809 int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
|
/Linux-v6.1/arch/powerpc/kvm/ |
D | e500_mmu_host.c | 323 u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, in kvmppc_e500_shadow_map() 560 u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, in kvmppc_e500_tlb1_map() 612 gfn_t gfn = gpaddr >> PAGE_SHIFT; in kvmppc_mmu_map()
|
/Linux-v6.1/arch/riscv/kvm/ |
D | vcpu_exit.c | 19 gfn_t gfn; in gstage_page_fault()
|
D | mmu.c | 395 gfn_t gfn_offset, in kvm_arch_mmu_enable_log_dirty_pt_masked() 625 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_riscv_gstage_map()
|
/Linux-v6.1/arch/arm64/kvm/ |
D | mmu.c | 912 gfn_t gfn_offset, unsigned long mask) in kvm_mmu_write_protect_pt_masked() 930 gfn_t gfn_offset, unsigned long mask) in kvm_arch_mmu_enable_log_dirty_pt_masked() 1136 gfn_t gfn; in user_mem_abort() 1389 gfn_t gfn; in kvm_handle_guest_abort()
|