/Linux-v5.4/arch/powerpc/kvm/ |
D | trace_hv.h | 285 __field(u64, base_gfn) 297 __entry->base_gfn = memslot ? memslot->base_gfn : -1UL; 305 __entry->base_gfn, __entry->slot_flags)
|
D | book3s_64_mmu_hv.c | 576 if (gfn_base < memslot->base_gfn) in kvmppc_book3s_hv_page_fault() 687 rmap = &memslot->arch.rmap[gfn_base - memslot->base_gfn]; in kvmppc_book3s_hv_page_fault() 861 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_unmap_rmapp() 907 gfn = memslot->base_gfn; in kvmppc_core_flush_memslot_hv() 936 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_age_rmapp() 999 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_test_age_rmapp() 1138 if (gfn < memslot->base_gfn || in kvmppc_harvest_vpa_dirty() 1139 gfn >= memslot->base_gfn + memslot->npages) in kvmppc_harvest_vpa_dirty() 1144 __set_bit_le(gfn - memslot->base_gfn, map); in kvmppc_harvest_vpa_dirty() 1219 set_bit_le(gfn - memslot->base_gfn, memslot->dirty_bitmap); in kvmppc_unpin_guest_page() [all …]
|
D | book3s_64_mmu_radix.c | 998 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_age_radix() 1026 unsigned long gfn = memslot->base_gfn + pagenum; in kvm_radix_test_clear_dirty() 1043 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_radix_test_clear_dirty() 1085 gpa = memslot->base_gfn << PAGE_SHIFT; in kvmppc_radix_flush_memslot()
|
D | book3s_hv_rm_mmu.c | 117 gfn -= memslot->base_gfn; in kvmppc_update_dirty_map() 155 rmap = real_vmalloc_addr(&memslot->arch.rmap[gfn - memslot->base_gfn]); in revmap_for_hpte() 246 slot_fn = gfn - memslot->base_gfn; in kvmppc_do_h_enter()
|
D | book3s_64_vio_hv.c | 92 *prmap = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvmppc_rm_tce_to_ua()
|
D | e500_mmu_host.c | 381 slot_start = pfn - (gfn - slot->base_gfn); in kvmppc_e500_shadow_map()
|
D | book3s_hv_nested.c | 887 gfn = (gpa >> PAGE_SHIFT) - memslot->base_gfn; in kvmhv_remove_nest_rmap_range() 1416 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in __kvmhv_nested_page_fault()
|
D | book3s_pr.c | 1882 ga = memslot->base_gfn << PAGE_SHIFT; in kvm_vm_ioctl_get_dirty_log_pr()
|
D | book3s_hv.c | 813 if ((from + len) >= ((from_memslot->base_gfn + from_memslot->npages) in kvmppc_copy_guest() 825 if ((to + len) >= ((to_memslot->base_gfn + to_memslot->npages) in kvmppc_copy_guest()
|
/Linux-v5.4/arch/x86/kvm/ |
D | page_track.c | 68 index = gfn_to_index(gfn, slot->base_gfn, PT_PAGE_TABLE_LEVEL); in update_gfn_track() 158 index = gfn_to_index(gfn, slot->base_gfn, PT_PAGE_TABLE_LEVEL); in kvm_page_track_is_active()
|
D | paging_tmpl.h | 624 gfn_t gfn, base_gfn; in FNAME() local 674 base_gfn = gfn; in FNAME() 687 base_gfn = gfn & ~(KVM_PAGES_PER_HPAGE(it.level) - 1); in FNAME() 696 sp = kvm_mmu_get_page(vcpu, base_gfn, addr, in FNAME() 705 it.level, base_gfn, pfn, prefault, map_writable); in FNAME()
|
D | mmu.c | 1182 idx = gfn_to_index(gfn, slot->base_gfn, level); in lpage_info_slot() 1469 idx = gfn_to_index(gfn, slot->base_gfn, level); in __gfn_to_rmap() 1753 rmap_head = __gfn_to_rmap(slot->base_gfn + gfn_offset + __ffs(mask), in kvm_mmu_write_protect_pt_masked() 1779 rmap_head = __gfn_to_rmap(slot->base_gfn + gfn_offset + __ffs(mask), in kvm_mmu_clear_dirty_pt_masked() 3328 gfn_t base_gfn = gfn; in __direct_map() local 3341 base_gfn = gfn & ~(KVM_PAGES_PER_HPAGE(it.level) - 1); in __direct_map() 3347 sp = kvm_mmu_get_page(vcpu, base_gfn, it.addr, in __direct_map() 3357 write, level, base_gfn, pfn, prefault, in __direct_map() 5713 end_level, memslot->base_gfn, in slot_handle_level() 5714 memslot->base_gfn + memslot->npages - 1, in slot_handle_level() [all …]
|
D | x86.c | 9657 lpages = gfn_to_index(slot->base_gfn + npages - 1, in kvm_arch_create_memslot() 9658 slot->base_gfn, level) + 1; in kvm_arch_create_memslot() 9674 if (slot->base_gfn & (KVM_PAGES_PER_HPAGE(level) - 1)) in kvm_arch_create_memslot() 9676 if ((slot->base_gfn + npages) & (KVM_PAGES_PER_HPAGE(level) - 1)) in kvm_arch_create_memslot() 9684 if ((slot->base_gfn ^ ugfn) & (KVM_PAGES_PER_HPAGE(level) - 1) || in kvm_arch_create_memslot()
|
/Linux-v5.4/include/linux/ |
D | kvm_host.h | 344 gfn_t base_gfn; member 1012 if (gfn >= memslots[slot].base_gfn && in search_memslots() 1013 gfn < memslots[slot].base_gfn + memslots[slot].npages) in search_memslots() 1019 if (gfn >= memslots[slot].base_gfn) in search_memslots() 1025 if (gfn >= memslots[start].base_gfn && in search_memslots() 1026 gfn < memslots[start].base_gfn + memslots[start].npages) { in search_memslots() 1043 return slot->userspace_addr + (gfn - slot->base_gfn) * PAGE_SIZE; in __gfn_to_hva_memslot() 1056 return slot->base_gfn + gfn_offset; in hva_to_gfn_memslot()
|
/Linux-v5.4/virt/kvm/arm/ |
D | mmu.c | 411 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_flush_memslot() 922 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_unmap_memslot() 1539 phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; in kvm_mmu_wp_memory_region() 1540 phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in kvm_mmu_wp_memory_region() 1563 phys_addr_t base_gfn = slot->base_gfn + gfn_offset; in kvm_mmu_write_protect_pt_masked() local 1564 phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT; in kvm_mmu_write_protect_pt_masked() 1565 phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT; in kvm_mmu_write_protect_pt_masked() 1617 gpa_start = memslot->base_gfn << PAGE_SHIFT; in fault_supports_stage2_huge_mapping() 2280 if (memslot->base_gfn + memslot->npages >= in kvm_arch_prepare_memory_region() 2379 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot()
|
/Linux-v5.4/virt/kvm/ |
D | kvm_main.c | 896 new->base_gfn <= mslots[i + 1].base_gfn) { in update_memslots() 915 new->base_gfn >= mslots[i - 1].base_gfn) { in update_memslots() 989 gfn_t base_gfn; in __kvm_set_memory_region() local 1022 base_gfn = mem->guest_phys_addr >> PAGE_SHIFT; in __kvm_set_memory_region() 1031 new.base_gfn = base_gfn; in __kvm_set_memory_region() 1044 if (base_gfn != old.base_gfn) in __kvm_set_memory_region() 1058 new.base_gfn = 0; in __kvm_set_memory_region() 1068 if (!((base_gfn + npages <= slot->base_gfn) || in __kvm_set_memory_region() 1069 (base_gfn >= slot->base_gfn + slot->npages))) in __kvm_set_memory_region() 1436 *nr_pages = slot->npages - (gfn - slot->base_gfn); in __gfn_to_hva_many() [all …]
|
/Linux-v5.4/arch/mips/kvm/ |
D | mmu.c | 445 gfn_t base_gfn = slot->base_gfn + gfn_offset; in kvm_arch_mmu_enable_log_dirty_pt_masked() local 446 gfn_t start = base_gfn + __ffs(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked() 447 gfn_t end = base_gfn + __fls(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked()
|
D | mips.c | 216 kvm_mips_flush_gpa_pt(kvm, slot->base_gfn, in kvm_arch_flush_shadow_memslot() 217 slot->base_gfn + slot->npages - 1); in kvm_arch_flush_shadow_memslot() 257 needs_flush = kvm_mips_mkclean_gpa_pt(kvm, new->base_gfn, in kvm_arch_commit_memory_region() 258 new->base_gfn + new->npages - 1); in kvm_arch_commit_memory_region()
|
/Linux-v5.4/arch/s390/kvm/ |
D | kvm-s390.c | 585 cur_gfn = memslot->base_gfn; in kvm_s390_sync_dirty_log() 586 last_gfn = memslot->base_gfn + memslot->npages; in kvm_s390_sync_dirty_log() 1922 if (gfn >= memslots[slot].base_gfn && in gfn_to_memslot_approx() 1923 gfn < memslots[slot].base_gfn + memslots[slot].npages) in gfn_to_memslot_approx() 1929 if (gfn >= memslots[slot].base_gfn) in gfn_to_memslot_approx() 1935 if (gfn >= memslots[start].base_gfn && in gfn_to_memslot_approx() 1936 gfn < memslots[start].base_gfn + memslots[start].npages) { in gfn_to_memslot_approx() 1971 unsigned long ofs = cur_gfn - ms->base_gfn; in kvm_s390_next_dirty_cmma() 1973 if (ms->base_gfn + ms->npages <= cur_gfn) { in kvm_s390_next_dirty_cmma() 1988 return ms->base_gfn + ofs; in kvm_s390_next_dirty_cmma() [all …]
|
D | priv.c | 1175 if (ms && !test_and_set_bit(gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms))) in __do_essa()
|
/Linux-v5.4/include/trace/events/ |
D | kvm.h | 272 __entry->hva = ((gfn - slot->base_gfn) <<
|
/Linux-v5.4/arch/powerpc/include/asm/ |
D | kvm_book3s_64.h | 493 return !(memslot->base_gfn & mask) && !(memslot->npages & mask); in slot_is_aligned()
|
/Linux-v5.4/arch/x86/include/asm/ |
D | kvm_host.h | 120 static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level) in gfn_to_index() argument 124 (base_gfn >> KVM_HPAGE_GFN_SHIFT(level)); in gfn_to_index()
|
/Linux-v5.4/drivers/gpu/drm/i915/gvt/ |
D | kvmgt.c | 1725 gfn = slot->base_gfn + i; in kvmgt_page_track_flush_slot()
|