Home
last modified time | relevance | path

Searched refs:gva (Results 1 – 25 of 28) sorted by relevance

12

/Linux-v5.4/include/trace/events/
Dkvm.h285 TP_PROTO(u64 gva, u64 gfn),
287 TP_ARGS(gva, gfn),
290 __field(__u64, gva)
295 __entry->gva = gva;
299 TP_printk("gva = %#llx, gfn = %#llx", __entry->gva, __entry->gfn)
304 TP_PROTO(u64 gva, u64 gfn),
306 TP_ARGS(gva, gfn)
311 TP_PROTO(u64 gva, u64 gfn),
313 TP_ARGS(gva, gfn)
318 TP_PROTO(u64 token, u64 gva),
[all …]
/Linux-v5.4/tools/testing/selftests/kvm/lib/s390x/
Dprocessor.c70 void virt_pg_map(struct kvm_vm *vm, uint64_t gva, uint64_t gpa, in virt_pg_map() argument
76 TEST_ASSERT((gva % vm->page_size) == 0, in virt_pg_map()
79 gva, vm->page_size); in virt_pg_map()
81 (gva >> vm->page_shift)), in virt_pg_map()
83 gva); in virt_pg_map()
87 gva, vm->page_size); in virt_pg_map()
91 gva, vm->max_gfn, vm->page_size); in virt_pg_map()
96 idx = (gva >> (64 - 11 * ri)) & 0x7ffu; in virt_pg_map()
103 idx = (gva >> 12) & 0x0ffu; /* page index */ in virt_pg_map()
130 vm_paddr_t addr_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) in addr_gva2gpa() argument
[all …]
/Linux-v5.4/tools/testing/selftests/kvm/lib/aarch64/
Dprocessor.c24 static uint64_t pgd_index(struct kvm_vm *vm, vm_vaddr_t gva) in pgd_index() argument
29 return (gva >> shift) & mask; in pgd_index()
32 static uint64_t pud_index(struct kvm_vm *vm, vm_vaddr_t gva) in pud_index() argument
40 return (gva >> shift) & mask; in pud_index()
43 static uint64_t pmd_index(struct kvm_vm *vm, vm_vaddr_t gva) in pmd_index() argument
51 return (gva >> shift) & mask; in pmd_index()
54 static uint64_t pte_index(struct kvm_vm *vm, vm_vaddr_t gva) in pte_index() argument
57 return (gva >> vm->page_shift) & mask; in pte_index()
148 vm_paddr_t addr_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) in addr_gva2gpa() argument
155 ptep = addr_gpa2hva(vm, vm->pgd) + pgd_index(vm, gva) * 8; in addr_gva2gpa()
[all …]
Ducall.c99 vm_vaddr_t gva; in get_ucall() local
103 memcpy(&gva, run->mmio.data, sizeof(gva)); in get_ucall()
104 memcpy(&ucall, addr_gva2hva(vm, gva), sizeof(ucall)); in get_ucall()
/Linux-v5.4/virt/kvm/
Dasync_pf.c67 gva_t gva = apf->gva; in async_pf_execute() local
95 trace_kvm_async_pf_completed(addr, gva); in async_pf_execute()
168 int kvm_setup_async_pf(struct kvm_vcpu *vcpu, gva_t gva, unsigned long hva, in kvm_setup_async_pf() argument
188 work->gva = gva; in kvm_setup_async_pf()
/Linux-v5.4/arch/x86/kvm/
Dx86.h187 gva_t gva, gfn_t gfn, unsigned access) in vcpu_cache_mmio_info() argument
198 vcpu->arch.mmio_gva = mmu_is_nested(vcpu) ? 0 : gva & PAGE_MASK; in vcpu_cache_mmio_info()
215 static inline void vcpu_clear_mmio_info(struct kvm_vcpu *vcpu, gva_t gva) in vcpu_clear_mmio_info() argument
217 if (gva != MMIO_GVA_ANY && vcpu->arch.mmio_gva != (gva & PAGE_MASK)) in vcpu_clear_mmio_info()
223 static inline bool vcpu_match_mmio_gva(struct kvm_vcpu *vcpu, unsigned long gva) in vcpu_match_mmio_gva() argument
226 vcpu->arch.mmio_gva == (gva & PAGE_MASK)) in vcpu_match_mmio_gva()
Dmmutrace.h252 TP_PROTO(struct kvm_vcpu *vcpu, gva_t gva, u32 error_code,
254 TP_ARGS(vcpu, gva, error_code, sptep, old_spte, retry),
258 __field(gva_t, gva)
268 __entry->gva = gva;
278 __entry->gva, __print_flags(__entry->error_code, "|",
Dmmu.c2225 static void nonpaging_invlpg(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root) in nonpaging_invlpg() argument
3426 static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, in handle_abnormal_pfn() argument
3436 vcpu_cache_mmio_info(vcpu, gva, gfn, in handle_abnormal_pfn()
3531 static bool fast_page_fault(struct kvm_vcpu *vcpu, gva_t gva, int level, in fast_page_fault() argument
3551 for_each_shadow_entry_lockless(vcpu, gva, iterator, spte) in fast_page_fault()
3629 trace_fast_page_fault(vcpu, gva, error_code, iterator.sptep, in fast_page_fault()
3637 gva_t gva, kvm_pfn_t *pfn, bool write, bool *writable);
4152 static int nonpaging_page_fault(struct kvm_vcpu *vcpu, gva_t gva, in nonpaging_page_fault() argument
4155 gfn_t gfn = gva >> PAGE_SHIFT; in nonpaging_page_fault()
4158 pgprintk("%s: gva %lx error %x\n", __func__, gva, error_code); in nonpaging_page_fault()
[all …]
Dtrace.h772 TP_PROTO(gva_t gva, gpa_t gpa, bool write, bool gpa_match),
773 TP_ARGS(gva, gpa, write, gpa_match),
776 __field(gva_t, gva)
783 __entry->gva = gva;
789 TP_printk("gva %#lx gpa %#llx %s %s", __entry->gva, __entry->gpa,
Dpaging_tmpl.h892 static void FNAME(invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa) in FNAME()
899 vcpu_clear_mmio_info(vcpu, gva); in FNAME()
913 for_each_shadow_entry_using_root(vcpu, root_hpa, gva, iterator) { in FNAME()
Dx86.c5281 gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_read() argument
5285 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_read()
5288 gpa_t kvm_mmu_gva_to_gpa_fetch(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_fetch() argument
5293 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_fetch()
5296 gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_write() argument
5301 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_write()
5305 gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_system() argument
5308 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, 0, exception); in kvm_mmu_gva_to_gpa_system()
5490 static int vcpu_is_mmio_gpa(struct kvm_vcpu *vcpu, unsigned long gva, in vcpu_is_mmio_gpa() argument
5498 trace_vcpu_match_mmio(gva, gpa, write, true); in vcpu_is_mmio_gpa()
[all …]
/Linux-v5.4/arch/s390/kvm/
Dgaccess.c493 static int trans_exc(struct kvm_vcpu *vcpu, int code, unsigned long gva, in trans_exc() argument
534 tec->addr = gva >> PAGE_SHIFT; in trans_exc()
614 static unsigned long guest_translate(struct kvm_vcpu *vcpu, unsigned long gva, in guest_translate() argument
618 union vaddress vaddr = {.addr = gva}; in guest_translate()
619 union raddress raddr = {.addr = gva}; in guest_translate()
904 int guest_translate_address(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar, in guest_translate_address() argument
912 gva = kvm_s390_logical_to_effective(vcpu, gva); in guest_translate_address()
913 rc = get_vcpu_asce(vcpu, &asce, gva, ar, mode); in guest_translate_address()
916 if (is_low_address(gva) && low_address_protection_enabled(vcpu, asce)) { in guest_translate_address()
918 return trans_exc(vcpu, PGM_PROTECTION, gva, 0, in guest_translate_address()
[all …]
Dgaccess.h161 int guest_translate_address(struct kvm_vcpu *vcpu, unsigned long gva,
163 int check_gva_range(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar,
/Linux-v5.4/arch/x86/kvm/vmx/
Dops.h19 void invvpid_error(unsigned long ext, u16 vpid, gva_t gva);
236 static inline void __invvpid(unsigned long ext, u16 vpid, gva_t gva) in __invvpid() argument
241 u64 gva; in __invvpid() member
242 } operand = { vpid, 0, gva }; in __invvpid()
244 vmx_asm2(invvpid, "r"(ext), "m"(operand), ext, vpid, gva); in __invvpid()
Dnested.c4332 gva_t gva; in nested_vmx_get_vmptr() local
4337 sizeof(*vmpointer), &gva)) in nested_vmx_get_vmptr()
4340 if (kvm_read_guest_virt(vcpu, gva, vmpointer, sizeof(*vmpointer), &e)) { in nested_vmx_get_vmptr()
4605 gva_t gva = 0; in handle_vmread() local
4653 vmx_instruction_info, true, len, &gva)) in handle_vmread()
4656 if (kvm_write_guest_virt_system(vcpu, gva, &field_value, len, &e)) in handle_vmread()
4691 gva_t gva; in handle_vmwrite() local
4719 vmx_instruction_info, false, len, &gva)) in handle_vmwrite()
4721 if (kvm_read_guest_virt(vcpu, gva, &field_value, len, &e)) { in handle_vmwrite()
4885 gva_t gva; in handle_vmptrst() local
[all …]
/Linux-v5.4/arch/mips/kvm/
Dmmu.c1042 unsigned long gva, in kvm_mips_handle_mapped_seg_tlb_fault() argument
1048 unsigned int idx = TLB_LO_IDX(*tlb, gva); in kvm_mips_handle_mapped_seg_tlb_fault()
1058 if (!((gva ^ KVM_GUEST_COMMPAGE_ADDR) & VPN2_MASK & (PAGE_MASK << 1))) in kvm_mips_handle_mapped_seg_tlb_fault()
1078 ptep_gva = kvm_trap_emul_pte_for_gva(vcpu, gva & ~PAGE_SIZE); in kvm_mips_handle_mapped_seg_tlb_fault()
1080 kvm_err("No ptep for gva %lx\n", gva); in kvm_mips_handle_mapped_seg_tlb_fault()
1089 kvm_mips_host_tlb_inv(vcpu, gva, !kernel, kernel); in kvm_mips_handle_mapped_seg_tlb_fault()
1197 unsigned long gva, in kvm_trap_emul_gva_fault() argument
1204 if (KVM_GUEST_KSEGX(gva) == KVM_GUEST_KSEG0) { in kvm_trap_emul_gva_fault()
1205 if (kvm_mips_handle_kseg0_tlb_fault(gva, vcpu, write) < 0) in kvm_trap_emul_gva_fault()
1207 } else if ((KVM_GUEST_KSEGX(gva) < KVM_GUEST_KSEG0) || in kvm_trap_emul_gva_fault()
[all …]
Dtlb.c304 int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva, in kvm_vz_guest_tlb_lookup() argument
322 write_gc0_entryhi((o_entryhi & 0x3ff) | (gva & ~0xfffl)); in kvm_vz_guest_tlb_lookup()
364 pa = entrylo[!!(gva & pagemaskbit)]; in kvm_vz_guest_tlb_lookup()
378 pa |= gva & ~(pagemask | pagemaskbit); in kvm_vz_guest_tlb_lookup()
Dtrap_emul.c23 static gpa_t kvm_trap_emul_gva_to_gpa_cb(gva_t gva) in kvm_trap_emul_gva_to_gpa_cb() argument
26 gva_t kseg = KSEGX(gva); in kvm_trap_emul_gva_to_gpa_cb()
27 gva_t gkseg = KVM_GUEST_KSEGX(gva); in kvm_trap_emul_gva_to_gpa_cb()
30 gpa = CPHYSADDR(gva); in kvm_trap_emul_gva_to_gpa_cb()
32 gpa = KVM_GUEST_CPHYSADDR(gva); in kvm_trap_emul_gva_to_gpa_cb()
34 kvm_err("%s: cannot find GPA for GVA: %#lx\n", __func__, gva); in kvm_trap_emul_gva_to_gpa_cb()
39 kvm_debug("%s: gva %#lx, gpa: %#llx\n", __func__, gva, gpa); in kvm_trap_emul_gva_to_gpa_cb()
Dvz.c183 static gpa_t kvm_vz_gva_to_gpa_cb(gva_t gva) in kvm_vz_gva_to_gpa_cb() argument
186 return gva; in kvm_vz_gva_to_gpa_cb()
714 static int kvm_vz_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva, in kvm_vz_gva_to_gpa() argument
717 u32 gva32 = gva; in kvm_vz_gva_to_gpa()
720 if ((long)gva == (s32)gva32) { in kvm_vz_gva_to_gpa()
775 } else if ((gva & 0xc000000000000000) == 0x8000000000000000) { in kvm_vz_gva_to_gpa()
783 if (segctl & (1ull << (56 + ((gva >> 59) & 0x7)))) { in kvm_vz_gva_to_gpa()
797 *gpa = gva & 0x07ffffffffffffff; in kvm_vz_gva_to_gpa()
803 return kvm_vz_guest_tlb_lookup(vcpu, gva, gpa); in kvm_vz_gva_to_gpa()
/Linux-v5.4/tools/testing/selftests/kvm/include/
Dkvm_util.h80 int kvm_memcmp_hva_gva(void *hva, struct kvm_vm *vm, const vm_vaddr_t gva,
109 void *addr_gva2hva(struct kvm_vm *vm, vm_vaddr_t gva);
111 vm_paddr_t addr_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva);
/Linux-v5.4/arch/x86/include/asm/
Dkvm_host.h383 int (*page_fault)(struct kvm_vcpu *vcpu, gva_t gva, u32 err,
387 gpa_t (*gva_to_gpa)(struct kvm_vcpu *vcpu, gva_t gva, u32 access,
393 void (*invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa);
1431 int kvm_mmu_unprotect_page_virt(struct kvm_vcpu *vcpu, gva_t gva);
1440 gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva,
1442 gpa_t kvm_mmu_gva_to_gpa_fetch(struct kvm_vcpu *vcpu, gva_t gva,
1444 gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva,
1446 gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva,
1453 int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gva_t gva, u64 error_code,
1455 void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva);
[all …]
/Linux-v5.4/tools/testing/selftests/kvm/lib/x86_64/
Dprocessor.c542 vm_paddr_t addr_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) in addr_gva2gpa() argument
553 index[0] = (gva >> 12) & 0x1ffu; in addr_gva2gpa()
554 index[1] = (gva >> 21) & 0x1ffu; in addr_gva2gpa()
555 index[2] = (gva >> 30) & 0x1ffu; in addr_gva2gpa()
556 index[3] = (gva >> 39) & 0x1ffu; in addr_gva2gpa()
576 return (pte[index[0]].address * vm->page_size) + (gva & 0xfffu); in addr_gva2gpa()
580 "gva: 0x%lx", gva); in addr_gva2gpa()
/Linux-v5.4/arch/mips/include/asm/
Dkvm_host.h796 gpa_t (*gva_to_gpa)(gva_t gva);
866 unsigned long gva,
885 int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva,
933 unsigned long gva,
/Linux-v5.4/tools/testing/selftests/kvm/lib/
Dkvm_util.c506 int kvm_memcmp_hva_gva(void *hva, struct kvm_vm *vm, vm_vaddr_t gva, size_t len) in kvm_memcmp_hva_gva() argument
521 uintptr_t ptr2 = (uintptr_t)addr_gva2hva(vm, gva + offset); in kvm_memcmp_hva_gva()
1615 void *addr_gva2hva(struct kvm_vm *vm, vm_vaddr_t gva) in addr_gva2hva() argument
1617 return addr_gpa2hva(vm, addr_gva2gpa(vm, gva)); in addr_gva2hva()
/Linux-v5.4/Documentation/virt/kvm/
Dmmu.txt36 gva guest virtual address
65 guest physical addresses, to host physical addresses (gva->gpa->hpa)
122 paging: gva->gpa->hpa
123 paging, tdp: (gva->)gpa->hpa
303 (gva->gpa or ngpa->gpa)

12