Home
last modified time | relevance | path

Searched refs:kvm_x86_ops (Results 1 – 14 of 14) sorted by relevance

/Linux-v4.19/arch/x86/kvm/
Dpmu.c166 config = kvm_x86_ops->pmu_ops->find_arch_event(pmc_to_pmu(pmc), in reprogram_gp_counter()
196 kvm_x86_ops->pmu_ops->find_fixed_event(idx), in reprogram_fixed_counter()
205 struct kvm_pmc *pmc = kvm_x86_ops->pmu_ops->pmc_idx_to_pmc(pmu, pmc_idx); in reprogram_counter()
230 struct kvm_pmc *pmc = kvm_x86_ops->pmu_ops->pmc_idx_to_pmc(pmu, bit); in kvm_pmu_handle_event()
244 return kvm_x86_ops->pmu_ops->is_valid_msr_idx(vcpu, idx); in kvm_pmu_is_valid_msr_idx()
290 pmc = kvm_x86_ops->pmu_ops->msr_idx_to_pmc(vcpu, idx); in kvm_pmu_rdpmc()
310 return kvm_x86_ops->pmu_ops->is_valid_msr(vcpu, msr); in kvm_pmu_is_valid_msr()
315 return kvm_x86_ops->pmu_ops->get_msr(vcpu, msr, data); in kvm_pmu_get_msr()
320 return kvm_x86_ops->pmu_ops->set_msr(vcpu, msr_info); in kvm_pmu_set_msr()
329 kvm_x86_ops->pmu_ops->refresh(vcpu); in kvm_pmu_refresh()
[all …]
Dkvm_cache_regs.h14 kvm_x86_ops->cache_reg(vcpu, reg); in kvm_register_read()
44 kvm_x86_ops->cache_reg(vcpu, (enum kvm_reg)VCPU_EXREG_PDPTR); in kvm_pdptr_read()
53 kvm_x86_ops->decache_cr0_guest_bits(vcpu); in kvm_read_cr0_bits()
66 kvm_x86_ops->decache_cr4_guest_bits(vcpu); in kvm_read_cr4_bits()
73 kvm_x86_ops->decache_cr3(vcpu); in kvm_read_cr3()
Dx86.c108 struct kvm_x86_ops *kvm_x86_ops __read_mostly;
109 EXPORT_SYMBOL_GPL(kvm_x86_ops);
536 if (kvm_x86_ops->get_cpl(vcpu) <= required_cpl) in kvm_require_cpl()
678 kvm_x86_ops->get_cs_db_l_bits(vcpu, &cs_db, &cs_l); in kvm_set_cr0()
691 kvm_x86_ops->set_cr0(vcpu, cr0); in kvm_set_cr0()
778 if (kvm_x86_ops->get_cpl(vcpu) != 0 || in kvm_set_xcr()
835 if (kvm_x86_ops->set_cr4(vcpu, cr4)) in kvm_set_cr4()
919 kvm_x86_ops->set_dr6(vcpu, vcpu->arch.dr6); in kvm_update_dr6()
930 kvm_x86_ops->set_dr7(vcpu, dr7); in kvm_update_dr7()
996 *val = kvm_x86_ops->get_dr6(vcpu); in kvm_get_dr()
[all …]
Dcpuid.c54 && kvm_x86_ops->mpx_supported()); in kvm_mpx_supported()
231 kvm_x86_ops->cpuid_update(vcpu); in kvm_vcpu_ioctl_set_cpuid()
254 kvm_x86_ops->cpuid_update(vcpu); in kvm_vcpu_ioctl_set_cpuid2()
328 unsigned f_gbpages = (kvm_x86_ops->get_lpage_level() == PT_PDPE_LEVEL) in __do_cpuid_ent()
335 unsigned f_rdtscp = kvm_x86_ops->rdtscp_supported() ? F(RDTSCP) : 0; in __do_cpuid_ent()
336 unsigned f_invpcid = kvm_x86_ops->invpcid_supported() ? F(INVPCID) : 0; in __do_cpuid_ent()
338 unsigned f_xsaves = kvm_x86_ops->xsaves_supported() ? F(XSAVES) : 0; in __do_cpuid_ent()
339 unsigned f_umip = kvm_x86_ops->umip_emulated() ? F(UMIP) : 0; in __do_cpuid_ent()
706 kvm_x86_ops->set_supported_cpuid(function, entry); in __do_cpuid_ent()
Dmmu.h165 int cpl = kvm_x86_ops->get_cpl(vcpu); in permission_fault()
166 unsigned long rflags = kvm_x86_ops->get_rflags(vcpu); in permission_fault()
Dlapic.c444 kvm_x86_ops->hwapic_irr_update(vcpu, in apic_clear_irr()
469 kvm_x86_ops->hwapic_isr_update(vcpu, vec); in apic_set_isr()
517 kvm_x86_ops->hwapic_isr_update(vcpu, in apic_clear_isr()
655 highest_irr = kvm_x86_ops->sync_pir_to_irr(apic->vcpu); in apic_has_interrupt_for_ppr()
1052 kvm_x86_ops->deliver_posted_interrupt(vcpu, vector); in __apic_accept_irq()
1638 kvm_x86_ops->cancel_hv_timer(apic->vcpu); in cancel_hv_timer()
1648 if (!kvm_x86_ops->set_hv_timer) in start_hv_timer()
1657 r = kvm_x86_ops->set_hv_timer(apic->vcpu, ktimer->tscdeadline); in start_hv_timer()
2078 kvm_x86_ops->set_virtual_apic_mode(vcpu); in kvm_lapic_set_base()
2148 kvm_x86_ops->apicv_post_state_restore(vcpu); in kvm_lapic_reset()
[all …]
Dpmu.h78 return kvm_x86_ops->pmu_ops->pmc_is_enabled(pmc); in pmc_is_enabled()
Dmmu.c1191 max_level = min(kvm_x86_ops->get_lpage_level(), host_level); in mapping_level()
1634 if (kvm_x86_ops->enable_log_dirty_pt_masked) in kvm_arch_mmu_enable_log_dirty_pt_masked()
1635 kvm_x86_ops->enable_log_dirty_pt_masked(kvm, slot, gfn_offset, in kvm_arch_mmu_enable_log_dirty_pt_masked()
1650 if (kvm_x86_ops->write_log_dirty) in kvm_arch_write_log_dirty()
1651 return kvm_x86_ops->write_log_dirty(vcpu); in kvm_arch_write_log_dirty()
2880 spte |= kvm_x86_ops->get_mt_mask(vcpu, gfn, in set_spte()
3963 return kvm_x86_ops->interrupt_allowed(vcpu); in kvm_can_do_async_pf()
4177 kvm_x86_ops->tlb_flush(vcpu, true); in fast_cr3_switch()
4734 role.level = kvm_x86_ops->get_tdp_level(vcpu); in kvm_calc_tdp_mmu_root_page_role()
4751 context->shadow_root_level = kvm_x86_ops->get_tdp_level(vcpu); in init_kvm_tdp_mmu()
[all …]
Dx86.h99 kvm_x86_ops->get_cs_db_l_bits(vcpu, &cs_db, &cs_l); in is_64_bit_mode()
Dvmx.c5332 if (kvm_x86_ops->tlb_remote_flush) { in vmx_set_cr3()
7919 kvm_x86_ops->set_apic_access_page_addr = NULL; in hardware_setup()
7922 kvm_x86_ops->update_cr8_intercept = NULL; in hardware_setup()
7930 kvm_x86_ops->tlb_remote_flush = vmx_hv_remote_flush_tlb; in hardware_setup()
7943 kvm_x86_ops->sync_pir_to_irr = NULL; in hardware_setup()
7960 kvm_x86_ops->get_nested_state = NULL; in hardware_setup()
7961 kvm_x86_ops->set_nested_state = NULL; in hardware_setup()
7972 kvm_x86_ops->slot_enable_log_dirty = NULL; in hardware_setup()
7973 kvm_x86_ops->slot_disable_log_dirty = NULL; in hardware_setup()
7974 kvm_x86_ops->flush_log_dirty = NULL; in hardware_setup()
[all …]
Dtrace.h241 kvm_x86_ops->get_exit_info(vcpu, &__entry->info1,
744 __entry->csbase = kvm_x86_ops->get_segment_base(vcpu, VCPU_SREG_CS);
Dhyperv.c984 kvm_x86_ops->patch_hypercall(vcpu, instructions); in kvm_hv_set_msr_pw()
1452 if (kvm_x86_ops->get_cpl(vcpu) != 0 || !is_protmode(vcpu)) { in kvm_hv_hypercall()
Dsvm.c7043 static struct kvm_x86_ops svm_x86_ops __ro_after_init = {
/Linux-v4.19/arch/x86/include/asm/
Dkvm_host.h939 struct kvm_x86_ops { struct
1145 extern struct kvm_x86_ops *kvm_x86_ops; argument
1150 return kvm_x86_ops->vm_alloc(); in kvm_arch_alloc_vm()
1155 return kvm_x86_ops->vm_free(kvm); in kvm_arch_free_vm()
1161 if (kvm_x86_ops->tlb_remote_flush && in kvm_arch_flush_remote_tlb()
1162 !kvm_x86_ops->tlb_remote_flush(kvm)) in kvm_arch_flush_remote_tlb()
1506 if (kvm_x86_ops->vcpu_blocking) in kvm_arch_vcpu_blocking()
1507 kvm_x86_ops->vcpu_blocking(vcpu); in kvm_arch_vcpu_blocking()
1512 if (kvm_x86_ops->vcpu_unblocking) in kvm_arch_vcpu_unblocking()
1513 kvm_x86_ops->vcpu_unblocking(vcpu); in kvm_arch_vcpu_unblocking()