Searched refs:kvm_x86_ops (Results 1 – 14 of 14) sorted by relevance
/Linux-v4.19/arch/x86/kvm/ |
D | pmu.c | 166 config = kvm_x86_ops->pmu_ops->find_arch_event(pmc_to_pmu(pmc), in reprogram_gp_counter() 196 kvm_x86_ops->pmu_ops->find_fixed_event(idx), in reprogram_fixed_counter() 205 struct kvm_pmc *pmc = kvm_x86_ops->pmu_ops->pmc_idx_to_pmc(pmu, pmc_idx); in reprogram_counter() 230 struct kvm_pmc *pmc = kvm_x86_ops->pmu_ops->pmc_idx_to_pmc(pmu, bit); in kvm_pmu_handle_event() 244 return kvm_x86_ops->pmu_ops->is_valid_msr_idx(vcpu, idx); in kvm_pmu_is_valid_msr_idx() 290 pmc = kvm_x86_ops->pmu_ops->msr_idx_to_pmc(vcpu, idx); in kvm_pmu_rdpmc() 310 return kvm_x86_ops->pmu_ops->is_valid_msr(vcpu, msr); in kvm_pmu_is_valid_msr() 315 return kvm_x86_ops->pmu_ops->get_msr(vcpu, msr, data); in kvm_pmu_get_msr() 320 return kvm_x86_ops->pmu_ops->set_msr(vcpu, msr_info); in kvm_pmu_set_msr() 329 kvm_x86_ops->pmu_ops->refresh(vcpu); in kvm_pmu_refresh() [all …]
|
D | kvm_cache_regs.h | 14 kvm_x86_ops->cache_reg(vcpu, reg); in kvm_register_read() 44 kvm_x86_ops->cache_reg(vcpu, (enum kvm_reg)VCPU_EXREG_PDPTR); in kvm_pdptr_read() 53 kvm_x86_ops->decache_cr0_guest_bits(vcpu); in kvm_read_cr0_bits() 66 kvm_x86_ops->decache_cr4_guest_bits(vcpu); in kvm_read_cr4_bits() 73 kvm_x86_ops->decache_cr3(vcpu); in kvm_read_cr3()
|
D | x86.c | 108 struct kvm_x86_ops *kvm_x86_ops __read_mostly; 109 EXPORT_SYMBOL_GPL(kvm_x86_ops); 536 if (kvm_x86_ops->get_cpl(vcpu) <= required_cpl) in kvm_require_cpl() 678 kvm_x86_ops->get_cs_db_l_bits(vcpu, &cs_db, &cs_l); in kvm_set_cr0() 691 kvm_x86_ops->set_cr0(vcpu, cr0); in kvm_set_cr0() 778 if (kvm_x86_ops->get_cpl(vcpu) != 0 || in kvm_set_xcr() 835 if (kvm_x86_ops->set_cr4(vcpu, cr4)) in kvm_set_cr4() 919 kvm_x86_ops->set_dr6(vcpu, vcpu->arch.dr6); in kvm_update_dr6() 930 kvm_x86_ops->set_dr7(vcpu, dr7); in kvm_update_dr7() 996 *val = kvm_x86_ops->get_dr6(vcpu); in kvm_get_dr() [all …]
|
D | cpuid.c | 54 && kvm_x86_ops->mpx_supported()); in kvm_mpx_supported() 231 kvm_x86_ops->cpuid_update(vcpu); in kvm_vcpu_ioctl_set_cpuid() 254 kvm_x86_ops->cpuid_update(vcpu); in kvm_vcpu_ioctl_set_cpuid2() 328 unsigned f_gbpages = (kvm_x86_ops->get_lpage_level() == PT_PDPE_LEVEL) in __do_cpuid_ent() 335 unsigned f_rdtscp = kvm_x86_ops->rdtscp_supported() ? F(RDTSCP) : 0; in __do_cpuid_ent() 336 unsigned f_invpcid = kvm_x86_ops->invpcid_supported() ? F(INVPCID) : 0; in __do_cpuid_ent() 338 unsigned f_xsaves = kvm_x86_ops->xsaves_supported() ? F(XSAVES) : 0; in __do_cpuid_ent() 339 unsigned f_umip = kvm_x86_ops->umip_emulated() ? F(UMIP) : 0; in __do_cpuid_ent() 706 kvm_x86_ops->set_supported_cpuid(function, entry); in __do_cpuid_ent()
|
D | mmu.h | 165 int cpl = kvm_x86_ops->get_cpl(vcpu); in permission_fault() 166 unsigned long rflags = kvm_x86_ops->get_rflags(vcpu); in permission_fault()
|
D | lapic.c | 444 kvm_x86_ops->hwapic_irr_update(vcpu, in apic_clear_irr() 469 kvm_x86_ops->hwapic_isr_update(vcpu, vec); in apic_set_isr() 517 kvm_x86_ops->hwapic_isr_update(vcpu, in apic_clear_isr() 655 highest_irr = kvm_x86_ops->sync_pir_to_irr(apic->vcpu); in apic_has_interrupt_for_ppr() 1052 kvm_x86_ops->deliver_posted_interrupt(vcpu, vector); in __apic_accept_irq() 1638 kvm_x86_ops->cancel_hv_timer(apic->vcpu); in cancel_hv_timer() 1648 if (!kvm_x86_ops->set_hv_timer) in start_hv_timer() 1657 r = kvm_x86_ops->set_hv_timer(apic->vcpu, ktimer->tscdeadline); in start_hv_timer() 2078 kvm_x86_ops->set_virtual_apic_mode(vcpu); in kvm_lapic_set_base() 2148 kvm_x86_ops->apicv_post_state_restore(vcpu); in kvm_lapic_reset() [all …]
|
D | pmu.h | 78 return kvm_x86_ops->pmu_ops->pmc_is_enabled(pmc); in pmc_is_enabled()
|
D | mmu.c | 1191 max_level = min(kvm_x86_ops->get_lpage_level(), host_level); in mapping_level() 1634 if (kvm_x86_ops->enable_log_dirty_pt_masked) in kvm_arch_mmu_enable_log_dirty_pt_masked() 1635 kvm_x86_ops->enable_log_dirty_pt_masked(kvm, slot, gfn_offset, in kvm_arch_mmu_enable_log_dirty_pt_masked() 1650 if (kvm_x86_ops->write_log_dirty) in kvm_arch_write_log_dirty() 1651 return kvm_x86_ops->write_log_dirty(vcpu); in kvm_arch_write_log_dirty() 2880 spte |= kvm_x86_ops->get_mt_mask(vcpu, gfn, in set_spte() 3963 return kvm_x86_ops->interrupt_allowed(vcpu); in kvm_can_do_async_pf() 4177 kvm_x86_ops->tlb_flush(vcpu, true); in fast_cr3_switch() 4734 role.level = kvm_x86_ops->get_tdp_level(vcpu); in kvm_calc_tdp_mmu_root_page_role() 4751 context->shadow_root_level = kvm_x86_ops->get_tdp_level(vcpu); in init_kvm_tdp_mmu() [all …]
|
D | x86.h | 99 kvm_x86_ops->get_cs_db_l_bits(vcpu, &cs_db, &cs_l); in is_64_bit_mode()
|
D | vmx.c | 5332 if (kvm_x86_ops->tlb_remote_flush) { in vmx_set_cr3() 7919 kvm_x86_ops->set_apic_access_page_addr = NULL; in hardware_setup() 7922 kvm_x86_ops->update_cr8_intercept = NULL; in hardware_setup() 7930 kvm_x86_ops->tlb_remote_flush = vmx_hv_remote_flush_tlb; in hardware_setup() 7943 kvm_x86_ops->sync_pir_to_irr = NULL; in hardware_setup() 7960 kvm_x86_ops->get_nested_state = NULL; in hardware_setup() 7961 kvm_x86_ops->set_nested_state = NULL; in hardware_setup() 7972 kvm_x86_ops->slot_enable_log_dirty = NULL; in hardware_setup() 7973 kvm_x86_ops->slot_disable_log_dirty = NULL; in hardware_setup() 7974 kvm_x86_ops->flush_log_dirty = NULL; in hardware_setup() [all …]
|
D | trace.h | 241 kvm_x86_ops->get_exit_info(vcpu, &__entry->info1, 744 __entry->csbase = kvm_x86_ops->get_segment_base(vcpu, VCPU_SREG_CS);
|
D | hyperv.c | 984 kvm_x86_ops->patch_hypercall(vcpu, instructions); in kvm_hv_set_msr_pw() 1452 if (kvm_x86_ops->get_cpl(vcpu) != 0 || !is_protmode(vcpu)) { in kvm_hv_hypercall()
|
D | svm.c | 7043 static struct kvm_x86_ops svm_x86_ops __ro_after_init = {
|
/Linux-v4.19/arch/x86/include/asm/ |
D | kvm_host.h | 939 struct kvm_x86_ops { struct 1145 extern struct kvm_x86_ops *kvm_x86_ops; argument 1150 return kvm_x86_ops->vm_alloc(); in kvm_arch_alloc_vm() 1155 return kvm_x86_ops->vm_free(kvm); in kvm_arch_free_vm() 1161 if (kvm_x86_ops->tlb_remote_flush && in kvm_arch_flush_remote_tlb() 1162 !kvm_x86_ops->tlb_remote_flush(kvm)) in kvm_arch_flush_remote_tlb() 1506 if (kvm_x86_ops->vcpu_blocking) in kvm_arch_vcpu_blocking() 1507 kvm_x86_ops->vcpu_blocking(vcpu); in kvm_arch_vcpu_blocking() 1512 if (kvm_x86_ops->vcpu_unblocking) in kvm_arch_vcpu_unblocking() 1513 kvm_x86_ops->vcpu_unblocking(vcpu); in kvm_arch_vcpu_unblocking()
|