Lines Matching refs:kvm_x86_ops
217 config = kvm_x86_ops.pmu_ops->find_arch_event(pmc_to_pmu(pmc), in reprogram_gp_counter()
271 kvm_x86_ops.pmu_ops->find_fixed_event(idx), in reprogram_fixed_counter()
280 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, pmc_idx); in reprogram_counter()
302 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, bit); in kvm_pmu_handle_event()
324 return kvm_x86_ops.pmu_ops->is_valid_rdpmc_ecx(vcpu, idx); in kvm_pmu_is_valid_rdpmc_ecx()
374 pmc = kvm_x86_ops.pmu_ops->rdpmc_ecx_to_pmc(vcpu, idx, &mask); in kvm_pmu_rdpmc()
390 if (kvm_x86_ops.pmu_ops->deliver_pmi) in kvm_pmu_deliver_pmi()
391 kvm_x86_ops.pmu_ops->deliver_pmi(vcpu); in kvm_pmu_deliver_pmi()
398 return kvm_x86_ops.pmu_ops->msr_idx_to_pmc(vcpu, msr) || in kvm_pmu_is_valid_msr()
399 kvm_x86_ops.pmu_ops->is_valid_msr(vcpu, msr); in kvm_pmu_is_valid_msr()
405 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->msr_idx_to_pmc(vcpu, msr); in kvm_pmu_mark_pmc_in_use()
413 return kvm_x86_ops.pmu_ops->get_msr(vcpu, msr_info); in kvm_pmu_get_msr()
419 return kvm_x86_ops.pmu_ops->set_msr(vcpu, msr_info); in kvm_pmu_set_msr()
428 kvm_x86_ops.pmu_ops->refresh(vcpu); in kvm_pmu_refresh()
436 kvm_x86_ops.pmu_ops->reset(vcpu); in kvm_pmu_reset()
444 kvm_x86_ops.pmu_ops->init(vcpu); in kvm_pmu_init()
476 pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, i); in kvm_pmu_cleanup()
482 if (kvm_x86_ops.pmu_ops->cleanup) in kvm_pmu_cleanup()
483 kvm_x86_ops.pmu_ops->cleanup(vcpu); in kvm_pmu_cleanup()