Lines Matching refs:kvm_x86_ops
214 config = kvm_x86_ops.pmu_ops->find_arch_event(pmc_to_pmu(pmc), in reprogram_gp_counter()
268 kvm_x86_ops.pmu_ops->find_fixed_event(idx), in reprogram_fixed_counter()
277 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, pmc_idx); in reprogram_counter()
299 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, bit); in kvm_pmu_handle_event()
321 return kvm_x86_ops.pmu_ops->is_valid_rdpmc_ecx(vcpu, idx); in kvm_pmu_is_valid_rdpmc_ecx()
371 pmc = kvm_x86_ops.pmu_ops->rdpmc_ecx_to_pmc(vcpu, idx, &mask); in kvm_pmu_rdpmc()
376 (kvm_x86_ops.get_cpl(vcpu) != 0) && in kvm_pmu_rdpmc()
392 return kvm_x86_ops.pmu_ops->msr_idx_to_pmc(vcpu, msr) || in kvm_pmu_is_valid_msr()
393 kvm_x86_ops.pmu_ops->is_valid_msr(vcpu, msr); in kvm_pmu_is_valid_msr()
399 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->msr_idx_to_pmc(vcpu, msr); in kvm_pmu_mark_pmc_in_use()
407 return kvm_x86_ops.pmu_ops->get_msr(vcpu, msr_info); in kvm_pmu_get_msr()
413 return kvm_x86_ops.pmu_ops->set_msr(vcpu, msr_info); in kvm_pmu_set_msr()
422 kvm_x86_ops.pmu_ops->refresh(vcpu); in kvm_pmu_refresh()
430 kvm_x86_ops.pmu_ops->reset(vcpu); in kvm_pmu_reset()
438 kvm_x86_ops.pmu_ops->init(vcpu); in kvm_pmu_init()
470 pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, i); in kvm_pmu_cleanup()