Searched refs:vcpu_to_pmu (Results 1 – 10 of 10) sorted by relevance
83 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_is_valid_rdpmc_ecx()94 return amd_pmc_idx_to_pmc(vcpu_to_pmu(vcpu), idx & ~(3u << 30)); in amd_rdpmc_ecx_to_pmc()99 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_msr_idx_to_pmc()110 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_is_valid_msr()133 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_get_msr()155 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_set_msr()183 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_refresh()222 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_init()238 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_reset()
130 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_is_valid_rdpmc_ecx()142 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_rdpmc_ecx_to_pmc()202 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_is_valid_msr()233 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_msr_idx_to_pmc()250 vcpu_to_pmu(vcpu)->event_count--; in intel_pmu_release_guest_lbr_event()257 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_create_guest_lbr_event()335 __set_bit(INTEL_PMC_IDX_FIXED_VLBR, vcpu_to_pmu(vcpu)->pmc_in_use); in intel_pmu_handle_lbr_msrs_access()339 clear_bit(INTEL_PMC_IDX_FIXED_VLBR, vcpu_to_pmu(vcpu)->pmc_in_use); in intel_pmu_handle_lbr_msrs_access()350 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_get_msr()393 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_set_msr()[all …]
2651 kvm_pmu_has_perf_global_ctrl(vcpu_to_pmu(vcpu)) && in prepare_vmcs02()2927 CC(!kvm_valid_perf_global_ctrl(vcpu_to_pmu(vcpu), in nested_vmx_check_host_state()3046 CC(!kvm_valid_perf_global_ctrl(vcpu_to_pmu(vcpu), in nested_vmx_check_guest_state()4526 kvm_pmu_has_perf_global_ctrl(vcpu_to_pmu(vcpu))) in load_vmcs12_host_state()
2424 if (data && !vcpu_to_pmu(vcpu)->version) in vmx_set_msr()7132 struct kvm_pmu *pmu = vcpu_to_pmu(&vmx->vcpu); in atomic_switch_perf_msrs()
202 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in pmu_ctr_read()265 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_incr_fw()282 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_read_hpm()316 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_num_ctrs()326 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_ctr_info()342 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_ctr_start()398 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_ctr_stop()465 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_ctr_cfg_match()548 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_init()613 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_deinit()
21 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_sbi_ext_pmu_handler()76 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_sbi_ext_pmu_probe()
438 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_handle_event()504 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_rdpmc()541 return kvm_pmu_has_perf_global_ctrl(vcpu_to_pmu(vcpu)); in kvm_pmu_is_valid_msr()551 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_mark_pmc_in_use()560 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_get_msr()585 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_set_msr()651 bitmap_zero(vcpu_to_pmu(vcpu)->all_valid_pmc_idx, X86_PMC_IDX_MAX); in kvm_pmu_refresh()662 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_init()674 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_cleanup()736 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_trigger_event()[all …]
7 #define vcpu_to_pmu(vcpu) (&(vcpu)->arch.pmu) macro
12293 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_arch_sched_in()
55 #define vcpu_to_pmu(vcpu) (&(vcpu)->arch.pmu_context) macro