Lines Matching refs:perf_event

127 	if (pmc->perf_event)  in kvm_pmu_get_pair_counter_value()
128 counter += perf_event_read_value(pmc->perf_event, &enabled, in kvm_pmu_get_pair_counter_value()
181 if (pmc->perf_event) { in kvm_pmu_release_perf_event()
182 perf_event_disable(pmc->perf_event); in kvm_pmu_release_perf_event()
183 perf_event_release_kernel(pmc->perf_event); in kvm_pmu_release_perf_event()
184 pmc->perf_event = NULL; in kvm_pmu_release_perf_event()
199 if (!pmc->perf_event) in kvm_pmu_stop_counter()
308 if (pmc->perf_event) { in kvm_pmu_enable_counter_mask()
309 perf_event_enable(pmc->perf_event); in kvm_pmu_enable_counter_mask()
310 if (pmc->perf_event->state != PERF_EVENT_STATE_ACTIVE) in kvm_pmu_enable_counter_mask()
349 if (pmc->perf_event) in kvm_pmu_disable_counter_mask()
350 perf_event_disable(pmc->perf_event); in kvm_pmu_disable_counter_mask()
441 static void kvm_pmu_perf_overflow(struct perf_event *perf_event, in kvm_pmu_perf_overflow() argument
445 struct kvm_pmc *pmc = perf_event->overflow_handler_context; in kvm_pmu_perf_overflow()
446 struct arm_pmu *cpu_pmu = to_arm_pmu(perf_event->pmu); in kvm_pmu_perf_overflow()
451 cpu_pmu->pmu.stop(perf_event, PERF_EF_UPDATE); in kvm_pmu_perf_overflow()
457 period = -(local64_read(&perf_event->count)); in kvm_pmu_perf_overflow()
462 local64_set(&perf_event->hw.period_left, 0); in kvm_pmu_perf_overflow()
463 perf_event->attr.sample_period = period; in kvm_pmu_perf_overflow()
464 perf_event->hw.sample_period = period; in kvm_pmu_perf_overflow()
473 cpu_pmu->pmu.start(perf_event, PERF_EF_RELOAD); in kvm_pmu_perf_overflow()
548 struct perf_event *event; in kvm_pmu_create_perf_event()
615 pmc->perf_event = event; in kvm_pmu_create_perf_event()