Lines Matching refs:pmu
185 if (type == event->pmu->type) in armpmu_map_event()
202 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period()
244 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_update()
274 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_stop()
290 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_start()
315 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_del()
331 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_add()
364 validate_event(struct pmu *pmu, struct pmu_hw_events *hw_events, in validate_event() argument
377 if (event->pmu != pmu) in validate_event()
386 armpmu = to_arm_pmu(event->pmu); in validate_event()
402 if (!validate_event(event->pmu, &fake_pmu, leader)) in validate_group()
409 if (!validate_event(event->pmu, &fake_pmu, sibling)) in validate_group()
413 if (!validate_event(event->pmu, &fake_pmu, event)) in validate_group()
446 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in __hw_perf_event_init()
502 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_init()
522 static void armpmu_enable(struct pmu *pmu) in armpmu_enable() argument
524 struct arm_pmu *armpmu = to_arm_pmu(pmu); in armpmu_enable()
536 static void armpmu_disable(struct pmu *pmu) in armpmu_disable() argument
538 struct arm_pmu *armpmu = to_arm_pmu(pmu); in armpmu_disable()
552 static bool armpmu_filter(struct pmu *pmu, int cpu) in armpmu_filter() argument
554 struct arm_pmu *armpmu = to_arm_pmu(pmu); in armpmu_filter()
686 static int armpmu_get_cpu_irq(struct arm_pmu *pmu, int cpu) in armpmu_get_cpu_irq() argument
688 struct pmu_hw_events __percpu *hw_events = pmu->hw_events; in armpmu_get_cpu_irq()
705 struct arm_pmu *pmu = hlist_entry_safe(node, struct arm_pmu, node); in arm_perf_starting_cpu() local
708 if (!cpumask_test_cpu(cpu, &pmu->supported_cpus)) in arm_perf_starting_cpu()
710 if (pmu->reset) in arm_perf_starting_cpu()
711 pmu->reset(pmu); in arm_perf_starting_cpu()
713 per_cpu(cpu_armpmu, cpu) = pmu; in arm_perf_starting_cpu()
715 irq = armpmu_get_cpu_irq(pmu, cpu); in arm_perf_starting_cpu()
724 struct arm_pmu *pmu = hlist_entry_safe(node, struct arm_pmu, node); in arm_perf_teardown_cpu() local
727 if (!cpumask_test_cpu(cpu, &pmu->supported_cpus)) in arm_perf_teardown_cpu()
730 irq = armpmu_get_cpu_irq(pmu, cpu); in arm_perf_teardown_cpu()
854 struct arm_pmu *pmu; in armpmu_alloc() local
857 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in armpmu_alloc()
858 if (!pmu) in armpmu_alloc()
861 pmu->hw_events = alloc_percpu_gfp(struct pmu_hw_events, GFP_KERNEL); in armpmu_alloc()
862 if (!pmu->hw_events) { in armpmu_alloc()
867 pmu->pmu = (struct pmu) { in armpmu_alloc()
877 .attr_groups = pmu->attr_groups, in armpmu_alloc()
889 pmu->attr_groups[ARMPMU_ATTR_GROUP_COMMON] = in armpmu_alloc()
895 events = per_cpu_ptr(pmu->hw_events, cpu); in armpmu_alloc()
897 events->percpu_pmu = pmu; in armpmu_alloc()
900 return pmu; in armpmu_alloc()
903 kfree(pmu); in armpmu_alloc()
908 void armpmu_free(struct arm_pmu *pmu) in armpmu_free() argument
910 free_percpu(pmu->hw_events); in armpmu_free()
911 kfree(pmu); in armpmu_free()
914 int armpmu_register(struct arm_pmu *pmu) in armpmu_register() argument
918 ret = cpu_pmu_init(pmu); in armpmu_register()
922 if (!pmu->set_event_filter) in armpmu_register()
923 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_EXCLUDE; in armpmu_register()
925 ret = perf_pmu_register(&pmu->pmu, pmu->name, -1); in armpmu_register()
930 pmu->name, pmu->num_events, in armpmu_register()
933 kvm_host_pmu_init(pmu); in armpmu_register()
938 cpu_pmu_destroy(pmu); in armpmu_register()