Lines Matching refs:event

31 static inline u64 arm_pmu_event_max_period(struct perf_event *event)  in arm_pmu_event_max_period()  argument
33 if (event->hw.flags & ARMPMU_EVT_64BIT) in arm_pmu_event_max_period()
93 armpmu_map_event(struct perf_event *event, in armpmu_map_event() argument
101 u64 config = event->attr.config; in armpmu_map_event()
102 int type = event->attr.type; in armpmu_map_event()
104 if (type == event->pmu->type) in armpmu_map_event()
119 int armpmu_event_set_period(struct perf_event *event) in armpmu_event_set_period() argument
121 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period()
122 struct hw_perf_event *hwc = &event->hw; in armpmu_event_set_period()
128 max_period = arm_pmu_event_max_period(event); in armpmu_event_set_period()
154 armpmu->write_counter(event, (u64)(-left) & max_period); in armpmu_event_set_period()
156 perf_event_update_userpage(event); in armpmu_event_set_period()
161 u64 armpmu_event_update(struct perf_event *event) in armpmu_event_update() argument
163 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_update()
164 struct hw_perf_event *hwc = &event->hw; in armpmu_event_update()
166 u64 max_period = arm_pmu_event_max_period(event); in armpmu_event_update()
170 new_raw_count = armpmu->read_counter(event); in armpmu_event_update()
178 local64_add(delta, &event->count); in armpmu_event_update()
185 armpmu_read(struct perf_event *event) in armpmu_read() argument
187 armpmu_event_update(event); in armpmu_read()
191 armpmu_stop(struct perf_event *event, int flags) in armpmu_stop() argument
193 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_stop()
194 struct hw_perf_event *hwc = &event->hw; in armpmu_stop()
201 armpmu->disable(event); in armpmu_stop()
202 armpmu_event_update(event); in armpmu_stop()
207 static void armpmu_start(struct perf_event *event, int flags) in armpmu_start() argument
209 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_start()
210 struct hw_perf_event *hwc = &event->hw; in armpmu_start()
227 armpmu_event_set_period(event); in armpmu_start()
228 armpmu->enable(event); in armpmu_start()
232 armpmu_del(struct perf_event *event, int flags) in armpmu_del() argument
234 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_del()
236 struct hw_perf_event *hwc = &event->hw; in armpmu_del()
239 armpmu_stop(event, PERF_EF_UPDATE); in armpmu_del()
241 armpmu->clear_event_idx(hw_events, event); in armpmu_del()
242 perf_event_update_userpage(event); in armpmu_del()
248 armpmu_add(struct perf_event *event, int flags) in armpmu_add() argument
250 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_add()
252 struct hw_perf_event *hwc = &event->hw; in armpmu_add()
260 idx = armpmu->get_event_idx(hw_events, event); in armpmu_add()
268 event->hw.idx = idx; in armpmu_add()
269 armpmu->disable(event); in armpmu_add()
270 hw_events->events[idx] = event; in armpmu_add()
274 armpmu_start(event, PERF_EF_RELOAD); in armpmu_add()
277 perf_event_update_userpage(event); in armpmu_add()
284 struct perf_event *event) in validate_event() argument
288 if (is_software_event(event)) in validate_event()
296 if (event->pmu != pmu) in validate_event()
299 if (event->state < PERF_EVENT_STATE_OFF) in validate_event()
302 if (event->state == PERF_EVENT_STATE_OFF && !event->attr.enable_on_exec) in validate_event()
305 armpmu = to_arm_pmu(event->pmu); in validate_event()
306 return armpmu->get_event_idx(hw_events, event) >= 0; in validate_event()
310 validate_group(struct perf_event *event) in validate_group() argument
312 struct perf_event *sibling, *leader = event->group_leader; in validate_group()
321 if (!validate_event(event->pmu, &fake_pmu, leader)) in validate_group()
325 if (!validate_event(event->pmu, &fake_pmu, sibling)) in validate_group()
329 if (!validate_event(event->pmu, &fake_pmu, event)) in validate_group()
367 __hw_perf_event_init(struct perf_event *event) in __hw_perf_event_init() argument
369 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in __hw_perf_event_init()
370 struct hw_perf_event *hwc = &event->hw; in __hw_perf_event_init()
374 mapping = armpmu->map_event(event); in __hw_perf_event_init()
377 pr_debug("event %x:%llx not supported\n", event->attr.type, in __hw_perf_event_init()
378 event->attr.config); in __hw_perf_event_init()
397 armpmu->set_event_filter(hwc, &event->attr)) && in __hw_perf_event_init()
398 event_requires_mode_exclusion(&event->attr)) { in __hw_perf_event_init()
409 if (!is_sampling_event(event)) { in __hw_perf_event_init()
416 hwc->sample_period = arm_pmu_event_max_period(event) >> 1; in __hw_perf_event_init()
421 if (event->group_leader != event) { in __hw_perf_event_init()
422 if (validate_group(event) != 0) in __hw_perf_event_init()
429 static int armpmu_event_init(struct perf_event *event) in armpmu_event_init() argument
431 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_init()
440 if (event->cpu != -1 && in armpmu_event_init()
441 !cpumask_test_cpu(event->cpu, &armpmu->supported_cpus)) in armpmu_event_init()
445 if (has_branch_stack(event)) in armpmu_event_init()
448 if (armpmu->map_event(event) == -ENOENT) in armpmu_event_init()
451 return __hw_perf_event_init(event); in armpmu_event_init()
484 static int armpmu_filter_match(struct perf_event *event) in armpmu_filter_match() argument
486 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_filter_match()
492 return armpmu->filter_match(event); in armpmu_filter_match()
670 struct perf_event *event; in cpu_pm_pmu_setup() local
674 event = hw_events->events[idx]; in cpu_pm_pmu_setup()
675 if (!event) in cpu_pm_pmu_setup()
683 armpmu_stop(event, PERF_EF_UPDATE); in cpu_pm_pmu_setup()
699 RCU_NONIDLE(armpmu_start(event, PERF_EF_RELOAD)); in cpu_pm_pmu_setup()