Lines Matching refs:evsel_list
102 static struct evlist *evsel_list; variable
104 .evlistp = &evsel_list,
238 evlist__reset_stats(evsel_list); in perf_stat__reset_stats()
318 int nthreads = perf_thread_map__nr(evsel_list->core.threads); in read_counter_cpu()
378 evlist__for_each_cpu(evlist_cpu_itr, evsel_list, affinity) { in read_affinity_counters()
400 evlist__for_each_entry(evsel_list, counter) { in read_bpf_map_counters()
425 evlist__for_each_entry(evsel_list, counter) { in process_counters()
433 perf_stat_merge_counters(&stat_config, evsel_list); in process_counters()
434 perf_stat_process_percore(&stat_config, evsel_list); in process_counters()
444 evlist__reset_aggr_stats(evsel_list); in process_interval()
474 evlist__for_each_entry(evsel_list, evsel) { in enable_counters()
485 evlist__enable(evsel_list); in enable_counters()
500 evlist__for_each_entry(evsel_list, counter) in disable_counters()
503 evlist__disable(evsel_list); in disable_counters()
607 child_exited = !is_target_alive(&target, evsel_list->core.threads) ? 1 : 0; in dispatch_events()
613 if (!(evlist__poll(evsel_list, time_to_sleep) > 0)) { /* poll timeout or EINTR */ in dispatch_events()
618 process_evlist(evsel_list, interval); in dispatch_events()
661 evsel_list->core.threads && in stat_handle_error()
662 evsel_list->core.threads->err_thread != -1) { in stat_handle_error()
667 if (!thread_map__remove(evsel_list->core.threads, in stat_handle_error()
668 evsel_list->core.threads->err_thread)) { in stat_handle_error()
669 evsel_list->core.threads->err_thread = -1; in stat_handle_error()
707 …if (evlist__prepare_workload(evsel_list, &target, argv, is_pipe, workload_exec_failed_signal) < 0)… in __run_perf_stat()
711 child_pid = evsel_list->workload.pid; in __run_perf_stat()
714 if (!cpu_map__is_dummy(evsel_list->core.user_requested_cpus)) { in __run_perf_stat()
720 evlist__for_each_entry(evsel_list, counter) { in __run_perf_stat()
728 evlist__reset_aggr_stats(evsel_list); in __run_perf_stat()
730 evlist__for_each_cpu(evlist_cpu_itr, evsel_list, affinity) { in __run_perf_stat()
758 evlist__reset_weak_group(evsel_list, counter, false); in __run_perf_stat()
786 evlist__for_each_cpu(evlist_cpu_itr, evsel_list, affinity) { in __run_perf_stat()
795 evlist__for_each_cpu(evlist_cpu_itr, evsel_list, affinity) { in __run_perf_stat()
821 evlist__for_each_entry(evsel_list, counter) { in __run_perf_stat()
832 evsel__store_ids(counter, evsel_list)) in __run_perf_stat()
836 if (evlist__apply_filters(evsel_list, &counter)) { in __run_perf_stat()
849 err = perf_session__write_header(perf_stat.session, evsel_list, in __run_perf_stat()
856 err = perf_event__synthesize_stat_events(&stat_config, NULL, evsel_list, in __run_perf_stat()
872 evlist__start_workload(evsel_list); in __run_perf_stat()
887 if (interval || timeout || evlist__ctlfd_initialized(evsel_list)) in __run_perf_stat()
920 evlist__copy_prev_raw_counts(evsel_list); in __run_perf_stat()
921 evlist__reset_prev_raw_counts(evsel_list); in __run_perf_stat()
922 evlist__reset_aggr_stats(evsel_list); in __run_perf_stat()
942 evlist__close(evsel_list); in __run_perf_stat()
981 evlist__print_counters(evsel_list, &stat_config, &target, ts, argc, argv); in print_counters()
1169 OPT_CALLBACK(0, "filter", &evsel_list, "filter",
1215 OPT_CALLBACK('G', "cgroup", &evsel_list, "name",
1265 OPT_CALLBACK('M', "metrics", &evsel_list, "metric/metric group list",
1284 OPT_CALLBACK(0, "cputype", &evsel_list, "hybrid cpu type",
1289 OPT_CALLBACK(0, "pfm-events", &evsel_list, "event",
1298 OPT_CALLBACK_OPTARG(0, "iostat", &evsel_list, &stat_config, "default",
1596 stat_config.aggr_map = cpu_aggr_map__new(evsel_list->core.user_requested_cpus, in perf_stat_init_aggr_mode()
1606 nr = perf_thread_map__nr(evsel_list->core.threads); in perf_stat_init_aggr_mode()
1625 if (evsel_list->core.user_requested_cpus) in perf_stat_init_aggr_mode()
1626 nr = perf_cpu_map__max(evsel_list->core.user_requested_cpus).cpu; in perf_stat_init_aggr_mode()
1894 int nr = perf_thread_map__nr(evsel_list->core.threads); in perf_stat_init_aggr_mode_file()
1912 stat_config.aggr_map = cpu_aggr_map__new(evsel_list->core.user_requested_cpus, in perf_stat_init_aggr_mode_file()
2058 return metricgroup__parse_groups(evsel_list, pmu, "transaction", in add_default_attributes()
2091 return metricgroup__parse_groups(evsel_list, pmu, "smi", in add_default_attributes()
2124 if (metricgroup__parse_groups(evsel_list, in add_default_attributes()
2138 if (!evsel_list->core.nr_entries) { in add_default_attributes()
2143 if (evlist__add_default_attrs(evsel_list, default_attrs0) < 0) in add_default_attributes()
2146 if (evlist__add_default_attrs(evsel_list, frontend_attrs) < 0) in add_default_attributes()
2150 if (evlist__add_default_attrs(evsel_list, backend_attrs) < 0) in add_default_attributes()
2153 if (evlist__add_default_attrs(evsel_list, default_attrs1) < 0) in add_default_attributes()
2179 evlist__splice_list_tail(evsel_list, &metric_evlist->core.entries); in add_default_attributes()
2184 if (evlist__add_default_attrs(evsel_list, default_null_attrs) < 0) in add_default_attributes()
2194 if (evlist__add_default_attrs(evsel_list, detailed_attrs) < 0) in add_default_attributes()
2201 if (evlist__add_default_attrs(evsel_list, very_detailed_attrs) < 0) in add_default_attributes()
2208 return evlist__add_default_attrs(evsel_list, very_very_detailed_attrs); in add_default_attributes()
2254 session->evlist = evsel_list; in __cmd_record()
2323 perf_evlist__set_maps(&evsel_list->core, st->cpus, st->threads); in set_maps()
2325 if (evlist__alloc_stats(&stat_config, evsel_list, /*alloc_raw=*/true)) in set_maps()
2433 evlist__delete(evsel_list); in __cmd_report()
2434 evsel_list = session->evlist; in __cmd_report()
2463 evlist__for_each_entry(evsel_list, counter) { in setup_system_wide()
2470 if (evsel_list->core.nr_entries) in setup_system_wide()
2490 evsel_list = evlist__new(); in cmd_stat()
2491 if (evsel_list == NULL) in cmd_stat()
2668 status = iostat_prepare(evsel_list, &stat_config); in cmd_stat()
2672 iostat_list(evsel_list, &stat_config); in cmd_stat()
2675 iostat_list(evsel_list, &stat_config); in cmd_stat()
2699 metricgroup__parse_groups(evsel_list, pmu, metrics, in cmd_stat()
2720 if (evlist__expand_cgroup(evsel_list, stat_config.cgroup_list, in cmd_stat()
2728 evlist__warn_user_requested_cpus(evsel_list, target.cpu_list); in cmd_stat()
2730 if (evlist__create_maps(evsel_list, &target) < 0) { in cmd_stat()
2743 evlist__check_cpu_maps(evsel_list); in cmd_stat()
2750 thread_map__read_comms(evsel_list->core.threads); in cmd_stat()
2786 if (evlist__alloc_stats(&stat_config, evsel_list, interval)) in cmd_stat()
2813 if (evlist__initialize_ctlfd(evsel_list, stat_config.ctl_fd, stat_config.ctl_fd_ack)) in cmd_stat()
2817 evlist__first(evsel_list)->ignore_missing_thread = target.pid; in cmd_stat()
2825 evlist__reset_prev_raw_counts(evsel_list); in cmd_stat()
2836 evlist__copy_res_stats(&stat_config, evsel_list); in cmd_stat()
2840 evlist__finalize_ctlfd(evsel_list); in cmd_stat()
2872 perf_session__write_header(perf_stat.session, evsel_list, fd, true); in cmd_stat()
2875 evlist__close(evsel_list); in cmd_stat()
2880 evlist__free_stats(evsel_list); in cmd_stat()
2883 iostat_release(evsel_list); in cmd_stat()
2891 evlist__delete(evsel_list); in cmd_stat()