1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (c) 2017, Intel Corporation.
4  */
5 
6 /* Manage metrics and groups of metrics from JSON files */
7 
8 #include "metricgroup.h"
9 #include "debug.h"
10 #include "evlist.h"
11 #include "evsel.h"
12 #include "strbuf.h"
13 #include "pmu.h"
14 #include "expr.h"
15 #include "rblist.h"
16 #include <string.h>
17 #include <errno.h>
18 #include "strlist.h"
19 #include <assert.h>
20 #include <linux/ctype.h>
21 #include <linux/string.h>
22 #include <linux/zalloc.h>
23 #include <subcmd/parse-options.h>
24 #include <api/fs/fs.h>
25 #include "util.h"
26 #include <asm/bug.h>
27 #include "cgroup.h"
28 
metricgroup__lookup(struct rblist * metric_events,struct evsel * evsel,bool create)29 struct metric_event *metricgroup__lookup(struct rblist *metric_events,
30 					 struct evsel *evsel,
31 					 bool create)
32 {
33 	struct rb_node *nd;
34 	struct metric_event me = {
35 		.evsel = evsel
36 	};
37 
38 	if (!metric_events)
39 		return NULL;
40 
41 	nd = rblist__find(metric_events, &me);
42 	if (nd)
43 		return container_of(nd, struct metric_event, nd);
44 	if (create) {
45 		rblist__add_node(metric_events, &me);
46 		nd = rblist__find(metric_events, &me);
47 		if (nd)
48 			return container_of(nd, struct metric_event, nd);
49 	}
50 	return NULL;
51 }
52 
metric_event_cmp(struct rb_node * rb_node,const void * entry)53 static int metric_event_cmp(struct rb_node *rb_node, const void *entry)
54 {
55 	struct metric_event *a = container_of(rb_node,
56 					      struct metric_event,
57 					      nd);
58 	const struct metric_event *b = entry;
59 
60 	if (a->evsel == b->evsel)
61 		return 0;
62 	if ((char *)a->evsel < (char *)b->evsel)
63 		return -1;
64 	return +1;
65 }
66 
metric_event_new(struct rblist * rblist __maybe_unused,const void * entry)67 static struct rb_node *metric_event_new(struct rblist *rblist __maybe_unused,
68 					const void *entry)
69 {
70 	struct metric_event *me = malloc(sizeof(struct metric_event));
71 
72 	if (!me)
73 		return NULL;
74 	memcpy(me, entry, sizeof(struct metric_event));
75 	me->evsel = ((struct metric_event *)entry)->evsel;
76 	INIT_LIST_HEAD(&me->head);
77 	return &me->nd;
78 }
79 
metric_event_delete(struct rblist * rblist __maybe_unused,struct rb_node * rb_node)80 static void metric_event_delete(struct rblist *rblist __maybe_unused,
81 				struct rb_node *rb_node)
82 {
83 	struct metric_event *me = container_of(rb_node, struct metric_event, nd);
84 	struct metric_expr *expr, *tmp;
85 
86 	list_for_each_entry_safe(expr, tmp, &me->head, nd) {
87 		free(expr->metric_refs);
88 		free(expr->metric_events);
89 		free(expr);
90 	}
91 
92 	free(me);
93 }
94 
metricgroup__rblist_init(struct rblist * metric_events)95 static void metricgroup__rblist_init(struct rblist *metric_events)
96 {
97 	rblist__init(metric_events);
98 	metric_events->node_cmp = metric_event_cmp;
99 	metric_events->node_new = metric_event_new;
100 	metric_events->node_delete = metric_event_delete;
101 }
102 
metricgroup__rblist_exit(struct rblist * metric_events)103 void metricgroup__rblist_exit(struct rblist *metric_events)
104 {
105 	rblist__exit(metric_events);
106 }
107 
108 /*
109  * A node in the list of referenced metrics. metric_expr
110  * is held as a convenience to avoid a search through the
111  * metric list.
112  */
113 struct metric_ref_node {
114 	const char *metric_name;
115 	const char *metric_expr;
116 	struct list_head list;
117 };
118 
119 struct metric {
120 	struct list_head nd;
121 	struct expr_parse_ctx pctx;
122 	const char *metric_name;
123 	const char *metric_expr;
124 	const char *metric_unit;
125 	struct list_head metric_refs;
126 	int metric_refs_cnt;
127 	int runtime;
128 	bool has_constraint;
129 };
130 
131 #define RECURSION_ID_MAX 1000
132 
133 struct expr_ids {
134 	struct expr_id	id[RECURSION_ID_MAX];
135 	int		cnt;
136 };
137 
expr_ids__alloc(struct expr_ids * ids)138 static struct expr_id *expr_ids__alloc(struct expr_ids *ids)
139 {
140 	if (ids->cnt >= RECURSION_ID_MAX)
141 		return NULL;
142 	return &ids->id[ids->cnt++];
143 }
144 
expr_ids__exit(struct expr_ids * ids)145 static void expr_ids__exit(struct expr_ids *ids)
146 {
147 	int i;
148 
149 	for (i = 0; i < ids->cnt; i++)
150 		free(ids->id[i].id);
151 }
152 
contains_event(struct evsel ** metric_events,int num_events,const char * event_name)153 static bool contains_event(struct evsel **metric_events, int num_events,
154 			const char *event_name)
155 {
156 	int i;
157 
158 	for (i = 0; i < num_events; i++) {
159 		if (!strcmp(metric_events[i]->name, event_name))
160 			return true;
161 	}
162 	return false;
163 }
164 
165 /**
166  * Find a group of events in perf_evlist that correspond to those from a parsed
167  * metric expression. Note, as find_evsel_group is called in the same order as
168  * perf_evlist was constructed, metric_no_merge doesn't need to test for
169  * underfilling a group.
170  * @perf_evlist: a list of events something like: {metric1 leader, metric1
171  * sibling, metric1 sibling}:W,duration_time,{metric2 leader, metric2 sibling,
172  * metric2 sibling}:W,duration_time
173  * @pctx: the parse context for the metric expression.
174  * @metric_no_merge: don't attempt to share events for the metric with other
175  * metrics.
176  * @has_constraint: is there a contraint on the group of events? In which case
177  * the events won't be grouped.
178  * @metric_events: out argument, null terminated array of evsel's associated
179  * with the metric.
180  * @evlist_used: in/out argument, bitmap tracking which evlist events are used.
181  * @return the first metric event or NULL on failure.
182  */
find_evsel_group(struct evlist * perf_evlist,struct expr_parse_ctx * pctx,bool metric_no_merge,bool has_constraint,struct evsel ** metric_events,unsigned long * evlist_used)183 static struct evsel *find_evsel_group(struct evlist *perf_evlist,
184 				      struct expr_parse_ctx *pctx,
185 				      bool metric_no_merge,
186 				      bool has_constraint,
187 				      struct evsel **metric_events,
188 				      unsigned long *evlist_used)
189 {
190 	struct evsel *ev, *current_leader = NULL;
191 	struct expr_id_data *val_ptr;
192 	int i = 0, matched_events = 0, events_to_match;
193 	const int idnum = (int)hashmap__size(&pctx->ids);
194 
195 	/*
196 	 * duration_time is always grouped separately, when events are grouped
197 	 * (ie has_constraint is false) then ignore it in the matching loop and
198 	 * add it to metric_events at the end.
199 	 */
200 	if (!has_constraint &&
201 	    hashmap__find(&pctx->ids, "duration_time", (void **)&val_ptr))
202 		events_to_match = idnum - 1;
203 	else
204 		events_to_match = idnum;
205 
206 	evlist__for_each_entry (perf_evlist, ev) {
207 		/*
208 		 * Events with a constraint aren't grouped and match the first
209 		 * events available.
210 		 */
211 		if (has_constraint && ev->weak_group)
212 			continue;
213 		/* Ignore event if already used and merging is disabled. */
214 		if (metric_no_merge && test_bit(ev->idx, evlist_used))
215 			continue;
216 		if (!has_constraint && ev->leader != current_leader) {
217 			/*
218 			 * Start of a new group, discard the whole match and
219 			 * start again.
220 			 */
221 			matched_events = 0;
222 			memset(metric_events, 0,
223 				sizeof(struct evsel *) * idnum);
224 			current_leader = ev->leader;
225 		}
226 		/*
227 		 * Check for duplicate events with the same name. For example,
228 		 * uncore_imc/cas_count_read/ will turn into 6 events per socket
229 		 * on skylakex. Only the first such event is placed in
230 		 * metric_events. If events aren't grouped then this also
231 		 * ensures that the same event in different sibling groups
232 		 * aren't both added to metric_events.
233 		 */
234 		if (contains_event(metric_events, matched_events, ev->name))
235 			continue;
236 		/* Does this event belong to the parse context? */
237 		if (hashmap__find(&pctx->ids, ev->name, (void **)&val_ptr))
238 			metric_events[matched_events++] = ev;
239 
240 		if (matched_events == events_to_match)
241 			break;
242 	}
243 
244 	if (events_to_match != idnum) {
245 		/* Add the first duration_time. */
246 		evlist__for_each_entry(perf_evlist, ev) {
247 			if (!strcmp(ev->name, "duration_time")) {
248 				metric_events[matched_events++] = ev;
249 				break;
250 			}
251 		}
252 	}
253 
254 	if (matched_events != idnum) {
255 		/* Not a whole match */
256 		return NULL;
257 	}
258 
259 	metric_events[idnum] = NULL;
260 
261 	for (i = 0; i < idnum; i++) {
262 		ev = metric_events[i];
263 		/* Don't free the used events. */
264 		set_bit(ev->idx, evlist_used);
265 		/*
266 		 * The metric leader points to the identically named event in
267 		 * metric_events.
268 		 */
269 		ev->metric_leader = ev;
270 		/*
271 		 * Mark two events with identical names in the same group (or
272 		 * globally) as being in use as uncore events may be duplicated
273 		 * for each pmu. Set the metric leader of such events to be the
274 		 * event that appears in metric_events.
275 		 */
276 		evlist__for_each_entry_continue(perf_evlist, ev) {
277 			/*
278 			 * If events are grouped then the search can terminate
279 			 * when then group is left.
280 			 */
281 			if (!has_constraint &&
282 			    ev->leader != metric_events[i]->leader)
283 				break;
284 			if (!strcmp(metric_events[i]->name, ev->name)) {
285 				set_bit(ev->idx, evlist_used);
286 				ev->metric_leader = metric_events[i];
287 			}
288 		}
289 	}
290 
291 	return metric_events[0];
292 }
293 
metricgroup__setup_events(struct list_head * groups,bool metric_no_merge,struct evlist * perf_evlist,struct rblist * metric_events_list)294 static int metricgroup__setup_events(struct list_head *groups,
295 				     bool metric_no_merge,
296 				     struct evlist *perf_evlist,
297 				     struct rblist *metric_events_list)
298 {
299 	struct metric_event *me;
300 	struct metric_expr *expr;
301 	int i = 0;
302 	int ret = 0;
303 	struct metric *m;
304 	struct evsel *evsel, *tmp;
305 	unsigned long *evlist_used;
306 
307 	evlist_used = bitmap_alloc(perf_evlist->core.nr_entries);
308 	if (!evlist_used)
309 		return -ENOMEM;
310 
311 	list_for_each_entry (m, groups, nd) {
312 		struct evsel **metric_events;
313 		struct metric_ref *metric_refs = NULL;
314 
315 		metric_events = calloc(sizeof(void *),
316 				hashmap__size(&m->pctx.ids) + 1);
317 		if (!metric_events) {
318 			ret = -ENOMEM;
319 			break;
320 		}
321 		evsel = find_evsel_group(perf_evlist, &m->pctx,
322 					 metric_no_merge,
323 					 m->has_constraint, metric_events,
324 					 evlist_used);
325 		if (!evsel) {
326 			pr_debug("Cannot resolve %s: %s\n",
327 					m->metric_name, m->metric_expr);
328 			free(metric_events);
329 			continue;
330 		}
331 		for (i = 0; metric_events[i]; i++)
332 			metric_events[i]->collect_stat = true;
333 		me = metricgroup__lookup(metric_events_list, evsel, true);
334 		if (!me) {
335 			ret = -ENOMEM;
336 			free(metric_events);
337 			break;
338 		}
339 		expr = malloc(sizeof(struct metric_expr));
340 		if (!expr) {
341 			ret = -ENOMEM;
342 			free(metric_events);
343 			break;
344 		}
345 
346 		/*
347 		 * Collect and store collected nested expressions
348 		 * for metric processing.
349 		 */
350 		if (m->metric_refs_cnt) {
351 			struct metric_ref_node *ref;
352 
353 			metric_refs = zalloc(sizeof(struct metric_ref) * (m->metric_refs_cnt + 1));
354 			if (!metric_refs) {
355 				ret = -ENOMEM;
356 				free(metric_events);
357 				free(expr);
358 				break;
359 			}
360 
361 			i = 0;
362 			list_for_each_entry(ref, &m->metric_refs, list) {
363 				/*
364 				 * Intentionally passing just const char pointers,
365 				 * originally from 'struct pmu_event' object.
366 				 * We don't need to change them, so there's no
367 				 * need to create our own copy.
368 				 */
369 				metric_refs[i].metric_name = ref->metric_name;
370 				metric_refs[i].metric_expr = ref->metric_expr;
371 				i++;
372 			}
373 		};
374 
375 		expr->metric_refs = metric_refs;
376 		expr->metric_expr = m->metric_expr;
377 		expr->metric_name = m->metric_name;
378 		expr->metric_unit = m->metric_unit;
379 		expr->metric_events = metric_events;
380 		expr->runtime = m->runtime;
381 		list_add(&expr->nd, &me->head);
382 	}
383 
384 	evlist__for_each_entry_safe(perf_evlist, tmp, evsel) {
385 		if (!test_bit(evsel->idx, evlist_used)) {
386 			evlist__remove(perf_evlist, evsel);
387 			evsel__delete(evsel);
388 		}
389 	}
390 	bitmap_free(evlist_used);
391 
392 	return ret;
393 }
394 
match_metric(const char * n,const char * list)395 static bool match_metric(const char *n, const char *list)
396 {
397 	int len;
398 	char *m;
399 
400 	if (!list)
401 		return false;
402 	if (!strcmp(list, "all"))
403 		return true;
404 	if (!n)
405 		return !strcasecmp(list, "No_group");
406 	len = strlen(list);
407 	m = strcasestr(n, list);
408 	if (!m)
409 		return false;
410 	if ((m == n || m[-1] == ';' || m[-1] == ' ') &&
411 	    (m[len] == 0 || m[len] == ';'))
412 		return true;
413 	return false;
414 }
415 
416 struct mep {
417 	struct rb_node nd;
418 	const char *name;
419 	struct strlist *metrics;
420 };
421 
mep_cmp(struct rb_node * rb_node,const void * entry)422 static int mep_cmp(struct rb_node *rb_node, const void *entry)
423 {
424 	struct mep *a = container_of(rb_node, struct mep, nd);
425 	struct mep *b = (struct mep *)entry;
426 
427 	return strcmp(a->name, b->name);
428 }
429 
mep_new(struct rblist * rl __maybe_unused,const void * entry)430 static struct rb_node *mep_new(struct rblist *rl __maybe_unused,
431 					const void *entry)
432 {
433 	struct mep *me = malloc(sizeof(struct mep));
434 
435 	if (!me)
436 		return NULL;
437 	memcpy(me, entry, sizeof(struct mep));
438 	me->name = strdup(me->name);
439 	if (!me->name)
440 		goto out_me;
441 	me->metrics = strlist__new(NULL, NULL);
442 	if (!me->metrics)
443 		goto out_name;
444 	return &me->nd;
445 out_name:
446 	zfree(&me->name);
447 out_me:
448 	free(me);
449 	return NULL;
450 }
451 
mep_lookup(struct rblist * groups,const char * name)452 static struct mep *mep_lookup(struct rblist *groups, const char *name)
453 {
454 	struct rb_node *nd;
455 	struct mep me = {
456 		.name = name
457 	};
458 	nd = rblist__find(groups, &me);
459 	if (nd)
460 		return container_of(nd, struct mep, nd);
461 	rblist__add_node(groups, &me);
462 	nd = rblist__find(groups, &me);
463 	if (nd)
464 		return container_of(nd, struct mep, nd);
465 	return NULL;
466 }
467 
mep_delete(struct rblist * rl __maybe_unused,struct rb_node * nd)468 static void mep_delete(struct rblist *rl __maybe_unused,
469 		       struct rb_node *nd)
470 {
471 	struct mep *me = container_of(nd, struct mep, nd);
472 
473 	strlist__delete(me->metrics);
474 	zfree(&me->name);
475 	free(me);
476 }
477 
metricgroup__print_strlist(struct strlist * metrics,bool raw)478 static void metricgroup__print_strlist(struct strlist *metrics, bool raw)
479 {
480 	struct str_node *sn;
481 	int n = 0;
482 
483 	strlist__for_each_entry (sn, metrics) {
484 		if (raw)
485 			printf("%s%s", n > 0 ? " " : "", sn->s);
486 		else
487 			printf("  %s\n", sn->s);
488 		n++;
489 	}
490 	if (raw)
491 		putchar('\n');
492 }
493 
metricgroup__print(bool metrics,bool metricgroups,char * filter,bool raw,bool details)494 void metricgroup__print(bool metrics, bool metricgroups, char *filter,
495 			bool raw, bool details)
496 {
497 	struct pmu_events_map *map = perf_pmu__find_map(NULL);
498 	struct pmu_event *pe;
499 	int i;
500 	struct rblist groups;
501 	struct rb_node *node, *next;
502 	struct strlist *metriclist = NULL;
503 
504 	if (!map)
505 		return;
506 
507 	if (!metricgroups) {
508 		metriclist = strlist__new(NULL, NULL);
509 		if (!metriclist)
510 			return;
511 	}
512 
513 	rblist__init(&groups);
514 	groups.node_new = mep_new;
515 	groups.node_cmp = mep_cmp;
516 	groups.node_delete = mep_delete;
517 	for (i = 0; ; i++) {
518 		const char *g;
519 		pe = &map->table[i];
520 
521 		if (!pe->name && !pe->metric_group && !pe->metric_name)
522 			break;
523 		if (!pe->metric_expr)
524 			continue;
525 		g = pe->metric_group;
526 		if (!g && pe->metric_name) {
527 			if (pe->name)
528 				continue;
529 			g = "No_group";
530 		}
531 		if (g) {
532 			char *omg;
533 			char *mg = strdup(g);
534 
535 			if (!mg)
536 				return;
537 			omg = mg;
538 			while ((g = strsep(&mg, ";")) != NULL) {
539 				struct mep *me;
540 				char *s;
541 
542 				g = skip_spaces(g);
543 				if (*g == 0)
544 					g = "No_group";
545 				if (filter && !strstr(g, filter))
546 					continue;
547 				if (raw)
548 					s = (char *)pe->metric_name;
549 				else {
550 					if (asprintf(&s, "%s\n%*s%s]",
551 						     pe->metric_name, 8, "[", pe->desc) < 0)
552 						return;
553 
554 					if (details) {
555 						if (asprintf(&s, "%s\n%*s%s]",
556 							     s, 8, "[", pe->metric_expr) < 0)
557 							return;
558 					}
559 				}
560 
561 				if (!s)
562 					continue;
563 
564 				if (!metricgroups) {
565 					strlist__add(metriclist, s);
566 				} else {
567 					me = mep_lookup(&groups, g);
568 					if (!me)
569 						continue;
570 					strlist__add(me->metrics, s);
571 				}
572 
573 				if (!raw)
574 					free(s);
575 			}
576 			free(omg);
577 		}
578 	}
579 
580 	if (!filter || !rblist__empty(&groups)) {
581 		if (metricgroups && !raw)
582 			printf("\nMetric Groups:\n\n");
583 		else if (metrics && !raw)
584 			printf("\nMetrics:\n\n");
585 	}
586 
587 	for (node = rb_first_cached(&groups.entries); node; node = next) {
588 		struct mep *me = container_of(node, struct mep, nd);
589 
590 		if (metricgroups)
591 			printf("%s%s%s", me->name, metrics && !raw ? ":" : "", raw ? " " : "\n");
592 		if (metrics)
593 			metricgroup__print_strlist(me->metrics, raw);
594 		next = rb_next(node);
595 		rblist__remove_node(&groups, node);
596 	}
597 	if (!metricgroups)
598 		metricgroup__print_strlist(metriclist, raw);
599 	strlist__delete(metriclist);
600 }
601 
metricgroup__add_metric_weak_group(struct strbuf * events,struct expr_parse_ctx * ctx)602 static void metricgroup__add_metric_weak_group(struct strbuf *events,
603 					       struct expr_parse_ctx *ctx)
604 {
605 	struct hashmap_entry *cur;
606 	size_t bkt;
607 	bool no_group = true, has_duration = false;
608 
609 	hashmap__for_each_entry((&ctx->ids), cur, bkt) {
610 		pr_debug("found event %s\n", (const char *)cur->key);
611 		/*
612 		 * Duration time maps to a software event and can make
613 		 * groups not count. Always use it outside a
614 		 * group.
615 		 */
616 		if (!strcmp(cur->key, "duration_time")) {
617 			has_duration = true;
618 			continue;
619 		}
620 		strbuf_addf(events, "%s%s",
621 			no_group ? "{" : ",",
622 			(const char *)cur->key);
623 		no_group = false;
624 	}
625 	if (!no_group) {
626 		strbuf_addf(events, "}:W");
627 		if (has_duration)
628 			strbuf_addf(events, ",duration_time");
629 	} else if (has_duration)
630 		strbuf_addf(events, "duration_time");
631 }
632 
metricgroup__add_metric_non_group(struct strbuf * events,struct expr_parse_ctx * ctx)633 static void metricgroup__add_metric_non_group(struct strbuf *events,
634 					      struct expr_parse_ctx *ctx)
635 {
636 	struct hashmap_entry *cur;
637 	size_t bkt;
638 	bool first = true;
639 
640 	hashmap__for_each_entry((&ctx->ids), cur, bkt) {
641 		if (!first)
642 			strbuf_addf(events, ",");
643 		strbuf_addf(events, "%s", (const char *)cur->key);
644 		first = false;
645 	}
646 }
647 
metricgroup___watchdog_constraint_hint(const char * name,bool foot)648 static void metricgroup___watchdog_constraint_hint(const char *name, bool foot)
649 {
650 	static bool violate_nmi_constraint;
651 
652 	if (!foot) {
653 		pr_warning("Splitting metric group %s into standalone metrics.\n", name);
654 		violate_nmi_constraint = true;
655 		return;
656 	}
657 
658 	if (!violate_nmi_constraint)
659 		return;
660 
661 	pr_warning("Try disabling the NMI watchdog to comply NO_NMI_WATCHDOG metric constraint:\n"
662 		   "    echo 0 > /proc/sys/kernel/nmi_watchdog\n"
663 		   "    perf stat ...\n"
664 		   "    echo 1 > /proc/sys/kernel/nmi_watchdog\n");
665 }
666 
metricgroup__has_constraint(struct pmu_event * pe)667 static bool metricgroup__has_constraint(struct pmu_event *pe)
668 {
669 	if (!pe->metric_constraint)
670 		return false;
671 
672 	if (!strcmp(pe->metric_constraint, "NO_NMI_WATCHDOG") &&
673 	    sysctl__nmi_watchdog_enabled()) {
674 		metricgroup___watchdog_constraint_hint(pe->metric_name, false);
675 		return true;
676 	}
677 
678 	return false;
679 }
680 
arch_get_runtimeparam(struct pmu_event * pe __maybe_unused)681 int __weak arch_get_runtimeparam(struct pmu_event *pe __maybe_unused)
682 {
683 	return 1;
684 }
685 
__add_metric(struct list_head * metric_list,struct pmu_event * pe,bool metric_no_group,int runtime,struct metric ** mp,struct expr_id * parent,struct expr_ids * ids)686 static int __add_metric(struct list_head *metric_list,
687 			struct pmu_event *pe,
688 			bool metric_no_group,
689 			int runtime,
690 			struct metric **mp,
691 			struct expr_id *parent,
692 			struct expr_ids *ids)
693 {
694 	struct metric_ref_node *ref;
695 	struct metric *m;
696 
697 	if (*mp == NULL) {
698 		/*
699 		 * We got in here for the parent group,
700 		 * allocate it and put it on the list.
701 		 */
702 		m = zalloc(sizeof(*m));
703 		if (!m)
704 			return -ENOMEM;
705 
706 		expr__ctx_init(&m->pctx);
707 		m->metric_name = pe->metric_name;
708 		m->metric_expr = pe->metric_expr;
709 		m->metric_unit = pe->unit;
710 		m->runtime = runtime;
711 		m->has_constraint = metric_no_group || metricgroup__has_constraint(pe);
712 		INIT_LIST_HEAD(&m->metric_refs);
713 		m->metric_refs_cnt = 0;
714 
715 		parent = expr_ids__alloc(ids);
716 		if (!parent) {
717 			free(m);
718 			return -EINVAL;
719 		}
720 
721 		parent->id = strdup(pe->metric_name);
722 		if (!parent->id) {
723 			free(m);
724 			return -ENOMEM;
725 		}
726 		*mp = m;
727 	} else {
728 		/*
729 		 * We got here for the referenced metric, via the
730 		 * recursive metricgroup__add_metric call, add
731 		 * it to the parent group.
732 		 */
733 		m = *mp;
734 
735 		ref = malloc(sizeof(*ref));
736 		if (!ref)
737 			return -ENOMEM;
738 
739 		/*
740 		 * Intentionally passing just const char pointers,
741 		 * from 'pe' object, so they never go away. We don't
742 		 * need to change them, so there's no need to create
743 		 * our own copy.
744 		 */
745 		ref->metric_name = pe->metric_name;
746 		ref->metric_expr = pe->metric_expr;
747 
748 		list_add(&ref->list, &m->metric_refs);
749 		m->metric_refs_cnt++;
750 	}
751 
752 	/* Force all found IDs in metric to have us as parent ID. */
753 	WARN_ON_ONCE(!parent);
754 	m->pctx.parent = parent;
755 
756 	/*
757 	 * For both the parent and referenced metrics, we parse
758 	 * all the metric's IDs and add it to the parent context.
759 	 */
760 	if (expr__find_other(pe->metric_expr, NULL, &m->pctx, runtime) < 0) {
761 		if (m->metric_refs_cnt == 0) {
762 			expr__ctx_clear(&m->pctx);
763 			free(m);
764 			*mp = NULL;
765 		}
766 		return -EINVAL;
767 	}
768 
769 	/*
770 	 * We add new group only in the 'parent' call,
771 	 * so bail out for referenced metric case.
772 	 */
773 	if (m->metric_refs_cnt)
774 		return 0;
775 
776 	if (list_empty(metric_list))
777 		list_add(&m->nd, metric_list);
778 	else {
779 		struct list_head *pos;
780 
781 		/* Place the largest groups at the front. */
782 		list_for_each_prev(pos, metric_list) {
783 			struct metric *old = list_entry(pos, struct metric, nd);
784 
785 			if (hashmap__size(&m->pctx.ids) <=
786 			    hashmap__size(&old->pctx.ids))
787 				break;
788 		}
789 		list_add(&m->nd, pos);
790 	}
791 
792 	return 0;
793 }
794 
795 #define map_for_each_event(__pe, __idx, __map)				\
796 	for (__idx = 0, __pe = &__map->table[__idx];			\
797 	     __pe->name || __pe->metric_group || __pe->metric_name;	\
798 	     __pe = &__map->table[++__idx])
799 
800 #define map_for_each_metric(__pe, __idx, __map, __metric)		\
801 	map_for_each_event(__pe, __idx, __map)				\
802 		if (__pe->metric_expr &&				\
803 		    (match_metric(__pe->metric_group, __metric) ||	\
804 		     match_metric(__pe->metric_name, __metric)))
805 
find_metric(const char * metric,struct pmu_events_map * map)806 static struct pmu_event *find_metric(const char *metric, struct pmu_events_map *map)
807 {
808 	struct pmu_event *pe;
809 	int i;
810 
811 	map_for_each_event(pe, i, map) {
812 		if (match_metric(pe->metric_name, metric))
813 			return pe;
814 	}
815 
816 	return NULL;
817 }
818 
recursion_check(struct metric * m,const char * id,struct expr_id ** parent,struct expr_ids * ids)819 static int recursion_check(struct metric *m, const char *id, struct expr_id **parent,
820 			   struct expr_ids *ids)
821 {
822 	struct expr_id_data *data;
823 	struct expr_id *p;
824 	int ret;
825 
826 	/*
827 	 * We get the parent referenced by 'id' argument and
828 	 * traverse through all the parent object IDs to check
829 	 * if we already processed 'id', if we did, it's recursion
830 	 * and we fail.
831 	 */
832 	ret = expr__get_id(&m->pctx, id, &data);
833 	if (ret)
834 		return ret;
835 
836 	p = data->parent;
837 
838 	while (p->parent) {
839 		if (!strcmp(p->id, id)) {
840 			pr_err("failed: recursion detected for %s\n", id);
841 			return -1;
842 		}
843 		p = p->parent;
844 	}
845 
846 	/*
847 	 * If we are over the limit of static entris, the metric
848 	 * is too difficult/nested to process, fail as well.
849 	 */
850 	p = expr_ids__alloc(ids);
851 	if (!p) {
852 		pr_err("failed: too many nested metrics\n");
853 		return -EINVAL;
854 	}
855 
856 	p->id     = strdup(id);
857 	p->parent = data->parent;
858 	*parent   = p;
859 
860 	return p->id ? 0 : -ENOMEM;
861 }
862 
863 static int add_metric(struct list_head *metric_list,
864 		      struct pmu_event *pe,
865 		      bool metric_no_group,
866 		      struct metric **mp,
867 		      struct expr_id *parent,
868 		      struct expr_ids *ids);
869 
__resolve_metric(struct metric * m,bool metric_no_group,struct list_head * metric_list,struct pmu_events_map * map,struct expr_ids * ids)870 static int __resolve_metric(struct metric *m,
871 			    bool metric_no_group,
872 			    struct list_head *metric_list,
873 			    struct pmu_events_map *map,
874 			    struct expr_ids *ids)
875 {
876 	struct hashmap_entry *cur;
877 	size_t bkt;
878 	bool all;
879 	int ret;
880 
881 	/*
882 	 * Iterate all the parsed IDs and if there's metric,
883 	 * add it to the context.
884 	 */
885 	do {
886 		all = true;
887 		hashmap__for_each_entry((&m->pctx.ids), cur, bkt) {
888 			struct expr_id *parent;
889 			struct pmu_event *pe;
890 
891 			pe = find_metric(cur->key, map);
892 			if (!pe)
893 				continue;
894 
895 			ret = recursion_check(m, cur->key, &parent, ids);
896 			if (ret)
897 				return ret;
898 
899 			all = false;
900 			/* The metric key itself needs to go out.. */
901 			expr__del_id(&m->pctx, cur->key);
902 
903 			/* ... and it gets resolved to the parent context. */
904 			ret = add_metric(metric_list, pe, metric_no_group, &m, parent, ids);
905 			if (ret)
906 				return ret;
907 
908 			/*
909 			 * We added new metric to hashmap, so we need
910 			 * to break the iteration and start over.
911 			 */
912 			break;
913 		}
914 	} while (!all);
915 
916 	return 0;
917 }
918 
resolve_metric(bool metric_no_group,struct list_head * metric_list,struct pmu_events_map * map,struct expr_ids * ids)919 static int resolve_metric(bool metric_no_group,
920 			  struct list_head *metric_list,
921 			  struct pmu_events_map *map,
922 			  struct expr_ids *ids)
923 {
924 	struct metric *m;
925 	int err;
926 
927 	list_for_each_entry(m, metric_list, nd) {
928 		err = __resolve_metric(m, metric_no_group, metric_list, map, ids);
929 		if (err)
930 			return err;
931 	}
932 	return 0;
933 }
934 
add_metric(struct list_head * metric_list,struct pmu_event * pe,bool metric_no_group,struct metric ** m,struct expr_id * parent,struct expr_ids * ids)935 static int add_metric(struct list_head *metric_list,
936 		      struct pmu_event *pe,
937 		      bool metric_no_group,
938 		      struct metric **m,
939 		      struct expr_id *parent,
940 		      struct expr_ids *ids)
941 {
942 	struct metric *orig = *m;
943 	int ret = 0;
944 
945 	pr_debug("metric expr %s for %s\n", pe->metric_expr, pe->metric_name);
946 
947 	if (!strstr(pe->metric_expr, "?")) {
948 		ret = __add_metric(metric_list, pe, metric_no_group, 1, m, parent, ids);
949 	} else {
950 		int j, count;
951 
952 		count = arch_get_runtimeparam(pe);
953 
954 		/* This loop is added to create multiple
955 		 * events depend on count value and add
956 		 * those events to metric_list.
957 		 */
958 
959 		for (j = 0; j < count && !ret; j++, *m = orig)
960 			ret = __add_metric(metric_list, pe, metric_no_group, j, m, parent, ids);
961 	}
962 
963 	return ret;
964 }
965 
metricgroup__add_metric(const char * metric,bool metric_no_group,struct strbuf * events,struct list_head * metric_list,struct pmu_events_map * map)966 static int metricgroup__add_metric(const char *metric, bool metric_no_group,
967 				   struct strbuf *events,
968 				   struct list_head *metric_list,
969 				   struct pmu_events_map *map)
970 {
971 	struct expr_ids ids = { .cnt = 0, };
972 	struct pmu_event *pe;
973 	struct metric *m;
974 	LIST_HEAD(list);
975 	int i, ret;
976 	bool has_match = false;
977 
978 	map_for_each_metric(pe, i, map, metric) {
979 		has_match = true;
980 		m = NULL;
981 
982 		ret = add_metric(&list, pe, metric_no_group, &m, NULL, &ids);
983 		if (ret)
984 			goto out;
985 
986 		/*
987 		 * Process any possible referenced metrics
988 		 * included in the expression.
989 		 */
990 		ret = resolve_metric(metric_no_group,
991 				     &list, map, &ids);
992 		if (ret)
993 			goto out;
994 	}
995 
996 	/* End of pmu events. */
997 	if (!has_match) {
998 		ret = -EINVAL;
999 		goto out;
1000 	}
1001 
1002 	list_for_each_entry(m, &list, nd) {
1003 		if (events->len > 0)
1004 			strbuf_addf(events, ",");
1005 
1006 		if (m->has_constraint) {
1007 			metricgroup__add_metric_non_group(events,
1008 							  &m->pctx);
1009 		} else {
1010 			metricgroup__add_metric_weak_group(events,
1011 							   &m->pctx);
1012 		}
1013 	}
1014 
1015 out:
1016 	/*
1017 	 * add to metric_list so that they can be released
1018 	 * even if it's failed
1019 	 */
1020 	list_splice(&list, metric_list);
1021 	expr_ids__exit(&ids);
1022 	return ret;
1023 }
1024 
metricgroup__add_metric_list(const char * list,bool metric_no_group,struct strbuf * events,struct list_head * metric_list,struct pmu_events_map * map)1025 static int metricgroup__add_metric_list(const char *list, bool metric_no_group,
1026 					struct strbuf *events,
1027 					struct list_head *metric_list,
1028 					struct pmu_events_map *map)
1029 {
1030 	char *llist, *nlist, *p;
1031 	int ret = -EINVAL;
1032 
1033 	nlist = strdup(list);
1034 	if (!nlist)
1035 		return -ENOMEM;
1036 	llist = nlist;
1037 
1038 	strbuf_init(events, 100);
1039 	strbuf_addf(events, "%s", "");
1040 
1041 	while ((p = strsep(&llist, ",")) != NULL) {
1042 		ret = metricgroup__add_metric(p, metric_no_group, events,
1043 					      metric_list, map);
1044 		if (ret == -EINVAL) {
1045 			fprintf(stderr, "Cannot find metric or group `%s'\n",
1046 					p);
1047 			break;
1048 		}
1049 	}
1050 	free(nlist);
1051 
1052 	if (!ret)
1053 		metricgroup___watchdog_constraint_hint(NULL, true);
1054 
1055 	return ret;
1056 }
1057 
metric__free_refs(struct metric * metric)1058 static void metric__free_refs(struct metric *metric)
1059 {
1060 	struct metric_ref_node *ref, *tmp;
1061 
1062 	list_for_each_entry_safe(ref, tmp, &metric->metric_refs, list) {
1063 		list_del(&ref->list);
1064 		free(ref);
1065 	}
1066 }
1067 
metricgroup__free_metrics(struct list_head * metric_list)1068 static void metricgroup__free_metrics(struct list_head *metric_list)
1069 {
1070 	struct metric *m, *tmp;
1071 
1072 	list_for_each_entry_safe (m, tmp, metric_list, nd) {
1073 		metric__free_refs(m);
1074 		expr__ctx_clear(&m->pctx);
1075 		list_del_init(&m->nd);
1076 		free(m);
1077 	}
1078 }
1079 
parse_groups(struct evlist * perf_evlist,const char * str,bool metric_no_group,bool metric_no_merge,struct perf_pmu * fake_pmu,struct rblist * metric_events,struct pmu_events_map * map)1080 static int parse_groups(struct evlist *perf_evlist, const char *str,
1081 			bool metric_no_group,
1082 			bool metric_no_merge,
1083 			struct perf_pmu *fake_pmu,
1084 			struct rblist *metric_events,
1085 			struct pmu_events_map *map)
1086 {
1087 	struct parse_events_error parse_error;
1088 	struct strbuf extra_events;
1089 	LIST_HEAD(metric_list);
1090 	int ret;
1091 
1092 	if (metric_events->nr_entries == 0)
1093 		metricgroup__rblist_init(metric_events);
1094 	ret = metricgroup__add_metric_list(str, metric_no_group,
1095 					   &extra_events, &metric_list, map);
1096 	if (ret)
1097 		goto out;
1098 	pr_debug("adding %s\n", extra_events.buf);
1099 	bzero(&parse_error, sizeof(parse_error));
1100 	ret = __parse_events(perf_evlist, extra_events.buf, &parse_error, fake_pmu);
1101 	if (ret) {
1102 		parse_events_print_error(&parse_error, extra_events.buf);
1103 		goto out;
1104 	}
1105 	ret = metricgroup__setup_events(&metric_list, metric_no_merge,
1106 					perf_evlist, metric_events);
1107 out:
1108 	metricgroup__free_metrics(&metric_list);
1109 	strbuf_release(&extra_events);
1110 	return ret;
1111 }
1112 
metricgroup__parse_groups(const struct option * opt,const char * str,bool metric_no_group,bool metric_no_merge,struct rblist * metric_events)1113 int metricgroup__parse_groups(const struct option *opt,
1114 			      const char *str,
1115 			      bool metric_no_group,
1116 			      bool metric_no_merge,
1117 			      struct rblist *metric_events)
1118 {
1119 	struct evlist *perf_evlist = *(struct evlist **)opt->value;
1120 	struct pmu_events_map *map = perf_pmu__find_map(NULL);
1121 
1122 	if (!map)
1123 		return 0;
1124 
1125 	return parse_groups(perf_evlist, str, metric_no_group,
1126 			    metric_no_merge, NULL, metric_events, map);
1127 }
1128 
metricgroup__parse_groups_test(struct evlist * evlist,struct pmu_events_map * map,const char * str,bool metric_no_group,bool metric_no_merge,struct rblist * metric_events)1129 int metricgroup__parse_groups_test(struct evlist *evlist,
1130 				   struct pmu_events_map *map,
1131 				   const char *str,
1132 				   bool metric_no_group,
1133 				   bool metric_no_merge,
1134 				   struct rblist *metric_events)
1135 {
1136 	return parse_groups(evlist, str, metric_no_group,
1137 			    metric_no_merge, &perf_pmu__fake, metric_events, map);
1138 }
1139 
metricgroup__has_metric(const char * metric)1140 bool metricgroup__has_metric(const char *metric)
1141 {
1142 	struct pmu_events_map *map = perf_pmu__find_map(NULL);
1143 	struct pmu_event *pe;
1144 	int i;
1145 
1146 	if (!map)
1147 		return false;
1148 
1149 	for (i = 0; ; i++) {
1150 		pe = &map->table[i];
1151 
1152 		if (!pe->name && !pe->metric_group && !pe->metric_name)
1153 			break;
1154 		if (!pe->metric_expr)
1155 			continue;
1156 		if (match_metric(pe->metric_name, metric))
1157 			return true;
1158 	}
1159 	return false;
1160 }
1161 
metricgroup__copy_metric_events(struct evlist * evlist,struct cgroup * cgrp,struct rblist * new_metric_events,struct rblist * old_metric_events)1162 int metricgroup__copy_metric_events(struct evlist *evlist, struct cgroup *cgrp,
1163 				    struct rblist *new_metric_events,
1164 				    struct rblist *old_metric_events)
1165 {
1166 	unsigned i;
1167 
1168 	for (i = 0; i < rblist__nr_entries(old_metric_events); i++) {
1169 		struct rb_node *nd;
1170 		struct metric_event *old_me, *new_me;
1171 		struct metric_expr *old_expr, *new_expr;
1172 		struct evsel *evsel;
1173 		size_t alloc_size;
1174 		int idx, nr;
1175 
1176 		nd = rblist__entry(old_metric_events, i);
1177 		old_me = container_of(nd, struct metric_event, nd);
1178 
1179 		evsel = evlist__find_evsel(evlist, old_me->evsel->idx);
1180 		if (!evsel)
1181 			return -EINVAL;
1182 		new_me = metricgroup__lookup(new_metric_events, evsel, true);
1183 		if (!new_me)
1184 			return -ENOMEM;
1185 
1186 		pr_debug("copying metric event for cgroup '%s': %s (idx=%d)\n",
1187 			 cgrp ? cgrp->name : "root", evsel->name, evsel->idx);
1188 
1189 		list_for_each_entry(old_expr, &old_me->head, nd) {
1190 			new_expr = malloc(sizeof(*new_expr));
1191 			if (!new_expr)
1192 				return -ENOMEM;
1193 
1194 			new_expr->metric_expr = old_expr->metric_expr;
1195 			new_expr->metric_name = old_expr->metric_name;
1196 			new_expr->metric_unit = old_expr->metric_unit;
1197 			new_expr->runtime = old_expr->runtime;
1198 
1199 			if (old_expr->metric_refs) {
1200 				/* calculate number of metric_events */
1201 				for (nr = 0; old_expr->metric_refs[nr].metric_name; nr++)
1202 					continue;
1203 				alloc_size = sizeof(*new_expr->metric_refs);
1204 				new_expr->metric_refs = calloc(nr + 1, alloc_size);
1205 				if (!new_expr->metric_refs) {
1206 					free(new_expr);
1207 					return -ENOMEM;
1208 				}
1209 
1210 				memcpy(new_expr->metric_refs, old_expr->metric_refs,
1211 				       nr * alloc_size);
1212 			} else {
1213 				new_expr->metric_refs = NULL;
1214 			}
1215 
1216 			/* calculate number of metric_events */
1217 			for (nr = 0; old_expr->metric_events[nr]; nr++)
1218 				continue;
1219 			alloc_size = sizeof(*new_expr->metric_events);
1220 			new_expr->metric_events = calloc(nr + 1, alloc_size);
1221 			if (!new_expr->metric_events) {
1222 				free(new_expr->metric_refs);
1223 				free(new_expr);
1224 				return -ENOMEM;
1225 			}
1226 
1227 			/* copy evsel in the same position */
1228 			for (idx = 0; idx < nr; idx++) {
1229 				evsel = old_expr->metric_events[idx];
1230 				evsel = evlist__find_evsel(evlist, evsel->idx);
1231 				if (evsel == NULL) {
1232 					free(new_expr->metric_events);
1233 					free(new_expr->metric_refs);
1234 					free(new_expr);
1235 					return -EINVAL;
1236 				}
1237 				new_expr->metric_events[idx] = evsel;
1238 			}
1239 
1240 			list_add(&new_expr->nd, &new_me->head);
1241 		}
1242 	}
1243 	return 0;
1244 }
1245