Commit 347c2f0a authored by Ian Rogers's avatar Ian Rogers Committed by Arnaldo Carvalho de Melo

perf parse-events: Sort and group parsed events

This change is intended to be a no-op for most current cases, the
default sort order is the order the events were parsed. Where it
varies is in how groups are handled. Previously an uncore and core
event that are grouped would most often cause the group to be removed:

```
$ perf stat -e '{instructions,uncore_imc_free_running_0/data_total/}' -a sleep 1
WARNING: grouped events cpus do not match, disabling group:
  anon group { instructions, uncore_imc_free_running_0/data_total/ }
...
```

However, when wildcards are used the events should be re-sorted and
re-grouped in parse_events__set_leader, but this currently fails for
simple examples:

```
$ perf stat -e '{uncore_imc_free_running/data_read/,uncore_imc_free_running/data_write/}' -a sleep 1

 Performance counter stats for 'system wide':

     <not counted> MiB  uncore_imc_free_running/data_read/
     <not counted> MiB  uncore_imc_free_running/data_write/

       1.000996992 seconds time elapsed
```

A futher failure mode, fixed in this patch, is to force topdown events
into a group.

This change moves sorting the evsels in the evlist after parsing. It
requires parsing to set up groups. First the evsels are sorted
respecting the existing groupings and parse order, but also reordering
to ensure evsels of the same PMU and group appear together. So that
software and aux events respect groups, their pmu_name is taken from
the group leader. The sorting is done with list_sort removing a memory
allocation.

After sorting a pass is done to correct the group leaders and for
topdown events ensuring they have a group leader.

This fixes the problems seen before:

```
$ perf stat -e '{uncore_imc_free_running/data_read/,uncore_imc_free_running/data_write/}' -a sleep 1

 Performance counter stats for 'system wide':

            727.42 MiB  uncore_imc_free_running/data_read/
             81.84 MiB  uncore_imc_free_running/data_write/

       1.000948615 seconds time elapsed
```

As well as making groups not fail for cases like:

```
$ perf stat -e '{imc_free_running_0/data_total/,imc_free_running_1/data_total/}' -a sleep 1

 Performance counter stats for 'system wide':

            256.47 MiB  imc_free_running_0/data_total/
            256.48 MiB  imc_free_running_1/data_total/

       1.001165442 seconds time elapsed
```
Signed-off-by: default avatarIan Rogers <irogers@google.com>
Cc: Adrian Hunter <adrian.hunter@intel.com>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Florian Fischer <florian.fischer@muhq.space>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: James Clark <james.clark@arm.com>
Cc: Jiri Olsa <jolsa@kernel.org>
Cc: John Garry <john.g.garry@oracle.com>
Cc: Kajol Jain <kjain@linux.ibm.com>
Cc: Kan Liang <kan.liang@linux.intel.com>
Cc: Kim Phillips <kim.phillips@amd.com>
Cc: Leo Yan <leo.yan@linaro.org>
Cc: Mark Rutland <mark.rutland@arm.com>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Ravi Bangoria <ravi.bangoria@amd.com>
Cc: Sean Christopherson <seanjc@google.com>
Cc: Steinar H. Gunderson <sesse@google.com>
Cc: Stephane Eranian <eranian@google.com>
Cc: Suzuki Poulouse <suzuki.poulose@arm.com>
Cc: Xing Zhengjun <zhengjun.xing@linux.intel.com>
Link: https://lore.kernel.org/r/20230312021543.3060328-2-irogers@google.comSigned-off-by: default avatarArnaldo Carvalho de Melo <acme@redhat.com>
parent 4bb311b2
...@@ -65,29 +65,22 @@ int arch_evlist__add_default_attrs(struct evlist *evlist, ...@@ -65,29 +65,22 @@ int arch_evlist__add_default_attrs(struct evlist *evlist,
return ___evlist__add_default_attrs(evlist, attrs, nr_attrs); return ___evlist__add_default_attrs(evlist, attrs, nr_attrs);
} }
struct evsel *arch_evlist__leader(struct list_head *list) int arch_evlist__cmp(const struct evsel *lhs, const struct evsel *rhs)
{ {
struct evsel *evsel, *first, *slots = NULL; if (topdown_sys_has_perf_metrics() &&
bool has_topdown = false; (!lhs->pmu_name || !strncmp(lhs->pmu_name, "cpu", 3))) {
/* Ensure the topdown slots comes first. */
first = list_first_entry(list, struct evsel, core.node); if (strcasestr(lhs->name, "slots"))
return -1;
if (!topdown_sys_has_perf_metrics()) if (strcasestr(rhs->name, "slots"))
return first; return 1;
/* Followed by topdown events. */
/* If there is a slots event and a topdown event then the slots event comes first. */ if (strcasestr(lhs->name, "topdown") && !strcasestr(rhs->name, "topdown"))
__evlist__for_each_entry(list, evsel) { return -1;
if (evsel->pmu_name && !strncmp(evsel->pmu_name, "cpu", 3) && evsel->name) { if (!strcasestr(lhs->name, "topdown") && strcasestr(rhs->name, "topdown"))
if (strcasestr(evsel->name, "slots")) { return 1;
slots = evsel;
if (slots == first)
return first;
}
if (strcasestr(evsel->name, "topdown"))
has_topdown = true;
if (slots && has_topdown)
return slots;
}
} }
return first;
/* Default ordering by insertion index. */
return lhs->core.idx - rhs->core.idx;
} }
...@@ -119,7 +119,7 @@ int arch_evlist__add_default_attrs(struct evlist *evlist, ...@@ -119,7 +119,7 @@ int arch_evlist__add_default_attrs(struct evlist *evlist,
#define evlist__add_default_attrs(evlist, array) \ #define evlist__add_default_attrs(evlist, array) \
arch_evlist__add_default_attrs(evlist, array, ARRAY_SIZE(array)) arch_evlist__add_default_attrs(evlist, array, ARRAY_SIZE(array))
struct evsel *arch_evlist__leader(struct list_head *list); int arch_evlist__cmp(const struct evsel *lhs, const struct evsel *rhs);
int evlist__add_dummy(struct evlist *evlist); int evlist__add_dummy(struct evlist *evlist);
struct evsel *evlist__add_aux_dummy(struct evlist *evlist, bool system_wide); struct evsel *evlist__add_aux_dummy(struct evlist *evlist, bool system_wide);
......
// SPDX-License-Identifier: GPL-2.0 // SPDX-License-Identifier: GPL-2.0
#include <linux/hw_breakpoint.h> #include <linux/hw_breakpoint.h>
#include <linux/err.h> #include <linux/err.h>
#include <linux/list_sort.h>
#include <linux/zalloc.h> #include <linux/zalloc.h>
#include <dirent.h> #include <dirent.h>
#include <errno.h> #include <errno.h>
...@@ -1655,125 +1656,7 @@ int parse_events__modifier_group(struct list_head *list, ...@@ -1655,125 +1656,7 @@ int parse_events__modifier_group(struct list_head *list,
return parse_events__modifier_event(list, event_mod, true); return parse_events__modifier_event(list, event_mod, true);
} }
/* void parse_events__set_leader(char *name, struct list_head *list)
* Check if the two uncore PMUs are from the same uncore block
* The format of the uncore PMU name is uncore_#blockname_#pmuidx
*/
static bool is_same_uncore_block(const char *pmu_name_a, const char *pmu_name_b)
{
char *end_a, *end_b;
end_a = strrchr(pmu_name_a, '_');
end_b = strrchr(pmu_name_b, '_');
if (!end_a || !end_b)
return false;
if ((end_a - pmu_name_a) != (end_b - pmu_name_b))
return false;
return (strncmp(pmu_name_a, pmu_name_b, end_a - pmu_name_a) == 0);
}
static int
parse_events__set_leader_for_uncore_aliase(char *name, struct list_head *list,
struct parse_events_state *parse_state)
{
struct evsel *evsel, *leader;
uintptr_t *leaders;
bool is_leader = true;
int i, nr_pmu = 0, total_members, ret = 0;
leader = list_first_entry(list, struct evsel, core.node);
evsel = list_last_entry(list, struct evsel, core.node);
total_members = evsel->core.idx - leader->core.idx + 1;
leaders = calloc(total_members, sizeof(uintptr_t));
if (WARN_ON(!leaders))
return 0;
/*
* Going through the whole group and doing sanity check.
* All members must use alias, and be from the same uncore block.
* Also, storing the leader events in an array.
*/
__evlist__for_each_entry(list, evsel) {
/* Only split the uncore group which members use alias */
if (!evsel->use_uncore_alias)
goto out;
/* The events must be from the same uncore block */
if (!is_same_uncore_block(leader->pmu_name, evsel->pmu_name))
goto out;
if (!is_leader)
continue;
/*
* If the event's PMU name starts to repeat, it must be a new
* event. That can be used to distinguish the leader from
* other members, even they have the same event name.
*/
if ((leader != evsel) &&
!strcmp(leader->pmu_name, evsel->pmu_name)) {
is_leader = false;
continue;
}
/* Store the leader event for each PMU */
leaders[nr_pmu++] = (uintptr_t) evsel;
}
/* only one event alias */
if (nr_pmu == total_members) {
parse_state->nr_groups--;
goto handled;
}
/*
* An uncore event alias is a joint name which means the same event
* runs on all PMUs of a block.
* Perf doesn't support mixed events from different PMUs in the same
* group. The big group has to be split into multiple small groups
* which only include the events from the same PMU.
*
* Here the uncore event aliases must be from the same uncore block.
* The number of PMUs must be same for each alias. The number of new
* small groups equals to the number of PMUs.
* Setting the leader event for corresponding members in each group.
*/
i = 0;
__evlist__for_each_entry(list, evsel) {
if (i >= nr_pmu)
i = 0;
evsel__set_leader(evsel, (struct evsel *) leaders[i++]);
}
/* The number of members and group name are same for each group */
for (i = 0; i < nr_pmu; i++) {
evsel = (struct evsel *) leaders[i];
evsel->core.nr_members = total_members / nr_pmu;
evsel->group_name = name ? strdup(name) : NULL;
}
/* Take the new small groups into account */
parse_state->nr_groups += nr_pmu - 1;
handled:
ret = 1;
free(name);
out:
free(leaders);
return ret;
}
__weak struct evsel *arch_evlist__leader(struct list_head *list)
{
return list_first_entry(list, struct evsel, core.node);
}
void parse_events__set_leader(char *name, struct list_head *list,
struct parse_events_state *parse_state)
{ {
struct evsel *leader; struct evsel *leader;
...@@ -1782,13 +1665,9 @@ void parse_events__set_leader(char *name, struct list_head *list, ...@@ -1782,13 +1665,9 @@ void parse_events__set_leader(char *name, struct list_head *list,
return; return;
} }
if (parse_events__set_leader_for_uncore_aliase(name, list, parse_state)) leader = list_first_entry(list, struct evsel, core.node);
return;
leader = arch_evlist__leader(list);
__perf_evlist__set_leader(list, &leader->core); __perf_evlist__set_leader(list, &leader->core);
leader->group_name = name; leader->group_name = name;
list_move(&leader->core.node, list);
} }
/* list_event is assumed to point to malloc'ed memory */ /* list_event is assumed to point to malloc'ed memory */
...@@ -2245,6 +2124,117 @@ static int parse_events__with_hybrid_pmu(struct parse_events_state *parse_state, ...@@ -2245,6 +2124,117 @@ static int parse_events__with_hybrid_pmu(struct parse_events_state *parse_state,
return ret; return ret;
} }
__weak int arch_evlist__cmp(const struct evsel *lhs, const struct evsel *rhs)
{
/* Order by insertion index. */
return lhs->core.idx - rhs->core.idx;
}
static int evlist__cmp(void *state, const struct list_head *l, const struct list_head *r)
{
const struct perf_evsel *lhs_core = container_of(l, struct perf_evsel, node);
const struct evsel *lhs = container_of(lhs_core, struct evsel, core);
const struct perf_evsel *rhs_core = container_of(r, struct perf_evsel, node);
const struct evsel *rhs = container_of(rhs_core, struct evsel, core);
int *leader_idx = state;
int lhs_leader_idx = *leader_idx, rhs_leader_idx = *leader_idx, ret;
const char *lhs_pmu_name, *rhs_pmu_name;
/*
* First sort by grouping/leader. Read the leader idx only if the evsel
* is part of a group, as -1 indicates no group.
*/
if (lhs_core->leader != lhs_core || lhs_core->nr_members > 1)
lhs_leader_idx = lhs_core->leader->idx;
if (rhs_core->leader != rhs_core || rhs_core->nr_members > 1)
rhs_leader_idx = rhs_core->leader->idx;
if (lhs_leader_idx != rhs_leader_idx)
return lhs_leader_idx - rhs_leader_idx;
/* Group by PMU. Groups can't span PMUs. */
lhs_pmu_name = evsel__group_pmu_name(lhs);
rhs_pmu_name = evsel__group_pmu_name(rhs);
ret = strcmp(lhs_pmu_name, rhs_pmu_name);
if (ret)
return ret;
/* Architecture specific sorting. */
return arch_evlist__cmp(lhs, rhs);
}
static void parse_events__sort_events_and_fix_groups(struct list_head *list)
{
int idx = -1;
struct evsel *pos, *cur_leader = NULL;
struct perf_evsel *cur_leaders_grp = NULL;
/*
* Compute index to insert ungrouped events at. Place them where the
* first ungrouped event appears.
*/
list_for_each_entry(pos, list, core.node) {
const struct evsel *pos_leader = evsel__leader(pos);
if (pos != pos_leader || pos->core.nr_members > 1)
continue;
idx = pos->core.idx;
break;
}
/* Sort events. */
list_sort(&idx, list, evlist__cmp);
/*
* Recompute groups, splitting for PMUs and adding groups for events
* that require them.
*/
idx = 0;
list_for_each_entry(pos, list, core.node) {
const struct evsel *pos_leader = evsel__leader(pos);
const char *pos_pmu_name = evsel__group_pmu_name(pos);
const char *cur_leader_pmu_name, *pos_leader_pmu_name;
bool force_grouped = arch_evsel__must_be_in_group(pos);
/* Reset index and nr_members. */
pos->core.idx = idx++;
pos->core.nr_members = 0;
/*
* Set the group leader respecting the given groupings and that
* groups can't span PMUs.
*/
if (!cur_leader)
cur_leader = pos;
cur_leader_pmu_name = evsel__group_pmu_name(cur_leader);
if ((cur_leaders_grp != pos->core.leader && !force_grouped) ||
strcmp(cur_leader_pmu_name, pos_pmu_name)) {
/* Event is for a different group/PMU than last. */
cur_leader = pos;
/*
* Remember the leader's group before it is overwritten,
* so that later events match as being in the same
* group.
*/
cur_leaders_grp = pos->core.leader;
}
pos_leader_pmu_name = evsel__group_pmu_name(pos_leader);
if (strcmp(pos_leader_pmu_name, pos_pmu_name) || force_grouped) {
/*
* Event's PMU differs from its leader's. Groups can't
* span PMUs, so update leader from the group/PMU
* tracker.
*/
evsel__set_leader(pos, cur_leader);
}
}
list_for_each_entry(pos, list, core.node) {
pos->core.leader->nr_members++;
}
}
int __parse_events(struct evlist *evlist, const char *str, int __parse_events(struct evlist *evlist, const char *str,
struct parse_events_error *err, struct perf_pmu *fake_pmu) struct parse_events_error *err, struct perf_pmu *fake_pmu)
{ {
...@@ -2266,6 +2256,8 @@ int __parse_events(struct evlist *evlist, const char *str, ...@@ -2266,6 +2256,8 @@ int __parse_events(struct evlist *evlist, const char *str,
return -1; return -1;
} }
parse_events__sort_events_and_fix_groups(&parse_state.list);
/* /*
* Add list to the evlist even with errors to allow callers to clean up. * Add list to the evlist even with errors to allow callers to clean up.
*/ */
......
...@@ -200,8 +200,7 @@ int parse_events_copy_term_list(struct list_head *old, ...@@ -200,8 +200,7 @@ int parse_events_copy_term_list(struct list_head *old,
enum perf_pmu_event_symbol_type enum perf_pmu_event_symbol_type
perf_pmu__parse_check(const char *name); perf_pmu__parse_check(const char *name);
void parse_events__set_leader(char *name, struct list_head *list, void parse_events__set_leader(char *name, struct list_head *list);
struct parse_events_state *parse_state);
void parse_events_update_lists(struct list_head *list_event, void parse_events_update_lists(struct list_head *list_event,
struct list_head *list_all); struct list_head *list_all);
void parse_events_evlist_error(struct parse_events_state *parse_state, void parse_events_evlist_error(struct parse_events_state *parse_state,
......
...@@ -203,7 +203,7 @@ PE_NAME '{' events '}' ...@@ -203,7 +203,7 @@ PE_NAME '{' events '}'
inc_group_count(list, _parse_state); inc_group_count(list, _parse_state);
/* Takes ownership of $1. */ /* Takes ownership of $1. */
parse_events__set_leader($1, list, _parse_state); parse_events__set_leader($1, list);
$$ = list; $$ = list;
} }
| |
...@@ -212,7 +212,7 @@ PE_NAME '{' events '}' ...@@ -212,7 +212,7 @@ PE_NAME '{' events '}'
struct list_head *list = $2; struct list_head *list = $2;
inc_group_count(list, _parse_state); inc_group_count(list, _parse_state);
parse_events__set_leader(NULL, list, _parse_state); parse_events__set_leader(NULL, list);
$$ = list; $$ = list;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment