Now aggr counts are ready for use. Convert the display routines to use
the aggr counts and update the shadow stat with them. It doesn't need
to aggregate counts or collect aliases anymore during the display. Get
rid of now unused struct perf_aggr_thread_value.
Note that there's a difference in the display order among the aggr mode.
For per-core/die/socket/node aggregation, it shows relevant events in
the same unit together, whereas global/thread/no aggregation it shows
the same events for different units together. So it still uses separate
codes to display them due to the ordering.
One more thing to note is that it breaks per-core event display for now.
The next patch will fix it to have identical output as of now.
Acked-by: Ian Rogers <[email protected]>
Signed-off-by: Namhyung Kim <[email protected]>
---
tools/perf/util/stat-display.c | 421 ++++-----------------------------
tools/perf/util/stat.c | 5 -
tools/perf/util/stat.h | 9 -
3 files changed, 49 insertions(+), 386 deletions(-)
diff --git a/tools/perf/util/stat-display.c b/tools/perf/util/stat-display.c
index 4113aa86772f..bfae2784609c 100644
--- a/tools/perf/util/stat-display.c
+++ b/tools/perf/util/stat-display.c
@@ -442,31 +442,6 @@ static void print_metric_header(struct perf_stat_config *config,
fprintf(os->fh, "%*s ", config->metric_only_len, unit);
}
-static int first_shadow_map_idx(struct perf_stat_config *config,
- struct evsel *evsel, const struct aggr_cpu_id *id)
-{
- struct perf_cpu_map *cpus = evsel__cpus(evsel);
- struct perf_cpu cpu;
- int idx;
-
- if (config->aggr_mode == AGGR_NONE)
- return perf_cpu_map__idx(cpus, id->cpu);
-
- if (config->aggr_mode == AGGR_THREAD)
- return id->thread_idx;
-
- if (!config->aggr_get_id)
- return 0;
-
- perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
- struct aggr_cpu_id cpu_id = config->aggr_get_id(config, cpu);
-
- if (aggr_cpu_id__equal(&cpu_id, id))
- return idx;
- }
- return 0;
-}
-
static void abs_printout(struct perf_stat_config *config,
struct aggr_cpu_id id, int nr, struct evsel *evsel, double avg)
{
@@ -537,7 +512,7 @@ static bool is_mixed_hw_group(struct evsel *counter)
static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int nr,
struct evsel *counter, double uval,
char *prefix, u64 run, u64 ena, double noise,
- struct runtime_stat *st)
+ struct runtime_stat *st, int map_idx)
{
struct perf_stat_output_ctx out;
struct outstate os = {
@@ -648,8 +623,7 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
print_running(config, run, ena);
}
- perf_stat__print_shadow_stats(config, counter, uval,
- first_shadow_map_idx(config, counter, &id),
+ perf_stat__print_shadow_stats(config, counter, uval, map_idx,
&out, &config->metric_events, st);
if (!config->csv_output && !config->metric_only && !config->json_output) {
print_noise(config, counter, noise);
@@ -657,34 +631,6 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
}
}
-static void aggr_update_shadow(struct perf_stat_config *config,
- struct evlist *evlist)
-{
- int idx, s;
- struct perf_cpu cpu;
- struct aggr_cpu_id s2, id;
- u64 val;
- struct evsel *counter;
- struct perf_cpu_map *cpus;
-
- for (s = 0; s < config->aggr_map->nr; s++) {
- id = config->aggr_map->map[s];
- evlist__for_each_entry(evlist, counter) {
- cpus = evsel__cpus(counter);
- val = 0;
- perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
- s2 = config->aggr_get_id(config, cpu);
- if (!aggr_cpu_id__equal(&s2, &id))
- continue;
- val += perf_counts(counter->counts, idx, 0)->val;
- }
- perf_stat__update_shadow_stats(counter, val,
- first_shadow_map_idx(config, counter, &id),
- &rt_stat);
- }
- }
-}
-
static void uniquify_event_name(struct evsel *counter)
{
char *new_name;
@@ -721,137 +667,51 @@ static void uniquify_event_name(struct evsel *counter)
counter->uniquified_name = true;
}
-static void collect_all_aliases(struct perf_stat_config *config, struct evsel *counter,
- void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
- bool first),
- void *data)
+static bool hybrid_uniquify(struct evsel *evsel, struct perf_stat_config *config)
{
- struct evlist *evlist = counter->evlist;
- struct evsel *alias;
-
- alias = list_prepare_entry(counter, &(evlist->core.entries), core.node);
- list_for_each_entry_continue (alias, &evlist->core.entries, core.node) {
- /* Merge events with the same name, etc. but on different PMUs. */
- if (!strcmp(evsel__name(alias), evsel__name(counter)) &&
- alias->scale == counter->scale &&
- alias->cgrp == counter->cgrp &&
- !strcmp(alias->unit, counter->unit) &&
- evsel__is_clock(alias) == evsel__is_clock(counter) &&
- strcmp(alias->pmu_name, counter->pmu_name)) {
- alias->merged_stat = true;
- cb(config, alias, data, false);
- }
- }
+ return evsel__is_hybrid(evsel) && !config->hybrid_merge;
}
-static bool hybrid_merge(struct evsel *counter, struct perf_stat_config *config,
- bool check)
+static void uniquify_counter(struct perf_stat_config *config, struct evsel *counter)
{
- if (evsel__is_hybrid(counter)) {
- if (check)
- return config->hybrid_merge;
- else
- return !config->hybrid_merge;
- }
-
- return false;
-}
-
-static bool collect_data(struct perf_stat_config *config, struct evsel *counter,
- void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
- bool first),
- void *data)
-{
- if (counter->merged_stat)
- return false;
- cb(config, counter, data, true);
- if (config->no_merge || hybrid_merge(counter, config, false))
+ if (config->no_merge || hybrid_uniquify(counter, config))
uniquify_event_name(counter);
- else if (counter->auto_merge_stats || hybrid_merge(counter, config, true))
- collect_all_aliases(config, counter, cb, data);
- return true;
-}
-
-struct aggr_data {
- u64 ena, run, val;
- struct aggr_cpu_id id;
- int nr;
- int cpu_map_idx;
-};
-
-static void aggr_cb(struct perf_stat_config *config,
- struct evsel *counter, void *data, bool first)
-{
- struct aggr_data *ad = data;
- int idx;
- struct perf_cpu cpu;
- struct perf_cpu_map *cpus;
- struct aggr_cpu_id s2;
-
- cpus = evsel__cpus(counter);
- perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
- struct perf_counts_values *counts;
-
- s2 = config->aggr_get_id(config, cpu);
- if (!aggr_cpu_id__equal(&s2, &ad->id))
- continue;
- if (first)
- ad->nr++;
- counts = perf_counts(counter->counts, idx, 0);
- /*
- * When any result is bad, make them all to give
- * consistent output in interval mode.
- */
- if (counts->ena == 0 || counts->run == 0 ||
- counter->counts->scaled == -1) {
- ad->ena = 0;
- ad->run = 0;
- break;
- }
- ad->val += counts->val;
- ad->ena += counts->ena;
- ad->run += counts->run;
- }
}
static void print_counter_aggrdata(struct perf_stat_config *config,
struct evsel *counter, int s,
char *prefix, bool metric_only,
- bool *first, struct perf_cpu cpu)
+ bool *first)
{
- struct aggr_data ad;
FILE *output = config->output;
u64 ena, run, val;
- int nr;
- struct aggr_cpu_id id;
double uval;
+ struct perf_stat_evsel *ps = counter->stats;
+ struct perf_stat_aggr *aggr = &ps->aggr[s];
+ struct aggr_cpu_id id = config->aggr_map->map[s];
+ double avg = aggr->counts.val;
- ad.id = id = config->aggr_map->map[s];
- ad.val = ad.ena = ad.run = 0;
- ad.nr = 0;
- if (!collect_data(config, counter, aggr_cb, &ad))
+ if (counter->supported && aggr->nr == 0)
return;
- if (perf_pmu__has_hybrid() && ad.ena == 0)
- return;
+ uniquify_counter(config, counter);
+
+ val = aggr->counts.val;
+ ena = aggr->counts.ena;
+ run = aggr->counts.run;
- nr = ad.nr;
- ena = ad.ena;
- run = ad.run;
- val = ad.val;
if (*first && metric_only) {
*first = false;
- aggr_printout(config, counter, id, nr);
+ aggr_printout(config, counter, id, aggr->nr);
}
if (prefix && !metric_only)
fprintf(output, "%s", prefix);
uval = val * counter->scale;
- if (cpu.cpu != -1)
- id = aggr_cpu_id__cpu(cpu, /*data=*/NULL);
- printout(config, id, nr, counter, uval,
- prefix, run, ena, 1.0, &rt_stat);
+ printout(config, id, aggr->nr, counter, uval,
+ prefix, run, ena, avg, &rt_stat, s);
+
if (!metric_only)
fputc('\n', output);
}
@@ -869,8 +729,6 @@ static void print_aggr(struct perf_stat_config *config,
if (!config->aggr_map || !config->aggr_get_id)
return;
- aggr_update_shadow(config, evlist);
-
/*
* With metric_only everything is on a single line.
* Without each counter has its own line.
@@ -881,188 +739,36 @@ static void print_aggr(struct perf_stat_config *config,
first = true;
evlist__for_each_entry(evlist, counter) {
+ if (counter->merged_stat)
+ continue;
+
print_counter_aggrdata(config, counter, s,
- prefix, metric_only,
- &first, (struct perf_cpu){ .cpu = -1 });
+ prefix, metric_only,
+ &first);
}
if (metric_only)
fputc('\n', output);
}
}
-static int cmp_val(const void *a, const void *b)
-{
- return ((struct perf_aggr_thread_value *)b)->val -
- ((struct perf_aggr_thread_value *)a)->val;
-}
-
-static struct perf_aggr_thread_value *sort_aggr_thread(
- struct evsel *counter,
- int *ret,
- struct target *_target)
-{
- int nthreads = perf_thread_map__nr(counter->core.threads);
- int i = 0;
- double uval;
- struct perf_aggr_thread_value *buf;
-
- buf = calloc(nthreads, sizeof(struct perf_aggr_thread_value));
- if (!buf)
- return NULL;
-
- for (int thread = 0; thread < nthreads; thread++) {
- int idx;
- u64 ena = 0, run = 0, val = 0;
-
- perf_cpu_map__for_each_idx(idx, evsel__cpus(counter)) {
- struct perf_counts_values *counts =
- perf_counts(counter->counts, idx, thread);
-
- val += counts->val;
- ena += counts->ena;
- run += counts->run;
- }
-
- uval = val * counter->scale;
-
- /*
- * Skip value 0 when enabling --per-thread globally,
- * otherwise too many 0 output.
- */
- if (uval == 0.0 && target__has_per_thread(_target))
- continue;
-
- buf[i].counter = counter;
- buf[i].id = aggr_cpu_id__empty();
- buf[i].id.thread_idx = thread;
- buf[i].uval = uval;
- buf[i].val = val;
- buf[i].run = run;
- buf[i].ena = ena;
- i++;
- }
-
- qsort(buf, i, sizeof(struct perf_aggr_thread_value), cmp_val);
-
- if (ret)
- *ret = i;
-
- return buf;
-}
-
-static void print_aggr_thread(struct perf_stat_config *config,
- struct target *_target,
- struct evsel *counter, char *prefix)
-{
- FILE *output = config->output;
- int thread, sorted_threads;
- struct aggr_cpu_id id;
- struct perf_aggr_thread_value *buf;
-
- buf = sort_aggr_thread(counter, &sorted_threads, _target);
- if (!buf) {
- perror("cannot sort aggr thread");
- return;
- }
-
- for (thread = 0; thread < sorted_threads; thread++) {
- if (prefix)
- fprintf(output, "%s", prefix);
-
- id = buf[thread].id;
- printout(config, id, 0, buf[thread].counter, buf[thread].uval,
- prefix, buf[thread].run, buf[thread].ena, 1.0,
- &rt_stat);
- fputc('\n', output);
- }
-
- free(buf);
-}
-
-struct caggr_data {
- double avg, avg_enabled, avg_running;
-};
-
-static void counter_aggr_cb(struct perf_stat_config *config __maybe_unused,
- struct evsel *counter, void *data,
- bool first __maybe_unused)
-{
- struct caggr_data *cd = data;
- struct perf_counts_values *aggr = &counter->counts->aggr;
-
- cd->avg += aggr->val;
- cd->avg_enabled += aggr->ena;
- cd->avg_running += aggr->run;
-}
-
-/*
- * Print out the results of a single counter:
- * aggregated counts in system-wide mode
- */
-static void print_counter_aggr(struct perf_stat_config *config,
- struct evsel *counter, char *prefix)
-{
- bool metric_only = config->metric_only;
- FILE *output = config->output;
- double uval;
- struct caggr_data cd = { .avg = 0.0 };
-
- if (!collect_data(config, counter, counter_aggr_cb, &cd))
- return;
-
- if (prefix && !metric_only)
- fprintf(output, "%s", prefix);
-
- uval = cd.avg * counter->scale;
- printout(config, aggr_cpu_id__empty(), 0, counter, uval, prefix, cd.avg_running,
- cd.avg_enabled, cd.avg, &rt_stat);
- if (!metric_only)
- fprintf(output, "\n");
-}
-
-static void counter_cb(struct perf_stat_config *config __maybe_unused,
- struct evsel *counter, void *data,
- bool first __maybe_unused)
-{
- struct aggr_data *ad = data;
-
- ad->val += perf_counts(counter->counts, ad->cpu_map_idx, 0)->val;
- ad->ena += perf_counts(counter->counts, ad->cpu_map_idx, 0)->ena;
- ad->run += perf_counts(counter->counts, ad->cpu_map_idx, 0)->run;
-}
-
-/*
- * Print out the results of a single counter:
- * does not use aggregated count in system-wide
- */
static void print_counter(struct perf_stat_config *config,
struct evsel *counter, char *prefix)
{
- FILE *output = config->output;
- u64 ena, run, val;
- double uval;
- int idx;
- struct perf_cpu cpu;
- struct aggr_cpu_id id;
-
- perf_cpu_map__for_each_cpu(cpu, idx, evsel__cpus(counter)) {
- struct aggr_data ad = { .cpu_map_idx = idx };
-
- if (!collect_data(config, counter, counter_cb, &ad))
- return;
- val = ad.val;
- ena = ad.ena;
- run = ad.run;
+ bool metric_only = config->metric_only;
+ bool first = false;
+ int s;
- if (prefix)
- fprintf(output, "%s", prefix);
+ /* AGGR_THREAD doesn't have config->aggr_get_id */
+ if (!config->aggr_map)
+ return;
- uval = val * counter->scale;
- id = aggr_cpu_id__cpu(cpu, /*data=*/NULL);
- printout(config, id, 0, counter, uval, prefix,
- run, ena, 1.0, &rt_stat);
+ if (counter->merged_stat)
+ return;
- fputc('\n', output);
+ for (s = 0; s < config->aggr_map->nr; s++) {
+ print_counter_aggrdata(config, counter, s,
+ prefix, metric_only,
+ &first);
}
}
@@ -1081,6 +787,7 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
u64 ena, run, val;
double uval;
struct aggr_cpu_id id;
+ struct perf_stat_evsel *ps = counter->stats;
int counter_idx = perf_cpu_map__idx(evsel__cpus(counter), cpu);
if (counter_idx < 0)
@@ -1093,13 +800,13 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
aggr_printout(config, counter, id, 0);
first = false;
}
- val = perf_counts(counter->counts, counter_idx, 0)->val;
- ena = perf_counts(counter->counts, counter_idx, 0)->ena;
- run = perf_counts(counter->counts, counter_idx, 0)->run;
+ val = ps->aggr[counter_idx].counts.val;
+ ena = ps->aggr[counter_idx].counts.ena;
+ run = ps->aggr[counter_idx].counts.run;
uval = val * counter->scale;
printout(config, id, 0, counter, uval, prefix,
- run, ena, 1.0, &rt_stat);
+ run, ena, 1.0, &rt_stat, counter_idx);
}
if (!first)
fputc('\n', config->output);
@@ -1135,8 +842,8 @@ static void print_metric_headers(struct perf_stat_config *config,
};
bool first = true;
- if (config->json_output && !config->interval)
- fprintf(config->output, "{");
+ if (config->json_output && !config->interval)
+ fprintf(config->output, "{");
if (prefix && !config->json_output)
fprintf(config->output, "%s", prefix);
@@ -1379,31 +1086,6 @@ static void print_footer(struct perf_stat_config *config)
"the same PMU. Try reorganizing the group.\n");
}
-static void print_percore_thread(struct perf_stat_config *config,
- struct evsel *counter, char *prefix)
-{
- int s;
- struct aggr_cpu_id s2, id;
- struct perf_cpu_map *cpus;
- bool first = true;
- int idx;
- struct perf_cpu cpu;
-
- cpus = evsel__cpus(counter);
- perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
- s2 = config->aggr_get_id(config, cpu);
- for (s = 0; s < config->aggr_map->nr; s++) {
- id = config->aggr_map->map[s];
- if (aggr_cpu_id__equal(&s2, &id))
- break;
- }
-
- print_counter_aggrdata(config, counter, s,
- prefix, false,
- &first, cpu);
- }
-}
-
static void print_percore(struct perf_stat_config *config,
struct evsel *counter, char *prefix)
{
@@ -1416,15 +1098,14 @@ static void print_percore(struct perf_stat_config *config,
return;
if (config->percore_show_thread)
- return print_percore_thread(config, counter, prefix);
+ return print_counter(config, counter, prefix);
for (s = 0; s < config->aggr_map->nr; s++) {
if (prefix && metric_only)
fprintf(output, "%s", prefix);
print_counter_aggrdata(config, counter, s,
- prefix, metric_only,
- &first, (struct perf_cpu){ .cpu = -1 });
+ prefix, metric_only, &first);
}
if (metric_only)
@@ -1469,17 +1150,13 @@ void evlist__print_counters(struct evlist *evlist, struct perf_stat_config *conf
print_aggr(config, evlist, prefix);
break;
case AGGR_THREAD:
- evlist__for_each_entry(evlist, counter) {
- print_aggr_thread(config, _target, counter, prefix);
- }
- break;
case AGGR_GLOBAL:
if (config->iostat_run)
iostat_print_counters(evlist, config, ts, prefix = buf,
- print_counter_aggr);
+ print_counter);
else {
evlist__for_each_entry(evlist, counter) {
- print_counter_aggr(config, counter, prefix);
+ print_counter(config, counter, prefix);
}
if (metric_only)
fputc('\n', config->output);
diff --git a/tools/perf/util/stat.c b/tools/perf/util/stat.c
index c0955a0427ab..0316557adce9 100644
--- a/tools/perf/util/stat.c
+++ b/tools/perf/util/stat.c
@@ -565,11 +565,6 @@ int perf_stat_process_counter(struct perf_stat_config *config,
evsel__name(counter), count[0], count[1], count[2]);
}
- /*
- * Save the full runtime - to allow normalization during printout:
- */
- perf_stat__update_shadow_stats(counter, *count, 0, &rt_stat);
-
return 0;
}
diff --git a/tools/perf/util/stat.h b/tools/perf/util/stat.h
index 3d413ba8c68a..382a1ab92ce1 100644
--- a/tools/perf/util/stat.h
+++ b/tools/perf/util/stat.h
@@ -224,15 +224,6 @@ static inline void update_rusage_stats(struct rusage_stats *ru_stats, struct rus
struct evsel;
struct evlist;
-struct perf_aggr_thread_value {
- struct evsel *counter;
- struct aggr_cpu_id id;
- double uval;
- u64 val;
- u64 run;
- u64 ena;
-};
-
bool __perf_stat_evsel__is(struct evsel *evsel, enum perf_stat_evsel_id id);
#define perf_stat_evsel__is(evsel, id) \
--
2.38.0.413.g74048e4d9e-goog
> On 18-Oct-2022, at 7:32 AM, Namhyung Kim <[email protected]> wrote:
>
> Now aggr counts are ready for use. Convert the display routines to use
> the aggr counts and update the shadow stat with them. It doesn't need
> to aggregate counts or collect aliases anymore during the display. Get
> rid of now unused struct perf_aggr_thread_value.
>
> Note that there's a difference in the display order among the aggr mode.
> For per-core/die/socket/node aggregation, it shows relevant events in
> the same unit together, whereas global/thread/no aggregation it shows
> the same events for different units together. So it still uses separate
> codes to display them due to the ordering.
>
> One more thing to note is that it breaks per-core event display for now.
> The next patch will fix it to have identical output as of now.
>
> Acked-by: Ian Rogers <[email protected]>
> Signed-off-by: Namhyung Kim <[email protected]>
> ---
> tools/perf/util/stat-display.c | 421 ++++-----------------------------
> tools/perf/util/stat.c | 5 -
> tools/perf/util/stat.h | 9 -
> 3 files changed, 49 insertions(+), 386 deletions(-)
>
> diff --git a/tools/perf/util/stat-display.c b/tools/perf/util/stat-display.c
> index 4113aa86772f..bfae2784609c 100644
> --- a/tools/perf/util/stat-display.c
> +++ b/tools/perf/util/stat-display.c
> @@ -442,31 +442,6 @@ static void print_metric_header(struct perf_stat_config *config,
> fprintf(os->fh, "%*s ", config->metric_only_len, unit);
> }
>
> -static int first_shadow_map_idx(struct perf_stat_config *config,
> - struct evsel *evsel, const struct aggr_cpu_id *id)
> -{
> - struct perf_cpu_map *cpus = evsel__cpus(evsel);
> - struct perf_cpu cpu;
> - int idx;
> -
> - if (config->aggr_mode == AGGR_NONE)
> - return perf_cpu_map__idx(cpus, id->cpu);
> -
> - if (config->aggr_mode == AGGR_THREAD)
> - return id->thread_idx;
> -
> - if (!config->aggr_get_id)
> - return 0;
> -
> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> - struct aggr_cpu_id cpu_id = config->aggr_get_id(config, cpu);
> -
> - if (aggr_cpu_id__equal(&cpu_id, id))
> - return idx;
> - }
> - return 0;
> -}
> -
> static void abs_printout(struct perf_stat_config *config,
> struct aggr_cpu_id id, int nr, struct evsel *evsel, double avg)
> {
> @@ -537,7 +512,7 @@ static bool is_mixed_hw_group(struct evsel *counter)
> static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int nr,
> struct evsel *counter, double uval,
> char *prefix, u64 run, u64 ena, double noise,
> - struct runtime_stat *st)
> + struct runtime_stat *st, int map_idx)
> {
> struct perf_stat_output_ctx out;
> struct outstate os = {
> @@ -648,8 +623,7 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
> print_running(config, run, ena);
> }
>
> - perf_stat__print_shadow_stats(config, counter, uval,
> - first_shadow_map_idx(config, counter, &id),
> + perf_stat__print_shadow_stats(config, counter, uval, map_idx,
> &out, &config->metric_events, st);
> if (!config->csv_output && !config->metric_only && !config->json_output) {
> print_noise(config, counter, noise);
> @@ -657,34 +631,6 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
> }
> }
>
> -static void aggr_update_shadow(struct perf_stat_config *config,
> - struct evlist *evlist)
> -{
> - int idx, s;
> - struct perf_cpu cpu;
> - struct aggr_cpu_id s2, id;
> - u64 val;
> - struct evsel *counter;
> - struct perf_cpu_map *cpus;
> -
> - for (s = 0; s < config->aggr_map->nr; s++) {
> - id = config->aggr_map->map[s];
> - evlist__for_each_entry(evlist, counter) {
> - cpus = evsel__cpus(counter);
> - val = 0;
> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> - s2 = config->aggr_get_id(config, cpu);
> - if (!aggr_cpu_id__equal(&s2, &id))
> - continue;
> - val += perf_counts(counter->counts, idx, 0)->val;
> - }
> - perf_stat__update_shadow_stats(counter, val,
> - first_shadow_map_idx(config, counter, &id),
> - &rt_stat);
> - }
> - }
> -}
> -
> static void uniquify_event_name(struct evsel *counter)
> {
> char *new_name;
> @@ -721,137 +667,51 @@ static void uniquify_event_name(struct evsel *counter)
> counter->uniquified_name = true;
> }
>
> -static void collect_all_aliases(struct perf_stat_config *config, struct evsel *counter,
> - void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
> - bool first),
> - void *data)
> +static bool hybrid_uniquify(struct evsel *evsel, struct perf_stat_config *config)
> {
> - struct evlist *evlist = counter->evlist;
> - struct evsel *alias;
> -
> - alias = list_prepare_entry(counter, &(evlist->core.entries), core.node);
> - list_for_each_entry_continue (alias, &evlist->core.entries, core.node) {
> - /* Merge events with the same name, etc. but on different PMUs. */
> - if (!strcmp(evsel__name(alias), evsel__name(counter)) &&
> - alias->scale == counter->scale &&
> - alias->cgrp == counter->cgrp &&
> - !strcmp(alias->unit, counter->unit) &&
> - evsel__is_clock(alias) == evsel__is_clock(counter) &&
> - strcmp(alias->pmu_name, counter->pmu_name)) {
> - alias->merged_stat = true;
> - cb(config, alias, data, false);
> - }
> - }
> + return evsel__is_hybrid(evsel) && !config->hybrid_merge;
> }
>
> -static bool hybrid_merge(struct evsel *counter, struct perf_stat_config *config,
> - bool check)
> +static void uniquify_counter(struct perf_stat_config *config, struct evsel *counter)
> {
> - if (evsel__is_hybrid(counter)) {
> - if (check)
> - return config->hybrid_merge;
> - else
> - return !config->hybrid_merge;
> - }
> -
> - return false;
> -}
> -
> -static bool collect_data(struct perf_stat_config *config, struct evsel *counter,
> - void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
> - bool first),
> - void *data)
> -{
> - if (counter->merged_stat)
> - return false;
> - cb(config, counter, data, true);
> - if (config->no_merge || hybrid_merge(counter, config, false))
> + if (config->no_merge || hybrid_uniquify(counter, config))
> uniquify_event_name(counter);
> - else if (counter->auto_merge_stats || hybrid_merge(counter, config, true))
> - collect_all_aliases(config, counter, cb, data);
> - return true;
> -}
> -
> -struct aggr_data {
> - u64 ena, run, val;
> - struct aggr_cpu_id id;
> - int nr;
> - int cpu_map_idx;
> -};
> -
> -static void aggr_cb(struct perf_stat_config *config,
> - struct evsel *counter, void *data, bool first)
> -{
> - struct aggr_data *ad = data;
> - int idx;
> - struct perf_cpu cpu;
> - struct perf_cpu_map *cpus;
> - struct aggr_cpu_id s2;
> -
> - cpus = evsel__cpus(counter);
> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> - struct perf_counts_values *counts;
> -
> - s2 = config->aggr_get_id(config, cpu);
> - if (!aggr_cpu_id__equal(&s2, &ad->id))
> - continue;
> - if (first)
> - ad->nr++;
> - counts = perf_counts(counter->counts, idx, 0);
> - /*
> - * When any result is bad, make them all to give
> - * consistent output in interval mode.
> - */
> - if (counts->ena == 0 || counts->run == 0 ||
> - counter->counts->scaled == -1) {
> - ad->ena = 0;
> - ad->run = 0;
> - break;
> - }
> - ad->val += counts->val;
> - ad->ena += counts->ena;
> - ad->run += counts->run;
> - }
> }
>
> static void print_counter_aggrdata(struct perf_stat_config *config,
> struct evsel *counter, int s,
> char *prefix, bool metric_only,
> - bool *first, struct perf_cpu cpu)
> + bool *first)
> {
> - struct aggr_data ad;
> FILE *output = config->output;
> u64 ena, run, val;
> - int nr;
> - struct aggr_cpu_id id;
> double uval;
> + struct perf_stat_evsel *ps = counter->stats;
> + struct perf_stat_aggr *aggr = &ps->aggr[s];
> + struct aggr_cpu_id id = config->aggr_map->map[s];
> + double avg = aggr->counts.val;
>
> - ad.id = id = config->aggr_map->map[s];
> - ad.val = ad.ena = ad.run = 0;
> - ad.nr = 0;
> - if (!collect_data(config, counter, aggr_cb, &ad))
> + if (counter->supported && aggr->nr == 0)
> return;
Hi Namhyung,
Observed one issue with this change in tmp.perf/core branch of git://git.kernel.org/pub/scm/linux/kernel/git/acme/linux.git
Patch link: https://git.kernel.org/pub/scm/linux/kernel/git/acme/linux.git/commit/?h=tmp.perf/core&id=91f85f98da7ab8c32105f42dd03884c01ec4498f
If we have multiple events in a group and if all events in that group can't go in at once, the group will fail to schedule.
The perf stat result used to have status of all events in group.
Example result from powerpc:
# perf stat -e '{r1001e,r401e0,r4c010}' <workload>
Performance counter stats for 'workload’:
<not counted> r1001e
<not counted> r401e0
<not supported> r4c010
But after this change, it shows only the event which broke the constraint.
# ./perf stat -e '{r1001e,r401e0,r4c010}' <workload>
Performance counter stats for 'workload’:
<not supported> r4c010
The new condition added by the patch is:
if (counter->supported && aggr->nr == 0)
return;
The first two events r1001e and r401e0 are good to go in this group, hence
counter->supported is true where as aggr->nr is zero since this group is
not scheduled. Since counter->supported is false for third event which
broke the group constraint, the condition won't return and hence only prints
this event in output.
Namhyung, is there any scenario because of which this condition was added ?
If not I would go ahead and sent a fix patch to remove this check.
Thanks
Athira
>
> - if (perf_pmu__has_hybrid() && ad.ena == 0)
> - return;
> + uniquify_counter(config, counter);
> +
> + val = aggr->counts.val;
> + ena = aggr->counts.ena;
> + run = aggr->counts.run;
>
> - nr = ad.nr;
> - ena = ad.ena;
> - run = ad.run;
> - val = ad.val;
> if (*first && metric_only) {
> *first = false;
> - aggr_printout(config, counter, id, nr);
> + aggr_printout(config, counter, id, aggr->nr);
> }
> if (prefix && !metric_only)
> fprintf(output, "%s", prefix);
>
> uval = val * counter->scale;
> - if (cpu.cpu != -1)
> - id = aggr_cpu_id__cpu(cpu, /*data=*/NULL);
>
> - printout(config, id, nr, counter, uval,
> - prefix, run, ena, 1.0, &rt_stat);
> + printout(config, id, aggr->nr, counter, uval,
> + prefix, run, ena, avg, &rt_stat, s);
> +
> if (!metric_only)
> fputc('\n', output);
> }
> @@ -869,8 +729,6 @@ static void print_aggr(struct perf_stat_config *config,
> if (!config->aggr_map || !config->aggr_get_id)
> return;
>
> - aggr_update_shadow(config, evlist);
> -
> /*
> * With metric_only everything is on a single line.
> * Without each counter has its own line.
> @@ -881,188 +739,36 @@ static void print_aggr(struct perf_stat_config *config,
>
> first = true;
> evlist__for_each_entry(evlist, counter) {
> + if (counter->merged_stat)
> + continue;
> +
> print_counter_aggrdata(config, counter, s,
> - prefix, metric_only,
> - &first, (struct perf_cpu){ .cpu = -1 });
> + prefix, metric_only,
> + &first);
> }
> if (metric_only)
> fputc('\n', output);
> }
> }
>
> -static int cmp_val(const void *a, const void *b)
> -{
> - return ((struct perf_aggr_thread_value *)b)->val -
> - ((struct perf_aggr_thread_value *)a)->val;
> -}
> -
> -static struct perf_aggr_thread_value *sort_aggr_thread(
> - struct evsel *counter,
> - int *ret,
> - struct target *_target)
> -{
> - int nthreads = perf_thread_map__nr(counter->core.threads);
> - int i = 0;
> - double uval;
> - struct perf_aggr_thread_value *buf;
> -
> - buf = calloc(nthreads, sizeof(struct perf_aggr_thread_value));
> - if (!buf)
> - return NULL;
> -
> - for (int thread = 0; thread < nthreads; thread++) {
> - int idx;
> - u64 ena = 0, run = 0, val = 0;
> -
> - perf_cpu_map__for_each_idx(idx, evsel__cpus(counter)) {
> - struct perf_counts_values *counts =
> - perf_counts(counter->counts, idx, thread);
> -
> - val += counts->val;
> - ena += counts->ena;
> - run += counts->run;
> - }
> -
> - uval = val * counter->scale;
> -
> - /*
> - * Skip value 0 when enabling --per-thread globally,
> - * otherwise too many 0 output.
> - */
> - if (uval == 0.0 && target__has_per_thread(_target))
> - continue;
> -
> - buf[i].counter = counter;
> - buf[i].id = aggr_cpu_id__empty();
> - buf[i].id.thread_idx = thread;
> - buf[i].uval = uval;
> - buf[i].val = val;
> - buf[i].run = run;
> - buf[i].ena = ena;
> - i++;
> - }
> -
> - qsort(buf, i, sizeof(struct perf_aggr_thread_value), cmp_val);
> -
> - if (ret)
> - *ret = i;
> -
> - return buf;
> -}
> -
> -static void print_aggr_thread(struct perf_stat_config *config,
> - struct target *_target,
> - struct evsel *counter, char *prefix)
> -{
> - FILE *output = config->output;
> - int thread, sorted_threads;
> - struct aggr_cpu_id id;
> - struct perf_aggr_thread_value *buf;
> -
> - buf = sort_aggr_thread(counter, &sorted_threads, _target);
> - if (!buf) {
> - perror("cannot sort aggr thread");
> - return;
> - }
> -
> - for (thread = 0; thread < sorted_threads; thread++) {
> - if (prefix)
> - fprintf(output, "%s", prefix);
> -
> - id = buf[thread].id;
> - printout(config, id, 0, buf[thread].counter, buf[thread].uval,
> - prefix, buf[thread].run, buf[thread].ena, 1.0,
> - &rt_stat);
> - fputc('\n', output);
> - }
> -
> - free(buf);
> -}
> -
> -struct caggr_data {
> - double avg, avg_enabled, avg_running;
> -};
> -
> -static void counter_aggr_cb(struct perf_stat_config *config __maybe_unused,
> - struct evsel *counter, void *data,
> - bool first __maybe_unused)
> -{
> - struct caggr_data *cd = data;
> - struct perf_counts_values *aggr = &counter->counts->aggr;
> -
> - cd->avg += aggr->val;
> - cd->avg_enabled += aggr->ena;
> - cd->avg_running += aggr->run;
> -}
> -
> -/*
> - * Print out the results of a single counter:
> - * aggregated counts in system-wide mode
> - */
> -static void print_counter_aggr(struct perf_stat_config *config,
> - struct evsel *counter, char *prefix)
> -{
> - bool metric_only = config->metric_only;
> - FILE *output = config->output;
> - double uval;
> - struct caggr_data cd = { .avg = 0.0 };
> -
> - if (!collect_data(config, counter, counter_aggr_cb, &cd))
> - return;
> -
> - if (prefix && !metric_only)
> - fprintf(output, "%s", prefix);
> -
> - uval = cd.avg * counter->scale;
> - printout(config, aggr_cpu_id__empty(), 0, counter, uval, prefix, cd.avg_running,
> - cd.avg_enabled, cd.avg, &rt_stat);
> - if (!metric_only)
> - fprintf(output, "\n");
> -}
> -
> -static void counter_cb(struct perf_stat_config *config __maybe_unused,
> - struct evsel *counter, void *data,
> - bool first __maybe_unused)
> -{
> - struct aggr_data *ad = data;
> -
> - ad->val += perf_counts(counter->counts, ad->cpu_map_idx, 0)->val;
> - ad->ena += perf_counts(counter->counts, ad->cpu_map_idx, 0)->ena;
> - ad->run += perf_counts(counter->counts, ad->cpu_map_idx, 0)->run;
> -}
> -
> -/*
> - * Print out the results of a single counter:
> - * does not use aggregated count in system-wide
> - */
> static void print_counter(struct perf_stat_config *config,
> struct evsel *counter, char *prefix)
> {
> - FILE *output = config->output;
> - u64 ena, run, val;
> - double uval;
> - int idx;
> - struct perf_cpu cpu;
> - struct aggr_cpu_id id;
> -
> - perf_cpu_map__for_each_cpu(cpu, idx, evsel__cpus(counter)) {
> - struct aggr_data ad = { .cpu_map_idx = idx };
> -
> - if (!collect_data(config, counter, counter_cb, &ad))
> - return;
> - val = ad.val;
> - ena = ad.ena;
> - run = ad.run;
> + bool metric_only = config->metric_only;
> + bool first = false;
> + int s;
>
> - if (prefix)
> - fprintf(output, "%s", prefix);
> + /* AGGR_THREAD doesn't have config->aggr_get_id */
> + if (!config->aggr_map)
> + return;
>
> - uval = val * counter->scale;
> - id = aggr_cpu_id__cpu(cpu, /*data=*/NULL);
> - printout(config, id, 0, counter, uval, prefix,
> - run, ena, 1.0, &rt_stat);
> + if (counter->merged_stat)
> + return;
>
> - fputc('\n', output);
> + for (s = 0; s < config->aggr_map->nr; s++) {
> + print_counter_aggrdata(config, counter, s,
> + prefix, metric_only,
> + &first);
> }
> }
>
> @@ -1081,6 +787,7 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
> u64 ena, run, val;
> double uval;
> struct aggr_cpu_id id;
> + struct perf_stat_evsel *ps = counter->stats;
> int counter_idx = perf_cpu_map__idx(evsel__cpus(counter), cpu);
>
> if (counter_idx < 0)
> @@ -1093,13 +800,13 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
> aggr_printout(config, counter, id, 0);
> first = false;
> }
> - val = perf_counts(counter->counts, counter_idx, 0)->val;
> - ena = perf_counts(counter->counts, counter_idx, 0)->ena;
> - run = perf_counts(counter->counts, counter_idx, 0)->run;
> + val = ps->aggr[counter_idx].counts.val;
> + ena = ps->aggr[counter_idx].counts.ena;
> + run = ps->aggr[counter_idx].counts.run;
>
> uval = val * counter->scale;
> printout(config, id, 0, counter, uval, prefix,
> - run, ena, 1.0, &rt_stat);
> + run, ena, 1.0, &rt_stat, counter_idx);
> }
> if (!first)
> fputc('\n', config->output);
> @@ -1135,8 +842,8 @@ static void print_metric_headers(struct perf_stat_config *config,
> };
> bool first = true;
>
> - if (config->json_output && !config->interval)
> - fprintf(config->output, "{");
> + if (config->json_output && !config->interval)
> + fprintf(config->output, "{");
>
> if (prefix && !config->json_output)
> fprintf(config->output, "%s", prefix);
> @@ -1379,31 +1086,6 @@ static void print_footer(struct perf_stat_config *config)
> "the same PMU. Try reorganizing the group.\n");
> }
>
> -static void print_percore_thread(struct perf_stat_config *config,
> - struct evsel *counter, char *prefix)
> -{
> - int s;
> - struct aggr_cpu_id s2, id;
> - struct perf_cpu_map *cpus;
> - bool first = true;
> - int idx;
> - struct perf_cpu cpu;
> -
> - cpus = evsel__cpus(counter);
> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> - s2 = config->aggr_get_id(config, cpu);
> - for (s = 0; s < config->aggr_map->nr; s++) {
> - id = config->aggr_map->map[s];
> - if (aggr_cpu_id__equal(&s2, &id))
> - break;
> - }
> -
> - print_counter_aggrdata(config, counter, s,
> - prefix, false,
> - &first, cpu);
> - }
> -}
> -
> static void print_percore(struct perf_stat_config *config,
> struct evsel *counter, char *prefix)
> {
> @@ -1416,15 +1098,14 @@ static void print_percore(struct perf_stat_config *config,
> return;
>
> if (config->percore_show_thread)
> - return print_percore_thread(config, counter, prefix);
> + return print_counter(config, counter, prefix);
>
> for (s = 0; s < config->aggr_map->nr; s++) {
> if (prefix && metric_only)
> fprintf(output, "%s", prefix);
>
> print_counter_aggrdata(config, counter, s,
> - prefix, metric_only,
> - &first, (struct perf_cpu){ .cpu = -1 });
> + prefix, metric_only, &first);
> }
>
> if (metric_only)
> @@ -1469,17 +1150,13 @@ void evlist__print_counters(struct evlist *evlist, struct perf_stat_config *conf
> print_aggr(config, evlist, prefix);
> break;
> case AGGR_THREAD:
> - evlist__for_each_entry(evlist, counter) {
> - print_aggr_thread(config, _target, counter, prefix);
> - }
> - break;
> case AGGR_GLOBAL:
> if (config->iostat_run)
> iostat_print_counters(evlist, config, ts, prefix = buf,
> - print_counter_aggr);
> + print_counter);
> else {
> evlist__for_each_entry(evlist, counter) {
> - print_counter_aggr(config, counter, prefix);
> + print_counter(config, counter, prefix);
> }
> if (metric_only)
> fputc('\n', config->output);
> diff --git a/tools/perf/util/stat.c b/tools/perf/util/stat.c
> index c0955a0427ab..0316557adce9 100644
> --- a/tools/perf/util/stat.c
> +++ b/tools/perf/util/stat.c
> @@ -565,11 +565,6 @@ int perf_stat_process_counter(struct perf_stat_config *config,
> evsel__name(counter), count[0], count[1], count[2]);
> }
>
> - /*
> - * Save the full runtime - to allow normalization during printout:
> - */
> - perf_stat__update_shadow_stats(counter, *count, 0, &rt_stat);
> -
> return 0;
> }
>
> diff --git a/tools/perf/util/stat.h b/tools/perf/util/stat.h
> index 3d413ba8c68a..382a1ab92ce1 100644
> --- a/tools/perf/util/stat.h
> +++ b/tools/perf/util/stat.h
> @@ -224,15 +224,6 @@ static inline void update_rusage_stats(struct rusage_stats *ru_stats, struct rus
> struct evsel;
> struct evlist;
>
> -struct perf_aggr_thread_value {
> - struct evsel *counter;
> - struct aggr_cpu_id id;
> - double uval;
> - u64 val;
> - u64 run;
> - u64 ena;
> -};
> -
> bool __perf_stat_evsel__is(struct evsel *evsel, enum perf_stat_evsel_id id);
>
> #define perf_stat_evsel__is(evsel, id) \
> --
> 2.38.0.413.g74048e4d9e-goog
>
Hello,
On Mon, Dec 5, 2022 at 2:52 AM Athira Rajeev
<[email protected]> wrote:
>
>
>
> > On 18-Oct-2022, at 7:32 AM, Namhyung Kim <[email protected]> wrote:
> >
> > Now aggr counts are ready for use. Convert the display routines to use
> > the aggr counts and update the shadow stat with them. It doesn't need
> > to aggregate counts or collect aliases anymore during the display. Get
> > rid of now unused struct perf_aggr_thread_value.
> >
> > Note that there's a difference in the display order among the aggr mode.
> > For per-core/die/socket/node aggregation, it shows relevant events in
> > the same unit together, whereas global/thread/no aggregation it shows
> > the same events for different units together. So it still uses separate
> > codes to display them due to the ordering.
> >
> > One more thing to note is that it breaks per-core event display for now.
> > The next patch will fix it to have identical output as of now.
> >
> > Acked-by: Ian Rogers <[email protected]>
> > Signed-off-by: Namhyung Kim <[email protected]>
> > ---
> > tools/perf/util/stat-display.c | 421 ++++-----------------------------
> > tools/perf/util/stat.c | 5 -
> > tools/perf/util/stat.h | 9 -
> > 3 files changed, 49 insertions(+), 386 deletions(-)
> >
> > diff --git a/tools/perf/util/stat-display.c b/tools/perf/util/stat-display.c
> > index 4113aa86772f..bfae2784609c 100644
> > --- a/tools/perf/util/stat-display.c
> > +++ b/tools/perf/util/stat-display.c
> > @@ -442,31 +442,6 @@ static void print_metric_header(struct perf_stat_config *config,
> > fprintf(os->fh, "%*s ", config->metric_only_len, unit);
> > }
> >
> > -static int first_shadow_map_idx(struct perf_stat_config *config,
> > - struct evsel *evsel, const struct aggr_cpu_id *id)
> > -{
> > - struct perf_cpu_map *cpus = evsel__cpus(evsel);
> > - struct perf_cpu cpu;
> > - int idx;
> > -
> > - if (config->aggr_mode == AGGR_NONE)
> > - return perf_cpu_map__idx(cpus, id->cpu);
> > -
> > - if (config->aggr_mode == AGGR_THREAD)
> > - return id->thread_idx;
> > -
> > - if (!config->aggr_get_id)
> > - return 0;
> > -
> > - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> > - struct aggr_cpu_id cpu_id = config->aggr_get_id(config, cpu);
> > -
> > - if (aggr_cpu_id__equal(&cpu_id, id))
> > - return idx;
> > - }
> > - return 0;
> > -}
> > -
> > static void abs_printout(struct perf_stat_config *config,
> > struct aggr_cpu_id id, int nr, struct evsel *evsel, double avg)
> > {
> > @@ -537,7 +512,7 @@ static bool is_mixed_hw_group(struct evsel *counter)
> > static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int nr,
> > struct evsel *counter, double uval,
> > char *prefix, u64 run, u64 ena, double noise,
> > - struct runtime_stat *st)
> > + struct runtime_stat *st, int map_idx)
> > {
> > struct perf_stat_output_ctx out;
> > struct outstate os = {
> > @@ -648,8 +623,7 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
> > print_running(config, run, ena);
> > }
> >
> > - perf_stat__print_shadow_stats(config, counter, uval,
> > - first_shadow_map_idx(config, counter, &id),
> > + perf_stat__print_shadow_stats(config, counter, uval, map_idx,
> > &out, &config->metric_events, st);
> > if (!config->csv_output && !config->metric_only && !config->json_output) {
> > print_noise(config, counter, noise);
> > @@ -657,34 +631,6 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
> > }
> > }
> >
> > -static void aggr_update_shadow(struct perf_stat_config *config,
> > - struct evlist *evlist)
> > -{
> > - int idx, s;
> > - struct perf_cpu cpu;
> > - struct aggr_cpu_id s2, id;
> > - u64 val;
> > - struct evsel *counter;
> > - struct perf_cpu_map *cpus;
> > -
> > - for (s = 0; s < config->aggr_map->nr; s++) {
> > - id = config->aggr_map->map[s];
> > - evlist__for_each_entry(evlist, counter) {
> > - cpus = evsel__cpus(counter);
> > - val = 0;
> > - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> > - s2 = config->aggr_get_id(config, cpu);
> > - if (!aggr_cpu_id__equal(&s2, &id))
> > - continue;
> > - val += perf_counts(counter->counts, idx, 0)->val;
> > - }
> > - perf_stat__update_shadow_stats(counter, val,
> > - first_shadow_map_idx(config, counter, &id),
> > - &rt_stat);
> > - }
> > - }
> > -}
> > -
> > static void uniquify_event_name(struct evsel *counter)
> > {
> > char *new_name;
> > @@ -721,137 +667,51 @@ static void uniquify_event_name(struct evsel *counter)
> > counter->uniquified_name = true;
> > }
> >
> > -static void collect_all_aliases(struct perf_stat_config *config, struct evsel *counter,
> > - void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
> > - bool first),
> > - void *data)
> > +static bool hybrid_uniquify(struct evsel *evsel, struct perf_stat_config *config)
> > {
> > - struct evlist *evlist = counter->evlist;
> > - struct evsel *alias;
> > -
> > - alias = list_prepare_entry(counter, &(evlist->core.entries), core.node);
> > - list_for_each_entry_continue (alias, &evlist->core.entries, core.node) {
> > - /* Merge events with the same name, etc. but on different PMUs. */
> > - if (!strcmp(evsel__name(alias), evsel__name(counter)) &&
> > - alias->scale == counter->scale &&
> > - alias->cgrp == counter->cgrp &&
> > - !strcmp(alias->unit, counter->unit) &&
> > - evsel__is_clock(alias) == evsel__is_clock(counter) &&
> > - strcmp(alias->pmu_name, counter->pmu_name)) {
> > - alias->merged_stat = true;
> > - cb(config, alias, data, false);
> > - }
> > - }
> > + return evsel__is_hybrid(evsel) && !config->hybrid_merge;
> > }
> >
> > -static bool hybrid_merge(struct evsel *counter, struct perf_stat_config *config,
> > - bool check)
> > +static void uniquify_counter(struct perf_stat_config *config, struct evsel *counter)
> > {
> > - if (evsel__is_hybrid(counter)) {
> > - if (check)
> > - return config->hybrid_merge;
> > - else
> > - return !config->hybrid_merge;
> > - }
> > -
> > - return false;
> > -}
> > -
> > -static bool collect_data(struct perf_stat_config *config, struct evsel *counter,
> > - void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
> > - bool first),
> > - void *data)
> > -{
> > - if (counter->merged_stat)
> > - return false;
> > - cb(config, counter, data, true);
> > - if (config->no_merge || hybrid_merge(counter, config, false))
> > + if (config->no_merge || hybrid_uniquify(counter, config))
> > uniquify_event_name(counter);
> > - else if (counter->auto_merge_stats || hybrid_merge(counter, config, true))
> > - collect_all_aliases(config, counter, cb, data);
> > - return true;
> > -}
> > -
> > -struct aggr_data {
> > - u64 ena, run, val;
> > - struct aggr_cpu_id id;
> > - int nr;
> > - int cpu_map_idx;
> > -};
> > -
> > -static void aggr_cb(struct perf_stat_config *config,
> > - struct evsel *counter, void *data, bool first)
> > -{
> > - struct aggr_data *ad = data;
> > - int idx;
> > - struct perf_cpu cpu;
> > - struct perf_cpu_map *cpus;
> > - struct aggr_cpu_id s2;
> > -
> > - cpus = evsel__cpus(counter);
> > - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> > - struct perf_counts_values *counts;
> > -
> > - s2 = config->aggr_get_id(config, cpu);
> > - if (!aggr_cpu_id__equal(&s2, &ad->id))
> > - continue;
> > - if (first)
> > - ad->nr++;
> > - counts = perf_counts(counter->counts, idx, 0);
> > - /*
> > - * When any result is bad, make them all to give
> > - * consistent output in interval mode.
> > - */
> > - if (counts->ena == 0 || counts->run == 0 ||
> > - counter->counts->scaled == -1) {
> > - ad->ena = 0;
> > - ad->run = 0;
> > - break;
> > - }
> > - ad->val += counts->val;
> > - ad->ena += counts->ena;
> > - ad->run += counts->run;
> > - }
> > }
> >
> > static void print_counter_aggrdata(struct perf_stat_config *config,
> > struct evsel *counter, int s,
> > char *prefix, bool metric_only,
> > - bool *first, struct perf_cpu cpu)
> > + bool *first)
> > {
> > - struct aggr_data ad;
> > FILE *output = config->output;
> > u64 ena, run, val;
> > - int nr;
> > - struct aggr_cpu_id id;
> > double uval;
> > + struct perf_stat_evsel *ps = counter->stats;
> > + struct perf_stat_aggr *aggr = &ps->aggr[s];
> > + struct aggr_cpu_id id = config->aggr_map->map[s];
> > + double avg = aggr->counts.val;
> >
> > - ad.id = id = config->aggr_map->map[s];
> > - ad.val = ad.ena = ad.run = 0;
> > - ad.nr = 0;
> > - if (!collect_data(config, counter, aggr_cb, &ad))
> > + if (counter->supported && aggr->nr == 0)
> > return;
>
> Hi Namhyung,
>
> Observed one issue with this change in tmp.perf/core branch of git://git.kernel.org/pub/scm/linux/kernel/git/acme/linux.git
> Patch link: https://git.kernel.org/pub/scm/linux/kernel/git/acme/linux.git/commit/?h=tmp.perf/core&id=91f85f98da7ab8c32105f42dd03884c01ec4498f
>
> If we have multiple events in a group and if all events in that group can't go in at once, the group will fail to schedule.
> The perf stat result used to have status of all events in group.
>
> Example result from powerpc:
> # perf stat -e '{r1001e,r401e0,r4c010}' <workload>
> Performance counter stats for 'workload’:
>
> <not counted> r1001e
> <not counted> r401e0
> <not supported> r4c010
>
> But after this change, it shows only the event which broke the constraint.
>
> # ./perf stat -e '{r1001e,r401e0,r4c010}' <workload>
> Performance counter stats for 'workload’:
> <not supported> r4c010
>
> The new condition added by the patch is:
>
> if (counter->supported && aggr->nr == 0)
> return;
>
>
> The first two events r1001e and r401e0 are good to go in this group, hence
> counter->supported is true where as aggr->nr is zero since this group is
> not scheduled. Since counter->supported is false for third event which
> broke the group constraint, the condition won't return and hence only prints
> this event in output.
>
> Namhyung, is there any scenario because of which this condition was added ?
> If not I would go ahead and sent a fix patch to remove this check.
I remember it's for AGGR_THREAD and merged (uncore) aliases and hybrid
events. In system-wide per-thread mode, many of them would get 0 results
and we'd like to skip them. And now I see that we have evsel->merged_stat
check already.
Maybe we could just add the aggr_mode check there.
Thanks,
Namhyung
> On 06-Dec-2022, at 12:30 AM, Namhyung Kim <[email protected]> wrote:
>
> Hello,
>
> On Mon, Dec 5, 2022 at 2:52 AM Athira Rajeev
> <[email protected]> wrote:
>>
>>
>>
>>> On 18-Oct-2022, at 7:32 AM, Namhyung Kim <[email protected]> wrote:
>>>
>>> Now aggr counts are ready for use. Convert the display routines to use
>>> the aggr counts and update the shadow stat with them. It doesn't need
>>> to aggregate counts or collect aliases anymore during the display. Get
>>> rid of now unused struct perf_aggr_thread_value.
>>>
>>> Note that there's a difference in the display order among the aggr mode.
>>> For per-core/die/socket/node aggregation, it shows relevant events in
>>> the same unit together, whereas global/thread/no aggregation it shows
>>> the same events for different units together. So it still uses separate
>>> codes to display them due to the ordering.
>>>
>>> One more thing to note is that it breaks per-core event display for now.
>>> The next patch will fix it to have identical output as of now.
>>>
>>> Acked-by: Ian Rogers <[email protected]>
>>> Signed-off-by: Namhyung Kim <[email protected]>
>>> ---
>>> tools/perf/util/stat-display.c | 421 ++++-----------------------------
>>> tools/perf/util/stat.c | 5 -
>>> tools/perf/util/stat.h | 9 -
>>> 3 files changed, 49 insertions(+), 386 deletions(-)
>>>
>>> diff --git a/tools/perf/util/stat-display.c b/tools/perf/util/stat-display.c
>>> index 4113aa86772f..bfae2784609c 100644
>>> --- a/tools/perf/util/stat-display.c
>>> +++ b/tools/perf/util/stat-display.c
>>> @@ -442,31 +442,6 @@ static void print_metric_header(struct perf_stat_config *config,
>>> fprintf(os->fh, "%*s ", config->metric_only_len, unit);
>>> }
>>>
>>> -static int first_shadow_map_idx(struct perf_stat_config *config,
>>> - struct evsel *evsel, const struct aggr_cpu_id *id)
>>> -{
>>> - struct perf_cpu_map *cpus = evsel__cpus(evsel);
>>> - struct perf_cpu cpu;
>>> - int idx;
>>> -
>>> - if (config->aggr_mode == AGGR_NONE)
>>> - return perf_cpu_map__idx(cpus, id->cpu);
>>> -
>>> - if (config->aggr_mode == AGGR_THREAD)
>>> - return id->thread_idx;
>>> -
>>> - if (!config->aggr_get_id)
>>> - return 0;
>>> -
>>> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
>>> - struct aggr_cpu_id cpu_id = config->aggr_get_id(config, cpu);
>>> -
>>> - if (aggr_cpu_id__equal(&cpu_id, id))
>>> - return idx;
>>> - }
>>> - return 0;
>>> -}
>>> -
>>> static void abs_printout(struct perf_stat_config *config,
>>> struct aggr_cpu_id id, int nr, struct evsel *evsel, double avg)
>>> {
>>> @@ -537,7 +512,7 @@ static bool is_mixed_hw_group(struct evsel *counter)
>>> static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int nr,
>>> struct evsel *counter, double uval,
>>> char *prefix, u64 run, u64 ena, double noise,
>>> - struct runtime_stat *st)
>>> + struct runtime_stat *st, int map_idx)
>>> {
>>> struct perf_stat_output_ctx out;
>>> struct outstate os = {
>>> @@ -648,8 +623,7 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
>>> print_running(config, run, ena);
>>> }
>>>
>>> - perf_stat__print_shadow_stats(config, counter, uval,
>>> - first_shadow_map_idx(config, counter, &id),
>>> + perf_stat__print_shadow_stats(config, counter, uval, map_idx,
>>> &out, &config->metric_events, st);
>>> if (!config->csv_output && !config->metric_only && !config->json_output) {
>>> print_noise(config, counter, noise);
>>> @@ -657,34 +631,6 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
>>> }
>>> }
>>>
>>> -static void aggr_update_shadow(struct perf_stat_config *config,
>>> - struct evlist *evlist)
>>> -{
>>> - int idx, s;
>>> - struct perf_cpu cpu;
>>> - struct aggr_cpu_id s2, id;
>>> - u64 val;
>>> - struct evsel *counter;
>>> - struct perf_cpu_map *cpus;
>>> -
>>> - for (s = 0; s < config->aggr_map->nr; s++) {
>>> - id = config->aggr_map->map[s];
>>> - evlist__for_each_entry(evlist, counter) {
>>> - cpus = evsel__cpus(counter);
>>> - val = 0;
>>> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
>>> - s2 = config->aggr_get_id(config, cpu);
>>> - if (!aggr_cpu_id__equal(&s2, &id))
>>> - continue;
>>> - val += perf_counts(counter->counts, idx, 0)->val;
>>> - }
>>> - perf_stat__update_shadow_stats(counter, val,
>>> - first_shadow_map_idx(config, counter, &id),
>>> - &rt_stat);
>>> - }
>>> - }
>>> -}
>>> -
>>> static void uniquify_event_name(struct evsel *counter)
>>> {
>>> char *new_name;
>>> @@ -721,137 +667,51 @@ static void uniquify_event_name(struct evsel *counter)
>>> counter->uniquified_name = true;
>>> }
>>>
>>> -static void collect_all_aliases(struct perf_stat_config *config, struct evsel *counter,
>>> - void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
>>> - bool first),
>>> - void *data)
>>> +static bool hybrid_uniquify(struct evsel *evsel, struct perf_stat_config *config)
>>> {
>>> - struct evlist *evlist = counter->evlist;
>>> - struct evsel *alias;
>>> -
>>> - alias = list_prepare_entry(counter, &(evlist->core.entries), core.node);
>>> - list_for_each_entry_continue (alias, &evlist->core.entries, core.node) {
>>> - /* Merge events with the same name, etc. but on different PMUs. */
>>> - if (!strcmp(evsel__name(alias), evsel__name(counter)) &&
>>> - alias->scale == counter->scale &&
>>> - alias->cgrp == counter->cgrp &&
>>> - !strcmp(alias->unit, counter->unit) &&
>>> - evsel__is_clock(alias) == evsel__is_clock(counter) &&
>>> - strcmp(alias->pmu_name, counter->pmu_name)) {
>>> - alias->merged_stat = true;
>>> - cb(config, alias, data, false);
>>> - }
>>> - }
>>> + return evsel__is_hybrid(evsel) && !config->hybrid_merge;
>>> }
>>>
>>> -static bool hybrid_merge(struct evsel *counter, struct perf_stat_config *config,
>>> - bool check)
>>> +static void uniquify_counter(struct perf_stat_config *config, struct evsel *counter)
>>> {
>>> - if (evsel__is_hybrid(counter)) {
>>> - if (check)
>>> - return config->hybrid_merge;
>>> - else
>>> - return !config->hybrid_merge;
>>> - }
>>> -
>>> - return false;
>>> -}
>>> -
>>> -static bool collect_data(struct perf_stat_config *config, struct evsel *counter,
>>> - void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
>>> - bool first),
>>> - void *data)
>>> -{
>>> - if (counter->merged_stat)
>>> - return false;
>>> - cb(config, counter, data, true);
>>> - if (config->no_merge || hybrid_merge(counter, config, false))
>>> + if (config->no_merge || hybrid_uniquify(counter, config))
>>> uniquify_event_name(counter);
>>> - else if (counter->auto_merge_stats || hybrid_merge(counter, config, true))
>>> - collect_all_aliases(config, counter, cb, data);
>>> - return true;
>>> -}
>>> -
>>> -struct aggr_data {
>>> - u64 ena, run, val;
>>> - struct aggr_cpu_id id;
>>> - int nr;
>>> - int cpu_map_idx;
>>> -};
>>> -
>>> -static void aggr_cb(struct perf_stat_config *config,
>>> - struct evsel *counter, void *data, bool first)
>>> -{
>>> - struct aggr_data *ad = data;
>>> - int idx;
>>> - struct perf_cpu cpu;
>>> - struct perf_cpu_map *cpus;
>>> - struct aggr_cpu_id s2;
>>> -
>>> - cpus = evsel__cpus(counter);
>>> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
>>> - struct perf_counts_values *counts;
>>> -
>>> - s2 = config->aggr_get_id(config, cpu);
>>> - if (!aggr_cpu_id__equal(&s2, &ad->id))
>>> - continue;
>>> - if (first)
>>> - ad->nr++;
>>> - counts = perf_counts(counter->counts, idx, 0);
>>> - /*
>>> - * When any result is bad, make them all to give
>>> - * consistent output in interval mode.
>>> - */
>>> - if (counts->ena == 0 || counts->run == 0 ||
>>> - counter->counts->scaled == -1) {
>>> - ad->ena = 0;
>>> - ad->run = 0;
>>> - break;
>>> - }
>>> - ad->val += counts->val;
>>> - ad->ena += counts->ena;
>>> - ad->run += counts->run;
>>> - }
>>> }
>>>
>>> static void print_counter_aggrdata(struct perf_stat_config *config,
>>> struct evsel *counter, int s,
>>> char *prefix, bool metric_only,
>>> - bool *first, struct perf_cpu cpu)
>>> + bool *first)
>>> {
>>> - struct aggr_data ad;
>>> FILE *output = config->output;
>>> u64 ena, run, val;
>>> - int nr;
>>> - struct aggr_cpu_id id;
>>> double uval;
>>> + struct perf_stat_evsel *ps = counter->stats;
>>> + struct perf_stat_aggr *aggr = &ps->aggr[s];
>>> + struct aggr_cpu_id id = config->aggr_map->map[s];
>>> + double avg = aggr->counts.val;
>>>
>>> - ad.id = id = config->aggr_map->map[s];
>>> - ad.val = ad.ena = ad.run = 0;
>>> - ad.nr = 0;
>>> - if (!collect_data(config, counter, aggr_cb, &ad))
>>> + if (counter->supported && aggr->nr == 0)
>>> return;
>>
>> Hi Namhyung,
>>
>> Observed one issue with this change in tmp.perf/core branch of git://git.kernel.org/pub/scm/linux/kernel/git/acme/linux.git
>> Patch link: https://git.kernel.org/pub/scm/linux/kernel/git/acme/linux.git/commit/?h=tmp.perf/core&id=91f85f98da7ab8c32105f42dd03884c01ec4498f
>>
>> If we have multiple events in a group and if all events in that group can't go in at once, the group will fail to schedule.
>> The perf stat result used to have status of all events in group.
>>
>> Example result from powerpc:
>> # perf stat -e '{r1001e,r401e0,r4c010}' <workload>
>> Performance counter stats for 'workload’:
>>
>> <not counted> r1001e
>> <not counted> r401e0
>> <not supported> r4c010
>>
>> But after this change, it shows only the event which broke the constraint.
>>
>> # ./perf stat -e '{r1001e,r401e0,r4c010}' <workload>
>> Performance counter stats for 'workload’:
>> <not supported> r4c010
>>
>> The new condition added by the patch is:
>>
>> if (counter->supported && aggr->nr == 0)
>> return;
>>
>>
>> The first two events r1001e and r401e0 are good to go in this group, hence
>> counter->supported is true where as aggr->nr is zero since this group is
>> not scheduled. Since counter->supported is false for third event which
>> broke the group constraint, the condition won't return and hence only prints
>> this event in output.
>>
>> Namhyung, is there any scenario because of which this condition was added ?
>> If not I would go ahead and sent a fix patch to remove this check.
>
> I remember it's for AGGR_THREAD and merged (uncore) aliases and hybrid
> events. In system-wide per-thread mode, many of them would get 0 results
> and we'd like to skip them. And now I see that we have evsel->merged_stat
> check already.
>
> Maybe we could just add the aggr_mode check there.
Thanks for the response Namhyung.
I will share my response in patch "[PATCH] perf stat: Update event skip condition”
Thanks
Athira
>
> Thanks,
> Namhyung
On Mon, Oct 17, 2022 at 07:02:25PM -0700, Namhyung Kim wrote:
> Now aggr counts are ready for use. Convert the display routines to use
> the aggr counts and update the shadow stat with them. It doesn't need
> to aggregate counts or collect aliases anymore during the display. Get
> rid of now unused struct perf_aggr_thread_value.
>
> Note that there's a difference in the display order among the aggr mode.
> For per-core/die/socket/node aggregation, it shows relevant events in
> the same unit together, whereas global/thread/no aggregation it shows
> the same events for different units together. So it still uses separate
> codes to display them due to the ordering.
>
> One more thing to note is that it breaks per-core event display for now.
> The next patch will fix it to have identical output as of now.
>
> Acked-by: Ian Rogers <[email protected]>
> Signed-off-by: Namhyung Kim <[email protected]>
hi,
this one seems to break 'perf stat -r X' not sure why so far..
final counts seems to be accumulated instead of displaying average, like:
with this patch:
Performance counter stats for './test_progs -n 103/1' (2 runs):
206,815,929 cycles:u ( +- 0.05% )
16,052,747,533 cycles:k ( +- 0.10% )
16,259,643,167 cycles ( +- 0.10% )
1.98093 +- 0.00586 seconds time elapsed ( +- 0.30% )
without this patch:
Performance counter stats for './test_progs -n 103/1' (2 runs):
103,300,812 cycles:u ( +- 0.37% )
8,016,856,866 cycles:k ( +- 0.32% )
8,120,200,572 cycles ( +- 0.32% )
1.97272 +- 0.00270 seconds time elapsed ( +- 0.14% )
any idea? ;-)
thanks,
jirka
> ---
> tools/perf/util/stat-display.c | 421 ++++-----------------------------
> tools/perf/util/stat.c | 5 -
> tools/perf/util/stat.h | 9 -
> 3 files changed, 49 insertions(+), 386 deletions(-)
>
> diff --git a/tools/perf/util/stat-display.c b/tools/perf/util/stat-display.c
> index 4113aa86772f..bfae2784609c 100644
> --- a/tools/perf/util/stat-display.c
> +++ b/tools/perf/util/stat-display.c
> @@ -442,31 +442,6 @@ static void print_metric_header(struct perf_stat_config *config,
> fprintf(os->fh, "%*s ", config->metric_only_len, unit);
> }
>
> -static int first_shadow_map_idx(struct perf_stat_config *config,
> - struct evsel *evsel, const struct aggr_cpu_id *id)
> -{
> - struct perf_cpu_map *cpus = evsel__cpus(evsel);
> - struct perf_cpu cpu;
> - int idx;
> -
> - if (config->aggr_mode == AGGR_NONE)
> - return perf_cpu_map__idx(cpus, id->cpu);
> -
> - if (config->aggr_mode == AGGR_THREAD)
> - return id->thread_idx;
> -
> - if (!config->aggr_get_id)
> - return 0;
> -
> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> - struct aggr_cpu_id cpu_id = config->aggr_get_id(config, cpu);
> -
> - if (aggr_cpu_id__equal(&cpu_id, id))
> - return idx;
> - }
> - return 0;
> -}
> -
> static void abs_printout(struct perf_stat_config *config,
> struct aggr_cpu_id id, int nr, struct evsel *evsel, double avg)
> {
> @@ -537,7 +512,7 @@ static bool is_mixed_hw_group(struct evsel *counter)
> static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int nr,
> struct evsel *counter, double uval,
> char *prefix, u64 run, u64 ena, double noise,
> - struct runtime_stat *st)
> + struct runtime_stat *st, int map_idx)
> {
> struct perf_stat_output_ctx out;
> struct outstate os = {
> @@ -648,8 +623,7 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
> print_running(config, run, ena);
> }
>
> - perf_stat__print_shadow_stats(config, counter, uval,
> - first_shadow_map_idx(config, counter, &id),
> + perf_stat__print_shadow_stats(config, counter, uval, map_idx,
> &out, &config->metric_events, st);
> if (!config->csv_output && !config->metric_only && !config->json_output) {
> print_noise(config, counter, noise);
> @@ -657,34 +631,6 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
> }
> }
>
> -static void aggr_update_shadow(struct perf_stat_config *config,
> - struct evlist *evlist)
> -{
> - int idx, s;
> - struct perf_cpu cpu;
> - struct aggr_cpu_id s2, id;
> - u64 val;
> - struct evsel *counter;
> - struct perf_cpu_map *cpus;
> -
> - for (s = 0; s < config->aggr_map->nr; s++) {
> - id = config->aggr_map->map[s];
> - evlist__for_each_entry(evlist, counter) {
> - cpus = evsel__cpus(counter);
> - val = 0;
> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> - s2 = config->aggr_get_id(config, cpu);
> - if (!aggr_cpu_id__equal(&s2, &id))
> - continue;
> - val += perf_counts(counter->counts, idx, 0)->val;
> - }
> - perf_stat__update_shadow_stats(counter, val,
> - first_shadow_map_idx(config, counter, &id),
> - &rt_stat);
> - }
> - }
> -}
> -
> static void uniquify_event_name(struct evsel *counter)
> {
> char *new_name;
> @@ -721,137 +667,51 @@ static void uniquify_event_name(struct evsel *counter)
> counter->uniquified_name = true;
> }
>
> -static void collect_all_aliases(struct perf_stat_config *config, struct evsel *counter,
> - void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
> - bool first),
> - void *data)
> +static bool hybrid_uniquify(struct evsel *evsel, struct perf_stat_config *config)
> {
> - struct evlist *evlist = counter->evlist;
> - struct evsel *alias;
> -
> - alias = list_prepare_entry(counter, &(evlist->core.entries), core.node);
> - list_for_each_entry_continue (alias, &evlist->core.entries, core.node) {
> - /* Merge events with the same name, etc. but on different PMUs. */
> - if (!strcmp(evsel__name(alias), evsel__name(counter)) &&
> - alias->scale == counter->scale &&
> - alias->cgrp == counter->cgrp &&
> - !strcmp(alias->unit, counter->unit) &&
> - evsel__is_clock(alias) == evsel__is_clock(counter) &&
> - strcmp(alias->pmu_name, counter->pmu_name)) {
> - alias->merged_stat = true;
> - cb(config, alias, data, false);
> - }
> - }
> + return evsel__is_hybrid(evsel) && !config->hybrid_merge;
> }
>
> -static bool hybrid_merge(struct evsel *counter, struct perf_stat_config *config,
> - bool check)
> +static void uniquify_counter(struct perf_stat_config *config, struct evsel *counter)
> {
> - if (evsel__is_hybrid(counter)) {
> - if (check)
> - return config->hybrid_merge;
> - else
> - return !config->hybrid_merge;
> - }
> -
> - return false;
> -}
> -
> -static bool collect_data(struct perf_stat_config *config, struct evsel *counter,
> - void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
> - bool first),
> - void *data)
> -{
> - if (counter->merged_stat)
> - return false;
> - cb(config, counter, data, true);
> - if (config->no_merge || hybrid_merge(counter, config, false))
> + if (config->no_merge || hybrid_uniquify(counter, config))
> uniquify_event_name(counter);
> - else if (counter->auto_merge_stats || hybrid_merge(counter, config, true))
> - collect_all_aliases(config, counter, cb, data);
> - return true;
> -}
> -
> -struct aggr_data {
> - u64 ena, run, val;
> - struct aggr_cpu_id id;
> - int nr;
> - int cpu_map_idx;
> -};
> -
> -static void aggr_cb(struct perf_stat_config *config,
> - struct evsel *counter, void *data, bool first)
> -{
> - struct aggr_data *ad = data;
> - int idx;
> - struct perf_cpu cpu;
> - struct perf_cpu_map *cpus;
> - struct aggr_cpu_id s2;
> -
> - cpus = evsel__cpus(counter);
> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> - struct perf_counts_values *counts;
> -
> - s2 = config->aggr_get_id(config, cpu);
> - if (!aggr_cpu_id__equal(&s2, &ad->id))
> - continue;
> - if (first)
> - ad->nr++;
> - counts = perf_counts(counter->counts, idx, 0);
> - /*
> - * When any result is bad, make them all to give
> - * consistent output in interval mode.
> - */
> - if (counts->ena == 0 || counts->run == 0 ||
> - counter->counts->scaled == -1) {
> - ad->ena = 0;
> - ad->run = 0;
> - break;
> - }
> - ad->val += counts->val;
> - ad->ena += counts->ena;
> - ad->run += counts->run;
> - }
> }
>
> static void print_counter_aggrdata(struct perf_stat_config *config,
> struct evsel *counter, int s,
> char *prefix, bool metric_only,
> - bool *first, struct perf_cpu cpu)
> + bool *first)
> {
> - struct aggr_data ad;
> FILE *output = config->output;
> u64 ena, run, val;
> - int nr;
> - struct aggr_cpu_id id;
> double uval;
> + struct perf_stat_evsel *ps = counter->stats;
> + struct perf_stat_aggr *aggr = &ps->aggr[s];
> + struct aggr_cpu_id id = config->aggr_map->map[s];
> + double avg = aggr->counts.val;
>
> - ad.id = id = config->aggr_map->map[s];
> - ad.val = ad.ena = ad.run = 0;
> - ad.nr = 0;
> - if (!collect_data(config, counter, aggr_cb, &ad))
> + if (counter->supported && aggr->nr == 0)
> return;
>
> - if (perf_pmu__has_hybrid() && ad.ena == 0)
> - return;
> + uniquify_counter(config, counter);
> +
> + val = aggr->counts.val;
> + ena = aggr->counts.ena;
> + run = aggr->counts.run;
>
> - nr = ad.nr;
> - ena = ad.ena;
> - run = ad.run;
> - val = ad.val;
> if (*first && metric_only) {
> *first = false;
> - aggr_printout(config, counter, id, nr);
> + aggr_printout(config, counter, id, aggr->nr);
> }
> if (prefix && !metric_only)
> fprintf(output, "%s", prefix);
>
> uval = val * counter->scale;
> - if (cpu.cpu != -1)
> - id = aggr_cpu_id__cpu(cpu, /*data=*/NULL);
>
> - printout(config, id, nr, counter, uval,
> - prefix, run, ena, 1.0, &rt_stat);
> + printout(config, id, aggr->nr, counter, uval,
> + prefix, run, ena, avg, &rt_stat, s);
> +
> if (!metric_only)
> fputc('\n', output);
> }
> @@ -869,8 +729,6 @@ static void print_aggr(struct perf_stat_config *config,
> if (!config->aggr_map || !config->aggr_get_id)
> return;
>
> - aggr_update_shadow(config, evlist);
> -
> /*
> * With metric_only everything is on a single line.
> * Without each counter has its own line.
> @@ -881,188 +739,36 @@ static void print_aggr(struct perf_stat_config *config,
>
> first = true;
> evlist__for_each_entry(evlist, counter) {
> + if (counter->merged_stat)
> + continue;
> +
> print_counter_aggrdata(config, counter, s,
> - prefix, metric_only,
> - &first, (struct perf_cpu){ .cpu = -1 });
> + prefix, metric_only,
> + &first);
> }
> if (metric_only)
> fputc('\n', output);
> }
> }
>
> -static int cmp_val(const void *a, const void *b)
> -{
> - return ((struct perf_aggr_thread_value *)b)->val -
> - ((struct perf_aggr_thread_value *)a)->val;
> -}
> -
> -static struct perf_aggr_thread_value *sort_aggr_thread(
> - struct evsel *counter,
> - int *ret,
> - struct target *_target)
> -{
> - int nthreads = perf_thread_map__nr(counter->core.threads);
> - int i = 0;
> - double uval;
> - struct perf_aggr_thread_value *buf;
> -
> - buf = calloc(nthreads, sizeof(struct perf_aggr_thread_value));
> - if (!buf)
> - return NULL;
> -
> - for (int thread = 0; thread < nthreads; thread++) {
> - int idx;
> - u64 ena = 0, run = 0, val = 0;
> -
> - perf_cpu_map__for_each_idx(idx, evsel__cpus(counter)) {
> - struct perf_counts_values *counts =
> - perf_counts(counter->counts, idx, thread);
> -
> - val += counts->val;
> - ena += counts->ena;
> - run += counts->run;
> - }
> -
> - uval = val * counter->scale;
> -
> - /*
> - * Skip value 0 when enabling --per-thread globally,
> - * otherwise too many 0 output.
> - */
> - if (uval == 0.0 && target__has_per_thread(_target))
> - continue;
> -
> - buf[i].counter = counter;
> - buf[i].id = aggr_cpu_id__empty();
> - buf[i].id.thread_idx = thread;
> - buf[i].uval = uval;
> - buf[i].val = val;
> - buf[i].run = run;
> - buf[i].ena = ena;
> - i++;
> - }
> -
> - qsort(buf, i, sizeof(struct perf_aggr_thread_value), cmp_val);
> -
> - if (ret)
> - *ret = i;
> -
> - return buf;
> -}
> -
> -static void print_aggr_thread(struct perf_stat_config *config,
> - struct target *_target,
> - struct evsel *counter, char *prefix)
> -{
> - FILE *output = config->output;
> - int thread, sorted_threads;
> - struct aggr_cpu_id id;
> - struct perf_aggr_thread_value *buf;
> -
> - buf = sort_aggr_thread(counter, &sorted_threads, _target);
> - if (!buf) {
> - perror("cannot sort aggr thread");
> - return;
> - }
> -
> - for (thread = 0; thread < sorted_threads; thread++) {
> - if (prefix)
> - fprintf(output, "%s", prefix);
> -
> - id = buf[thread].id;
> - printout(config, id, 0, buf[thread].counter, buf[thread].uval,
> - prefix, buf[thread].run, buf[thread].ena, 1.0,
> - &rt_stat);
> - fputc('\n', output);
> - }
> -
> - free(buf);
> -}
> -
> -struct caggr_data {
> - double avg, avg_enabled, avg_running;
> -};
> -
> -static void counter_aggr_cb(struct perf_stat_config *config __maybe_unused,
> - struct evsel *counter, void *data,
> - bool first __maybe_unused)
> -{
> - struct caggr_data *cd = data;
> - struct perf_counts_values *aggr = &counter->counts->aggr;
> -
> - cd->avg += aggr->val;
> - cd->avg_enabled += aggr->ena;
> - cd->avg_running += aggr->run;
> -}
> -
> -/*
> - * Print out the results of a single counter:
> - * aggregated counts in system-wide mode
> - */
> -static void print_counter_aggr(struct perf_stat_config *config,
> - struct evsel *counter, char *prefix)
> -{
> - bool metric_only = config->metric_only;
> - FILE *output = config->output;
> - double uval;
> - struct caggr_data cd = { .avg = 0.0 };
> -
> - if (!collect_data(config, counter, counter_aggr_cb, &cd))
> - return;
> -
> - if (prefix && !metric_only)
> - fprintf(output, "%s", prefix);
> -
> - uval = cd.avg * counter->scale;
> - printout(config, aggr_cpu_id__empty(), 0, counter, uval, prefix, cd.avg_running,
> - cd.avg_enabled, cd.avg, &rt_stat);
> - if (!metric_only)
> - fprintf(output, "\n");
> -}
> -
> -static void counter_cb(struct perf_stat_config *config __maybe_unused,
> - struct evsel *counter, void *data,
> - bool first __maybe_unused)
> -{
> - struct aggr_data *ad = data;
> -
> - ad->val += perf_counts(counter->counts, ad->cpu_map_idx, 0)->val;
> - ad->ena += perf_counts(counter->counts, ad->cpu_map_idx, 0)->ena;
> - ad->run += perf_counts(counter->counts, ad->cpu_map_idx, 0)->run;
> -}
> -
> -/*
> - * Print out the results of a single counter:
> - * does not use aggregated count in system-wide
> - */
> static void print_counter(struct perf_stat_config *config,
> struct evsel *counter, char *prefix)
> {
> - FILE *output = config->output;
> - u64 ena, run, val;
> - double uval;
> - int idx;
> - struct perf_cpu cpu;
> - struct aggr_cpu_id id;
> -
> - perf_cpu_map__for_each_cpu(cpu, idx, evsel__cpus(counter)) {
> - struct aggr_data ad = { .cpu_map_idx = idx };
> -
> - if (!collect_data(config, counter, counter_cb, &ad))
> - return;
> - val = ad.val;
> - ena = ad.ena;
> - run = ad.run;
> + bool metric_only = config->metric_only;
> + bool first = false;
> + int s;
>
> - if (prefix)
> - fprintf(output, "%s", prefix);
> + /* AGGR_THREAD doesn't have config->aggr_get_id */
> + if (!config->aggr_map)
> + return;
>
> - uval = val * counter->scale;
> - id = aggr_cpu_id__cpu(cpu, /*data=*/NULL);
> - printout(config, id, 0, counter, uval, prefix,
> - run, ena, 1.0, &rt_stat);
> + if (counter->merged_stat)
> + return;
>
> - fputc('\n', output);
> + for (s = 0; s < config->aggr_map->nr; s++) {
> + print_counter_aggrdata(config, counter, s,
> + prefix, metric_only,
> + &first);
> }
> }
>
> @@ -1081,6 +787,7 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
> u64 ena, run, val;
> double uval;
> struct aggr_cpu_id id;
> + struct perf_stat_evsel *ps = counter->stats;
> int counter_idx = perf_cpu_map__idx(evsel__cpus(counter), cpu);
>
> if (counter_idx < 0)
> @@ -1093,13 +800,13 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
> aggr_printout(config, counter, id, 0);
> first = false;
> }
> - val = perf_counts(counter->counts, counter_idx, 0)->val;
> - ena = perf_counts(counter->counts, counter_idx, 0)->ena;
> - run = perf_counts(counter->counts, counter_idx, 0)->run;
> + val = ps->aggr[counter_idx].counts.val;
> + ena = ps->aggr[counter_idx].counts.ena;
> + run = ps->aggr[counter_idx].counts.run;
>
> uval = val * counter->scale;
> printout(config, id, 0, counter, uval, prefix,
> - run, ena, 1.0, &rt_stat);
> + run, ena, 1.0, &rt_stat, counter_idx);
> }
> if (!first)
> fputc('\n', config->output);
> @@ -1135,8 +842,8 @@ static void print_metric_headers(struct perf_stat_config *config,
> };
> bool first = true;
>
> - if (config->json_output && !config->interval)
> - fprintf(config->output, "{");
> + if (config->json_output && !config->interval)
> + fprintf(config->output, "{");
>
> if (prefix && !config->json_output)
> fprintf(config->output, "%s", prefix);
> @@ -1379,31 +1086,6 @@ static void print_footer(struct perf_stat_config *config)
> "the same PMU. Try reorganizing the group.\n");
> }
>
> -static void print_percore_thread(struct perf_stat_config *config,
> - struct evsel *counter, char *prefix)
> -{
> - int s;
> - struct aggr_cpu_id s2, id;
> - struct perf_cpu_map *cpus;
> - bool first = true;
> - int idx;
> - struct perf_cpu cpu;
> -
> - cpus = evsel__cpus(counter);
> - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> - s2 = config->aggr_get_id(config, cpu);
> - for (s = 0; s < config->aggr_map->nr; s++) {
> - id = config->aggr_map->map[s];
> - if (aggr_cpu_id__equal(&s2, &id))
> - break;
> - }
> -
> - print_counter_aggrdata(config, counter, s,
> - prefix, false,
> - &first, cpu);
> - }
> -}
> -
> static void print_percore(struct perf_stat_config *config,
> struct evsel *counter, char *prefix)
> {
> @@ -1416,15 +1098,14 @@ static void print_percore(struct perf_stat_config *config,
> return;
>
> if (config->percore_show_thread)
> - return print_percore_thread(config, counter, prefix);
> + return print_counter(config, counter, prefix);
>
> for (s = 0; s < config->aggr_map->nr; s++) {
> if (prefix && metric_only)
> fprintf(output, "%s", prefix);
>
> print_counter_aggrdata(config, counter, s,
> - prefix, metric_only,
> - &first, (struct perf_cpu){ .cpu = -1 });
> + prefix, metric_only, &first);
> }
>
> if (metric_only)
> @@ -1469,17 +1150,13 @@ void evlist__print_counters(struct evlist *evlist, struct perf_stat_config *conf
> print_aggr(config, evlist, prefix);
> break;
> case AGGR_THREAD:
> - evlist__for_each_entry(evlist, counter) {
> - print_aggr_thread(config, _target, counter, prefix);
> - }
> - break;
> case AGGR_GLOBAL:
> if (config->iostat_run)
> iostat_print_counters(evlist, config, ts, prefix = buf,
> - print_counter_aggr);
> + print_counter);
> else {
> evlist__for_each_entry(evlist, counter) {
> - print_counter_aggr(config, counter, prefix);
> + print_counter(config, counter, prefix);
> }
> if (metric_only)
> fputc('\n', config->output);
> diff --git a/tools/perf/util/stat.c b/tools/perf/util/stat.c
> index c0955a0427ab..0316557adce9 100644
> --- a/tools/perf/util/stat.c
> +++ b/tools/perf/util/stat.c
> @@ -565,11 +565,6 @@ int perf_stat_process_counter(struct perf_stat_config *config,
> evsel__name(counter), count[0], count[1], count[2]);
> }
>
> - /*
> - * Save the full runtime - to allow normalization during printout:
> - */
> - perf_stat__update_shadow_stats(counter, *count, 0, &rt_stat);
> -
> return 0;
> }
>
> diff --git a/tools/perf/util/stat.h b/tools/perf/util/stat.h
> index 3d413ba8c68a..382a1ab92ce1 100644
> --- a/tools/perf/util/stat.h
> +++ b/tools/perf/util/stat.h
> @@ -224,15 +224,6 @@ static inline void update_rusage_stats(struct rusage_stats *ru_stats, struct rus
> struct evsel;
> struct evlist;
>
> -struct perf_aggr_thread_value {
> - struct evsel *counter;
> - struct aggr_cpu_id id;
> - double uval;
> - u64 val;
> - u64 run;
> - u64 ena;
> -};
> -
> bool __perf_stat_evsel__is(struct evsel *evsel, enum perf_stat_evsel_id id);
>
> #define perf_stat_evsel__is(evsel, id) \
> --
> 2.38.0.413.g74048e4d9e-goog
>
On Wed, Jun 14, 2023 at 6:40 AM Jiri Olsa <[email protected]> wrote:
>
> On Mon, Oct 17, 2022 at 07:02:25PM -0700, Namhyung Kim wrote:
> > Now aggr counts are ready for use. Convert the display routines to use
> > the aggr counts and update the shadow stat with them. It doesn't need
> > to aggregate counts or collect aliases anymore during the display. Get
> > rid of now unused struct perf_aggr_thread_value.
> >
> > Note that there's a difference in the display order among the aggr mode.
> > For per-core/die/socket/node aggregation, it shows relevant events in
> > the same unit together, whereas global/thread/no aggregation it shows
> > the same events for different units together. So it still uses separate
> > codes to display them due to the ordering.
> >
> > One more thing to note is that it breaks per-core event display for now.
> > The next patch will fix it to have identical output as of now.
> >
> > Acked-by: Ian Rogers <[email protected]>
> > Signed-off-by: Namhyung Kim <[email protected]>
>
> hi,
> this one seems to break 'perf stat -r X' not sure why so far..
>
> final counts seems to be accumulated instead of displaying average, like:
>
>
> with this patch:
>
> Performance counter stats for './test_progs -n 103/1' (2 runs):
>
> 206,815,929 cycles:u ( +- 0.05% )
> 16,052,747,533 cycles:k ( +- 0.10% )
> 16,259,643,167 cycles ( +- 0.10% )
>
> 1.98093 +- 0.00586 seconds time elapsed ( +- 0.30% )
>
>
> without this patch:
>
> Performance counter stats for './test_progs -n 103/1' (2 runs):
>
> 103,300,812 cycles:u ( +- 0.37% )
> 8,016,856,866 cycles:k ( +- 0.32% )
> 8,120,200,572 cycles ( +- 0.32% )
>
> 1.97272 +- 0.00270 seconds time elapsed ( +- 0.14% )
>
>
> any idea? ;-)
Is this still broken in perf-tools-next? The patch is quite old and
there's been work in this area. I'm assuming yes, but thought it was
worth checking.
Thanks,
Ian
> thanks,
> jirka
>
>
>
>
> > ---
> > tools/perf/util/stat-display.c | 421 ++++-----------------------------
> > tools/perf/util/stat.c | 5 -
> > tools/perf/util/stat.h | 9 -
> > 3 files changed, 49 insertions(+), 386 deletions(-)
> >
> > diff --git a/tools/perf/util/stat-display.c b/tools/perf/util/stat-display.c
> > index 4113aa86772f..bfae2784609c 100644
> > --- a/tools/perf/util/stat-display.c
> > +++ b/tools/perf/util/stat-display.c
> > @@ -442,31 +442,6 @@ static void print_metric_header(struct perf_stat_config *config,
> > fprintf(os->fh, "%*s ", config->metric_only_len, unit);
> > }
> >
> > -static int first_shadow_map_idx(struct perf_stat_config *config,
> > - struct evsel *evsel, const struct aggr_cpu_id *id)
> > -{
> > - struct perf_cpu_map *cpus = evsel__cpus(evsel);
> > - struct perf_cpu cpu;
> > - int idx;
> > -
> > - if (config->aggr_mode == AGGR_NONE)
> > - return perf_cpu_map__idx(cpus, id->cpu);
> > -
> > - if (config->aggr_mode == AGGR_THREAD)
> > - return id->thread_idx;
> > -
> > - if (!config->aggr_get_id)
> > - return 0;
> > -
> > - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> > - struct aggr_cpu_id cpu_id = config->aggr_get_id(config, cpu);
> > -
> > - if (aggr_cpu_id__equal(&cpu_id, id))
> > - return idx;
> > - }
> > - return 0;
> > -}
> > -
> > static void abs_printout(struct perf_stat_config *config,
> > struct aggr_cpu_id id, int nr, struct evsel *evsel, double avg)
> > {
> > @@ -537,7 +512,7 @@ static bool is_mixed_hw_group(struct evsel *counter)
> > static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int nr,
> > struct evsel *counter, double uval,
> > char *prefix, u64 run, u64 ena, double noise,
> > - struct runtime_stat *st)
> > + struct runtime_stat *st, int map_idx)
> > {
> > struct perf_stat_output_ctx out;
> > struct outstate os = {
> > @@ -648,8 +623,7 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
> > print_running(config, run, ena);
> > }
> >
> > - perf_stat__print_shadow_stats(config, counter, uval,
> > - first_shadow_map_idx(config, counter, &id),
> > + perf_stat__print_shadow_stats(config, counter, uval, map_idx,
> > &out, &config->metric_events, st);
> > if (!config->csv_output && !config->metric_only && !config->json_output) {
> > print_noise(config, counter, noise);
> > @@ -657,34 +631,6 @@ static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int
> > }
> > }
> >
> > -static void aggr_update_shadow(struct perf_stat_config *config,
> > - struct evlist *evlist)
> > -{
> > - int idx, s;
> > - struct perf_cpu cpu;
> > - struct aggr_cpu_id s2, id;
> > - u64 val;
> > - struct evsel *counter;
> > - struct perf_cpu_map *cpus;
> > -
> > - for (s = 0; s < config->aggr_map->nr; s++) {
> > - id = config->aggr_map->map[s];
> > - evlist__for_each_entry(evlist, counter) {
> > - cpus = evsel__cpus(counter);
> > - val = 0;
> > - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> > - s2 = config->aggr_get_id(config, cpu);
> > - if (!aggr_cpu_id__equal(&s2, &id))
> > - continue;
> > - val += perf_counts(counter->counts, idx, 0)->val;
> > - }
> > - perf_stat__update_shadow_stats(counter, val,
> > - first_shadow_map_idx(config, counter, &id),
> > - &rt_stat);
> > - }
> > - }
> > -}
> > -
> > static void uniquify_event_name(struct evsel *counter)
> > {
> > char *new_name;
> > @@ -721,137 +667,51 @@ static void uniquify_event_name(struct evsel *counter)
> > counter->uniquified_name = true;
> > }
> >
> > -static void collect_all_aliases(struct perf_stat_config *config, struct evsel *counter,
> > - void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
> > - bool first),
> > - void *data)
> > +static bool hybrid_uniquify(struct evsel *evsel, struct perf_stat_config *config)
> > {
> > - struct evlist *evlist = counter->evlist;
> > - struct evsel *alias;
> > -
> > - alias = list_prepare_entry(counter, &(evlist->core.entries), core.node);
> > - list_for_each_entry_continue (alias, &evlist->core.entries, core.node) {
> > - /* Merge events with the same name, etc. but on different PMUs. */
> > - if (!strcmp(evsel__name(alias), evsel__name(counter)) &&
> > - alias->scale == counter->scale &&
> > - alias->cgrp == counter->cgrp &&
> > - !strcmp(alias->unit, counter->unit) &&
> > - evsel__is_clock(alias) == evsel__is_clock(counter) &&
> > - strcmp(alias->pmu_name, counter->pmu_name)) {
> > - alias->merged_stat = true;
> > - cb(config, alias, data, false);
> > - }
> > - }
> > + return evsel__is_hybrid(evsel) && !config->hybrid_merge;
> > }
> >
> > -static bool hybrid_merge(struct evsel *counter, struct perf_stat_config *config,
> > - bool check)
> > +static void uniquify_counter(struct perf_stat_config *config, struct evsel *counter)
> > {
> > - if (evsel__is_hybrid(counter)) {
> > - if (check)
> > - return config->hybrid_merge;
> > - else
> > - return !config->hybrid_merge;
> > - }
> > -
> > - return false;
> > -}
> > -
> > -static bool collect_data(struct perf_stat_config *config, struct evsel *counter,
> > - void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
> > - bool first),
> > - void *data)
> > -{
> > - if (counter->merged_stat)
> > - return false;
> > - cb(config, counter, data, true);
> > - if (config->no_merge || hybrid_merge(counter, config, false))
> > + if (config->no_merge || hybrid_uniquify(counter, config))
> > uniquify_event_name(counter);
> > - else if (counter->auto_merge_stats || hybrid_merge(counter, config, true))
> > - collect_all_aliases(config, counter, cb, data);
> > - return true;
> > -}
> > -
> > -struct aggr_data {
> > - u64 ena, run, val;
> > - struct aggr_cpu_id id;
> > - int nr;
> > - int cpu_map_idx;
> > -};
> > -
> > -static void aggr_cb(struct perf_stat_config *config,
> > - struct evsel *counter, void *data, bool first)
> > -{
> > - struct aggr_data *ad = data;
> > - int idx;
> > - struct perf_cpu cpu;
> > - struct perf_cpu_map *cpus;
> > - struct aggr_cpu_id s2;
> > -
> > - cpus = evsel__cpus(counter);
> > - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> > - struct perf_counts_values *counts;
> > -
> > - s2 = config->aggr_get_id(config, cpu);
> > - if (!aggr_cpu_id__equal(&s2, &ad->id))
> > - continue;
> > - if (first)
> > - ad->nr++;
> > - counts = perf_counts(counter->counts, idx, 0);
> > - /*
> > - * When any result is bad, make them all to give
> > - * consistent output in interval mode.
> > - */
> > - if (counts->ena == 0 || counts->run == 0 ||
> > - counter->counts->scaled == -1) {
> > - ad->ena = 0;
> > - ad->run = 0;
> > - break;
> > - }
> > - ad->val += counts->val;
> > - ad->ena += counts->ena;
> > - ad->run += counts->run;
> > - }
> > }
> >
> > static void print_counter_aggrdata(struct perf_stat_config *config,
> > struct evsel *counter, int s,
> > char *prefix, bool metric_only,
> > - bool *first, struct perf_cpu cpu)
> > + bool *first)
> > {
> > - struct aggr_data ad;
> > FILE *output = config->output;
> > u64 ena, run, val;
> > - int nr;
> > - struct aggr_cpu_id id;
> > double uval;
> > + struct perf_stat_evsel *ps = counter->stats;
> > + struct perf_stat_aggr *aggr = &ps->aggr[s];
> > + struct aggr_cpu_id id = config->aggr_map->map[s];
> > + double avg = aggr->counts.val;
> >
> > - ad.id = id = config->aggr_map->map[s];
> > - ad.val = ad.ena = ad.run = 0;
> > - ad.nr = 0;
> > - if (!collect_data(config, counter, aggr_cb, &ad))
> > + if (counter->supported && aggr->nr == 0)
> > return;
> >
> > - if (perf_pmu__has_hybrid() && ad.ena == 0)
> > - return;
> > + uniquify_counter(config, counter);
> > +
> > + val = aggr->counts.val;
> > + ena = aggr->counts.ena;
> > + run = aggr->counts.run;
> >
> > - nr = ad.nr;
> > - ena = ad.ena;
> > - run = ad.run;
> > - val = ad.val;
> > if (*first && metric_only) {
> > *first = false;
> > - aggr_printout(config, counter, id, nr);
> > + aggr_printout(config, counter, id, aggr->nr);
> > }
> > if (prefix && !metric_only)
> > fprintf(output, "%s", prefix);
> >
> > uval = val * counter->scale;
> > - if (cpu.cpu != -1)
> > - id = aggr_cpu_id__cpu(cpu, /*data=*/NULL);
> >
> > - printout(config, id, nr, counter, uval,
> > - prefix, run, ena, 1.0, &rt_stat);
> > + printout(config, id, aggr->nr, counter, uval,
> > + prefix, run, ena, avg, &rt_stat, s);
> > +
> > if (!metric_only)
> > fputc('\n', output);
> > }
> > @@ -869,8 +729,6 @@ static void print_aggr(struct perf_stat_config *config,
> > if (!config->aggr_map || !config->aggr_get_id)
> > return;
> >
> > - aggr_update_shadow(config, evlist);
> > -
> > /*
> > * With metric_only everything is on a single line.
> > * Without each counter has its own line.
> > @@ -881,188 +739,36 @@ static void print_aggr(struct perf_stat_config *config,
> >
> > first = true;
> > evlist__for_each_entry(evlist, counter) {
> > + if (counter->merged_stat)
> > + continue;
> > +
> > print_counter_aggrdata(config, counter, s,
> > - prefix, metric_only,
> > - &first, (struct perf_cpu){ .cpu = -1 });
> > + prefix, metric_only,
> > + &first);
> > }
> > if (metric_only)
> > fputc('\n', output);
> > }
> > }
> >
> > -static int cmp_val(const void *a, const void *b)
> > -{
> > - return ((struct perf_aggr_thread_value *)b)->val -
> > - ((struct perf_aggr_thread_value *)a)->val;
> > -}
> > -
> > -static struct perf_aggr_thread_value *sort_aggr_thread(
> > - struct evsel *counter,
> > - int *ret,
> > - struct target *_target)
> > -{
> > - int nthreads = perf_thread_map__nr(counter->core.threads);
> > - int i = 0;
> > - double uval;
> > - struct perf_aggr_thread_value *buf;
> > -
> > - buf = calloc(nthreads, sizeof(struct perf_aggr_thread_value));
> > - if (!buf)
> > - return NULL;
> > -
> > - for (int thread = 0; thread < nthreads; thread++) {
> > - int idx;
> > - u64 ena = 0, run = 0, val = 0;
> > -
> > - perf_cpu_map__for_each_idx(idx, evsel__cpus(counter)) {
> > - struct perf_counts_values *counts =
> > - perf_counts(counter->counts, idx, thread);
> > -
> > - val += counts->val;
> > - ena += counts->ena;
> > - run += counts->run;
> > - }
> > -
> > - uval = val * counter->scale;
> > -
> > - /*
> > - * Skip value 0 when enabling --per-thread globally,
> > - * otherwise too many 0 output.
> > - */
> > - if (uval == 0.0 && target__has_per_thread(_target))
> > - continue;
> > -
> > - buf[i].counter = counter;
> > - buf[i].id = aggr_cpu_id__empty();
> > - buf[i].id.thread_idx = thread;
> > - buf[i].uval = uval;
> > - buf[i].val = val;
> > - buf[i].run = run;
> > - buf[i].ena = ena;
> > - i++;
> > - }
> > -
> > - qsort(buf, i, sizeof(struct perf_aggr_thread_value), cmp_val);
> > -
> > - if (ret)
> > - *ret = i;
> > -
> > - return buf;
> > -}
> > -
> > -static void print_aggr_thread(struct perf_stat_config *config,
> > - struct target *_target,
> > - struct evsel *counter, char *prefix)
> > -{
> > - FILE *output = config->output;
> > - int thread, sorted_threads;
> > - struct aggr_cpu_id id;
> > - struct perf_aggr_thread_value *buf;
> > -
> > - buf = sort_aggr_thread(counter, &sorted_threads, _target);
> > - if (!buf) {
> > - perror("cannot sort aggr thread");
> > - return;
> > - }
> > -
> > - for (thread = 0; thread < sorted_threads; thread++) {
> > - if (prefix)
> > - fprintf(output, "%s", prefix);
> > -
> > - id = buf[thread].id;
> > - printout(config, id, 0, buf[thread].counter, buf[thread].uval,
> > - prefix, buf[thread].run, buf[thread].ena, 1.0,
> > - &rt_stat);
> > - fputc('\n', output);
> > - }
> > -
> > - free(buf);
> > -}
> > -
> > -struct caggr_data {
> > - double avg, avg_enabled, avg_running;
> > -};
> > -
> > -static void counter_aggr_cb(struct perf_stat_config *config __maybe_unused,
> > - struct evsel *counter, void *data,
> > - bool first __maybe_unused)
> > -{
> > - struct caggr_data *cd = data;
> > - struct perf_counts_values *aggr = &counter->counts->aggr;
> > -
> > - cd->avg += aggr->val;
> > - cd->avg_enabled += aggr->ena;
> > - cd->avg_running += aggr->run;
> > -}
> > -
> > -/*
> > - * Print out the results of a single counter:
> > - * aggregated counts in system-wide mode
> > - */
> > -static void print_counter_aggr(struct perf_stat_config *config,
> > - struct evsel *counter, char *prefix)
> > -{
> > - bool metric_only = config->metric_only;
> > - FILE *output = config->output;
> > - double uval;
> > - struct caggr_data cd = { .avg = 0.0 };
> > -
> > - if (!collect_data(config, counter, counter_aggr_cb, &cd))
> > - return;
> > -
> > - if (prefix && !metric_only)
> > - fprintf(output, "%s", prefix);
> > -
> > - uval = cd.avg * counter->scale;
> > - printout(config, aggr_cpu_id__empty(), 0, counter, uval, prefix, cd.avg_running,
> > - cd.avg_enabled, cd.avg, &rt_stat);
> > - if (!metric_only)
> > - fprintf(output, "\n");
> > -}
> > -
> > -static void counter_cb(struct perf_stat_config *config __maybe_unused,
> > - struct evsel *counter, void *data,
> > - bool first __maybe_unused)
> > -{
> > - struct aggr_data *ad = data;
> > -
> > - ad->val += perf_counts(counter->counts, ad->cpu_map_idx, 0)->val;
> > - ad->ena += perf_counts(counter->counts, ad->cpu_map_idx, 0)->ena;
> > - ad->run += perf_counts(counter->counts, ad->cpu_map_idx, 0)->run;
> > -}
> > -
> > -/*
> > - * Print out the results of a single counter:
> > - * does not use aggregated count in system-wide
> > - */
> > static void print_counter(struct perf_stat_config *config,
> > struct evsel *counter, char *prefix)
> > {
> > - FILE *output = config->output;
> > - u64 ena, run, val;
> > - double uval;
> > - int idx;
> > - struct perf_cpu cpu;
> > - struct aggr_cpu_id id;
> > -
> > - perf_cpu_map__for_each_cpu(cpu, idx, evsel__cpus(counter)) {
> > - struct aggr_data ad = { .cpu_map_idx = idx };
> > -
> > - if (!collect_data(config, counter, counter_cb, &ad))
> > - return;
> > - val = ad.val;
> > - ena = ad.ena;
> > - run = ad.run;
> > + bool metric_only = config->metric_only;
> > + bool first = false;
> > + int s;
> >
> > - if (prefix)
> > - fprintf(output, "%s", prefix);
> > + /* AGGR_THREAD doesn't have config->aggr_get_id */
> > + if (!config->aggr_map)
> > + return;
> >
> > - uval = val * counter->scale;
> > - id = aggr_cpu_id__cpu(cpu, /*data=*/NULL);
> > - printout(config, id, 0, counter, uval, prefix,
> > - run, ena, 1.0, &rt_stat);
> > + if (counter->merged_stat)
> > + return;
> >
> > - fputc('\n', output);
> > + for (s = 0; s < config->aggr_map->nr; s++) {
> > + print_counter_aggrdata(config, counter, s,
> > + prefix, metric_only,
> > + &first);
> > }
> > }
> >
> > @@ -1081,6 +787,7 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
> > u64 ena, run, val;
> > double uval;
> > struct aggr_cpu_id id;
> > + struct perf_stat_evsel *ps = counter->stats;
> > int counter_idx = perf_cpu_map__idx(evsel__cpus(counter), cpu);
> >
> > if (counter_idx < 0)
> > @@ -1093,13 +800,13 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
> > aggr_printout(config, counter, id, 0);
> > first = false;
> > }
> > - val = perf_counts(counter->counts, counter_idx, 0)->val;
> > - ena = perf_counts(counter->counts, counter_idx, 0)->ena;
> > - run = perf_counts(counter->counts, counter_idx, 0)->run;
> > + val = ps->aggr[counter_idx].counts.val;
> > + ena = ps->aggr[counter_idx].counts.ena;
> > + run = ps->aggr[counter_idx].counts.run;
> >
> > uval = val * counter->scale;
> > printout(config, id, 0, counter, uval, prefix,
> > - run, ena, 1.0, &rt_stat);
> > + run, ena, 1.0, &rt_stat, counter_idx);
> > }
> > if (!first)
> > fputc('\n', config->output);
> > @@ -1135,8 +842,8 @@ static void print_metric_headers(struct perf_stat_config *config,
> > };
> > bool first = true;
> >
> > - if (config->json_output && !config->interval)
> > - fprintf(config->output, "{");
> > + if (config->json_output && !config->interval)
> > + fprintf(config->output, "{");
> >
> > if (prefix && !config->json_output)
> > fprintf(config->output, "%s", prefix);
> > @@ -1379,31 +1086,6 @@ static void print_footer(struct perf_stat_config *config)
> > "the same PMU. Try reorganizing the group.\n");
> > }
> >
> > -static void print_percore_thread(struct perf_stat_config *config,
> > - struct evsel *counter, char *prefix)
> > -{
> > - int s;
> > - struct aggr_cpu_id s2, id;
> > - struct perf_cpu_map *cpus;
> > - bool first = true;
> > - int idx;
> > - struct perf_cpu cpu;
> > -
> > - cpus = evsel__cpus(counter);
> > - perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
> > - s2 = config->aggr_get_id(config, cpu);
> > - for (s = 0; s < config->aggr_map->nr; s++) {
> > - id = config->aggr_map->map[s];
> > - if (aggr_cpu_id__equal(&s2, &id))
> > - break;
> > - }
> > -
> > - print_counter_aggrdata(config, counter, s,
> > - prefix, false,
> > - &first, cpu);
> > - }
> > -}
> > -
> > static void print_percore(struct perf_stat_config *config,
> > struct evsel *counter, char *prefix)
> > {
> > @@ -1416,15 +1098,14 @@ static void print_percore(struct perf_stat_config *config,
> > return;
> >
> > if (config->percore_show_thread)
> > - return print_percore_thread(config, counter, prefix);
> > + return print_counter(config, counter, prefix);
> >
> > for (s = 0; s < config->aggr_map->nr; s++) {
> > if (prefix && metric_only)
> > fprintf(output, "%s", prefix);
> >
> > print_counter_aggrdata(config, counter, s,
> > - prefix, metric_only,
> > - &first, (struct perf_cpu){ .cpu = -1 });
> > + prefix, metric_only, &first);
> > }
> >
> > if (metric_only)
> > @@ -1469,17 +1150,13 @@ void evlist__print_counters(struct evlist *evlist, struct perf_stat_config *conf
> > print_aggr(config, evlist, prefix);
> > break;
> > case AGGR_THREAD:
> > - evlist__for_each_entry(evlist, counter) {
> > - print_aggr_thread(config, _target, counter, prefix);
> > - }
> > - break;
> > case AGGR_GLOBAL:
> > if (config->iostat_run)
> > iostat_print_counters(evlist, config, ts, prefix = buf,
> > - print_counter_aggr);
> > + print_counter);
> > else {
> > evlist__for_each_entry(evlist, counter) {
> > - print_counter_aggr(config, counter, prefix);
> > + print_counter(config, counter, prefix);
> > }
> > if (metric_only)
> > fputc('\n', config->output);
> > diff --git a/tools/perf/util/stat.c b/tools/perf/util/stat.c
> > index c0955a0427ab..0316557adce9 100644
> > --- a/tools/perf/util/stat.c
> > +++ b/tools/perf/util/stat.c
> > @@ -565,11 +565,6 @@ int perf_stat_process_counter(struct perf_stat_config *config,
> > evsel__name(counter), count[0], count[1], count[2]);
> > }
> >
> > - /*
> > - * Save the full runtime - to allow normalization during printout:
> > - */
> > - perf_stat__update_shadow_stats(counter, *count, 0, &rt_stat);
> > -
> > return 0;
> > }
> >
> > diff --git a/tools/perf/util/stat.h b/tools/perf/util/stat.h
> > index 3d413ba8c68a..382a1ab92ce1 100644
> > --- a/tools/perf/util/stat.h
> > +++ b/tools/perf/util/stat.h
> > @@ -224,15 +224,6 @@ static inline void update_rusage_stats(struct rusage_stats *ru_stats, struct rus
> > struct evsel;
> > struct evlist;
> >
> > -struct perf_aggr_thread_value {
> > - struct evsel *counter;
> > - struct aggr_cpu_id id;
> > - double uval;
> > - u64 val;
> > - u64 run;
> > - u64 ena;
> > -};
> > -
> > bool __perf_stat_evsel__is(struct evsel *evsel, enum perf_stat_evsel_id id);
> >
> > #define perf_stat_evsel__is(evsel, id) \
> > --
> > 2.38.0.413.g74048e4d9e-goog
> >
On Wed, Jun 14, 2023 at 09:20:53AM -0700, Ian Rogers wrote:
> On Wed, Jun 14, 2023 at 6:40 AM Jiri Olsa <[email protected]> wrote:
> >
> > On Mon, Oct 17, 2022 at 07:02:25PM -0700, Namhyung Kim wrote:
> > > Now aggr counts are ready for use. Convert the display routines to use
> > > the aggr counts and update the shadow stat with them. It doesn't need
> > > to aggregate counts or collect aliases anymore during the display. Get
> > > rid of now unused struct perf_aggr_thread_value.
> > >
> > > Note that there's a difference in the display order among the aggr mode.
> > > For per-core/die/socket/node aggregation, it shows relevant events in
> > > the same unit together, whereas global/thread/no aggregation it shows
> > > the same events for different units together. So it still uses separate
> > > codes to display them due to the ordering.
> > >
> > > One more thing to note is that it breaks per-core event display for now.
> > > The next patch will fix it to have identical output as of now.
> > >
> > > Acked-by: Ian Rogers <[email protected]>
> > > Signed-off-by: Namhyung Kim <[email protected]>
> >
> > hi,
> > this one seems to break 'perf stat -r X' not sure why so far..
> >
> > final counts seems to be accumulated instead of displaying average, like:
> >
> >
> > with this patch:
> >
> > Performance counter stats for './test_progs -n 103/1' (2 runs):
> >
> > 206,815,929 cycles:u ( +- 0.05% )
> > 16,052,747,533 cycles:k ( +- 0.10% )
> > 16,259,643,167 cycles ( +- 0.10% )
> >
> > 1.98093 +- 0.00586 seconds time elapsed ( +- 0.30% )
> >
> >
> > without this patch:
> >
> > Performance counter stats for './test_progs -n 103/1' (2 runs):
> >
> > 103,300,812 cycles:u ( +- 0.37% )
> > 8,016,856,866 cycles:k ( +- 0.32% )
> > 8,120,200,572 cycles ( +- 0.32% )
> >
> > 1.97272 +- 0.00270 seconds time elapsed ( +- 0.14% )
> >
> >
> > any idea? ;-)
>
> Is this still broken in perf-tools-next? The patch is quite old and
> there's been work in this area. I'm assuming yes, but thought it was
> worth checking.
yes
single run:
[root@krava perf]# ./perf stat -e cycles:u ./perf bench sched pipe
# Running 'sched/pipe' benchmark:
# Executed 1000000 pipe operations between two processes
Total time: 4.725 [sec]
4.725795 usecs/op
211604 ops/sec
Performance counter stats for './perf bench sched pipe':
398,096,363 cycles:u
4.737638715 seconds time elapsed
0.227961000 seconds user
4.348895000 seconds sys
3 runs (with verbose):
[root@krava perf]# ./perf stat -v -r 3 -e cycles:u ./perf bench sched pipe
Using CPUID GenuineIntel-6-8C-1
Control descriptor is not initialized
[ perf stat: executing run #1 ... ]
# Running 'sched/pipe' benchmark:
# Executed 1000000 pipe operations between two processes
Total time: 4.659 [sec]
4.659396 usecs/op
214620 ops/sec
cycles:u: 399150620 3866604490 3866604490
[ perf stat: executing run #2 ... ]
# Running 'sched/pipe' benchmark:
# Executed 1000000 pipe operations between two processes
Total time: 4.656 [sec]
4.656820 usecs/op
214738 ops/sec
cycles:u: 795910540 7700776638 7700776638
[ perf stat: executing run #3 ... ]
# Running 'sched/pipe' benchmark:
# Executed 1000000 pipe operations between two processes
Total time: 4.634 [sec]
4.634090 usecs/op
215792 ops/sec
cycles:u: 1189927957 11522039559 11522039559
Performance counter stats for './perf bench sched pipe' (3 runs):
1,189,927,957 cycles:u ( +- 19.18% )
4.6611 +- 0.0102 seconds time elapsed ( +- 0.22% )
jirka
Hi Jiri,
On Thu, Jun 15, 2023 at 1:10 AM Jiri Olsa <[email protected]> wrote:
>
> On Wed, Jun 14, 2023 at 09:20:53AM -0700, Ian Rogers wrote:
> > On Wed, Jun 14, 2023 at 6:40 AM Jiri Olsa <[email protected]> wrote:
> > >
> > > On Mon, Oct 17, 2022 at 07:02:25PM -0700, Namhyung Kim wrote:
> > > > Now aggr counts are ready for use. Convert the display routines to use
> > > > the aggr counts and update the shadow stat with them. It doesn't need
> > > > to aggregate counts or collect aliases anymore during the display. Get
> > > > rid of now unused struct perf_aggr_thread_value.
> > > >
> > > > Note that there's a difference in the display order among the aggr mode.
> > > > For per-core/die/socket/node aggregation, it shows relevant events in
> > > > the same unit together, whereas global/thread/no aggregation it shows
> > > > the same events for different units together. So it still uses separate
> > > > codes to display them due to the ordering.
> > > >
> > > > One more thing to note is that it breaks per-core event display for now.
> > > > The next patch will fix it to have identical output as of now.
> > > >
> > > > Acked-by: Ian Rogers <[email protected]>
> > > > Signed-off-by: Namhyung Kim <[email protected]>
> > >
> > > hi,
> > > this one seems to break 'perf stat -r X' not sure why so far..
> > >
> > > final counts seems to be accumulated instead of displaying average, like:
> > >
> > >
> > > with this patch:
> > >
> > > Performance counter stats for './test_progs -n 103/1' (2 runs):
> > >
> > > 206,815,929 cycles:u ( +- 0.05% )
> > > 16,052,747,533 cycles:k ( +- 0.10% )
> > > 16,259,643,167 cycles ( +- 0.10% )
> > >
> > > 1.98093 +- 0.00586 seconds time elapsed ( +- 0.30% )
> > >
> > >
> > > without this patch:
> > >
> > > Performance counter stats for './test_progs -n 103/1' (2 runs):
> > >
> > > 103,300,812 cycles:u ( +- 0.37% )
> > > 8,016,856,866 cycles:k ( +- 0.32% )
> > > 8,120,200,572 cycles ( +- 0.32% )
> > >
> > > 1.97272 +- 0.00270 seconds time elapsed ( +- 0.14% )
> > >
> > >
> > > any idea? ;-)
> >
> > Is this still broken in perf-tools-next? The patch is quite old and
> > there's been work in this area. I'm assuming yes, but thought it was
> > worth checking.
>
> yes
I'll take a look.