From: Kan Liang <[email protected]>
Provide a new solution to replace the reverted commit ac2dc29edd21
("perf stat: Add default hybrid events").
For the default software attrs, nothing is changed.
For the default hardware attrs, create a new evsel for each hybrid pmu.
With the new solution, adding a new default attr will not require the
special support for the hybrid platform anymore.
Also, the "--detailed" is supported on the hybrid platform
With the patch,
./perf stat -a -ddd sleep 1
Performance counter stats for 'system wide':
32,231.06 msec cpu-clock # 32.056 CPUs utilized
529 context-switches # 16.413 /sec
32 cpu-migrations # 0.993 /sec
69 page-faults # 2.141 /sec
176,754,151 cpu_core/cycles/ # 5.484 M/sec (41.65%)
161,695,280 cpu_atom/cycles/ # 5.017 M/sec (49.92%)
48,595,992 cpu_core/instructions/ # 1.508 M/sec (49.98%)
32,363,337 cpu_atom/instructions/ # 1.004 M/sec (58.26%)
10,088,639 cpu_core/branches/ # 313.010 K/sec (58.31%)
6,390,582 cpu_atom/branches/ # 198.274 K/sec (58.26%)
846,201 cpu_core/branch-misses/ # 26.254 K/sec (66.65%)
676,477 cpu_atom/branch-misses/ # 20.988 K/sec (58.27%)
14,290,070 cpu_core/L1-dcache-loads/ # 443.363 K/sec (66.66%)
9,983,532 cpu_atom/L1-dcache-loads/ # 309.749 K/sec (58.27%)
740,725 cpu_core/L1-dcache-load-misses/ # 22.982 K/sec (66.66%)
<not supported> cpu_atom/L1-dcache-load-misses/
480,441 cpu_core/LLC-loads/ # 14.906 K/sec (66.67%)
326,570 cpu_atom/LLC-loads/ # 10.132 K/sec (58.27%)
329 cpu_core/LLC-load-misses/ # 10.208 /sec (66.68%)
0 cpu_atom/LLC-load-misses/ # 0.000 /sec (58.32%)
<not supported> cpu_core/L1-icache-loads/
21,982,491 cpu_atom/L1-icache-loads/ # 682.028 K/sec (58.43%)
4,493,189 cpu_core/L1-icache-load-misses/ # 139.406 K/sec (33.34%)
4,711,404 cpu_atom/L1-icache-load-misses/ # 146.176 K/sec (50.08%)
13,713,090 cpu_core/dTLB-loads/ # 425.462 K/sec (33.34%)
9,384,727 cpu_atom/dTLB-loads/ # 291.170 K/sec (50.08%)
157,387 cpu_core/dTLB-load-misses/ # 4.883 K/sec (33.33%)
108,328 cpu_atom/dTLB-load-misses/ # 3.361 K/sec (50.08%)
<not supported> cpu_core/iTLB-loads/
<not supported> cpu_atom/iTLB-loads/
37,655 cpu_core/iTLB-load-misses/ # 1.168 K/sec (33.32%)
61,661 cpu_atom/iTLB-load-misses/ # 1.913 K/sec (50.03%)
<not supported> cpu_core/L1-dcache-prefetches/
<not supported> cpu_atom/L1-dcache-prefetches/
<not supported> cpu_core/L1-dcache-prefetch-misses/
<not supported> cpu_atom/L1-dcache-prefetch-misses/
1.005466919 seconds time elapsed
Signed-off-by: Kan Liang <[email protected]>
Signed-off-by: Zhengjun Xing <[email protected]>
---
tools/perf/arch/x86/util/evlist.c | 52 ++++++++++++++++++++++++++++++-
tools/perf/util/evlist.c | 2 +-
tools/perf/util/evlist.h | 2 ++
3 files changed, 54 insertions(+), 2 deletions(-)
diff --git a/tools/perf/arch/x86/util/evlist.c b/tools/perf/arch/x86/util/evlist.c
index 777bdf182a58..1b3f9e1a2287 100644
--- a/tools/perf/arch/x86/util/evlist.c
+++ b/tools/perf/arch/x86/util/evlist.c
@@ -4,16 +4,66 @@
#include "util/evlist.h"
#include "util/parse-events.h"
#include "topdown.h"
+#include "util/event.h"
+#include "util/pmu-hybrid.h"
#define TOPDOWN_L1_EVENTS "{slots,topdown-retiring,topdown-bad-spec,topdown-fe-bound,topdown-be-bound}"
#define TOPDOWN_L2_EVENTS "{slots,topdown-retiring,topdown-bad-spec,topdown-fe-bound,topdown-be-bound,topdown-heavy-ops,topdown-br-mispredict,topdown-fetch-lat,topdown-mem-bound}"
+static int ___evlist__add_default_attrs(struct evlist *evlist, struct perf_event_attr *attrs, size_t nr_attrs)
+{
+ struct perf_cpu_map *cpus;
+ struct evsel *evsel, *n;
+ struct perf_pmu *pmu;
+ LIST_HEAD(head);
+ size_t i, j = 0;
+
+ for (i = 0; i < nr_attrs; i++)
+ event_attr_init(attrs + i);
+
+ if (!perf_pmu__has_hybrid())
+ return evlist__add_attrs(evlist, attrs, nr_attrs);
+
+ for (i = 0; i < nr_attrs; i++) {
+ if (attrs[i].type == PERF_TYPE_SOFTWARE) {
+ evsel = evsel__new_idx(attrs + i, evlist->core.nr_entries + j);
+ if (evsel == NULL)
+ goto out_delete_partial_list;
+ j++;
+ list_add_tail(&evsel->core.node, &head);
+ continue;
+ }
+
+ perf_pmu__for_each_hybrid_pmu(pmu) {
+ evsel = evsel__new_idx(attrs + i, evlist->core.nr_entries + j);
+ if (evsel == NULL)
+ goto out_delete_partial_list;
+ j++;
+ evsel->core.attr.config |= (__u64)pmu->type << PERF_PMU_TYPE_SHIFT;
+ cpus = perf_cpu_map__get(pmu->cpus);
+ evsel->core.cpus = cpus;
+ evsel->core.own_cpus = perf_cpu_map__get(cpus);
+ evsel->pmu_name = strdup(pmu->name);
+ list_add_tail(&evsel->core.node, &head);
+ }
+ }
+
+ evlist__splice_list_tail(evlist, &head);
+
+ return 0;
+
+out_delete_partial_list:
+ __evlist__for_each_entry_safe(&head, n, evsel)
+ evsel__delete(evsel);
+ return -1;
+}
+
int arch_evlist__add_default_attrs(struct evlist *evlist,
struct perf_event_attr *attrs,
size_t nr_attrs)
{
if (nr_attrs)
- return __evlist__add_default_attrs(evlist, attrs, nr_attrs);
+ return ___evlist__add_default_attrs(evlist, attrs, nr_attrs);
if (!pmu_have_event("cpu", "slots"))
return 0;
diff --git a/tools/perf/util/evlist.c b/tools/perf/util/evlist.c
index efa5f006b5c6..5ff4b9504828 100644
--- a/tools/perf/util/evlist.c
+++ b/tools/perf/util/evlist.c
@@ -309,7 +309,7 @@ struct evsel *evlist__add_aux_dummy(struct evlist *evlist, bool system_wide)
return evsel;
}
-static int evlist__add_attrs(struct evlist *evlist, struct perf_event_attr *attrs, size_t nr_attrs)
+int evlist__add_attrs(struct evlist *evlist, struct perf_event_attr *attrs, size_t nr_attrs)
{
struct evsel *evsel, *n;
LIST_HEAD(head);
diff --git a/tools/perf/util/evlist.h b/tools/perf/util/evlist.h
index 129095c0fe6d..351ba2887a79 100644
--- a/tools/perf/util/evlist.h
+++ b/tools/perf/util/evlist.h
@@ -104,6 +104,8 @@ static inline int evlist__add_default(struct evlist *evlist)
return __evlist__add_default(evlist, true);
}
+int evlist__add_attrs(struct evlist *evlist, struct perf_event_attr *attrs, size_t nr_attrs);
+
int __evlist__add_default_attrs(struct evlist *evlist,
struct perf_event_attr *attrs, size_t nr_attrs);
--
2.25.1
Hello,
On Tue, Jun 7, 2022 at 12:31 AM <[email protected]> wrote:
>
> From: Kan Liang <[email protected]>
>
> Provide a new solution to replace the reverted commit ac2dc29edd21
> ("perf stat: Add default hybrid events").
>
> For the default software attrs, nothing is changed.
> For the default hardware attrs, create a new evsel for each hybrid pmu.
>
> With the new solution, adding a new default attr will not require the
> special support for the hybrid platform anymore.
>
> Also, the "--detailed" is supported on the hybrid platform
>
> With the patch,
>
> ./perf stat -a -ddd sleep 1
>
> Performance counter stats for 'system wide':
>
> 32,231.06 msec cpu-clock # 32.056 CPUs utilized
> 529 context-switches # 16.413 /sec
> 32 cpu-migrations # 0.993 /sec
> 69 page-faults # 2.141 /sec
> 176,754,151 cpu_core/cycles/ # 5.484 M/sec (41.65%)
> 161,695,280 cpu_atom/cycles/ # 5.017 M/sec (49.92%)
> 48,595,992 cpu_core/instructions/ # 1.508 M/sec (49.98%)
> 32,363,337 cpu_atom/instructions/ # 1.004 M/sec (58.26%)
> 10,088,639 cpu_core/branches/ # 313.010 K/sec (58.31%)
> 6,390,582 cpu_atom/branches/ # 198.274 K/sec (58.26%)
> 846,201 cpu_core/branch-misses/ # 26.254 K/sec (66.65%)
> 676,477 cpu_atom/branch-misses/ # 20.988 K/sec (58.27%)
> 14,290,070 cpu_core/L1-dcache-loads/ # 443.363 K/sec (66.66%)
> 9,983,532 cpu_atom/L1-dcache-loads/ # 309.749 K/sec (58.27%)
> 740,725 cpu_core/L1-dcache-load-misses/ # 22.982 K/sec (66.66%)
> <not supported> cpu_atom/L1-dcache-load-misses/
> 480,441 cpu_core/LLC-loads/ # 14.906 K/sec (66.67%)
> 326,570 cpu_atom/LLC-loads/ # 10.132 K/sec (58.27%)
> 329 cpu_core/LLC-load-misses/ # 10.208 /sec (66.68%)
> 0 cpu_atom/LLC-load-misses/ # 0.000 /sec (58.32%)
> <not supported> cpu_core/L1-icache-loads/
> 21,982,491 cpu_atom/L1-icache-loads/ # 682.028 K/sec (58.43%)
> 4,493,189 cpu_core/L1-icache-load-misses/ # 139.406 K/sec (33.34%)
> 4,711,404 cpu_atom/L1-icache-load-misses/ # 146.176 K/sec (50.08%)
> 13,713,090 cpu_core/dTLB-loads/ # 425.462 K/sec (33.34%)
> 9,384,727 cpu_atom/dTLB-loads/ # 291.170 K/sec (50.08%)
> 157,387 cpu_core/dTLB-load-misses/ # 4.883 K/sec (33.33%)
> 108,328 cpu_atom/dTLB-load-misses/ # 3.361 K/sec (50.08%)
> <not supported> cpu_core/iTLB-loads/
> <not supported> cpu_atom/iTLB-loads/
> 37,655 cpu_core/iTLB-load-misses/ # 1.168 K/sec (33.32%)
> 61,661 cpu_atom/iTLB-load-misses/ # 1.913 K/sec (50.03%)
> <not supported> cpu_core/L1-dcache-prefetches/
> <not supported> cpu_atom/L1-dcache-prefetches/
> <not supported> cpu_core/L1-dcache-prefetch-misses/
> <not supported> cpu_atom/L1-dcache-prefetch-misses/
>
> 1.005466919 seconds time elapsed
>
> Signed-off-by: Kan Liang <[email protected]>
> Signed-off-by: Zhengjun Xing <[email protected]>
> ---
> tools/perf/arch/x86/util/evlist.c | 52 ++++++++++++++++++++++++++++++-
> tools/perf/util/evlist.c | 2 +-
> tools/perf/util/evlist.h | 2 ++
> 3 files changed, 54 insertions(+), 2 deletions(-)
>
> diff --git a/tools/perf/arch/x86/util/evlist.c b/tools/perf/arch/x86/util/evlist.c
> index 777bdf182a58..1b3f9e1a2287 100644
> --- a/tools/perf/arch/x86/util/evlist.c
> +++ b/tools/perf/arch/x86/util/evlist.c
> @@ -4,16 +4,66 @@
> #include "util/evlist.h"
> #include "util/parse-events.h"
> #include "topdown.h"
> +#include "util/event.h"
> +#include "util/pmu-hybrid.h"
>
> #define TOPDOWN_L1_EVENTS "{slots,topdown-retiring,topdown-bad-spec,topdown-fe-bound,topdown-be-bound}"
> #define TOPDOWN_L2_EVENTS "{slots,topdown-retiring,topdown-bad-spec,topdown-fe-bound,topdown-be-bound,topdown-heavy-ops,topdown-br-mispredict,topdown-fetch-lat,topdown-mem-bound}"
>
> +static int ___evlist__add_default_attrs(struct evlist *evlist, struct perf_event_attr *attrs, size_t nr_attrs)
> +{
> + struct perf_cpu_map *cpus;
> + struct evsel *evsel, *n;
> + struct perf_pmu *pmu;
> + LIST_HEAD(head);
> + size_t i, j = 0;
> +
> + for (i = 0; i < nr_attrs; i++)
> + event_attr_init(attrs + i);
> +
> + if (!perf_pmu__has_hybrid())
> + return evlist__add_attrs(evlist, attrs, nr_attrs);
> +
> + for (i = 0; i < nr_attrs; i++) {
> + if (attrs[i].type == PERF_TYPE_SOFTWARE) {
> + evsel = evsel__new_idx(attrs + i, evlist->core.nr_entries + j);
Probably no need to calculate index (j) as it's updated
later when it goes to the evlist...
> + if (evsel == NULL)
> + goto out_delete_partial_list;
> + j++;
> + list_add_tail(&evsel->core.node, &head);
> + continue;
> + }
> +
> + perf_pmu__for_each_hybrid_pmu(pmu) {
> + evsel = evsel__new_idx(attrs + i, evlist->core.nr_entries + j);
> + if (evsel == NULL)
> + goto out_delete_partial_list;
> + j++;
> + evsel->core.attr.config |= (__u64)pmu->type << PERF_PMU_TYPE_SHIFT;
> + cpus = perf_cpu_map__get(pmu->cpus);
> + evsel->core.cpus = cpus;
> + evsel->core.own_cpus = perf_cpu_map__get(cpus);
> + evsel->pmu_name = strdup(pmu->name);
> + list_add_tail(&evsel->core.node, &head);
> + }
> + }
> +
> + evlist__splice_list_tail(evlist, &head);
... like here.
Thanks,
Namhyung
> +
> + return 0;
> +
> +out_delete_partial_list:
> + __evlist__for_each_entry_safe(&head, n, evsel)
> + evsel__delete(evsel);
> + return -1;
> +}
On 6/8/2022 8:04 PM, Namhyung Kim wrote:
> Hello,
>
> On Tue, Jun 7, 2022 at 12:31 AM <[email protected]> wrote:
>>
>> From: Kan Liang <[email protected]>
>>
>> Provide a new solution to replace the reverted commit ac2dc29edd21
>> ("perf stat: Add default hybrid events").
>>
>> For the default software attrs, nothing is changed.
>> For the default hardware attrs, create a new evsel for each hybrid pmu.
>>
>> With the new solution, adding a new default attr will not require the
>> special support for the hybrid platform anymore.
>>
>> Also, the "--detailed" is supported on the hybrid platform
>>
>> With the patch,
>>
>> ./perf stat -a -ddd sleep 1
>>
>> Performance counter stats for 'system wide':
>>
>> 32,231.06 msec cpu-clock # 32.056 CPUs utilized
>> 529 context-switches # 16.413 /sec
>> 32 cpu-migrations # 0.993 /sec
>> 69 page-faults # 2.141 /sec
>> 176,754,151 cpu_core/cycles/ # 5.484 M/sec (41.65%)
>> 161,695,280 cpu_atom/cycles/ # 5.017 M/sec (49.92%)
>> 48,595,992 cpu_core/instructions/ # 1.508 M/sec (49.98%)
>> 32,363,337 cpu_atom/instructions/ # 1.004 M/sec (58.26%)
>> 10,088,639 cpu_core/branches/ # 313.010 K/sec (58.31%)
>> 6,390,582 cpu_atom/branches/ # 198.274 K/sec (58.26%)
>> 846,201 cpu_core/branch-misses/ # 26.254 K/sec (66.65%)
>> 676,477 cpu_atom/branch-misses/ # 20.988 K/sec (58.27%)
>> 14,290,070 cpu_core/L1-dcache-loads/ # 443.363 K/sec (66.66%)
>> 9,983,532 cpu_atom/L1-dcache-loads/ # 309.749 K/sec (58.27%)
>> 740,725 cpu_core/L1-dcache-load-misses/ # 22.982 K/sec (66.66%)
>> <not supported> cpu_atom/L1-dcache-load-misses/
>> 480,441 cpu_core/LLC-loads/ # 14.906 K/sec (66.67%)
>> 326,570 cpu_atom/LLC-loads/ # 10.132 K/sec (58.27%)
>> 329 cpu_core/LLC-load-misses/ # 10.208 /sec (66.68%)
>> 0 cpu_atom/LLC-load-misses/ # 0.000 /sec (58.32%)
>> <not supported> cpu_core/L1-icache-loads/
>> 21,982,491 cpu_atom/L1-icache-loads/ # 682.028 K/sec (58.43%)
>> 4,493,189 cpu_core/L1-icache-load-misses/ # 139.406 K/sec (33.34%)
>> 4,711,404 cpu_atom/L1-icache-load-misses/ # 146.176 K/sec (50.08%)
>> 13,713,090 cpu_core/dTLB-loads/ # 425.462 K/sec (33.34%)
>> 9,384,727 cpu_atom/dTLB-loads/ # 291.170 K/sec (50.08%)
>> 157,387 cpu_core/dTLB-load-misses/ # 4.883 K/sec (33.33%)
>> 108,328 cpu_atom/dTLB-load-misses/ # 3.361 K/sec (50.08%)
>> <not supported> cpu_core/iTLB-loads/
>> <not supported> cpu_atom/iTLB-loads/
>> 37,655 cpu_core/iTLB-load-misses/ # 1.168 K/sec (33.32%)
>> 61,661 cpu_atom/iTLB-load-misses/ # 1.913 K/sec (50.03%)
>> <not supported> cpu_core/L1-dcache-prefetches/
>> <not supported> cpu_atom/L1-dcache-prefetches/
>> <not supported> cpu_core/L1-dcache-prefetch-misses/
>> <not supported> cpu_atom/L1-dcache-prefetch-misses/
>>
>> 1.005466919 seconds time elapsed
>>
>> Signed-off-by: Kan Liang <[email protected]>
>> Signed-off-by: Zhengjun Xing <[email protected]>
>> ---
>> tools/perf/arch/x86/util/evlist.c | 52 ++++++++++++++++++++++++++++++-
>> tools/perf/util/evlist.c | 2 +-
>> tools/perf/util/evlist.h | 2 ++
>> 3 files changed, 54 insertions(+), 2 deletions(-)
>>
>> diff --git a/tools/perf/arch/x86/util/evlist.c b/tools/perf/arch/x86/util/evlist.c
>> index 777bdf182a58..1b3f9e1a2287 100644
>> --- a/tools/perf/arch/x86/util/evlist.c
>> +++ b/tools/perf/arch/x86/util/evlist.c
>> @@ -4,16 +4,66 @@
>> #include "util/evlist.h"
>> #include "util/parse-events.h"
>> #include "topdown.h"
>> +#include "util/event.h"
>> +#include "util/pmu-hybrid.h"
>>
>> #define TOPDOWN_L1_EVENTS "{slots,topdown-retiring,topdown-bad-spec,topdown-fe-bound,topdown-be-bound}"
>> #define TOPDOWN_L2_EVENTS "{slots,topdown-retiring,topdown-bad-spec,topdown-fe-bound,topdown-be-bound,topdown-heavy-ops,topdown-br-mispredict,topdown-fetch-lat,topdown-mem-bound}"
>>
>> +static int ___evlist__add_default_attrs(struct evlist *evlist, struct perf_event_attr *attrs, size_t nr_attrs)
>> +{
>> + struct perf_cpu_map *cpus;
>> + struct evsel *evsel, *n;
>> + struct perf_pmu *pmu;
>> + LIST_HEAD(head);
>> + size_t i, j = 0;
>> +
>> + for (i = 0; i < nr_attrs; i++)
>> + event_attr_init(attrs + i);
>> +
>> + if (!perf_pmu__has_hybrid())
>> + return evlist__add_attrs(evlist, attrs, nr_attrs);
>> +
>> + for (i = 0; i < nr_attrs; i++) {
>> + if (attrs[i].type == PERF_TYPE_SOFTWARE) {
>> + evsel = evsel__new_idx(attrs + i, evlist->core.nr_entries + j);
>
> Probably no need to calculate index (j) as it's updated
> later when it goes to the evlist...
>
>
>> + if (evsel == NULL)
>> + goto out_delete_partial_list;
>> + j++;
>> + list_add_tail(&evsel->core.node, &head);
>> + continue;
>> + }
>> +
>> + perf_pmu__for_each_hybrid_pmu(pmu) {
>> + evsel = evsel__new_idx(attrs + i, evlist->core.nr_entries + j);
>> + if (evsel == NULL)
>> + goto out_delete_partial_list;
>> + j++;
>> + evsel->core.attr.config |= (__u64)pmu->type << PERF_PMU_TYPE_SHIFT;
>> + cpus = perf_cpu_map__get(pmu->cpus);
>> + evsel->core.cpus = cpus;
>> + evsel->core.own_cpus = perf_cpu_map__get(cpus);
>> + evsel->pmu_name = strdup(pmu->name);
>> + list_add_tail(&evsel->core.node, &head);
>> + }
>> + }
>> +
>> + evlist__splice_list_tail(evlist, &head);
>
> ... like here.
Yes, the index of all new evsel will be updated when adding to the evlist.
Zhengjun, could you please handle the patch? Just set 0 idx for the new
evsel should be good enough.
Thanks,
Kan
>
> Thanks,
> Namhyung
>
>
>> +
>> + return 0;
>> +
>> +out_delete_partial_list:
>> + __evlist__for_each_entry_safe(&head, n, evsel)
>> + evsel__delete(evsel);
>> + return -1;
>> +}
On 6/9/2022 8:47 PM, Liang, Kan wrote:
>
>
> On 6/8/2022 8:04 PM, Namhyung Kim wrote:
>> Hello,
>>
>> On Tue, Jun 7, 2022 at 12:31 AM <[email protected]> wrote:
>>>
>>> From: Kan Liang <[email protected]>
>>>
>>> Provide a new solution to replace the reverted commit ac2dc29edd21
>>> ("perf stat: Add default hybrid events").
>>>
>>> For the default software attrs, nothing is changed.
>>> For the default hardware attrs, create a new evsel for each hybrid pmu.
>>>
>>> With the new solution, adding a new default attr will not require the
>>> special support for the hybrid platform anymore.
>>>
>>> Also, the "--detailed" is supported on the hybrid platform
>>>
>>> With the patch,
>>>
>>> ./perf stat -a -ddd sleep 1
>>>
>>> Performance counter stats for 'system wide':
>>>
>>> 32,231.06 msec cpu-clock # 32.056 CPUs
>>> utilized
>>> 529 context-switches # 16.413 /sec
>>> 32 cpu-migrations # 0.993 /sec
>>> 69 page-faults # 2.141 /sec
>>> 176,754,151 cpu_core/cycles/ # 5.484
>>> M/sec (41.65%)
>>> 161,695,280 cpu_atom/cycles/ # 5.017
>>> M/sec (49.92%)
>>> 48,595,992 cpu_core/instructions/ # 1.508
>>> M/sec (49.98%)
>>> 32,363,337 cpu_atom/instructions/ # 1.004
>>> M/sec (58.26%)
>>> 10,088,639 cpu_core/branches/ # 313.010
>>> K/sec (58.31%)
>>> 6,390,582 cpu_atom/branches/ # 198.274
>>> K/sec (58.26%)
>>> 846,201 cpu_core/branch-misses/ # 26.254
>>> K/sec (66.65%)
>>> 676,477 cpu_atom/branch-misses/ # 20.988
>>> K/sec (58.27%)
>>> 14,290,070 cpu_core/L1-dcache-loads/ # 443.363
>>> K/sec (66.66%)
>>> 9,983,532 cpu_atom/L1-dcache-loads/ # 309.749
>>> K/sec (58.27%)
>>> 740,725 cpu_core/L1-dcache-load-misses/ # 22.982
>>> K/sec (66.66%)
>>> <not supported> cpu_atom/L1-dcache-load-misses/
>>> 480,441 cpu_core/LLC-loads/ # 14.906
>>> K/sec (66.67%)
>>> 326,570 cpu_atom/LLC-loads/ # 10.132
>>> K/sec (58.27%)
>>> 329 cpu_core/LLC-load-misses/ # 10.208
>>> /sec (66.68%)
>>> 0 cpu_atom/LLC-load-misses/ # 0.000
>>> /sec (58.32%)
>>> <not supported> cpu_core/L1-icache-loads/
>>> 21,982,491 cpu_atom/L1-icache-loads/ # 682.028
>>> K/sec (58.43%)
>>> 4,493,189 cpu_core/L1-icache-load-misses/ # 139.406
>>> K/sec (33.34%)
>>> 4,711,404 cpu_atom/L1-icache-load-misses/ # 146.176
>>> K/sec (50.08%)
>>> 13,713,090 cpu_core/dTLB-loads/ # 425.462
>>> K/sec (33.34%)
>>> 9,384,727 cpu_atom/dTLB-loads/ # 291.170
>>> K/sec (50.08%)
>>> 157,387 cpu_core/dTLB-load-misses/ # 4.883
>>> K/sec (33.33%)
>>> 108,328 cpu_atom/dTLB-load-misses/ # 3.361
>>> K/sec (50.08%)
>>> <not supported> cpu_core/iTLB-loads/
>>> <not supported> cpu_atom/iTLB-loads/
>>> 37,655 cpu_core/iTLB-load-misses/ # 1.168
>>> K/sec (33.32%)
>>> 61,661 cpu_atom/iTLB-load-misses/ # 1.913
>>> K/sec (50.03%)
>>> <not supported> cpu_core/L1-dcache-prefetches/
>>> <not supported> cpu_atom/L1-dcache-prefetches/
>>> <not supported> cpu_core/L1-dcache-prefetch-misses/
>>> <not supported> cpu_atom/L1-dcache-prefetch-misses/
>>>
>>> 1.005466919 seconds time elapsed
>>>
>>> Signed-off-by: Kan Liang <[email protected]>
>>> Signed-off-by: Zhengjun Xing <[email protected]>
>>> ---
>>> tools/perf/arch/x86/util/evlist.c | 52 ++++++++++++++++++++++++++++++-
>>> tools/perf/util/evlist.c | 2 +-
>>> tools/perf/util/evlist.h | 2 ++
>>> 3 files changed, 54 insertions(+), 2 deletions(-)
>>>
>>> diff --git a/tools/perf/arch/x86/util/evlist.c
>>> b/tools/perf/arch/x86/util/evlist.c
>>> index 777bdf182a58..1b3f9e1a2287 100644
>>> --- a/tools/perf/arch/x86/util/evlist.c
>>> +++ b/tools/perf/arch/x86/util/evlist.c
>>> @@ -4,16 +4,66 @@
>>> #include "util/evlist.h"
>>> #include "util/parse-events.h"
>>> #include "topdown.h"
>>> +#include "util/event.h"
>>> +#include "util/pmu-hybrid.h"
>>>
>>> #define TOPDOWN_L1_EVENTS
>>> "{slots,topdown-retiring,topdown-bad-spec,topdown-fe-bound,topdown-be-bound}"
>>>
>>> #define TOPDOWN_L2_EVENTS
>>> "{slots,topdown-retiring,topdown-bad-spec,topdown-fe-bound,topdown-be-bound,topdown-heavy-ops,topdown-br-mispredict,topdown-fetch-lat,topdown-mem-bound}"
>>>
>>>
>>> +static int ___evlist__add_default_attrs(struct evlist *evlist,
>>> struct perf_event_attr *attrs, size_t nr_attrs)
>>> +{
>>> + struct perf_cpu_map *cpus;
>>> + struct evsel *evsel, *n;
>>> + struct perf_pmu *pmu;
>>> + LIST_HEAD(head);
>>> + size_t i, j = 0;
>>> +
>>> + for (i = 0; i < nr_attrs; i++)
>>> + event_attr_init(attrs + i);
>>> +
>>> + if (!perf_pmu__has_hybrid())
>>> + return evlist__add_attrs(evlist, attrs, nr_attrs);
>>> +
>>> + for (i = 0; i < nr_attrs; i++) {
>>> + if (attrs[i].type == PERF_TYPE_SOFTWARE) {
>>> + evsel = evsel__new_idx(attrs + i,
>>> evlist->core.nr_entries + j);
>>
>> Probably no need to calculate index (j) as it's updated
>> later when it goes to the evlist...
>>
>>
>>> + if (evsel == NULL)
>>> + goto out_delete_partial_list;
>>> + j++;
>>> + list_add_tail(&evsel->core.node, &head);
>>> + continue;
>>> + }
>>> +
>>> + perf_pmu__for_each_hybrid_pmu(pmu) {
>>> + evsel = evsel__new_idx(attrs + i,
>>> evlist->core.nr_entries + j);
>>> + if (evsel == NULL)
>>> + goto out_delete_partial_list;
>>> + j++;
>>> + evsel->core.attr.config |= (__u64)pmu->type
>>> << PERF_PMU_TYPE_SHIFT;
>>> + cpus = perf_cpu_map__get(pmu->cpus);
>>> + evsel->core.cpus = cpus;
>>> + evsel->core.own_cpus = perf_cpu_map__get(cpus);
>>> + evsel->pmu_name = strdup(pmu->name);
>>> + list_add_tail(&evsel->core.node, &head);
>>> + }
>>> + }
>>> +
>>> + evlist__splice_list_tail(evlist, &head);
>>
>> ... like here.
>
> Yes, the index of all new evsel will be updated when adding to the evlist.
>
> Zhengjun, could you please handle the patch? Just set 0 idx for the new
> evsel should be good enough.
>
>
Ok, I will update it in the new version.
> Thanks,
> Kan
>
>>
>> Thanks,
>> Namhyung
>>
>>
>>> +
>>> + return 0;
>>> +
>>> +out_delete_partial_list:
>>> + __evlist__for_each_entry_safe(&head, n, evsel)
>>> + evsel__delete(evsel);
>>> + return -1;
>>> +}
--
Zhengjun Xing