Received: by 2002:a05:6a10:9848:0:0:0:0 with SMTP id x8csp127610pxf; Wed, 10 Mar 2021 23:10:53 -0800 (PST) X-Google-Smtp-Source: ABdhPJy/HkfazhbCbowtYzQIUd2Cfv9iT2Ql7EYA9UTAVg/acTKcB+MjbrCk5RcG0T3QtN6xG4Xw X-Received: by 2002:a17:906:a8a:: with SMTP id y10mr1738585ejf.288.1615446653366; Wed, 10 Mar 2021 23:10:53 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1615446653; cv=none; d=google.com; s=arc-20160816; b=UMY77yUglbbp/jlku7JIEfQQO0m2H/k9eFJW4KZG6bH8nZPEAanJTUBn797dWY5Aie KEjEfsgwtLBXTBCM1ZpM+e/4eEpCXWYb+84IndVGthv+cK+R1uxcbSepZSkihsKakx0X ePrBH3zrjcCu8tzE0O5kYi9c0NI4L9ofonkqvPpnMxcmUCgQHP/jvBVUqwVoVs4pidZQ ZiaoaU0Bc9Ym8HQOXVqCbrJ4Z67y5UYJoXlLvdRGOMJM9MB/z/cu3g3mOhQPUrgpeZHx xSdpIlqn2avvLNuxaexoDt3nzQFGjrVyV3J0WBPj6yN6SXNHM/j5sbHh3c5TdOJQQ5L+ nh8A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:references:in-reply-to:message-id:date:subject :cc:to:from:ironport-sdr:ironport-sdr; bh=5fMWScDrh9Ud0P4MHMTPcXFoxTtLuZOgvHY3EZ+F5BA=; b=pl/CchVG0FLsITWxhM7WAi1Um9FAi3Faea8ShFPlNM38jAWfWmd1bIhyz8t1tBj8r8 S8PxQKl7rMg9mImaN6v1rGSx2chttBhtx2aKzG4PSrmwWua1Z5iUdEyhjEs3jDw692m5 auEd6I003AXnaiwZaRAOuESr0qiwaosm34IS41hnM4NBFE9HxN65wIo7mb/SSIZa1jQH 5/Z0QYhpj6C7sNIeTjBu6r8j1IyfJG6Xkti2AmLZXuARGNlQz9jF7wDZIbZqNIzHEeNP YHiGwag73wrkJBfxz+jQMvFh9M4q44Zf9/si2OnSn+nf2hqWS6KpRmYJ5mh0cnxHs3e5 h4nw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id l11si1229725ejq.118.2021.03.10.23.10.31; Wed, 10 Mar 2021 23:10:53 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231755AbhCKHJC (ORCPT + 99 others); Thu, 11 Mar 2021 02:09:02 -0500 Received: from mga04.intel.com ([192.55.52.120]:22599 "EHLO mga04.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231633AbhCKHIg (ORCPT ); Thu, 11 Mar 2021 02:08:36 -0500 IronPort-SDR: bSvvpyMiAtwzlG/q2GLFq1GtEOc8xRjyqR4Xl11VcR7KEL4fMQp6H1aRP1vCiJNyPZqqxF2mLx Hy8tMhgXnMNw== X-IronPort-AV: E=McAfee;i="6000,8403,9919"; a="186246042" X-IronPort-AV: E=Sophos;i="5.81,239,1610438400"; d="scan'208";a="186246042" Received: from fmsmga001.fm.intel.com ([10.253.24.23]) by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 10 Mar 2021 23:08:36 -0800 IronPort-SDR: +kwOJwSS40JmlNf62kZLbpyvGt6SwE5hzxBOz+B236QqmGKYJCNdJLBFk68ORyzdO90wx6UYbE 2XMEfdbjjoQA== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.81,239,1610438400"; d="scan'208";a="509937861" Received: from kbl-ppc.sh.intel.com ([10.239.159.163]) by fmsmga001.fm.intel.com with ESMTP; 10 Mar 2021 23:08:34 -0800 From: Jin Yao To: acme@kernel.org, jolsa@kernel.org, peterz@infradead.org, mingo@redhat.com, alexander.shishkin@linux.intel.com Cc: Linux-kernel@vger.kernel.org, ak@linux.intel.com, kan.liang@intel.com, yao.jin@intel.com, Jin Yao Subject: [PATCH v2 07/27] perf evlist: Hybrid event uses its own cpus Date: Thu, 11 Mar 2021 15:07:22 +0800 Message-Id: <20210311070742.9318-8-yao.jin@linux.intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20210311070742.9318-1-yao.jin@linux.intel.com> References: <20210311070742.9318-1-yao.jin@linux.intel.com> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On hybrid platform, atom events can be only enabled on atom CPUs. Core events can be only enabled on core CPUs. So for a hybrid event, it can be only enabled on it's own CPUs. But the problem for current perf is, the cpus for evsel (via PMU sysfs) have been merged to evsel_list->core.all_cpus. It might be all CPUs. So we need to figure out one way to let the hybrid event only use it's own CPUs. The idea is to create a new evlist__invalidate_all_cpus to invalidate the evsel_list->core.all_cpus then evlist__for_each_cpu returns cpu -1 for hybrid evsel. If cpu is -1, hybrid evsel will use it's own cpus. We will see following code piece in patch. if (cpu == -1 && !evlist->thread_mode) evsel__enable_cpus(pos); It lets the event be enabled on event's own cpus. Signed-off-by: Jin Yao --- tools/perf/builtin-stat.c | 37 ++++++++++++++- tools/perf/util/evlist.c | 72 ++++++++++++++++++++++++++++-- tools/perf/util/evlist.h | 4 ++ tools/perf/util/evsel.h | 8 ++++ tools/perf/util/python-ext-sources | 2 + 5 files changed, 117 insertions(+), 6 deletions(-) diff --git a/tools/perf/builtin-stat.c b/tools/perf/builtin-stat.c index 2e2e4a8345ea..68ecf68699a9 100644 --- a/tools/perf/builtin-stat.c +++ b/tools/perf/builtin-stat.c @@ -393,6 +393,18 @@ static int read_counter_cpu(struct evsel *counter, struct timespec *rs, int cpu) return 0; } +static int read_counter_cpus(struct evsel *counter, struct timespec *rs) +{ + int cpu, nr_cpus, err = 0; + struct perf_cpu_map *cpus = evsel__cpus(counter); + + nr_cpus = cpus ? cpus->nr : 1; + for (cpu = 0; cpu < nr_cpus; cpu++) + err = read_counter_cpu(counter, rs, cpu); + + return err; +} + static int read_affinity_counters(struct timespec *rs) { struct evsel *counter; @@ -414,8 +426,14 @@ static int read_affinity_counters(struct timespec *rs) if (evsel__cpu_iter_skip(counter, cpu)) continue; if (!counter->err) { - counter->err = read_counter_cpu(counter, rs, - counter->cpu_iter - 1); + if (cpu == -1 && !evsel_list->thread_mode) { + counter->err = read_counter_cpus(counter, rs); + } else if (evsel_list->thread_mode) { + counter->err = read_counter_cpu(counter, rs, 0); + } else { + counter->err = read_counter_cpu(counter, rs, + counter->cpu_iter - 1); + } } } } @@ -781,6 +799,21 @@ static int __run_perf_stat(int argc, const char **argv, int run_idx) if (group) evlist__set_leader(evsel_list); + /* + * On hybrid platform, the cpus for evsel (via PMU sysfs) have been + * merged to evsel_list->core.all_cpus. We use evlist__invalidate_all_cpus + * to invalidate the evsel_list->core.all_cpus then evlist__for_each_cpu + * returns cpu -1 for hybrid evsel. If cpu is -1, hybrid evsel will + * use it's own cpus. + */ + if (evlist__has_hybrid_events(evsel_list)) { + evlist__invalidate_all_cpus(evsel_list); + if (!target__has_cpu(&target) || + target__has_per_thread(&target)) { + evsel_list->thread_mode = true; + } + } + if (affinity__setup(&affinity) < 0) return -1; diff --git a/tools/perf/util/evlist.c b/tools/perf/util/evlist.c index 882cd1f721d9..3ee12fcd0c9f 100644 --- a/tools/perf/util/evlist.c +++ b/tools/perf/util/evlist.c @@ -381,7 +381,8 @@ bool evsel__cpu_iter_skip_no_inc(struct evsel *ev, int cpu) bool evsel__cpu_iter_skip(struct evsel *ev, int cpu) { if (!evsel__cpu_iter_skip_no_inc(ev, cpu)) { - ev->cpu_iter++; + if (cpu != -1) + ev->cpu_iter++; return false; } return true; @@ -410,6 +411,16 @@ static int evlist__is_enabled(struct evlist *evlist) return false; } +static void evsel__disable_cpus(struct evsel *evsel) +{ + int cpu, nr_cpus; + struct perf_cpu_map *cpus = evsel__cpus(evsel); + + nr_cpus = cpus ? cpus->nr : 1; + for (cpu = 0; cpu < nr_cpus; cpu++) + evsel__disable_cpu(evsel, cpu); +} + static void __evlist__disable(struct evlist *evlist, char *evsel_name) { struct evsel *pos; @@ -436,7 +447,12 @@ static void __evlist__disable(struct evlist *evlist, char *evsel_name) has_imm = true; if (pos->immediate != imm) continue; - evsel__disable_cpu(pos, pos->cpu_iter - 1); + if (cpu == -1 && !evlist->thread_mode) + evsel__disable_cpus(pos); + else if (evlist->thread_mode) + evsel__disable_cpu(pos, 0); + else + evsel__disable_cpu(pos, pos->cpu_iter - 1); } } if (!has_imm) @@ -472,6 +488,15 @@ void evlist__disable_evsel(struct evlist *evlist, char *evsel_name) __evlist__disable(evlist, evsel_name); } +static void evsel__enable_cpus(struct evsel *evsel) +{ + int cpu, nr_cpus; + struct perf_cpu_map *cpus = evsel__cpus(evsel); + + nr_cpus = cpus ? cpus->nr : 1; + for (cpu = 0; cpu < nr_cpus; cpu++) + evsel__enable_cpu(evsel, cpu); +} static void __evlist__enable(struct evlist *evlist, char *evsel_name) { struct evsel *pos; @@ -491,7 +516,12 @@ static void __evlist__enable(struct evlist *evlist, char *evsel_name) continue; if (!evsel__is_group_leader(pos) || !pos->core.fd) continue; - evsel__enable_cpu(pos, pos->cpu_iter - 1); + if (cpu == -1 && !evlist->thread_mode) + evsel__enable_cpus(pos); + else if (evlist->thread_mode) + evsel__enable_cpu(pos, 0); + else + evsel__enable_cpu(pos, pos->cpu_iter - 1); } } affinity__cleanup(&affinity); @@ -1274,6 +1304,16 @@ void evlist__set_selected(struct evlist *evlist, struct evsel *evsel) evlist->selected = evsel; } +static void evsel__close_cpus(struct evsel *evsel) +{ + int cpu, nr_cpus; + struct perf_cpu_map *cpus = evsel__cpus(evsel); + + nr_cpus = cpus ? cpus->nr : 1; + for (cpu = 0; cpu < nr_cpus; cpu++) + perf_evsel__close_cpu(&evsel->core, cpu); +} + void evlist__close(struct evlist *evlist) { struct evsel *evsel; @@ -1298,7 +1338,13 @@ void evlist__close(struct evlist *evlist) evlist__for_each_entry_reverse(evlist, evsel) { if (evsel__cpu_iter_skip(evsel, cpu)) continue; - perf_evsel__close_cpu(&evsel->core, evsel->cpu_iter - 1); + + if (cpu == -1 && !evlist->thread_mode) + evsel__close_cpus(evsel); + else if (evlist->thread_mode) + perf_evsel__close_cpu(&evsel->core, 0); + else + perf_evsel__close_cpu(&evsel->core, evsel->cpu_iter - 1); } } affinity__cleanup(&affinity); @@ -2130,3 +2176,21 @@ struct evsel *evlist__find_evsel(struct evlist *evlist, int idx) } return NULL; } + +bool evlist__has_hybrid_events(struct evlist *evlist) +{ + struct evsel *evsel; + + evlist__for_each_entry(evlist, evsel) { + if (evsel__is_hybrid_event(evsel)) + return true; + } + + return false; +} + +void evlist__invalidate_all_cpus(struct evlist *evlist) +{ + perf_cpu_map__put(evlist->core.all_cpus); + evlist->core.all_cpus = perf_cpu_map__empty_new(1); +} diff --git a/tools/perf/util/evlist.h b/tools/perf/util/evlist.h index b695ffaae519..0da683511d98 100644 --- a/tools/perf/util/evlist.h +++ b/tools/perf/util/evlist.h @@ -52,6 +52,7 @@ struct evlist { struct perf_evlist core; int nr_groups; bool enabled; + bool thread_mode; int id_pos; int is_pos; u64 combined_sample_type; @@ -365,4 +366,7 @@ int evlist__ctlfd_ack(struct evlist *evlist); #define EVLIST_DISABLED_MSG "Events disabled\n" struct evsel *evlist__find_evsel(struct evlist *evlist, int idx); +void evlist__invalidate_all_cpus(struct evlist *evlist); + +bool evlist__has_hybrid_events(struct evlist *evlist); #endif /* __PERF_EVLIST_H */ diff --git a/tools/perf/util/evsel.h b/tools/perf/util/evsel.h index 6026487353dd..69aadc52c1bd 100644 --- a/tools/perf/util/evsel.h +++ b/tools/perf/util/evsel.h @@ -7,9 +7,11 @@ #include #include #include +#include #include #include #include "symbol_conf.h" +#include "pmu-hybrid.h" #include struct bpf_object; @@ -435,4 +437,10 @@ struct perf_env *evsel__env(struct evsel *evsel); int evsel__store_ids(struct evsel *evsel, struct evlist *evlist); void evsel__zero_per_pkg(struct evsel *evsel); + +static inline bool evsel__is_hybrid_event(struct evsel *evsel) +{ + return evsel->pmu_name && perf_pmu__is_hybrid(evsel->pmu_name); +} + #endif /* __PERF_EVSEL_H */ diff --git a/tools/perf/util/python-ext-sources b/tools/perf/util/python-ext-sources index 845dd46e3c61..d7c976671e3a 100644 --- a/tools/perf/util/python-ext-sources +++ b/tools/perf/util/python-ext-sources @@ -37,3 +37,5 @@ util/units.c util/affinity.c util/rwsem.c util/hashmap.c +util/pmu-hybrid.c +util/fncache.c -- 2.17.1