Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1758128Ab3CYN52 (ORCPT ); Mon, 25 Mar 2013 09:57:28 -0400 Received: from mail-qe0-f48.google.com ([209.85.128.48]:39341 "EHLO mail-qe0-f48.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1757974Ab3CYN51 (ORCPT ); Mon, 25 Mar 2013 09:57:27 -0400 MIME-Version: 1.0 In-Reply-To: References: <1360846649-6411-1-git-send-email-eranian@google.com> Date: Mon, 25 Mar 2013 14:57:26 +0100 Message-ID: Subject: Re: [PATCH v2 0/3] perf stat: add per-core count aggregation From: Stephane Eranian To: LKML Cc: Peter Zijlstra , "mingo@elte.hu" , "ak@linux.intel.com" , Arnaldo Carvalho de Melo , Jiri Olsa , Namhyung Kim Content-Type: text/plain; charset=UTF-8 Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2759 Lines: 69 Arnaldo, Where are we with this one? On Thu, Mar 7, 2013 at 5:22 PM, Stephane Eranian wrote: > Arnaldo, > > Any comments on this series? > > > On Thu, Feb 14, 2013 at 1:57 PM, Stephane Eranian wrote: >> This patch series contains improvement to the aggregation support >> in perf stat. >> >> First, the aggregation code is refactored and a aggr_mode enum >> is defined. There is also an important bug fix for the existing >> per-socket aggregation. >> >> Second, the option --aggr-socket is renamed --per-socket. >> >> Third, the patch adds a new --per-core option to perf stat. >> It aggregates counts per physical core and becomes useful on >> systems with hyper-threading. The cores are presented per >> socket: S0-C1, means socket 0 core 1. Note that the core number >> represents its physical core id. As such, numbers may not always >> be contiguous. All of this is based on topology information available >> in sysfs. >> >> Per-core aggregation can be combined with interval printing: >> >> # perf stat -a --per-core -I 1000 -e cycles sleep 100 >> # time core cpus counts events >> 1.000101160 S0-C0 2 6,051,254,899 cycles >> 1.000101160 S0-C1 2 6,379,230,776 cycles >> 1.000101160 S0-C2 2 6,480,268,471 cycles >> 1.000101160 S0-C3 2 6,110,514,321 cycles >> 2.000663750 S0-C0 2 6,572,533,016 cycles >> 2.000663750 S0-C1 2 6,378,623,674 cycles >> 2.000663750 S0-C2 2 6,264,127,589 cycles >> 2.000663750 S0-C3 2 6,305,346,613 cycles >> >> For instance here on this SNB machine, we can see that the load >> is evenly balanced across all 4 physical core (HT is on). >> >> In v2, we print events across all cores or socket and we renamed >> --aggr-socket to --per-socket and renamed --aggr-core to --per-core >> >> Signed-off-by: Stephane Eranian >> >> Stephane Eranian (3): >> perf stat: refactor aggregation code >> perf stat: rename --aggr-socket to --per-socket >> perf stat: add per-core aggregation >> >> tools/perf/Documentation/perf-stat.txt | 10 +- >> tools/perf/builtin-stat.c | 237 ++++++++++++++++++++------------ >> tools/perf/util/cpumap.c | 86 ++++++++++-- >> tools/perf/util/cpumap.h | 12 ++ >> 4 files changed, 241 insertions(+), 104 deletions(-) >> >> -- >> 1.7.9.5 >> -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/