Received: by 2002:a05:6a10:8c0a:0:0:0:0 with SMTP id go10csp4190807pxb; Mon, 8 Feb 2021 10:02:01 -0800 (PST) X-Google-Smtp-Source: ABdhPJytklwBHJxB78IhpuzaTu66GlEwQaGsEESyy5M9iIHjLSSkqXWOzbEBBO7rh+yh3D4mxP8I X-Received: by 2002:ac2:5095:: with SMTP id f21mr10632734lfm.289.1612807321099; Mon, 08 Feb 2021 10:02:01 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1612807321; cv=none; d=google.com; s=arc-20160816; b=VNGw6gjfKrbtNxdG36QYy9YtpJ7zSZos8/cEV4Yx8QUXYE3zRuGoW1/loUB7crfaos WQc+0MradN0jCi0S4DjHd2VX7j8HzSsyovxPKfCsvNRpEaUu95v4wPq7PCatHFDre7NG VyQMUG5XrUSHjcIk31hPUUV0KixdGAe/uWkpjNPDOFTbKuQl7qRBI7s50DIkH8qGitec 8A09dERlOX3dtemGv8khO6cSRqAX9U9GP7Q1j6QIFbo9Jqrwuf5qGJFhsiHGobHP+4Rs 7UuaWVCNIhl1ZI5lwOLXv9oK3CaRvVGcMjClDtAt+3c0WZ6Ph9T1DuU4IroBB5q5jJsw JLKA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:references:in-reply-to:message-id:date:subject :cc:to:from:ironport-sdr:ironport-sdr; bh=GNIxZ8rKCVOexjrXrl3Sjw+aMzq0pxTI3PrEf6k6rIU=; b=iV9ZjW6TL1v+Rg1RTUezuLuRAZfIrrU+DHW1JbI6YUD3yhB5o7Cg4TRti8FI1Lyl+a d8nm38Wr1EQMmIBi8f2Cf1kcjUwj7sBN9Ee8RFz4h18USekcefyJxgxpJO4eoEgNfDez 9F0oxZqlnBHoP7bm/RTeD6DINNEwXwhSDrOwnMwem9eGETKq7TUloAs00qXGm38aOrQe 6BgEdtffmoMCpOYcxBTIewnuoMpFrhCny6riQS+BFiUNjkilQC/PrKnf5dlpNKn+tLff FUvPDdUlFfJI+0nVElgDI1fQrxQxnugVt3XdEjoqRJN9vuHJ1HDOz/UsL11DEfoMsH6L zCjA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id u23si14006747edi.71.2021.02.08.10.01.36; Mon, 08 Feb 2021 10:02:01 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S235103AbhBHR7g (ORCPT + 99 others); Mon, 8 Feb 2021 12:59:36 -0500 Received: from mga14.intel.com ([192.55.52.115]:62781 "EHLO mga14.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S233966AbhBHPfK (ORCPT ); Mon, 8 Feb 2021 10:35:10 -0500 IronPort-SDR: mUf51565TqM8VmrwaSnD0IBKxj5c4wiwRo7ijTpYYLG4pi3w1weLa4hNXkjpcCdbxYiVabJ5Mo +oBGu2I0d9uA== X-IronPort-AV: E=McAfee;i="6000,8403,9889"; a="180951903" X-IronPort-AV: E=Sophos;i="5.81,162,1610438400"; d="scan'208";a="180951903" Received: from orsmga003.jf.intel.com ([10.7.209.27]) by fmsmga103.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 08 Feb 2021 07:30:02 -0800 IronPort-SDR: wxJWNEo5UxwaAZwDC4VhTt20CfaBS++JbvASN2NLpRuUhdk1AnGwA5ysl176rw9/dARkTjvcM2 kpgpY1R9UhfQ== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.81,162,1610438400"; d="scan'208";a="358820533" Received: from otc-lr-04.jf.intel.com ([10.54.39.41]) by orsmga003.jf.intel.com with ESMTP; 08 Feb 2021 07:30:02 -0800 From: kan.liang@linux.intel.com To: peterz@infradead.org, acme@kernel.org, mingo@kernel.org, linux-kernel@vger.kernel.org Cc: tglx@linutronix.de, bp@alien8.de, namhyung@kernel.org, jolsa@redhat.com, ak@linux.intel.com, yao.jin@linux.intel.com, alexander.shishkin@linux.intel.com, adrian.hunter@intel.com, Kan Liang Subject: [PATCH 05/49] perf/x86: Hybrid PMU support for counters Date: Mon, 8 Feb 2021 07:25:02 -0800 Message-Id: <1612797946-18784-6-git-send-email-kan.liang@linux.intel.com> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1612797946-18784-1-git-send-email-kan.liang@linux.intel.com> References: <1612797946-18784-1-git-send-email-kan.liang@linux.intel.com> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Kan Liang The number of GP and fixed counters are different among hybrid PMUs. Each hybrid PMU should use its own counter related information. When handling a certain hybrid PMU, apply the number of counters from the corresponding hybrid PMU. When reserving the counters in the initialization of a new event, reserve all possible counters. Add a hybrid_pmu_bitmap to indicate the possible hybrid PMUs. For the non-hybrid architecture, it's empty. The number of counter recored in the global x86_pmu is for the architecture counters which are available for all hybrid PMUs. KVM doesn't support the hybrid PMU yet. Return the number of the architecture counters for now. For the functions only available for the old platforms, e.g., intel_pmu_drain_pebs_nhm(), nothing is changed. Reviewed-by: Andi Kleen Signed-off-by: Kan Liang --- arch/x86/events/core.c | 55 ++++++++++++++++++++++++++++++-------------- arch/x86/events/intel/core.c | 8 ++++--- arch/x86/events/intel/ds.c | 14 +++++++---- arch/x86/events/perf_event.h | 5 ++++ 4 files changed, 57 insertions(+), 25 deletions(-) diff --git a/arch/x86/events/core.c b/arch/x86/events/core.c index 170acbf..5f79b37 100644 --- a/arch/x86/events/core.c +++ b/arch/x86/events/core.c @@ -179,16 +179,29 @@ static DEFINE_MUTEX(pmc_reserve_mutex); #ifdef CONFIG_X86_LOCAL_APIC +static inline int get_possible_num_counters(void) +{ + int bit, num_counters = 0; + + if (!IS_X86_HYBRID) + return x86_pmu.num_counters; + + for_each_set_bit(bit, &x86_pmu.hybrid_pmu_bitmap, X86_HYBRID_PMU_MAX_INDEX) + num_counters = max_t(int, num_counters, x86_pmu.hybrid_pmu[bit].num_counters); + + return num_counters; +} + static bool reserve_pmc_hardware(void) { - int i; + int i, num_counters = get_possible_num_counters(); - for (i = 0; i < x86_pmu.num_counters; i++) { + for (i = 0; i < num_counters; i++) { if (!reserve_perfctr_nmi(x86_pmu_event_addr(i))) goto perfctr_fail; } - for (i = 0; i < x86_pmu.num_counters; i++) { + for (i = 0; i < num_counters; i++) { if (!reserve_evntsel_nmi(x86_pmu_config_addr(i))) goto eventsel_fail; } @@ -199,7 +212,7 @@ static bool reserve_pmc_hardware(void) for (i--; i >= 0; i--) release_evntsel_nmi(x86_pmu_config_addr(i)); - i = x86_pmu.num_counters; + i = num_counters; perfctr_fail: for (i--; i >= 0; i--) @@ -210,9 +223,9 @@ static bool reserve_pmc_hardware(void) static void release_pmc_hardware(void) { - int i; + int i, num_counters = get_possible_num_counters(); - for (i = 0; i < x86_pmu.num_counters; i++) { + for (i = 0; i < num_counters; i++) { release_perfctr_nmi(x86_pmu_event_addr(i)); release_evntsel_nmi(x86_pmu_config_addr(i)); } @@ -933,6 +946,7 @@ EXPORT_SYMBOL_GPL(perf_assign_events); int x86_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign) { + int num_counters = X86_HYBRID_READ_FROM_CPUC(num_counters, cpuc); struct event_constraint *c; struct perf_event *e; int n0, i, wmin, wmax, unsched = 0; @@ -1008,7 +1022,7 @@ int x86_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign) /* slow path */ if (i != n) { - int gpmax = x86_pmu.num_counters; + int gpmax = num_counters; /* * Do not allow scheduling of more than half the available @@ -1029,7 +1043,7 @@ int x86_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign) * the extra Merge events needed by large increment events. */ if (x86_pmu.flags & PMU_FL_PAIR) { - gpmax = x86_pmu.num_counters - cpuc->n_pair; + gpmax = num_counters - cpuc->n_pair; WARN_ON(gpmax <= 0); } @@ -1116,10 +1130,12 @@ static int collect_event(struct cpu_hw_events *cpuc, struct perf_event *event, */ static int collect_events(struct cpu_hw_events *cpuc, struct perf_event *leader, bool dogrp) { + int num_counters = X86_HYBRID_READ_FROM_CPUC(num_counters, cpuc); + int num_counters_fixed = X86_HYBRID_READ_FROM_CPUC(num_counters_fixed, cpuc); struct perf_event *event; int n, max_count; - max_count = x86_pmu.num_counters + x86_pmu.num_counters_fixed; + max_count = num_counters + num_counters_fixed; /* current number of events already accepted */ n = cpuc->n_events; @@ -1487,18 +1503,18 @@ void perf_event_print_debug(void) { u64 ctrl, status, overflow, pmc_ctrl, pmc_count, prev_left, fixed; u64 pebs, debugctl; - struct cpu_hw_events *cpuc; + int cpu = smp_processor_id(); + struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); + int num_counters = X86_HYBRID_READ_FROM_CPUC(num_counters, cpuc); + int num_counters_fixed = X86_HYBRID_READ_FROM_CPUC(num_counters_fixed, cpuc); unsigned long flags; - int cpu, idx; + int idx; - if (!x86_pmu.num_counters) + if (!num_counters) return; local_irq_save(flags); - cpu = smp_processor_id(); - cpuc = &per_cpu(cpu_hw_events, cpu); - if (x86_pmu.version >= 2) { rdmsrl(MSR_CORE_PERF_GLOBAL_CTRL, ctrl); rdmsrl(MSR_CORE_PERF_GLOBAL_STATUS, status); @@ -1521,7 +1537,7 @@ void perf_event_print_debug(void) } pr_info("CPU#%d: active: %016llx\n", cpu, *(u64 *)cpuc->active_mask); - for (idx = 0; idx < x86_pmu.num_counters; idx++) { + for (idx = 0; idx < num_counters; idx++) { rdmsrl(x86_pmu_config_addr(idx), pmc_ctrl); rdmsrl(x86_pmu_event_addr(idx), pmc_count); @@ -1534,7 +1550,7 @@ void perf_event_print_debug(void) pr_info("CPU#%d: gen-PMC%d left: %016llx\n", cpu, idx, prev_left); } - for (idx = 0; idx < x86_pmu.num_counters_fixed; idx++) { + for (idx = 0; idx < num_counters_fixed; idx++) { if (fixed_counter_disabled(idx, cpuc)) continue; rdmsrl(MSR_ARCH_PERFMON_FIXED_CTR0 + idx, pmc_count); @@ -2776,6 +2792,11 @@ unsigned long perf_misc_flags(struct pt_regs *regs) void perf_get_x86_pmu_capability(struct x86_pmu_capability *cap) { cap->version = x86_pmu.version; + /* + * KVM doesn't support the hybrid PMU yet. + * Return the common value in global x86_pmu, + * which available for all cores. + */ cap->num_counters_gp = x86_pmu.num_counters; cap->num_counters_fixed = x86_pmu.num_counters_fixed; cap->bit_width_gp = x86_pmu.cntval_bits; diff --git a/arch/x86/events/intel/core.c b/arch/x86/events/intel/core.c index 1b9563c..3b8d728 100644 --- a/arch/x86/events/intel/core.c +++ b/arch/x86/events/intel/core.c @@ -2711,21 +2711,23 @@ static void intel_pmu_reset(void) { struct debug_store *ds = __this_cpu_read(cpu_hw_events.ds); struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); + int num_counters = X86_HYBRID_READ_FROM_CPUC(num_counters, cpuc); + int num_counters_fixed = X86_HYBRID_READ_FROM_CPUC(num_counters_fixed, cpuc); unsigned long flags; int idx; - if (!x86_pmu.num_counters) + if (!num_counters) return; local_irq_save(flags); pr_info("clearing PMU state on CPU#%d\n", smp_processor_id()); - for (idx = 0; idx < x86_pmu.num_counters; idx++) { + for (idx = 0; idx < num_counters; idx++) { wrmsrl_safe(x86_pmu_config_addr(idx), 0ull); wrmsrl_safe(x86_pmu_event_addr(idx), 0ull); } - for (idx = 0; idx < x86_pmu.num_counters_fixed; idx++) { + for (idx = 0; idx < num_counters_fixed; idx++) { if (fixed_counter_disabled(idx, cpuc)) continue; wrmsrl_safe(MSR_ARCH_PERFMON_FIXED_CTR0 + idx, 0ull); diff --git a/arch/x86/events/intel/ds.c b/arch/x86/events/intel/ds.c index ba7cf05..a528966 100644 --- a/arch/x86/events/intel/ds.c +++ b/arch/x86/events/intel/ds.c @@ -1007,6 +1007,8 @@ void intel_pmu_pebs_sched_task(struct perf_event_context *ctx, bool sched_in) static inline void pebs_update_threshold(struct cpu_hw_events *cpuc) { struct debug_store *ds = cpuc->ds; + int max_pebs_events = X86_HYBRID_READ_FROM_CPUC(max_pebs_events, cpuc); + int num_counters_fixed = X86_HYBRID_READ_FROM_CPUC(num_counters_fixed, cpuc); u64 threshold; int reserved; @@ -1014,9 +1016,9 @@ static inline void pebs_update_threshold(struct cpu_hw_events *cpuc) return; if (x86_pmu.flags & PMU_FL_PEBS_ALL) - reserved = x86_pmu.max_pebs_events + x86_pmu.num_counters_fixed; + reserved = max_pebs_events + num_counters_fixed; else - reserved = x86_pmu.max_pebs_events; + reserved = max_pebs_events; if (cpuc->n_pebs == cpuc->n_large_pebs) { threshold = ds->pebs_absolute_maximum - @@ -2072,6 +2074,8 @@ static void intel_pmu_drain_pebs_icl(struct pt_regs *iregs, struct perf_sample_d { short counts[INTEL_PMC_IDX_FIXED + MAX_FIXED_PEBS_EVENTS] = {}; struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); + int max_pebs_events = X86_HYBRID_READ_FROM_CPUC(max_pebs_events, cpuc); + int num_counters_fixed = X86_HYBRID_READ_FROM_CPUC(num_counters_fixed, cpuc); struct debug_store *ds = cpuc->ds; struct perf_event *event; void *base, *at, *top; @@ -2086,9 +2090,9 @@ static void intel_pmu_drain_pebs_icl(struct pt_regs *iregs, struct perf_sample_d ds->pebs_index = ds->pebs_buffer_base; - mask = ((1ULL << x86_pmu.max_pebs_events) - 1) | - (((1ULL << x86_pmu.num_counters_fixed) - 1) << INTEL_PMC_IDX_FIXED); - size = INTEL_PMC_IDX_FIXED + x86_pmu.num_counters_fixed; + mask = ((1ULL << max_pebs_events) - 1) | + (((1ULL << num_counters_fixed) - 1) << INTEL_PMC_IDX_FIXED); + size = INTEL_PMC_IDX_FIXED + num_counters_fixed; if (unlikely(base >= top)) { intel_pmu_pebs_event_update_no_drain(cpuc, size); diff --git a/arch/x86/events/perf_event.h b/arch/x86/events/perf_event.h index b939784..bda4bdc 100644 --- a/arch/x86/events/perf_event.h +++ b/arch/x86/events/perf_event.h @@ -648,6 +648,9 @@ struct x86_hybrid_pmu { struct pmu pmu; union perf_capabilities intel_cap; u64 intel_ctrl; + int max_pebs_events; + int num_counters; + int num_counters_fixed; }; #define IS_X86_HYBRID cpu_feature_enabled(X86_FEATURE_HYBRID_CPU) @@ -853,7 +856,9 @@ struct x86_pmu { * Most PMU capabilities are the same among different hybrid PMUs. The * global x86_pmu saves the architecture capabilities, which are available * for all PMUs. The hybrid_pmu only includes the unique capabilities. + * The hybrid_pmu_bitmap is the bits map of the possible hybrid_pmu. */ + unsigned long hybrid_pmu_bitmap; struct x86_hybrid_pmu hybrid_pmu[X86_HYBRID_PMU_MAX_INDEX]; }; -- 2.7.4