Received: by 2002:a05:6358:701b:b0:131:369:b2a3 with SMTP id 27csp4124530rwo; Tue, 25 Jul 2023 00:35:06 -0700 (PDT) X-Google-Smtp-Source: APBJJlHi2Ipn1aaGGz7H/dAaa958PFyyqJe2twbT26WVWtI6oV6uzL42ZF+TSIwNdxPVSTi6z2DJ X-Received: by 2002:ac2:5dee:0:b0:4fb:99c6:8533 with SMTP id z14-20020ac25dee000000b004fb99c68533mr6711803lfq.33.1690270506149; Tue, 25 Jul 2023 00:35:06 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1690270506; cv=none; d=google.com; s=arc-20160816; b=lRY6GU03ch+IZHm4/frWJFghDC9ovY3fIH0isAepUFwwnzNScdXnPThKQrGCwjPBAo V67huJ6IbJWldHU4Cm6jT04MUVW3+e1R2CTM6Yv6sayDnUS1rsgN8RMFLTGlfs0HJ+tD 4q7lduh9Lb6cGRS6L8TchCmOxUMSh0lqFZu3EZgqg5VN2SD2GCMnNe+5qtZwfwDjpmzm t8eZ4I+RYFErbMFKDrkUSRQhWfBkxc1keBR7M7oh/UObn7qOFiDNFnk1vovKbbohTaW4 PX0hSDlu74adjSToyxQwjJp8oqrxkAzBJ9ZC/O/iUKczABVWzPcFENdg31cmtF1j6Ey+ 7t6Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:from :references:cc:to:subject:user-agent:mime-version:date:message-id; bh=cCYAVapXyDaS21aXr1z47JGyPfkW2VnKNqHX7hA/Jkw=; fh=7xEwuYx9eGi65/G5H9OxoJBHb4FxLcOc/i68Jo06TXM=; b=MTndccyk2n7a0y73oiaNi7Eemrp1F1GCImKo8s/ZlImlmnjZmGCu8Vj8d5gvNyOMvY KCXWP62zgHM9k/G5DKP69gMrbP0oum3IjA4nObz1bJpY/OmCe3QTDJpCfl2y5t5gIwKN /DBpLi1zTO+HP1eT4FxHCTHXt4G18JVdUY57DZaF+N+Mrdp97kQefcDK8zg8oxcy3V8H VRdY361Ilj8T0DcBlP49oef5mt1AqTHW0XbjLAYhfgmRmV6aplLSyYs+BipZJhQ+9gq8 6OVN5bkQRXD9Ts1c8sLf8QYOJBLbz3V4kOGSUwUL26g/8fEMWkmHaKqN1tELATomw939 CkyA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=huawei.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id a1-20020aa7d901000000b0051dd50eb02bsi8092929edr.5.2023.07.25.00.34.41; Tue, 25 Jul 2023 00:35:06 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=huawei.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232206AbjGYHNI (ORCPT + 99 others); Tue, 25 Jul 2023 03:13:08 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:38628 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232284AbjGYHM4 (ORCPT ); Tue, 25 Jul 2023 03:12:56 -0400 Received: from szxga02-in.huawei.com (szxga02-in.huawei.com [45.249.212.188]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 36AD5BF; Tue, 25 Jul 2023 00:12:52 -0700 (PDT) Received: from dggpemm500005.china.huawei.com (unknown [172.30.72.53]) by szxga02-in.huawei.com (SkyGuard) with ESMTP id 4R97SR1G2PzNmZX; Tue, 25 Jul 2023 15:09:27 +0800 (CST) Received: from [10.67.103.232] (10.67.103.232) by dggpemm500005.china.huawei.com (7.185.36.74) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2507.27; Tue, 25 Jul 2023 15:12:49 +0800 Message-ID: <5c7c1ff3-1e2a-1258-7fa0-c82a9ab62646@huawei.com> Date: Tue, 25 Jul 2023 15:12:49 +0800 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:102.0) Gecko/20100101 Thunderbird/102.11.0 Subject: Re: [PATCH V13 - RESEND 06/10] arm64/perf: Enable branch stack events via FEAT_BRBE To: Anshuman Khandual , , , , , CC: Mark Brown , James Clark , Rob Herring , Marc Zyngier , Suzuki Poulose , Peter Zijlstra , Ingo Molnar , Arnaldo Carvalho de Melo , References: <20230711082455.215983-1-anshuman.khandual@arm.com> <20230711082455.215983-7-anshuman.khandual@arm.com> From: Yang Shen In-Reply-To: <20230711082455.215983-7-anshuman.khandual@arm.com> Content-Type: text/plain; charset="UTF-8"; format=flowed Content-Transfer-Encoding: 8bit X-Originating-IP: [10.67.103.232] X-ClientProxiedBy: dggems705-chm.china.huawei.com (10.3.19.182) To dggpemm500005.china.huawei.com (7.185.36.74) X-CFilter-Loop: Reflected X-Spam-Status: No, score=-4.3 required=5.0 tests=BAYES_00,NICE_REPLY_A, RCVD_IN_DNSWL_MED,RCVD_IN_MSPIKE_H5,RCVD_IN_MSPIKE_WL,SPF_HELO_NONE, SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org 在 2023/7/11 16:24, Anshuman Khandual 写道: > This enables branch stack sampling events in ARMV8 PMU, via an architecture > feature FEAT_BRBE aka branch record buffer extension. This defines required > branch helper functions pmuv8pmu_branch_XXXXX() and the implementation here > is wrapped with a new config option CONFIG_ARM64_BRBE. > > Cc: Catalin Marinas > Cc: Will Deacon > Cc: Mark Rutland > Cc: linux-arm-kernel@lists.infradead.org > Cc: linux-kernel@vger.kernel.org > Tested-by: James Clark > Signed-off-by: Anshuman Khandual > --- > arch/arm64/include/asm/perf_event.h | 9 + > drivers/perf/Kconfig | 11 + > drivers/perf/Makefile | 1 + > drivers/perf/arm_brbe.c | 549 ++++++++++++++++++++++++++++ > drivers/perf/arm_brbe.h | 257 +++++++++++++ > drivers/perf/arm_pmuv3.c | 4 + > 6 files changed, 831 insertions(+) > create mode 100644 drivers/perf/arm_brbe.c > create mode 100644 drivers/perf/arm_brbe.h > > diff --git a/arch/arm64/include/asm/perf_event.h b/arch/arm64/include/asm/perf_event.h > index ebc392ba3559..49a973571415 100644 > --- a/arch/arm64/include/asm/perf_event.h > +++ b/arch/arm64/include/asm/perf_event.h > @@ -31,6 +31,14 @@ struct perf_event; > #ifdef CONFIG_PERF_EVENTS > static inline bool has_branch_stack(struct perf_event *event); > > +#ifdef CONFIG_ARM64_BRBE > +void armv8pmu_branch_read(struct pmu_hw_events *cpuc, struct perf_event *event); > +bool armv8pmu_branch_attr_valid(struct perf_event *event); > +void armv8pmu_branch_enable(struct perf_event *event); > +void armv8pmu_branch_disable(struct perf_event *event); > +void armv8pmu_branch_probe(struct arm_pmu *arm_pmu); > +void armv8pmu_branch_reset(void); > +#else > static inline void armv8pmu_branch_read(struct pmu_hw_events *cpuc, struct perf_event *event) > { > WARN_ON_ONCE(!has_branch_stack(event)); > @@ -56,3 +64,4 @@ static inline void armv8pmu_branch_probe(struct arm_pmu *arm_pmu) { } > static inline void armv8pmu_branch_reset(void) { } > #endif > #endif > +#endif > diff --git a/drivers/perf/Kconfig b/drivers/perf/Kconfig > index f4572a5cca72..7c8448051741 100644 > --- a/drivers/perf/Kconfig > +++ b/drivers/perf/Kconfig > @@ -180,6 +180,17 @@ config ARM_SPE_PMU > Extension, which provides periodic sampling of operations in > the CPU pipeline and reports this via the perf AUX interface. > > +config ARM64_BRBE > + bool "Enable support for Branch Record Buffer Extension (BRBE)" > + depends on PERF_EVENTS && ARM64 && ARM_PMU > + default y > + help > + Enable perf support for Branch Record Buffer Extension (BRBE) which > + records all branches taken in an execution path. This supports some > + branch types and privilege based filtering. It captured additional > + relevant information such as cycle count, misprediction and branch > + type, branch privilege level etc. > + > config ARM_DMC620_PMU > tristate "Enable PMU support for the ARM DMC-620 memory controller" > depends on (ARM64 && ACPI) || COMPILE_TEST > diff --git a/drivers/perf/Makefile b/drivers/perf/Makefile > index 16b3ec4db916..a8b7bc22e3d6 100644 > --- a/drivers/perf/Makefile > +++ b/drivers/perf/Makefile > @@ -18,6 +18,7 @@ obj-$(CONFIG_RISCV_PMU_SBI) += riscv_pmu_sbi.o > obj-$(CONFIG_THUNDERX2_PMU) += thunderx2_pmu.o > obj-$(CONFIG_XGENE_PMU) += xgene_pmu.o > obj-$(CONFIG_ARM_SPE_PMU) += arm_spe_pmu.o > +obj-$(CONFIG_ARM64_BRBE) += arm_brbe.o > obj-$(CONFIG_ARM_DMC620_PMU) += arm_dmc620_pmu.o > obj-$(CONFIG_MARVELL_CN10K_TAD_PMU) += marvell_cn10k_tad_pmu.o > obj-$(CONFIG_MARVELL_CN10K_DDR_PMU) += marvell_cn10k_ddr_pmu.o > diff --git a/drivers/perf/arm_brbe.c b/drivers/perf/arm_brbe.c > new file mode 100644 > index 000000000000..79106300cf2e > --- /dev/null > +++ b/drivers/perf/arm_brbe.c > @@ -0,0 +1,549 @@ > +// SPDX-License-Identifier: GPL-2.0-only > +/* > + * Branch Record Buffer Extension Driver. > + * > + * Copyright (C) 2022 ARM Limited > + * > + * Author: Anshuman Khandual > + */ > +#include "arm_brbe.h" > + > +static bool valid_brbe_nr(int brbe_nr) > +{ > + return brbe_nr == BRBIDR0_EL1_NUMREC_8 || > + brbe_nr == BRBIDR0_EL1_NUMREC_16 || > + brbe_nr == BRBIDR0_EL1_NUMREC_32 || > + brbe_nr == BRBIDR0_EL1_NUMREC_64; > +} > + > +static bool valid_brbe_cc(int brbe_cc) > +{ > + return brbe_cc == BRBIDR0_EL1_CC_20_BIT; > +} > + > +static bool valid_brbe_format(int brbe_format) > +{ > + return brbe_format == BRBIDR0_EL1_FORMAT_0; > +} > + > +static bool valid_brbe_version(int brbe_version) > +{ > + return brbe_version == ID_AA64DFR0_EL1_BRBE_IMP || > + brbe_version == ID_AA64DFR0_EL1_BRBE_BRBE_V1P1; > +} > + > +static void select_brbe_bank(int bank) > +{ > + u64 brbfcr; > + > + WARN_ON(bank > BRBE_BANK_IDX_1); > + brbfcr = read_sysreg_s(SYS_BRBFCR_EL1); > + brbfcr &= ~BRBFCR_EL1_BANK_MASK; > + brbfcr |= SYS_FIELD_PREP(BRBFCR_EL1, BANK, bank); > + write_sysreg_s(brbfcr, SYS_BRBFCR_EL1); > + isb(); > +} > + > +/* > + * Generic perf branch filters supported on BRBE > + * > + * New branch filters need to be evaluated whether they could be supported on > + * BRBE. This ensures that such branch filters would not just be accepted, to > + * fail silently. PERF_SAMPLE_BRANCH_HV is a special case that is selectively > + * supported only on platforms where kernel is in hyp mode. > + */ > +#define BRBE_EXCLUDE_BRANCH_FILTERS (PERF_SAMPLE_BRANCH_ABORT_TX | \ > + PERF_SAMPLE_BRANCH_IN_TX | \ > + PERF_SAMPLE_BRANCH_NO_TX | \ > + PERF_SAMPLE_BRANCH_CALL_STACK) > + > +#define BRBE_ALLOWED_BRANCH_FILTERS (PERF_SAMPLE_BRANCH_USER | \ > + PERF_SAMPLE_BRANCH_KERNEL | \ > + PERF_SAMPLE_BRANCH_HV | \ > + PERF_SAMPLE_BRANCH_ANY | \ > + PERF_SAMPLE_BRANCH_ANY_CALL | \ > + PERF_SAMPLE_BRANCH_ANY_RETURN | \ > + PERF_SAMPLE_BRANCH_IND_CALL | \ > + PERF_SAMPLE_BRANCH_COND | \ > + PERF_SAMPLE_BRANCH_IND_JUMP | \ > + PERF_SAMPLE_BRANCH_CALL | \ > + PERF_SAMPLE_BRANCH_NO_FLAGS | \ > + PERF_SAMPLE_BRANCH_NO_CYCLES | \ > + PERF_SAMPLE_BRANCH_TYPE_SAVE | \ > + PERF_SAMPLE_BRANCH_HW_INDEX | \ > + PERF_SAMPLE_BRANCH_PRIV_SAVE) > + > +#define BRBE_PERF_BRANCH_FILTERS (BRBE_ALLOWED_BRANCH_FILTERS | \ > + BRBE_EXCLUDE_BRANCH_FILTERS) > + > +bool armv8pmu_branch_attr_valid(struct perf_event *event) > +{ > + u64 branch_type = event->attr.branch_sample_type; > + > + /* > + * Ensure both perf branch filter allowed and exclude > + * masks are always in sync with the generic perf ABI. > + */ > + BUILD_BUG_ON(BRBE_PERF_BRANCH_FILTERS != (PERF_SAMPLE_BRANCH_MAX - 1)); > + > + if (branch_type & ~BRBE_ALLOWED_BRANCH_FILTERS) { > + pr_debug_once("requested branch filter not supported 0x%llx\n", branch_type); > + return false; > + } > + > + /* > + * If the event does not have at least one of the privilege > + * branch filters as in PERF_SAMPLE_BRANCH_PLM_ALL, the core > + * perf will adjust its value based on perf event's existing > + * privilege level via attr.exclude_[user|kernel|hv]. > + * > + * As event->attr.branch_sample_type might have been changed > + * when the event reaches here, it is not possible to figure > + * out whether the event originally had HV privilege request > + * or got added via the core perf. Just report this situation > + * once and continue ignoring if there are other instances. > + */ > + if ((branch_type & PERF_SAMPLE_BRANCH_HV) && !is_kernel_in_hyp_mode()) > + pr_debug_once("hypervisor privilege filter not supported 0x%llx\n", branch_type); > + > + return true; > +} > + > +static int brbe_attributes_probe(struct arm_pmu *armpmu, u32 brbe) > +{ > + u64 brbidr = read_sysreg_s(SYS_BRBIDR0_EL1); > + int brbe_version, brbe_format, brbe_cc, brbe_nr; > + > + brbe_version = brbe; > + brbe_format = brbe_get_format(brbidr); > + brbe_cc = brbe_get_cc_bits(brbidr); > + brbe_nr = brbe_get_numrec(brbidr); > + armpmu->reg_brbidr = brbidr; > + > + if (!valid_brbe_version(brbe_version) || > + !valid_brbe_format(brbe_format) || > + !valid_brbe_cc(brbe_cc) || > + !valid_brbe_nr(brbe_nr)) > + return -EOPNOTSUPP; > + return 0; > +} > + > +void armv8pmu_branch_probe(struct arm_pmu *armpmu) > +{ > + u64 aa64dfr0 = read_sysreg_s(SYS_ID_AA64DFR0_EL1); > + u32 brbe; > + > + brbe = cpuid_feature_extract_unsigned_field(aa64dfr0, ID_AA64DFR0_EL1_BRBE_SHIFT); > + if (!brbe) > + return; > + > + if (brbe_attributes_probe(armpmu, brbe)) > + return; > + > + armpmu->has_branch_stack = 1; > +} > + > +static u64 branch_type_to_brbfcr(int branch_type) > +{ > + u64 brbfcr = 0; > + > + if (branch_type & PERF_SAMPLE_BRANCH_ANY) { > + brbfcr |= BRBFCR_EL1_BRANCH_FILTERS; > + return brbfcr; > + } > + > + if (branch_type & PERF_SAMPLE_BRANCH_ANY_CALL) { > + brbfcr |= BRBFCR_EL1_INDCALL; > + brbfcr |= BRBFCR_EL1_DIRCALL; > + } > + > + if (branch_type & PERF_SAMPLE_BRANCH_ANY_RETURN) > + brbfcr |= BRBFCR_EL1_RTN; > + > + if (branch_type & PERF_SAMPLE_BRANCH_IND_CALL) > + brbfcr |= BRBFCR_EL1_INDCALL; > + > + if (branch_type & PERF_SAMPLE_BRANCH_COND) > + brbfcr |= BRBFCR_EL1_CONDDIR; > + > + if (branch_type & PERF_SAMPLE_BRANCH_IND_JUMP) > + brbfcr |= BRBFCR_EL1_INDIRECT; > + > + if (branch_type & PERF_SAMPLE_BRANCH_CALL) > + brbfcr |= BRBFCR_EL1_DIRCALL; > + > + return brbfcr; > +} > + > +static u64 branch_type_to_brbcr(int branch_type) > +{ > + u64 brbcr = BRBCR_EL1_DEFAULT_TS; > + > + /* > + * BRBE should be paused on PMU interrupt while tracing kernel > + * space to stop capturing further branch records. Otherwise > + * interrupt handler branch records might get into the samples > + * which is not desired. > + * > + * BRBE need not be paused on PMU interrupt while tracing only > + * the user space, because it will automatically be inside the > + * prohibited region. But even after PMU overflow occurs, the > + * interrupt could still take much more cycles, before it can > + * be taken and by that time BRBE will have been overwritten. > + * Hence enable pause on PMU interrupt mechanism even for user > + * only traces as well. > + */ > + brbcr |= BRBCR_EL1_FZP; > + > + /* > + * When running in the hyp mode, writing into BRBCR_EL1 > + * actually writes into BRBCR_EL2 instead. Field E2BRE > + * is also at the same position as E1BRE. > + */ > + if (branch_type & PERF_SAMPLE_BRANCH_USER) > + brbcr |= BRBCR_EL1_E0BRE; > + > + if (branch_type & PERF_SAMPLE_BRANCH_KERNEL) > + brbcr |= BRBCR_EL1_E1BRE; > + > + if (branch_type & PERF_SAMPLE_BRANCH_HV) { > + if (is_kernel_in_hyp_mode()) > + brbcr |= BRBCR_EL1_E1BRE; > + } > + > + if (!(branch_type & PERF_SAMPLE_BRANCH_NO_CYCLES)) > + brbcr |= BRBCR_EL1_CC; Hi Anshuman, Here is problem about enable CYCLES_COUNT. The SPEC defines that the CYCLES_COUNT is only valid when the BRECR_EL1.CC & BRBCR_EL2.CC is true. And here the SPEC also defines that when PSTATE.EL == EL2 and HCR_EL2.E2h == '1', 'MSR BRBCR_EL1, ' means writing to BRBCR_EL2 actually. So 'armv8pmu_branch_enable' can only set the BRBCR_EL2.CC, while the BRECR_EL1.CC is still 0. The CYCLES_COUNT will be always 0 in records. As a solution, maybe BRBCR_EL12 should be added for driver according to the registers definition. Or, do you have a more standard solution? Thanks, Yang > + > + if (!(branch_type & PERF_SAMPLE_BRANCH_NO_FLAGS)) > + brbcr |= BRBCR_EL1_MPRED; > + > + /* > + * The exception and exception return branches could be > + * captured, irrespective of the perf event's privilege. > + * If the perf event does not have enough privilege for > + * a given exception level, then addresses which falls > + * under that exception level will be reported as zero > + * for the captured branch record, creating source only > + * or target only records. > + */ > + if (branch_type & PERF_SAMPLE_BRANCH_ANY) { > + brbcr |= BRBCR_EL1_EXCEPTION; > + brbcr |= BRBCR_EL1_ERTN; > + } > + > + if (branch_type & PERF_SAMPLE_BRANCH_ANY_CALL) > + brbcr |= BRBCR_EL1_EXCEPTION; > + > + if (branch_type & PERF_SAMPLE_BRANCH_ANY_RETURN) > + brbcr |= BRBCR_EL1_ERTN; > + > + return brbcr & BRBCR_EL1_DEFAULT_CONFIG; > +} > + > +void armv8pmu_branch_enable(struct perf_event *event) > +{ > + u64 branch_type = event->attr.branch_sample_type; > + u64 brbfcr, brbcr; > + > + brbfcr = read_sysreg_s(SYS_BRBFCR_EL1); > + brbfcr &= ~BRBFCR_EL1_DEFAULT_CONFIG; > + brbfcr |= branch_type_to_brbfcr(branch_type); > + write_sysreg_s(brbfcr, SYS_BRBFCR_EL1); > + isb(); > + > + brbcr = read_sysreg_s(SYS_BRBCR_EL1); > + brbcr &= ~BRBCR_EL1_DEFAULT_CONFIG; > + brbcr |= branch_type_to_brbcr(branch_type); > + write_sysreg_s(brbcr, SYS_BRBCR_EL1); > + isb(); > + armv8pmu_branch_reset(); > +} > + > +void armv8pmu_branch_disable(struct perf_event *event) > +{ > + u64 brbfcr = read_sysreg_s(SYS_BRBFCR_EL1); > + u64 brbcr = read_sysreg_s(SYS_BRBCR_EL1); > + > + brbcr &= ~(BRBCR_EL1_E0BRE | BRBCR_EL1_E1BRE); > + brbfcr |= BRBFCR_EL1_PAUSED; > + write_sysreg_s(brbcr, SYS_BRBCR_EL1); > + write_sysreg_s(brbfcr, SYS_BRBFCR_EL1); > + isb(); > +} > + > +static void brbe_set_perf_entry_type(struct perf_branch_entry *entry, u64 brbinf) > +{ > + int brbe_type = brbe_get_type(brbinf); > + > + switch (brbe_type) { > + case BRBINFx_EL1_TYPE_UNCOND_DIRECT: > + entry->type = PERF_BR_UNCOND; > + break; > + case BRBINFx_EL1_TYPE_INDIRECT: > + entry->type = PERF_BR_IND; > + break; > + case BRBINFx_EL1_TYPE_DIRECT_LINK: > + entry->type = PERF_BR_CALL; > + break; > + case BRBINFx_EL1_TYPE_INDIRECT_LINK: > + entry->type = PERF_BR_IND_CALL; > + break; > + case BRBINFx_EL1_TYPE_RET: > + entry->type = PERF_BR_RET; > + break; > + case BRBINFx_EL1_TYPE_COND_DIRECT: > + entry->type = PERF_BR_COND; > + break; > + case BRBINFx_EL1_TYPE_CALL: > + entry->type = PERF_BR_CALL; > + break; > + case BRBINFx_EL1_TYPE_TRAP: > + entry->type = PERF_BR_SYSCALL; > + break; > + case BRBINFx_EL1_TYPE_ERET: > + entry->type = PERF_BR_ERET; > + break; > + case BRBINFx_EL1_TYPE_IRQ: > + entry->type = PERF_BR_IRQ; > + break; > + case BRBINFx_EL1_TYPE_DEBUG_HALT: > + entry->type = PERF_BR_EXTEND_ABI; > + entry->new_type = PERF_BR_ARM64_DEBUG_HALT; > + break; > + case BRBINFx_EL1_TYPE_SERROR: > + entry->type = PERF_BR_SERROR; > + break; > + case BRBINFx_EL1_TYPE_INSN_DEBUG: > + entry->type = PERF_BR_EXTEND_ABI; > + entry->new_type = PERF_BR_ARM64_DEBUG_INST; > + break; > + case BRBINFx_EL1_TYPE_DATA_DEBUG: > + entry->type = PERF_BR_EXTEND_ABI; > + entry->new_type = PERF_BR_ARM64_DEBUG_DATA; > + break; > + case BRBINFx_EL1_TYPE_ALIGN_FAULT: > + entry->type = PERF_BR_EXTEND_ABI; > + entry->new_type = PERF_BR_NEW_FAULT_ALGN; > + break; > + case BRBINFx_EL1_TYPE_INSN_FAULT: > + entry->type = PERF_BR_EXTEND_ABI; > + entry->new_type = PERF_BR_NEW_FAULT_INST; > + break; > + case BRBINFx_EL1_TYPE_DATA_FAULT: > + entry->type = PERF_BR_EXTEND_ABI; > + entry->new_type = PERF_BR_NEW_FAULT_DATA; > + break; > + case BRBINFx_EL1_TYPE_FIQ: > + entry->type = PERF_BR_EXTEND_ABI; > + entry->new_type = PERF_BR_ARM64_FIQ; > + break; > + case BRBINFx_EL1_TYPE_DEBUG_EXIT: > + entry->type = PERF_BR_EXTEND_ABI; > + entry->new_type = PERF_BR_ARM64_DEBUG_EXIT; > + break; > + default: > + pr_warn_once("%d - unknown branch type captured\n", brbe_type); > + entry->type = PERF_BR_UNKNOWN; > + break; > + } > +} > + > +static int brbe_get_perf_priv(u64 brbinf) > +{ > + int brbe_el = brbe_get_el(brbinf); > + > + switch (brbe_el) { > + case BRBINFx_EL1_EL_EL0: > + return PERF_BR_PRIV_USER; > + case BRBINFx_EL1_EL_EL1: > + return PERF_BR_PRIV_KERNEL; > + case BRBINFx_EL1_EL_EL2: > + if (is_kernel_in_hyp_mode()) > + return PERF_BR_PRIV_KERNEL; > + return PERF_BR_PRIV_HV; > + default: > + pr_warn_once("%d - unknown branch privilege captured\n", brbe_el); > + return PERF_BR_PRIV_UNKNOWN; > + } > +} > + > +static void capture_brbe_flags(struct perf_branch_entry *entry, struct perf_event *event, > + u64 brbinf) > +{ > + if (branch_sample_type(event)) > + brbe_set_perf_entry_type(entry, brbinf); > + > + if (!branch_sample_no_cycles(event)) > + entry->cycles = brbe_get_cycles(brbinf); > + > + if (!branch_sample_no_flags(event)) { > + /* > + * BRBINFx_EL1.LASTFAILED indicates that a TME transaction failed (or > + * was cancelled) prior to this record, and some number of records > + * prior to this one, may have been generated during an attempt to > + * execute the transaction. > + * > + * We will remove such entries later in process_branch_aborts(). > + */ > + entry->abort = brbe_get_lastfailed(brbinf); > + > + /* > + * All these information (i.e transaction state and mispredicts) > + * are available for source only and complete branch records. > + */ > + if (brbe_record_is_complete(brbinf) || > + brbe_record_is_source_only(brbinf)) { > + entry->mispred = brbe_get_mispredict(brbinf); > + entry->predicted = !entry->mispred; > + entry->in_tx = brbe_get_in_tx(brbinf); > + } > + } > + > + if (branch_sample_priv(event)) { > + /* > + * All these information (i.e branch privilege level) are > + * available for target only and complete branch records. > + */ > + if (brbe_record_is_complete(brbinf) || > + brbe_record_is_target_only(brbinf)) > + entry->priv = brbe_get_perf_priv(brbinf); > + } > +} > + > +/* > + * A branch record with BRBINFx_EL1.LASTFAILED set, implies that all > + * preceding consecutive branch records, that were in a transaction > + * (i.e their BRBINFx_EL1.TX set) have been aborted. > + * > + * Similarly BRBFCR_EL1.LASTFAILED set, indicate that all preceding > + * consecutive branch records up to the last record, which were in a > + * transaction (i.e their BRBINFx_EL1.TX set) have been aborted. > + * > + * --------------------------------- ------------------- > + * | 00 | BRBSRC | BRBTGT | BRBINF | | TX = 1 | LF = 0 | [TX success] > + * --------------------------------- ------------------- > + * | 01 | BRBSRC | BRBTGT | BRBINF | | TX = 1 | LF = 0 | [TX success] > + * --------------------------------- ------------------- > + * | 02 | BRBSRC | BRBTGT | BRBINF | | TX = 0 | LF = 0 | > + * --------------------------------- ------------------- > + * | 03 | BRBSRC | BRBTGT | BRBINF | | TX = 1 | LF = 0 | [TX failed] > + * --------------------------------- ------------------- > + * | 04 | BRBSRC | BRBTGT | BRBINF | | TX = 1 | LF = 0 | [TX failed] > + * --------------------------------- ------------------- > + * | 05 | BRBSRC | BRBTGT | BRBINF | | TX = 0 | LF = 1 | > + * --------------------------------- ------------------- > + * | .. | BRBSRC | BRBTGT | BRBINF | | TX = 0 | LF = 0 | > + * --------------------------------- ------------------- > + * | 61 | BRBSRC | BRBTGT | BRBINF | | TX = 1 | LF = 0 | [TX failed] > + * --------------------------------- ------------------- > + * | 62 | BRBSRC | BRBTGT | BRBINF | | TX = 1 | LF = 0 | [TX failed] > + * --------------------------------- ------------------- > + * | 63 | BRBSRC | BRBTGT | BRBINF | | TX = 1 | LF = 0 | [TX failed] > + * --------------------------------- ------------------- > + * > + * BRBFCR_EL1.LASTFAILED == 1 > + * > + * BRBFCR_EL1.LASTFAILED fails all those consecutive, in transaction > + * branches records near the end of the BRBE buffer. > + * > + * Architecture does not guarantee a non transaction (TX = 0) branch > + * record between two different transactions. So it is possible that > + * a subsequent lastfailed record (TX = 0, LF = 1) might erroneously > + * mark more than required transactions as aborted. > + */ > +static void process_branch_aborts(struct pmu_hw_events *cpuc) > +{ > + u64 brbfcr = read_sysreg_s(SYS_BRBFCR_EL1); > + bool lastfailed = !!(brbfcr & BRBFCR_EL1_LASTFAILED); > + int idx = brbe_get_numrec(cpuc->percpu_pmu->reg_brbidr) - 1; > + struct perf_branch_entry *entry; > + > + do { > + entry = &cpuc->branches->branch_entries[idx]; > + if (entry->in_tx) { > + entry->abort = lastfailed; > + } else { > + lastfailed = entry->abort; > + entry->abort = false; > + } > + } while (idx--, idx >= 0); > +} > + > +void armv8pmu_branch_reset(void) > +{ > + asm volatile(BRB_IALL); > + isb(); > +} > + > +static bool capture_branch_entry(struct pmu_hw_events *cpuc, > + struct perf_event *event, int idx) > +{ > + struct perf_branch_entry *entry = &cpuc->branches->branch_entries[idx]; > + u64 brbinf = get_brbinf_reg(idx); > + > + /* > + * There are no valid entries anymore on the buffer. > + * Abort the branch record processing to save some > + * cycles and also reduce the capture/process load > + * for the user space as well. > + */ > + if (brbe_invalid(brbinf)) > + return false; > + > + perf_clear_branch_entry_bitfields(entry); > + if (brbe_record_is_complete(brbinf)) { > + entry->from = get_brbsrc_reg(idx); > + entry->to = get_brbtgt_reg(idx); > + } else if (brbe_record_is_source_only(brbinf)) { > + entry->from = get_brbsrc_reg(idx); > + entry->to = 0; > + } else if (brbe_record_is_target_only(brbinf)) { > + entry->from = 0; > + entry->to = get_brbtgt_reg(idx); > + } > + capture_brbe_flags(entry, event, brbinf); > + return true; > +} > + > +void armv8pmu_branch_read(struct pmu_hw_events *cpuc, struct perf_event *event) > +{ > + int nr_hw_entries = brbe_get_numrec(cpuc->percpu_pmu->reg_brbidr); > + u64 brbfcr, brbcr; > + int idx = 0; > + > + brbcr = read_sysreg_s(SYS_BRBCR_EL1); > + brbfcr = read_sysreg_s(SYS_BRBFCR_EL1); > + > + /* Ensure pause on PMU interrupt is enabled */ > + WARN_ON_ONCE(!(brbcr & BRBCR_EL1_FZP)); > + > + /* Pause the buffer */ > + write_sysreg_s(brbfcr | BRBFCR_EL1_PAUSED, SYS_BRBFCR_EL1); > + isb(); > + > + /* Loop through bank 0 */ > + select_brbe_bank(BRBE_BANK_IDX_0); > + while (idx < nr_hw_entries && idx <= BRBE_BANK0_IDX_MAX) { > + if (!capture_branch_entry(cpuc, event, idx)) > + goto skip_bank_1; > + idx++; > + } > + > + /* Loop through bank 1 */ > + select_brbe_bank(BRBE_BANK_IDX_1); > + while (idx < nr_hw_entries && idx <= BRBE_BANK1_IDX_MAX) { > + if (!capture_branch_entry(cpuc, event, idx)) > + break; > + idx++; > + } > + > +skip_bank_1: > + cpuc->branches->branch_stack.nr = idx; > + cpuc->branches->branch_stack.hw_idx = -1ULL; > + process_branch_aborts(cpuc); > + > + /* Unpause the buffer */ > + write_sysreg_s(brbfcr & ~BRBFCR_EL1_PAUSED, SYS_BRBFCR_EL1); > + isb(); > + armv8pmu_branch_reset(); > +} > diff --git a/drivers/perf/arm_brbe.h b/drivers/perf/arm_brbe.h > new file mode 100644 > index 000000000000..a47480eec070 > --- /dev/null > +++ b/drivers/perf/arm_brbe.h > @@ -0,0 +1,257 @@ > +/* SPDX-License-Identifier: GPL-2.0-only */ > +/* > + * Branch Record Buffer Extension Helpers. > + * > + * Copyright (C) 2022 ARM Limited > + * > + * Author: Anshuman Khandual > + */ > +#define pr_fmt(fmt) "brbe: " fmt > + > +#include > + > +#define BRBFCR_EL1_BRANCH_FILTERS (BRBFCR_EL1_DIRECT | \ > + BRBFCR_EL1_INDIRECT | \ > + BRBFCR_EL1_RTN | \ > + BRBFCR_EL1_INDCALL | \ > + BRBFCR_EL1_DIRCALL | \ > + BRBFCR_EL1_CONDDIR) > + > +#define BRBFCR_EL1_DEFAULT_CONFIG (BRBFCR_EL1_BANK_MASK | \ > + BRBFCR_EL1_PAUSED | \ > + BRBFCR_EL1_EnI | \ > + BRBFCR_EL1_BRANCH_FILTERS) > + > +/* > + * BRBTS_EL1 is currently not used for branch stack implementation > + * purpose but BRBCR_EL1.TS needs to have a valid value from all > + * available options. BRBCR_EL1_TS_VIRTUAL is selected for this. > + */ > +#define BRBCR_EL1_DEFAULT_TS FIELD_PREP(BRBCR_EL1_TS_MASK, BRBCR_EL1_TS_VIRTUAL) > + > +#define BRBCR_EL1_DEFAULT_CONFIG (BRBCR_EL1_EXCEPTION | \ > + BRBCR_EL1_ERTN | \ > + BRBCR_EL1_CC | \ > + BRBCR_EL1_MPRED | \ > + BRBCR_EL1_E1BRE | \ > + BRBCR_EL1_E0BRE | \ > + BRBCR_EL1_FZP | \ > + BRBCR_EL1_DEFAULT_TS) > +/* > + * BRBE Instructions > + * > + * BRB_IALL : Invalidate the entire buffer > + * BRB_INJ : Inject latest branch record derived from [BRBSRCINJ, BRBTGTINJ, BRBINFINJ] > + */ > +#define BRB_IALL __emit_inst(0xD5000000 | sys_insn(1, 1, 7, 2, 4) | (0x1f)) > +#define BRB_INJ __emit_inst(0xD5000000 | sys_insn(1, 1, 7, 2, 5) | (0x1f)) > + > +/* > + * BRBE Buffer Organization > + * > + * BRBE buffer is arranged as multiple banks of 32 branch record > + * entries each. An individual branch record in a given bank could > + * be accessed, after selecting the bank in BRBFCR_EL1.BANK and > + * accessing the registers i.e [BRBSRC, BRBTGT, BRBINF] set with > + * indices [0..31]. > + * > + * Bank 0 > + * > + * --------------------------------- ------ > + * | 00 | BRBSRC | BRBTGT | BRBINF | | 00 | > + * --------------------------------- ------ > + * | 01 | BRBSRC | BRBTGT | BRBINF | | 01 | > + * --------------------------------- ------ > + * | .. | BRBSRC | BRBTGT | BRBINF | | .. | > + * --------------------------------- ------ > + * | 31 | BRBSRC | BRBTGT | BRBINF | | 31 | > + * --------------------------------- ------ > + * > + * Bank 1 > + * > + * --------------------------------- ------ > + * | 32 | BRBSRC | BRBTGT | BRBINF | | 00 | > + * --------------------------------- ------ > + * | 33 | BRBSRC | BRBTGT | BRBINF | | 01 | > + * --------------------------------- ------ > + * | .. | BRBSRC | BRBTGT | BRBINF | | .. | > + * --------------------------------- ------ > + * | 63 | BRBSRC | BRBTGT | BRBINF | | 31 | > + * --------------------------------- ------ > + */ > +#define BRBE_BANK_MAX_ENTRIES 32 > + > +#define BRBE_BANK0_IDX_MIN 0 > +#define BRBE_BANK0_IDX_MAX 31 > +#define BRBE_BANK1_IDX_MIN 32 > +#define BRBE_BANK1_IDX_MAX 63 > + > +struct brbe_hw_attr { > + int brbe_version; > + int brbe_cc; > + int brbe_nr; > + int brbe_format; > +}; > + > +enum brbe_bank_idx { > + BRBE_BANK_IDX_INVALID = -1, > + BRBE_BANK_IDX_0, > + BRBE_BANK_IDX_1, > + BRBE_BANK_IDX_MAX > +}; > + > +#define RETURN_READ_BRBSRCN(n) \ > + read_sysreg_s(SYS_BRBSRC##n##_EL1) > + > +#define RETURN_READ_BRBTGTN(n) \ > + read_sysreg_s(SYS_BRBTGT##n##_EL1) > + > +#define RETURN_READ_BRBINFN(n) \ > + read_sysreg_s(SYS_BRBINF##n##_EL1) > + > +#define BRBE_REGN_CASE(n, case_macro) \ > + case n: return case_macro(n); break > + > +#define BRBE_REGN_SWITCH(x, case_macro) \ > + do { \ > + switch (x) { \ > + BRBE_REGN_CASE(0, case_macro); \ > + BRBE_REGN_CASE(1, case_macro); \ > + BRBE_REGN_CASE(2, case_macro); \ > + BRBE_REGN_CASE(3, case_macro); \ > + BRBE_REGN_CASE(4, case_macro); \ > + BRBE_REGN_CASE(5, case_macro); \ > + BRBE_REGN_CASE(6, case_macro); \ > + BRBE_REGN_CASE(7, case_macro); \ > + BRBE_REGN_CASE(8, case_macro); \ > + BRBE_REGN_CASE(9, case_macro); \ > + BRBE_REGN_CASE(10, case_macro); \ > + BRBE_REGN_CASE(11, case_macro); \ > + BRBE_REGN_CASE(12, case_macro); \ > + BRBE_REGN_CASE(13, case_macro); \ > + BRBE_REGN_CASE(14, case_macro); \ > + BRBE_REGN_CASE(15, case_macro); \ > + BRBE_REGN_CASE(16, case_macro); \ > + BRBE_REGN_CASE(17, case_macro); \ > + BRBE_REGN_CASE(18, case_macro); \ > + BRBE_REGN_CASE(19, case_macro); \ > + BRBE_REGN_CASE(20, case_macro); \ > + BRBE_REGN_CASE(21, case_macro); \ > + BRBE_REGN_CASE(22, case_macro); \ > + BRBE_REGN_CASE(23, case_macro); \ > + BRBE_REGN_CASE(24, case_macro); \ > + BRBE_REGN_CASE(25, case_macro); \ > + BRBE_REGN_CASE(26, case_macro); \ > + BRBE_REGN_CASE(27, case_macro); \ > + BRBE_REGN_CASE(28, case_macro); \ > + BRBE_REGN_CASE(29, case_macro); \ > + BRBE_REGN_CASE(30, case_macro); \ > + BRBE_REGN_CASE(31, case_macro); \ > + default: \ > + pr_warn("unknown register index\n"); \ > + return -1; \ > + } \ > + } while (0) > + > +static inline int buffer_to_brbe_idx(int buffer_idx) > +{ > + return buffer_idx % BRBE_BANK_MAX_ENTRIES; > +} > + > +static inline u64 get_brbsrc_reg(int buffer_idx) > +{ > + int brbe_idx = buffer_to_brbe_idx(buffer_idx); > + > + BRBE_REGN_SWITCH(brbe_idx, RETURN_READ_BRBSRCN); > +} > + > +static inline u64 get_brbtgt_reg(int buffer_idx) > +{ > + int brbe_idx = buffer_to_brbe_idx(buffer_idx); > + > + BRBE_REGN_SWITCH(brbe_idx, RETURN_READ_BRBTGTN); > +} > + > +static inline u64 get_brbinf_reg(int buffer_idx) > +{ > + int brbe_idx = buffer_to_brbe_idx(buffer_idx); > + > + BRBE_REGN_SWITCH(brbe_idx, RETURN_READ_BRBINFN); > +} > + > +static inline u64 brbe_record_valid(u64 brbinf) > +{ > + return FIELD_GET(BRBINFx_EL1_VALID_MASK, brbinf); > +} > + > +static inline bool brbe_invalid(u64 brbinf) > +{ > + return brbe_record_valid(brbinf) == BRBINFx_EL1_VALID_NONE; > +} > + > +static inline bool brbe_record_is_complete(u64 brbinf) > +{ > + return brbe_record_valid(brbinf) == BRBINFx_EL1_VALID_FULL; > +} > + > +static inline bool brbe_record_is_source_only(u64 brbinf) > +{ > + return brbe_record_valid(brbinf) == BRBINFx_EL1_VALID_SOURCE; > +} > + > +static inline bool brbe_record_is_target_only(u64 brbinf) > +{ > + return brbe_record_valid(brbinf) == BRBINFx_EL1_VALID_TARGET; > +} > + > +static inline int brbe_get_in_tx(u64 brbinf) > +{ > + return FIELD_GET(BRBINFx_EL1_T_MASK, brbinf); > +} > + > +static inline int brbe_get_mispredict(u64 brbinf) > +{ > + return FIELD_GET(BRBINFx_EL1_MPRED_MASK, brbinf); > +} > + > +static inline int brbe_get_lastfailed(u64 brbinf) > +{ > + return FIELD_GET(BRBINFx_EL1_LASTFAILED_MASK, brbinf); > +} > + > +static inline int brbe_get_cycles(u64 brbinf) > +{ > + /* > + * Captured cycle count is unknown and hence > + * should not be passed on to the user space. > + */ > + if (brbinf & BRBINFx_EL1_CCU) > + return 0; > + > + return FIELD_GET(BRBINFx_EL1_CC_MASK, brbinf); > +} > + > +static inline int brbe_get_type(u64 brbinf) > +{ > + return FIELD_GET(BRBINFx_EL1_TYPE_MASK, brbinf); > +} > + > +static inline int brbe_get_el(u64 brbinf) > +{ > + return FIELD_GET(BRBINFx_EL1_EL_MASK, brbinf); > +} > + > +static inline int brbe_get_numrec(u64 brbidr) > +{ > + return FIELD_GET(BRBIDR0_EL1_NUMREC_MASK, brbidr); > +} > + > +static inline int brbe_get_format(u64 brbidr) > +{ > + return FIELD_GET(BRBIDR0_EL1_FORMAT_MASK, brbidr); > +} > + > +static inline int brbe_get_cc_bits(u64 brbidr) > +{ > + return FIELD_GET(BRBIDR0_EL1_CC_MASK, brbidr); > +} > diff --git a/drivers/perf/arm_pmuv3.c b/drivers/perf/arm_pmuv3.c > index 10bbe02f4079..7c9e9045c24e 100644 > --- a/drivers/perf/arm_pmuv3.c > +++ b/drivers/perf/arm_pmuv3.c > @@ -813,6 +813,10 @@ static irqreturn_t armv8pmu_handle_irq(struct arm_pmu *cpu_pmu) > if (!armpmu_event_set_period(event)) > continue; > > + /* > + * PMU IRQ should remain asserted until all branch records > + * are captured and processed into struct perf_sample_data. > + */ > if (has_branch_stack(event) && !WARN_ON(!cpuc->branches)) { > armv8pmu_branch_read(cpuc, event); > perf_sample_save_brstack(&data, event, &cpuc->branches->branch_stack);