Received: by 2002:ac0:aed5:0:0:0:0:0 with SMTP id t21csp253381imb; Thu, 28 Feb 2019 23:03:13 -0800 (PST) X-Google-Smtp-Source: APXvYqy2/1y5KhFq6s7QTXA6ndRHtijyUu8V56APAdq4FEb0NxkDbi3TpfJclggfMTVH3N0vVwkD X-Received: by 2002:a63:5d48:: with SMTP id o8mr3400045pgm.297.1551423793494; Thu, 28 Feb 2019 23:03:13 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1551423793; cv=none; d=google.com; s=arc-20160816; b=f59qKMDALbLMR2HeCUHIshb0AugQVF5FudedYr7S7QRnqJYEQEIo5daadisjxC4K/2 5kwAY5yLFvjkpFM31X/iecI+CM6BOQW847lzNrbL2EWEHOIBGTdoVzS7ITh5lv6kTpMC 6D0scUMhIaMx5P+xobz/FKJhuoHQxvFXAmoGUOem8gBLgDqJfCnvMx5BfixpW2Or8+NT aOefD9gxWkFBbr/OJ1P7AG4+ARQVmkBFnOqHRpamLaOYI9TP8NbUcz4wl8f3aaTmSbwM Lf/nLmyfLKUqBUTk66tSNMNBiUjE0VAY/nDVK4f5iaaWXHIqDAu3xkuju9QI1l9W4SNR 6OUg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=i7hoFXYdh2+wPoNqjaflQCBwWI9vGemK2bSJWz39zJo=; b=tszYpETCSgojOz11rruAJc3nADDd5wXdBh7a0BZ9rbQ4HzzBM4s658XMGhyvPqDkRu uj6XLjV6qKrvFrXSbf9EoAmfD0YThl05TtGERu0uYtLTU3ObnuVcRimAV+EVGUjBptQt ETTqGZI6Kn7r+tmHy6rJuZBa2o1SNQDKkvcGLO2T3kuxBPLnxQ2sLm0rZrtDPKwW6nCr Z9juWcQdtboQe/IXK6i8dypeigD74Ta+TVxgaa4bUM9FK2eydHqvj/O/esulGFKDz5wZ 3wuxk9y6ki61jowN4eJ+miKoOE9B0uuQhuHfdAZRHYCKZMcf4FJAvkMMRY0A9BNEGoVg /UZg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id d1si19302195pll.283.2019.02.28.23.02.57; Thu, 28 Feb 2019 23:03:13 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728047AbfCAHCM (ORCPT + 99 others); Fri, 1 Mar 2019 02:02:12 -0500 Received: from usa-sjc-mx-foss1.foss.arm.com ([217.140.101.70]:58752 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725982AbfCAHCL (ORCPT ); Fri, 1 Mar 2019 02:02:11 -0500 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.72.51.249]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 2772AEBD; Thu, 28 Feb 2019 23:02:11 -0800 (PST) Received: from [172.20.0.134] (usa-sjc-mx-foss1.foss.arm.com [217.140.101.70]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 4CE8A3F5C1; Thu, 28 Feb 2019 23:02:10 -0800 (PST) Subject: Re: [PATCH v5 09/10] arm64: add sysfs vulnerability show for speculative store bypass To: Jeremy Linton , linux-arm-kernel@lists.infradead.org Cc: catalin.marinas@arm.com, will.deacon@arm.com, marc.zyngier@arm.com, suzuki.poulose@arm.com, Dave.Martin@arm.com, shankerd@codeaurora.org, julien.thierry@arm.com, mlangsdo@redhat.com, stefan.wahren@i2e.com, linux-kernel@vger.kernel.org References: <20190227010544.597579-1-jeremy.linton@arm.com> <20190227010544.597579-10-jeremy.linton@arm.com> From: Andre Przywara Message-ID: <04eb2b5d-3754-0e39-3c6e-54b45741da83@foss.arm.com> Date: Fri, 1 Mar 2019 01:02:07 -0600 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.5.1 MIME-Version: 1.0 In-Reply-To: <20190227010544.597579-10-jeremy.linton@arm.com> Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi, On 2/26/19 7:05 PM, Jeremy Linton wrote: > Return status based on ssbd_state and the arm64 SSBS feature. If > the mitigation is disabled, or the firmware isn't responding then > return the expected machine state based on a new blacklist of known > vulnerable cores. > > Signed-off-by: Jeremy Linton > --- > arch/arm64/kernel/cpu_errata.c | 43 ++++++++++++++++++++++++++++++++++ > 1 file changed, 43 insertions(+) > > diff --git a/arch/arm64/kernel/cpu_errata.c b/arch/arm64/kernel/cpu_errata.c > index 5f5611d17dc1..e1b03f643799 100644 > --- a/arch/arm64/kernel/cpu_errata.c > +++ b/arch/arm64/kernel/cpu_errata.c > @@ -279,6 +279,7 @@ static int detect_harden_bp_fw(void) > DEFINE_PER_CPU_READ_MOSTLY(u64, arm64_ssbd_callback_required); > > int ssbd_state __read_mostly = ARM64_SSBD_KERNEL; > +static bool __ssb_safe = true; > > static const struct ssbd_options { > const char *str; > @@ -387,6 +388,9 @@ static bool has_ssbd_mitigation(const struct arm64_cpu_capabilities *entry, > > WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible()); > > + if (is_midr_in_range_list(read_cpuid_id(), entry->midr_range_list)) > + __ssb_safe = false; Is that the only place where we set it to false? What about if firmware reports that (at least one core) is vulnerable? > + > if (this_cpu_has_cap(ARM64_SSBS)) { > required = false; > goto out_printmsg; > @@ -420,6 +424,7 @@ static bool has_ssbd_mitigation(const struct arm64_cpu_capabilities *entry, > ssbd_state = ARM64_SSBD_UNKNOWN; > return false; > > + /* machines with mixed mitigation requirements must not return this */ > case SMCCC_RET_NOT_REQUIRED: > pr_info_once("%s mitigation not required\n", entry->desc); > ssbd_state = ARM64_SSBD_MITIGATED; > @@ -475,6 +480,16 @@ static bool has_ssbd_mitigation(const struct arm64_cpu_capabilities *entry, > return required; > } > > +/* known vulnerable cores */ > +static const struct midr_range arm64_ssb_cpus[] = { > + MIDR_ALL_VERSIONS(MIDR_CORTEX_A57), > + MIDR_ALL_VERSIONS(MIDR_CORTEX_A72), > + MIDR_ALL_VERSIONS(MIDR_CORTEX_A73), > + MIDR_ALL_VERSIONS(MIDR_CORTEX_A75), > + MIDR_ALL_VERSIONS(MIDR_CORTEX_A76), > + {}, > +}; > + > static void __maybe_unused > cpu_enable_cache_maint_trap(const struct arm64_cpu_capabilities *__unused) > { > @@ -770,6 +785,7 @@ const struct arm64_cpu_capabilities arm64_errata[] = { > .capability = ARM64_SSBD, > .type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM, > .matches = has_ssbd_mitigation, > + .midr_range_list = arm64_ssb_cpus, > }, > #ifdef CONFIG_ARM64_ERRATUM_1188873 > { > @@ -808,3 +824,30 @@ ssize_t cpu_show_spectre_v2(struct device *dev, struct device_attribute *attr, > > return sprintf(buf, "Vulnerable\n"); > } > + > +ssize_t cpu_show_spec_store_bypass(struct device *dev, > + struct device_attribute *attr, char *buf) > +{ > + /* > + * Two assumptions: First, ssbd_state reflects the worse case > + * for hetrogenous machines, and that if SSBS is supported its heterogeneous Cheers, Andre. > + * supported by all cores. > + */ > + switch (ssbd_state) { > + case ARM64_SSBD_MITIGATED: > + return sprintf(buf, "Not affected\n"); > + > + case ARM64_SSBD_KERNEL: > + case ARM64_SSBD_FORCE_ENABLE: > + if (cpus_have_cap(ARM64_SSBS)) > + return sprintf(buf, "Not affected\n"); > + if (IS_ENABLED(CONFIG_ARM64_SSBD)) > + return sprintf(buf, > + "Mitigation: Speculative Store Bypass disabled\n"); > + } > + > + if (__ssb_safe) > + return sprintf(buf, "Not affected\n"); > + > + return sprintf(buf, "Vulnerable\n"); > +} >