Received: by 2002:a6b:500f:0:0:0:0:0 with SMTP id e15csp1186382iob; Thu, 19 May 2022 00:52:05 -0700 (PDT) X-Google-Smtp-Source: ABdhPJy+GDjYqtSwIIeik532bp/d2QD3MtmePd6sYw3gcjFrFZlWuLd5FpxCpEjX/tV6MUJDyFq8 X-Received: by 2002:a17:90b:388f:b0:1dc:6e0f:372b with SMTP id mu15-20020a17090b388f00b001dc6e0f372bmr3771658pjb.93.1652946725588; Thu, 19 May 2022 00:52:05 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1652946725; cv=none; d=google.com; s=arc-20160816; b=V6zbs95JRRFGAy4pw5u+JlDwSf3wmPOlBDkWVUJWZ3pCOgTX/V7xdWoWOAHqfncydS 95RzoNP1V/hWNX/+eJMw1oS5XoN+rEppsHDYq18jhMbvKemnP8SDcrnh/6QZfU0G0QW7 JNE6PaZV02EDRCl42TRIifDFFhkLp06MWQxbO3SL+1SveLIco9KWioTcu+7BRdKbZODH z4UoNuIrWPLNDK8xJhlHAeLhVIFaiJiCCrqyhy1bqeYfj0U2yQgyTwEW5s97za6nwHhB eoyC0Py+ew2XlUrONdDFSgbpvKndSvtrIq5SCVfs3V5js1OpPN31Tf3ChnVjpj+ypkZn UOOQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:from :references:cc:to:subject:user-agent:mime-version:date:message-id; bh=rrG1n5WbhJSTrUW9tL6Zm2ACHPmNw7nc6OA8g58P9sI=; b=uMqTV8F/wfYu4IArcjeYumKotE8DrWwfI3PoypuTq7uugAJbjzN8YL7vuZWLC5vYKu 7YndaG83X7BXrnq/Go1cToNM4i8xDY0Wbacpo5fTMofojKvqnla7BU/dOAzpxIsySq0T TFf7pStKrMjIf6QluECOo8bGEkCbLt5gmcnvwPguu/TANdAJuQbz54iBKf66pMtPtyL4 8sPzZrudVf0/welsg9GOqi7MG42Gp8DM6oeNUNfC0ho/mEh7mv3+M0Nxg1S4yJlvmjpp 3AHDrGbMC6Rra2J75mjeZ4n57IV1HKVXAZwjGKwacL8ljkgBKD9Z/OHPE8ohFnzdbH0r 3mXA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=huawei.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id d6-20020a056a0010c600b0050d606c1c3csi5884569pfu.207.2022.05.19.00.51.47; Thu, 19 May 2022 00:52:05 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=huawei.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S233509AbiESGaH (ORCPT + 99 others); Thu, 19 May 2022 02:30:07 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:49552 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234179AbiESGaC (ORCPT ); Thu, 19 May 2022 02:30:02 -0400 Received: from szxga01-in.huawei.com (szxga01-in.huawei.com [45.249.212.187]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 2A766AEE09 for ; Wed, 18 May 2022 23:30:00 -0700 (PDT) Received: from kwepemi100004.china.huawei.com (unknown [172.30.72.56]) by szxga01-in.huawei.com (SkyGuard) with ESMTP id 4L3g1X259DzhZBY; Thu, 19 May 2022 14:29:20 +0800 (CST) Received: from kwepemm600017.china.huawei.com (7.193.23.234) by kwepemi100004.china.huawei.com (7.221.188.70) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2375.24; Thu, 19 May 2022 14:29:56 +0800 Received: from [10.174.179.234] (10.174.179.234) by kwepemm600017.china.huawei.com (7.193.23.234) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2375.24; Thu, 19 May 2022 14:29:55 +0800 Message-ID: <46e5954c-a9a8-f4a8-07cc-de42e2753051@huawei.com> Date: Thu, 19 May 2022 14:29:54 +0800 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:91.0) Gecko/20100101 Thunderbird/91.8.0 Subject: Re: [PATCH -next v4 3/7] arm64: add support for machine check error safe To: Mark Rutland CC: James Morse , Andrew Morton , Thomas Gleixner , "Ingo Molnar" , Borislav Petkov , Robin Murphy , Dave Hansen , "Catalin Marinas" , Will Deacon , "Alexander Viro" , Michael Ellerman , Benjamin Herrenschmidt , Paul Mackerras , , "H . Peter Anvin" , , , , , Kefeng Wang , Xie XiuQi , Guohanjun References: <20220420030418.3189040-1-tongtiangen@huawei.com> <20220420030418.3189040-4-tongtiangen@huawei.com> From: Tong Tiangen In-Reply-To: Content-Type: text/plain; charset="UTF-8"; format=flowed Content-Transfer-Encoding: 8bit X-Originating-IP: [10.174.179.234] X-ClientProxiedBy: dggems701-chm.china.huawei.com (10.3.19.178) To kwepemm600017.china.huawei.com (7.193.23.234) X-CFilter-Loop: Reflected X-Spam-Status: No, score=-6.3 required=5.0 tests=BAYES_00,NICE_REPLY_A, RCVD_IN_DNSWL_MED,SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org 在 2022/5/13 23:26, Mark Rutland 写道: > On Wed, Apr 20, 2022 at 03:04:14AM +0000, Tong Tiangen wrote: >> During the processing of arm64 kernel hardware memory errors(do_sea()), if >> the errors is consumed in the kernel, the current processing is panic. >> However, it is not optimal. >> >> Take uaccess for example, if the uaccess operation fails due to memory >> error, only the user process will be affected, kill the user process >> and isolate the user page with hardware memory errors is a better choice. > > Conceptually, I'm fine with the idea of constraining what we do for a > true uaccess, but I don't like the implementation of this at all, and I > think we first need to clean up the arm64 extable usage to clearly > distinguish a uaccess from another access. OK,using EX_TYPE_UACCESS and this extable type could be recover, this is more reasonable. For EX_TYPE_UACCESS_ERR_ZERO, today we use it for kernel accesses in a couple of cases, such as get_user/futex/__user_cache_maint()/__user_swpX_asm(), your suggestion is: get_user continues to use EX_TYPE_UACCESS_ERR_ZERO and the other cases use new type EX_TYPE_FIXUP_ERR_ZERO? Thanks, Tong. > >> This patch only enable machine error check framework, it add exception >> fixup before kernel panic in do_sea() and only limit the consumption of >> hardware memory errors in kernel mode triggered by user mode processes. >> If fixup successful, panic can be avoided. >> >> Consistent with PPC/x86, it is implemented by CONFIG_ARCH_HAS_COPY_MC. >> >> Also add copy_mc_to_user() in include/linux/uaccess.h, this helper is >> called when CONFIG_ARCH_HAS_COPOY_MC is open. >> >> Signed-off-by: Tong Tiangen >> --- >> arch/arm64/Kconfig | 1 + >> arch/arm64/include/asm/extable.h | 1 + >> arch/arm64/mm/extable.c | 17 +++++++++++++++++ >> arch/arm64/mm/fault.c | 27 ++++++++++++++++++++++++++- >> include/linux/uaccess.h | 9 +++++++++ >> 5 files changed, 54 insertions(+), 1 deletion(-) >> >> diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig >> index d9325dd95eba..012e38309955 100644 >> --- a/arch/arm64/Kconfig >> +++ b/arch/arm64/Kconfig >> @@ -19,6 +19,7 @@ config ARM64 >> select ARCH_ENABLE_SPLIT_PMD_PTLOCK if PGTABLE_LEVELS > 2 >> select ARCH_ENABLE_THP_MIGRATION if TRANSPARENT_HUGEPAGE >> select ARCH_HAS_CACHE_LINE_SIZE >> + select ARCH_HAS_COPY_MC if ACPI_APEI_GHES >> select ARCH_HAS_CURRENT_STACK_POINTER >> select ARCH_HAS_DEBUG_VIRTUAL >> select ARCH_HAS_DEBUG_VM_PGTABLE >> diff --git a/arch/arm64/include/asm/extable.h b/arch/arm64/include/asm/extable.h >> index 72b0e71cc3de..f80ebd0addfd 100644 >> --- a/arch/arm64/include/asm/extable.h >> +++ b/arch/arm64/include/asm/extable.h >> @@ -46,4 +46,5 @@ bool ex_handler_bpf(const struct exception_table_entry *ex, >> #endif /* !CONFIG_BPF_JIT */ >> >> bool fixup_exception(struct pt_regs *regs); >> +bool fixup_exception_mc(struct pt_regs *regs); >> #endif >> diff --git a/arch/arm64/mm/extable.c b/arch/arm64/mm/extable.c >> index 489455309695..4f0083a550d4 100644 >> --- a/arch/arm64/mm/extable.c >> +++ b/arch/arm64/mm/extable.c >> @@ -9,6 +9,7 @@ >> >> #include >> #include >> +#include >> >> static inline unsigned long >> get_ex_fixup(const struct exception_table_entry *ex) >> @@ -84,3 +85,19 @@ bool fixup_exception(struct pt_regs *regs) >> >> BUG(); >> } >> + >> +bool fixup_exception_mc(struct pt_regs *regs) >> +{ >> + const struct exception_table_entry *ex; >> + >> + ex = search_exception_tables(instruction_pointer(regs)); >> + if (!ex) >> + return false; >> + >> + /* >> + * This is not complete, More Machine check safe extable type can >> + * be processed here. >> + */ >> + >> + return false; >> +} > > This is at best misnamed; It doesn't actually apply the fixup, it just > searches for one. Yeah, you're right about the current logic, so i added notes to explain the scenarios that will be added later. > >> diff --git a/arch/arm64/mm/fault.c b/arch/arm64/mm/fault.c >> index 77341b160aca..a9e6fb1999d1 100644 >> --- a/arch/arm64/mm/fault.c >> +++ b/arch/arm64/mm/fault.c >> @@ -695,6 +695,29 @@ static int do_bad(unsigned long far, unsigned int esr, struct pt_regs *regs) >> return 1; /* "fault" */ >> } >> >> +static bool arm64_do_kernel_sea(unsigned long addr, unsigned int esr, >> + struct pt_regs *regs, int sig, int code) >> +{ >> + if (!IS_ENABLED(CONFIG_ARCH_HAS_COPY_MC)) >> + return false; >> + >> + if (user_mode(regs) || !current->mm) >> + return false; >> + >> + if (apei_claim_sea(regs) < 0) >> + return false; >> + >> + if (!fixup_exception_mc(regs)) >> + return false; >> + >> + set_thread_esr(0, esr); >> + >> + arm64_force_sig_fault(sig, code, addr, >> + "Uncorrected hardware memory error in kernel-access\n"); >> + >> + return true; >> +} >> + >> static int do_sea(unsigned long far, unsigned int esr, struct pt_regs *regs) >> { >> const struct fault_info *inf; >> @@ -720,7 +743,9 @@ static int do_sea(unsigned long far, unsigned int esr, struct pt_regs *regs) >> */ >> siaddr = untagged_addr(far); >> } >> - arm64_notify_die(inf->name, regs, inf->sig, inf->code, siaddr, esr); >> + >> + if (!arm64_do_kernel_sea(siaddr, esr, regs, inf->sig, inf->code)) >> + arm64_notify_die(inf->name, regs, inf->sig, inf->code, siaddr, esr); >> >> return 0; >> } >> diff --git a/include/linux/uaccess.h b/include/linux/uaccess.h >> index 546179418ffa..884661b29c17 100644 >> --- a/include/linux/uaccess.h >> +++ b/include/linux/uaccess.h >> @@ -174,6 +174,15 @@ copy_mc_to_kernel(void *dst, const void *src, size_t cnt) >> } >> #endif >> >> +#ifndef copy_mc_to_user >> +static inline unsigned long __must_check >> +copy_mc_to_user(void *dst, const void *src, size_t cnt) >> +{ >> + check_object_size(src, cnt, true); >> + return raw_copy_to_user(dst, src, cnt); >> +} >> +#endif > > Why do we need a special copy_mc_to_user() ? > > Why are we not making *every* true uaccess recoverable? That way the > regular copy_to_user() would just work. Agreed, will fixed next version. Thanks, Tong. > > Thanks, > Mark. > .