Received: by 2002:a05:6358:3188:b0:123:57c1:9b43 with SMTP id q8csp1952357rwd; Fri, 2 Jun 2023 02:35:13 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ7Qtc2WHWMJbCAGeAh8qwoh2VODxN27vMB6DRJZyPKcc7wZt+phSSAL6jZupCY/2CVrxArf X-Received: by 2002:a17:90a:1b0e:b0:256:9b5c:2aa2 with SMTP id q14-20020a17090a1b0e00b002569b5c2aa2mr1843394pjq.6.1685698513497; Fri, 02 Jun 2023 02:35:13 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1685698513; cv=none; d=google.com; s=arc-20160816; b=h4vRLYgpIugmIUkdGwATb9IKJNiVJem19zZt/Metbi7l4yfazLKRfpySSRSm7ukntg wiCnaFAmu5j6/eQsuJyvFooM6twrqSRbmytQJohhmAS5pM2J9NpVtEe9EPbkAJHgLzsb 4XdbzhcXi3Ko6HjNSX2agF46eT11Y4bAZ40ijLVYpgx2yBPrwTdiPIT2seP221sJGT5/ adg8Pkxr8sz8RjuT/lWGu6S2GcEiGmNt4Fk3tKCbPh6jPWFnXAfGJ9UMGZ/zHH5KYBWE p33BbZ60/KPuJtdFMBPCsNIVxtTQ8iH0MHnUwLdl/4fbN8/IGSHO3iapta4t70KJ/7ji TxWQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from; bh=vnQhUZRz4Dbd2zDC3OaU0O6/6crJ/of1VRi4XZSdyoM=; b=V+0BdTAP4xCvcFqCK1QnHM40FlnK1uQqdXr/g2murGRvoq5FeLSThBL48iFcKtZE62 EmaCOhwbqm1NapHy6i1iYTgSZS97nytE7atb/Yb2BP+Kd+ThMFcPYY+ovyChczYQ3scy iXlKFCxF9eTHjxUU7xXfDW7trBIAxVzJ+GVRc4Y3aF+XHBT2vQdch9S4oQgrWBqAmBSd 20mMxI3bR+2a69jmxE9Ji/xp7uwu68IpnFl6Uk3DVtZRCHopGjw4lTU5wwg28KoZSp8V fswIcDuHs9F/7q94jBaiKl1pu7dwHFQfpE+AUE+wAhsDW6CQh0n5pM9V9Uya+EpAdgQs ctrg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id mn13-20020a17090b188d00b00256c6c76723si768522pjb.69.2023.06.02.02.34.58; Fri, 02 Jun 2023 02:35:13 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S234533AbjFBJZk (ORCPT + 99 others); Fri, 2 Jun 2023 05:25:40 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:44624 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235081AbjFBJYV (ORCPT ); Fri, 2 Jun 2023 05:24:21 -0400 Received: from dggsgout12.his.huawei.com (dggsgout12.his.huawei.com [45.249.212.56]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 7136210C8; Fri, 2 Jun 2023 02:22:34 -0700 (PDT) Received: from mail02.huawei.com (unknown [172.30.67.153]) by dggsgout12.his.huawei.com (SkyGuard) with ESMTP id 4QXcwP42qZz4f3v7C; Fri, 2 Jun 2023 17:22:29 +0800 (CST) Received: from huaweicloud.com (unknown [10.175.104.67]) by APP4 (Coremail) with SMTP id gCh0CgD3X7PVtHlkS1cKKw--.42263S5; Fri, 02 Jun 2023 17:22:31 +0800 (CST) From: linan666@huaweicloud.com To: song@kernel.org, neilb@suse.de Cc: linux-raid@vger.kernel.org, linux-kernel@vger.kernel.org, linan122@huawei.com, yukuai3@huawei.com, yi.zhang@huawei.com, houtao1@huawei.com, yangerkun@huawei.com Subject: [PATCH v7 1/2] md/raid10: Do not add spare disk when recovery fails Date: Fri, 2 Jun 2023 17:18:38 +0800 Message-Id: <20230602091839.743798-2-linan666@huaweicloud.com> X-Mailer: git-send-email 2.39.2 In-Reply-To: <20230602091839.743798-1-linan666@huaweicloud.com> References: <20230602091839.743798-1-linan666@huaweicloud.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-CM-TRANSID: gCh0CgD3X7PVtHlkS1cKKw--.42263S5 X-Coremail-Antispam: 1UD129KBjvJXoW7Cw17JF4kCrykXr13Kw48WFg_yoW8Kr15pa nrJF93tr98Xw1fA3Z8AryUXFWFy3y8Ja47tryxW34j93y3tryDZFWxXF4UXFWUXF1FqFWY qw4DJr45CF1kKaDanT9S1TB71UUUUUUqnTZGkaVYY2UrUUUUjbIjqfuFe4nvWSU5nxnvy2 9KBjDU0xBIdaVrnRJUUUBlb4IE77IF4wAFF20E14v26ryj6rWUM7CY07I20VC2zVCF04k2 6cxKx2IYs7xG6rWj6s0DM7CIcVAFz4kK6r1j6r18M28IrcIa0xkI8VA2jI8067AKxVWUGw A2048vs2IY020Ec7CjxVAFwI0_JFI_Gr1l8cAvFVAK0II2c7xJM28CjxkF64kEwVA0rcxS w2x7M28EF7xvwVC0I7IYx2IY67AKxVWDJVCq3wA2z4x0Y4vE2Ix0cI8IcVCY1x0267AKxV W8Jr0_Cr1UM28EF7xvwVC2z280aVAFwI0_GcCE3s1l84ACjcxK6I8E87Iv6xkF7I0E14v2 6rxl6s0DM2vYz4IE04k24VAvwVAKI4IrM2AIxVAIcxkEcVAq07x20xvEncxIr21l5I8CrV ACY4xI64kE6c02F40Ex7xfMcIj6xIIjxv20xvE14v26r1j6r18McIj6I8E87Iv67AKxVWU JVW8JwAm72CE4IkC6x0Yz7v_Jr0_Gr1lF7xvr2IYc2Ij64vIr41lw4CEc2x0rVAKj4xxMx AIw28IcxkI7VAKI48JMxC20s026xCaFVCjc4AY6r1j6r4UMI8I3I0E5I8CrVAFwI0_Jr0_ Jr4lx2IqxVCjr7xvwVAFwI0_JrI_JrWlx4CE17CEb7AF67AKxVWUtVW8ZwCIc40Y0x0EwI xGrwCI42IY6xIIjxv20xvE14v26r1j6r1xMIIF0xvE2Ix0cI8IcVCY1x0267AKxVW8JVWx JwCI42IY6xAIw20EY4v20xvaj40_Jr0_JF4lIxAIcVC2z280aVAFwI0_Jr0_Gr1lIxAIcV C2z280aVCY1x0267AKxVW8JVW8JrUvcSsGvfC2KfnxnUUI43ZEXa7IU858n5UUUUU== X-CM-SenderInfo: polqt0awwwqx5xdzvxpfor3voofrz/ X-CFilter-Loop: Reflected X-Spam-Status: No, score=-1.9 required=5.0 tests=BAYES_00,SPF_HELO_NONE, SPF_NONE,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Li Nan In raid10_sync_request(), if data cannot be read from any disk for recovery, it will go to 'giveup' and let 'chunks_skipped' + 1. After multiple 'giveup', when 'chunks_skipped >= geo.raid_disks', it will return 'max_sector', indicating that the recovery has been completed. However, the recovery is just aborted and the data remains inconsistent. Fix it by setting mirror->recovery_disabled, which will prevent the spare disk from being added to this mirror. The same issue also exists during resync, it will be fixed afterwards. Signed-off-by: Li Nan --- drivers/md/raid10.c | 20 ++++++++++++++++++-- 1 file changed, 18 insertions(+), 2 deletions(-) diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c index d93d8cb2b620..3d52fb618571 100644 --- a/drivers/md/raid10.c +++ b/drivers/md/raid10.c @@ -3303,6 +3303,7 @@ static sector_t raid10_sync_request(struct mddev *mddev, sector_t sector_nr, int chunks_skipped = 0; sector_t chunk_mask = conf->geo.chunk_mask; int page_idx = 0; + int error_disk = -1; /* * Allow skipping a full rebuild for incremental assembly @@ -3386,8 +3387,21 @@ static sector_t raid10_sync_request(struct mddev *mddev, sector_t sector_nr, return reshape_request(mddev, sector_nr, skipped); if (chunks_skipped >= conf->geo.raid_disks) { - /* if there has been nothing to do on any drive, - * then there is nothing to do at all.. + pr_err("md/raid10:%s: %s fails\n", mdname(mddev), + test_bit(MD_RECOVERY_SYNC, &mddev->recovery) ? "resync" : "recovery"); + if (error_disk >= 0 && + !test_bit(MD_RECOVERY_SYNC, &mddev->recovery)) { + /* + * recovery fails, set mirrors.recovery_disabled, + * device shouldn't be added to there. + */ + conf->mirrors[error_disk].recovery_disabled = + mddev->recovery_disabled; + return 0; + } + /* + * if there has been nothing to do on any drive, + * then there is nothing to do at all. */ *skipped = 1; return (max_sector - sector_nr) + sectors_skipped; @@ -3638,6 +3652,8 @@ static sector_t raid10_sync_request(struct mddev *mddev, sector_t sector_nr, mdname(mddev)); mirror->recovery_disabled = mddev->recovery_disabled; + } else { + error_disk = i; } put_buf(r10_bio); if (rb2) -- 2.39.2