Received: by 2002:a05:6358:3188:b0:123:57c1:9b43 with SMTP id q8csp2068777rwd; Fri, 26 May 2023 00:56:33 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ5WIUCYLTPDgkNcmaLcix7dYmxIgwbADNuG3944rhzYvWTmtZyZwzXbtbnbC+4+y5uqtTON X-Received: by 2002:a17:90a:cd05:b0:24d:ec16:6f8c with SMTP id d5-20020a17090acd0500b0024dec166f8cmr1562478pju.20.1685087793516; Fri, 26 May 2023 00:56:33 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1685087793; cv=none; d=google.com; s=arc-20160816; b=xuE6ZAPOeTc2wwnOfTWpQcdvZgBIt06AsERDBTnn/87PbWg7U3oQeIg5BOBQYUhb62 4pGRtu6LXl17TxzPuY10zRqRstzp94pNFgkUgx7HPMFXS8wfYQhhVjWCQOs2JMvWm4zX Rfwn4YpTsXaPkWe+DflV4MoTitAXmXevK3qhD3sv13XeIyE7r8STYMtKvJBq8er974c3 FR083jFK1w/1ONqTn60NYxaC3BDydXMp/labC3qJTYW94cCjTcfhwvL+SAWey4ZWW2DI gbBwD/gqZfYE8vHb2aSyo7TeWZ+LBMmDpb2tfn3skXAi7ZEjUt/bS42ONn2dS+eSasTm d2rw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from; bh=BE0dcualMF6I8Xh2d7sh3ALrOAgsEL6EXxNMIxsRsRQ=; b=Rm8nTmHwq+yTkQF+PK0EfQrmPzRZYWZnbC+EhJGZsLrPXoO7KEd8o4c1QCYtx00MOO JuhG7j5H5iZq24d9ha/Bku08Nf5zsgQ00VOpriypropfBM3v/VuFuEa7QXeQbFAsm/X8 uE/VIr6KBOcSKnvGqhYajlTxsXqGViy2cBmDAw1ITJUGx3TwBcn+qp8enxl061M4yTj1 yR0HHdf3MV2+0bKgH5fYqlBVNSjR4JyHOHpXl+BGeTot8VSxWqg09LeDjc+KGZzol6Bd UFxirjYVw4OH1Cck3MF/6PjVP92ucZd72uGoUcoNy2o2TMlaXYXUDGf0ZJBwowO+rYIv Y3Yg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id a5-20020a17090a8c0500b0025627c2ddeasi1232654pjo.163.2023.05.26.00.56.19; Fri, 26 May 2023 00:56:33 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S242411AbjEZHrv (ORCPT + 99 others); Fri, 26 May 2023 03:47:51 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:42450 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S236611AbjEZHrj (ORCPT ); Fri, 26 May 2023 03:47:39 -0400 Received: from dggsgout12.his.huawei.com (unknown [45.249.212.56]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D5EB0134; Fri, 26 May 2023 00:47:37 -0700 (PDT) Received: from mail02.huawei.com (unknown [172.30.67.153]) by dggsgout12.his.huawei.com (SkyGuard) with ESMTP id 4QSH84536tz4f3nBx; Fri, 26 May 2023 15:47:32 +0800 (CST) Received: from huaweicloud.com (unknown [10.175.127.227]) by APP4 (Coremail) with SMTP id gCh0CgCX_7INZHBko+n3KA--.7888S7; Fri, 26 May 2023 15:47:34 +0800 (CST) From: linan666@huaweicloud.com To: song@kernel.org, bingjingc@synology.com, allenpeng@synology.com, alexwu@synology.com, shli@fb.com, neilb@suse.de Cc: linux-raid@vger.kernel.org, linux-kernel@vger.kernel.org, linan122@huawei.com, yukuai3@huawei.com, yi.zhang@huawei.com, houtao1@huawei.com, yangerkun@huawei.com Subject: [PATCH v2 3/4] md/raid10: fix incorrect done of recovery Date: Fri, 26 May 2023 15:45:50 +0800 Message-Id: <20230526074551.669792-4-linan666@huaweicloud.com> X-Mailer: git-send-email 2.31.1 In-Reply-To: <20230526074551.669792-1-linan666@huaweicloud.com> References: <20230526074551.669792-1-linan666@huaweicloud.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-CM-TRANSID: gCh0CgCX_7INZHBko+n3KA--.7888S7 X-Coremail-Antispam: 1UD129KBjvJXoW7AFyktF13tF17GFWDXr17GFg_yoW8ur4fpa n3JFZ3t398WayfJ3WDAryUXFWFy348Ja4jkr47W34Y9F90qrWDZFWfXFWUXF9rWryFqayY qw4kJrW5CFyxKaDanT9S1TB71UUUUUUqnTZGkaVYY2UrUUUUjbIjqfuFe4nvWSU5nxnvy2 9KBjDU0xBIdaVrnRJUUUm0b4IE77IF4wAFF20E14v26rWj6s0DM7CY07I20VC2zVCF04k2 6cxKx2IYs7xG6rWj6s0DM7CIcVAFz4kK6r1j6r18M28IrcIa0xkI8VA2jI8067AKxVWUWw A2048vs2IY020Ec7CjxVAFwI0_Xr0E3s1l8cAvFVAK0II2c7xJM28CjxkF64kEwVA0rcxS w2x7M28EF7xvwVC0I7IYx2IY67AKxVWDJVCq3wA2z4x0Y4vE2Ix0cI8IcVCY1x0267AKxV W8Jr0_Cr1UM28EF7xvwVC2z280aVAFwI0_GcCE3s1l84ACjcxK6I8E87Iv6xkF7I0E14v2 6rxl6s0DM2vYz4IE04k24VAvwVAKI4IrM2AIxVAIcxkEcVAq07x20xvEncxIr21l5I8CrV ACY4xI64kE6c02F40Ex7xfMcIj6xIIjxv20xvE14v26r126r1DMcIj6I8E87Iv67AKxVWU JVW8JwAm72CE4IkC6x0Yz7v_Jr0_Gr1lF7xvr2IYc2Ij64vIr41lFIxGxcIEc7CjxVA2Y2 ka0xkIwI1lw4CEc2x0rVAKj4xxMxAIw28IcxkI7VAKI48JMxC20s026xCaFVCjc4AY6r1j 6r4UMI8I3I0E5I8CrVAFwI0_Jr0_Jr4lx2IqxVCjr7xvwVAFwI0_JrI_JrWlx4CE17CEb7 AF67AKxVWUtVW8ZwCIc40Y0x0EwIxGrwCI42IY6xIIjxv20xvE14v26r1j6r1xMIIF0xvE 2Ix0cI8IcVCY1x0267AKxVWxJVW8Jr1lIxAIcVCF04k26cxKx2IYs7xG6r1j6r1xMIIF0x vEx4A2jsIE14v26r1j6r4UMIIF0xvEx4A2jsIEc7CjxVAFwI0_Gr0_Gr1UYxBIdaVFxhVj vjDU0xZFpf9x07Ut5rxUUUUU= X-CM-SenderInfo: polqt0awwwqx5xdzvxpfor3voofrz/ X-CFilter-Loop: Reflected X-Spam-Status: No, score=-1.5 required=5.0 tests=BAYES_00,KHOP_HELO_FCRDNS, SPF_HELO_NONE,SPF_NONE,T_SCC_BODY_TEXT_LINE autolearn=no autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Li Nan Recovery will go to giveup and let chunks_skipped++ in raid10_sync_request() if there are some bad_blocks, and it will return max_sector when chunks_skipped >= geo.raid_disks. Now, recovery fail and data is inconsistent but user think recovery is done, it is wrong. Fix it by set mirror's recovery_disabled, spare device will not be added to here. The same issue alos exists on resync, it will be fixd in future. Signed-off-by: Li Nan --- drivers/md/raid10.c | 18 +++++++++++++++++- 1 file changed, 17 insertions(+), 1 deletion(-) diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c index 9de9eabff209..aa22782a7330 100644 --- a/drivers/md/raid10.c +++ b/drivers/md/raid10.c @@ -3303,6 +3303,7 @@ static sector_t raid10_sync_request(struct mddev *mddev, sector_t sector_nr, int chunks_skipped = 0; sector_t chunk_mask = conf->geo.chunk_mask; int page_idx = 0; + int error_disk = -1; /* * Allow skipping a full rebuild for incremental assembly @@ -3386,7 +3387,20 @@ static sector_t raid10_sync_request(struct mddev *mddev, sector_t sector_nr, return reshape_request(mddev, sector_nr, skipped); if (chunks_skipped >= conf->geo.raid_disks) { - /* if there has been nothing to do on any drive, + pr_err("md/raid10:%s: %s fail\n", mdname(mddev), + test_bit(MD_RECOVERY_SYNC, &mddev->recovery) ? "resync" : "recovery"); + if (error_disk >= 0 && + !test_bit(MD_RECOVERY_SYNC, &mddev->recovery)) { + /* + * recovery fail, set mirrors.recovory_disabled, + * device shouldn't be added to there. + */ + conf->mirrors[error_disk].recovery_disabled = + mddev->recovery_disabled; + return 0; + } + /* + * if there has been nothing to do on any drive, * then there is nothing to do at all.. */ *skipped = 1; @@ -3640,6 +3654,8 @@ static sector_t raid10_sync_request(struct mddev *mddev, sector_t sector_nr, mdname(mddev)); mirror->recovery_disabled = mddev->recovery_disabled; + } else { + error_disk = i; } put_buf(r10_bio); if (rb2) -- 2.31.1