Received: by 2002:a05:6a10:8c0a:0:0:0:0 with SMTP id go10csp3388379pxb; Mon, 25 Jan 2021 15:01:17 -0800 (PST) X-Google-Smtp-Source: ABdhPJwXoM4NE8QTxWrN6IoGg+S7MJTIMGGhaqx2L+HEwFI5avSMeP+LbB5Cc+A0dXBtmDTE4eGC X-Received: by 2002:a17:906:7f83:: with SMTP id f3mr1657223ejr.282.1611615676996; Mon, 25 Jan 2021 15:01:16 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1611615676; cv=none; d=google.com; s=arc-20160816; b=Rhw0p95Svyyx+a3NNjRCTjQHKE95ypifs1kpvO4p4zeOz8wazt/9t8ZMr+0tDmfCCj yCPcI8fBxT++nWwD+Dz14u+/ltIj+bZ6Kw4yKMuiU4Y4eRUXVMhG3ez6q2Cd/0LcJGiK P4PYRoM1XNe9TffoeRunUlgATQu/xOFUDwdrHdh8VXX3S7Xzcx26wHeWrzVVvPQ0IWxp cBr9VqXW2ENduDxBzdarj0xNs9IKnoIGBVLkQfK0aCFvAlZncxC/R5lLAZLnZlpLzEJA GRIXS4msHu0vuhB3YdMZRGljaQtYo5YL6D5JmW20Gp7X0wN0woEHOqHKIPG/axf9hChB v9RQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from; bh=1WGqMYhlqCLkofPg3McDAL7OcZxQrP0y+CUz/AH9v2Y=; b=W9UKY9qgGeoXbUrnB87Ryli3AUKKUvBO0VEVCLZBP/JzK3FICZ1UVTR7hP9ag7tsec FGt5KAjTwuk4pXrJ4advSvCyt/oLNY4Ayx3wh1wlAf7xwXf+Qi26rrkkMT5yc5Agx4b6 qhpaRpisk1Bc+fdFKBohGnvqCX0DDk7nSjOSeSZAVNHrYujk5q3EMLfHUd9zMOKJCI2Q 1KLRG2L2Ci/qGdnb73KI+Qthp277IcBqA6YVTP2KycIq7CfPJR+JUFrgARN7wHaITzlq N2qyHQkTuG5xSIv+VKE958L0MdTnZ9MzAtaozu25sOyWikvpbt90kamaDlfF9G9nclGw fgxw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id s9si6367680ejc.752.2021.01.25.15.00.53; Mon, 25 Jan 2021 15:01:16 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1732671AbhAYW6D (ORCPT + 99 others); Mon, 25 Jan 2021 17:58:03 -0500 Received: from mail.cn.fujitsu.com ([183.91.158.132]:41443 "EHLO heian.cn.fujitsu.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1732118AbhAYW5b (ORCPT ); Mon, 25 Jan 2021 17:57:31 -0500 X-IronPort-AV: E=Sophos;i="5.79,374,1602518400"; d="scan'208";a="103820574" Received: from unknown (HELO cn.fujitsu.com) ([10.167.33.5]) by heian.cn.fujitsu.com with ESMTP; 26 Jan 2021 06:55:38 +0800 Received: from G08CNEXMBPEKD05.g08.fujitsu.local (unknown [10.167.33.204]) by cn.fujitsu.com (Postfix) with ESMTP id 5E4BE4CE6781; Tue, 26 Jan 2021 06:55:36 +0800 (CST) Received: from G08CNEXCHPEKD04.g08.fujitsu.local (10.167.33.200) by G08CNEXMBPEKD05.g08.fujitsu.local (10.167.33.204) with Microsoft SMTP Server (TLS) id 15.0.1497.2; Tue, 26 Jan 2021 06:55:37 +0800 Received: from irides.mr.mr.mr (10.167.225.141) by G08CNEXCHPEKD04.g08.fujitsu.local (10.167.33.209) with Microsoft SMTP Server id 15.0.1497.2 via Frontend Transport; Tue, 26 Jan 2021 06:55:34 +0800 From: Shiyang Ruan To: , , , CC: , , , , , , , , , Subject: [PATCH v2 05/10] mm, pmem: Implement ->memory_failure() in pmem driver Date: Tue, 26 Jan 2021 06:55:21 +0800 Message-ID: <20210125225526.1048877-6-ruansy.fnst@cn.fujitsu.com> X-Mailer: git-send-email 2.30.0 In-Reply-To: <20210125225526.1048877-1-ruansy.fnst@cn.fujitsu.com> References: <20210125225526.1048877-1-ruansy.fnst@cn.fujitsu.com> MIME-Version: 1.0 Content-Transfer-Encoding: 7BIT Content-Type: text/plain; charset=US-ASCII X-yoursite-MailScanner-ID: 5E4BE4CE6781.AB943 X-yoursite-MailScanner: Found to be clean X-yoursite-MailScanner-From: ruansy.fnst@cn.fujitsu.com X-Spam-Status: No Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Call the ->memory_failure() which is implemented by pmem driver, in order to finally notify filesystem to handle the corrupted data. The handler which collects and kills processes are moved into mf_dax_mapping_kill_procs(), which will be called by filesystem. Keep the old handler in order to roll back if driver/device/filesystem does not support ->memory_failure()/->corrupted_range(). Signed-off-by: Shiyang Ruan --- drivers/nvdimm/pmem.c | 25 +++++++++++ mm/memory-failure.c | 102 +++++++++++++++++++++++++----------------- 2 files changed, 86 insertions(+), 41 deletions(-) diff --git a/drivers/nvdimm/pmem.c b/drivers/nvdimm/pmem.c index 875076b0ea6c..c9e4fb38f94a 100644 --- a/drivers/nvdimm/pmem.c +++ b/drivers/nvdimm/pmem.c @@ -363,9 +363,34 @@ static void pmem_release_disk(void *__pmem) put_disk(pmem->disk); } +static int pmem_pagemap_memory_failure(struct dev_pagemap *pgmap, + unsigned long pfn, int flags) +{ + struct pmem_device *pdev; + struct gendisk *disk; + loff_t disk_offset; + int rc = 0; + unsigned long size = page_size(pfn_to_page(pfn)); + + pdev = container_of(pgmap, struct pmem_device, pgmap); + disk = pdev->disk; + if (!disk) + return -ENXIO; + + disk_offset = PFN_PHYS(pfn) - pdev->phys_addr - pdev->data_offset; + if (disk->fops->corrupted_range) { + rc = disk->fops->corrupted_range(disk, NULL, disk_offset, size, &flags); + if (rc == -ENODEV) + rc = -ENXIO; + } else + rc = -EOPNOTSUPP; + return rc; +} + static const struct dev_pagemap_ops fsdax_pagemap_ops = { .kill = pmem_pagemap_kill, .cleanup = pmem_pagemap_cleanup, + .memory_failure = pmem_pagemap_memory_failure, }; static int pmem_attach_disk(struct device *dev, diff --git a/mm/memory-failure.c b/mm/memory-failure.c index 158fe0c8e602..670e29cd263e 100644 --- a/mm/memory-failure.c +++ b/mm/memory-failure.c @@ -1219,6 +1219,54 @@ static int try_to_split_thp_page(struct page *page, const char *msg) return 0; } +int mf_generic_kill_procs(unsigned long long pfn, int flags) +{ + struct page *page = pfn_to_page(pfn); + const bool unmap_success = true; + unsigned long size = 0; + struct to_kill *tk; + LIST_HEAD(to_kill); + loff_t start; + dax_entry_t cookie; + + /* + * Prevent the inode from being freed while we are interrogating + * the address_space, typically this would be handled by + * lock_page(), but dax pages do not use the page lock. This + * also prevents changes to the mapping of this pfn until + * poison signaling is complete. + */ + cookie = dax_lock_page(page); + if (!cookie) + return -EBUSY; + /* + * Unlike System-RAM there is no possibility to swap in a + * different physical page at a given virtual address, so all + * userspace consumption of ZONE_DEVICE memory necessitates + * SIGBUS (i.e. MF_MUST_KILL) + */ + flags |= MF_ACTION_REQUIRED | MF_MUST_KILL; + collect_procs(page, &to_kill, flags & MF_ACTION_REQUIRED); + + list_for_each_entry(tk, &to_kill, nd) + if (tk->size_shift) + size = max(size, 1UL << tk->size_shift); + if (size) { + /* + * Unmap the largest mapping to avoid breaking up + * device-dax mappings which are constant size. The + * actual size of the mapping being torn down is + * communicated in siginfo, see kill_proc() + */ + start = (page->index << PAGE_SHIFT) & ~(size - 1); + unmap_mapping_range(page->mapping, start, start + size, 0); + } + kill_procs(&to_kill, flags & MF_MUST_KILL, !unmap_success, pfn, flags); + + dax_unlock_page(page, cookie); + return 0; +} + int mf_dax_mapping_kill_procs(struct address_space *mapping, pgoff_t index, int flags) { const bool unmap_success = true; @@ -1343,13 +1391,7 @@ static int memory_failure_dev_pagemap(unsigned long pfn, int flags, struct dev_pagemap *pgmap) { struct page *page = pfn_to_page(pfn); - const bool unmap_success = true; - unsigned long size = 0; - struct to_kill *tk; - LIST_HEAD(to_kill); int rc = -EBUSY; - loff_t start; - dax_entry_t cookie; if (flags & MF_COUNT_INCREASED) /* @@ -1357,20 +1399,9 @@ static int memory_failure_dev_pagemap(unsigned long pfn, int flags, */ put_page(page); - /* - * Prevent the inode from being freed while we are interrogating - * the address_space, typically this would be handled by - * lock_page(), but dax pages do not use the page lock. This - * also prevents changes to the mapping of this pfn until - * poison signaling is complete. - */ - cookie = dax_lock_page(page); - if (!cookie) - goto out; - if (hwpoison_filter(page)) { rc = 0; - goto unlock; + goto out; } if (pgmap->type == MEMORY_DEVICE_PRIVATE) { @@ -1378,7 +1409,7 @@ static int memory_failure_dev_pagemap(unsigned long pfn, int flags, * TODO: Handle HMM pages which may need coordination * with device-side memory. */ - goto unlock; + goto out; } /* @@ -1388,32 +1419,21 @@ static int memory_failure_dev_pagemap(unsigned long pfn, int flags, SetPageHWPoison(page); /* - * Unlike System-RAM there is no possibility to swap in a - * different physical page at a given virtual address, so all - * userspace consumption of ZONE_DEVICE memory necessitates - * SIGBUS (i.e. MF_MUST_KILL) + * Call driver's implementation to handle the memory failure, + * otherwise roll back to generic handler. */ - flags |= MF_ACTION_REQUIRED | MF_MUST_KILL; - collect_procs_file(page, page->mapping, page->index, &to_kill, - flags & MF_ACTION_REQUIRED); - - list_for_each_entry(tk, &to_kill, nd) - if (tk->size_shift) - size = max(size, 1UL << tk->size_shift); - if (size) { + if (pgmap->ops->memory_failure) { + rc = pgmap->ops->memory_failure(pgmap, pfn, flags); /* - * Unmap the largest mapping to avoid breaking up - * device-dax mappings which are constant size. The - * actual size of the mapping being torn down is - * communicated in siginfo, see kill_proc() + * Roll back to generic handler too if operation is not + * supported inside the driver/device/filesystem. */ - start = (page->index << PAGE_SHIFT) & ~(size - 1); - unmap_mapping_range(page->mapping, start, start + size, 0); + if (rc != EOPNOTSUPP) + goto out; } - kill_procs(&to_kill, flags & MF_MUST_KILL, !unmap_success, pfn, flags); - rc = 0; -unlock: - dax_unlock_page(page, cookie); + + rc = mf_generic_kill_procs(pfn, flags); + out: /* drop pgmap ref acquired in caller */ put_dev_pagemap(pgmap); -- 2.30.0