Received: by 2002:ab2:620c:0:b0:1ef:ffd0:ce49 with SMTP id o12csp1248726lqt; Tue, 19 Mar 2024 18:45:30 -0700 (PDT) X-Forwarded-Encrypted: i=3; AJvYcCUiOt1fsG5ovfelGBzws7G4cU1Tw0XZSxGrGwxnrW/crT05cn35j7TjxwDlw9c8Knfjs6419R9KvAA4iFlwKtCaO8JGsYJzHS1X2oX8qw== X-Google-Smtp-Source: AGHT+IGYgb1vmawUvKwEeMbA/K6/GUdSk2SoxQtI5O7lulVEE0+su0B60jcsp/KfteDTOTEbFI+b X-Received: by 2002:a37:e313:0:b0:789:d55a:d1d3 with SMTP id y19-20020a37e313000000b00789d55ad1d3mr4599955qki.4.1710899130424; Tue, 19 Mar 2024 18:45:30 -0700 (PDT) ARC-Seal: i=2; a=rsa-sha256; t=1710899130; cv=pass; d=google.com; s=arc-20160816; b=JgiAIszG30mynOufEcc1/8bDGBZURN66Wzr/uGt3Eu6A5KdKjxaYRK8QnmzuMetUBf B9Ohc/frCet4sDS0qA3Hwb2Nvn62g8/PzLT+9sV2hTkx6aT6ej1lhOhT4IrxlVvmMRcU JgB1cDqiGi2XFrZS5sJIye7eDcXLYhTsW6j7QHLseHcjLTGVE1FTJoDXt1nGvSlX9oCC u+Pih8D1a9PISVWmi7W/B0VGE0v95xWwNf2/AyIo+VkD7AnWMJU0yAYQO3YbKY6xjeO3 SPmieEIWfBqj6NibCf9cz9gR/5Zfge/y4Vlx04gPNmf3oq7C/praI8RBf8z0IsYS73uj BE8w== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:reply-to:message-id:date:subject :cc:to:from:feedback-id:dkim-signature:dkim-signature; bh=OllGBd/e2P4hWoSOipzEy0ACyR7T+q17LVRPTxf0DMQ=; fh=0SnJ87RJni787+3mD8GPeuSJMeVnrb05uDR0gj1ch8Y=; b=BX09a6+C9t4bE44VHMhq+Hcqcq2TMrcJOeZshi4t3R+qcfxIuBr577T29MOxiGx/fF jlqD4CIT/vFRqFb/r0mQkZv2hWHVtg4oBDADuFRXW5kpPDY9nqMcGTDQ+CjoU1epk+3z Jvmq0ZuaAmzc7CUz74mE0pYr7zmxjFsAaH1cuR99Ojg+4y8KigmDk+SZj0iCmu9GABJY hG3RBDiJ9qJyMTcIEWLZsFg5UbbMbkANCeT/kfNcUKrFI2v2422WVbnuJIaXiMt4LfU0 sBq1RVoqdj4niPT/8/9/KXNjQR5tlIqE50hzHP0A+365sZsFYm+12z9/QmRMoL3kSbTH 1y/w==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@sent.com header.s=fm2 header.b=MZgnlphE; dkim=pass header.i=@messagingengine.com header.s=fm2 header.b=vaMgqZNG; arc=pass (i=1 spf=pass spfdomain=sent.com dkim=pass dkdomain=sent.com dkim=pass dkdomain=messagingengine.com dmarc=pass fromdomain=sent.com); spf=pass (google.com: domain of linux-kernel+bounces-108328-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-108328-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=sent.com Return-Path: Received: from ny.mirrors.kernel.org (ny.mirrors.kernel.org. [2604:1380:45d1:ec00::1]) by mx.google.com with ESMTPS id m22-20020a05620a215600b00789f242c63bsi7483843qkm.715.2024.03.19.18.45.30 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 19 Mar 2024 18:45:30 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel+bounces-108328-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) client-ip=2604:1380:45d1:ec00::1; Authentication-Results: mx.google.com; dkim=pass header.i=@sent.com header.s=fm2 header.b=MZgnlphE; dkim=pass header.i=@messagingengine.com header.s=fm2 header.b=vaMgqZNG; arc=pass (i=1 spf=pass spfdomain=sent.com dkim=pass dkdomain=sent.com dkim=pass dkdomain=messagingengine.com dmarc=pass fromdomain=sent.com); spf=pass (google.com: domain of linux-kernel+bounces-108328-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-108328-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=sent.com Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ny.mirrors.kernel.org (Postfix) with ESMTPS id 1A5CE1C220CF for ; Wed, 20 Mar 2024 01:45:30 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 27A0F7489; Wed, 20 Mar 2024 01:45:28 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=sent.com header.i=@sent.com header.b="MZgnlphE"; dkim=pass (2048-bit key) header.d=messagingengine.com header.i=@messagingengine.com header.b="vaMgqZNG" Received: from fhigh6-smtp.messagingengine.com (fhigh6-smtp.messagingengine.com [103.168.172.157]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 7A64E7464 for ; Wed, 20 Mar 2024 01:45:24 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=103.168.172.157 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1710899126; cv=none; b=UE+kUweNGwi7Jeg8qCOKP1x7AJamot7vn7MWYPINfyvH5E/m0tgSmj94/py2MscL+kXE496Wz+0UySJKTQgy2BBypbG6iJFjMPZ8efajaUOxhV1AeqhGfGVTjhfwTVA9yHy5Y52Nfc04bbaojTO7A4P+KkD6SkI+ynoZsg76ECs= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1710899126; c=relaxed/simple; bh=Hr+AikhkC1pDqOhI1uyGIhBY7nBLv2Fwzn0O/Au6RXY=; h=From:To:Cc:Subject:Date:Message-ID:MIME-Version:Content-Type; b=oNt6e+W2xJBEhwPHYBtxNBIL5DjkxkoN4/N1F0D6A3CeHjOj+01xLune6HszNipQh3YwlDIUNuULQAuR72BpIeWwj9yDJsT+JhymY2Dnrv/HBlBqcKbCbKTOW4KrVig5B5goo/AluuS3Qc75028NS2GizypmkSUMgMsv9M9mI+M= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=sent.com; spf=pass smtp.mailfrom=sent.com; dkim=pass (2048-bit key) header.d=sent.com header.i=@sent.com header.b=MZgnlphE; dkim=pass (2048-bit key) header.d=messagingengine.com header.i=@messagingengine.com header.b=vaMgqZNG; arc=none smtp.client-ip=103.168.172.157 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=sent.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=sent.com Received: from compute5.internal (compute5.nyi.internal [10.202.2.45]) by mailfhigh.nyi.internal (Postfix) with ESMTP id 8B156114011F; Tue, 19 Mar 2024 21:45:23 -0400 (EDT) Received: from mailfrontend1 ([10.202.2.162]) by compute5.internal (MEProxy); Tue, 19 Mar 2024 21:45:23 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=sent.com; h=cc :cc:content-transfer-encoding:content-type:content-type:date :date:from:from:in-reply-to:message-id:mime-version:reply-to :reply-to:subject:subject:to:to; s=fm2; t=1710899123; x= 1710985523; bh=OllGBd/e2P4hWoSOipzEy0ACyR7T+q17LVRPTxf0DMQ=; b=M ZgnlphEblWkAjWrxCxWUh6gDkJ5GAShRjT40ftx3ETe8GbZ0T7ZpKNkw8VkB8WH1 rfcXw6cW5XxutEDJBKUBvHQphk+w3I5bpnuAqN6v3fh4Slg3IzsCs2T80b5PaLJv 6TG3ADK1rCUdlWx5GD8ugznonTzm0Zea/8hu+m8JJ0ZVOipdO3qJRFSuf5Cr8yuW OCtCBkp+LMGYKABMJNs2/eWnRR6vFxJTZIFOrcaj+mIfDF/79uXURqxbMM/G2bdG r46AZ6xBsM6TXwRpM3O2SluGTbwRbLezWpBcL/ND9yrBeEV7RxLk3fblNULCFBMg oQRSFsrs60o0C9EeRZbYA== DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d= messagingengine.com; h=cc:cc:content-transfer-encoding :content-type:content-type:date:date:feedback-id:feedback-id :from:from:in-reply-to:message-id:mime-version:reply-to:reply-to :subject:subject:to:to:x-me-proxy:x-me-proxy:x-me-sender :x-me-sender:x-sasl-enc; s=fm2; t=1710899123; x=1710985523; bh=O llGBd/e2P4hWoSOipzEy0ACyR7T+q17LVRPTxf0DMQ=; b=vaMgqZNGFpNqXGQoF y1LPPJaBw97LhzZ7VLVWAAYKoc96ems8x6Y+b5r0zig/5T/XYml4QJLFmubuKTTc /n4vLq12ylEYTomaJEK76+lBXv2+zdNd/+2P8bWLv55q+U790AB+gDBHIQY293J5 zeU7DdSUK5xPMXyXOes5WwqzwLNvsWkCvO9OrEFc/iBVQoTUhRaeqokz1yjuRQH8 Y5EyKcc/3XTdnKgfQ3XzHilAfxSVljDzqPzjRe3uRfFqf4aoX7kvv6jl+lNC1WAY DKKkTZ3AihZd4RprQ0aJ4u57Dc6EsicRnXuwXIWSXDgJjy0xWtYeIOK4tDLFfQUq x8ipg== X-ME-Sender: X-ME-Received: X-ME-Proxy-Cause: gggruggvucftvghtrhhoucdtuddrgedvledrledvgddtlecutefuodetggdotefrodftvf curfhrohhfihhlvgemucfhrghsthforghilhdpqfgfvfdpuffrtefokffrpgfnqfghnecu uegrihhlohhuthemuceftddtnecusecvtfgvtghiphhivghnthhsucdlqddutddtmdenuc fjughrpefhvfevufffkfforhggtgfgsehtkeertdertdejnecuhfhrohhmpegkihcujggr nhcuoeiiihdrhigrnhesshgvnhhtrdgtohhmqeenucggtffrrghtthgvrhhnpeetudevhe etheehieetiefhjeevjeeltdfgvdeiueeiudetffdtvedthfetvedtffenucffohhmrghi nhepkhgvrhhnvghlrdhorhhgnecuvehluhhsthgvrhfuihiivgeptdenucfrrghrrghmpe hmrghilhhfrhhomhepiihirdihrghnsehsvghnthdrtghomh X-ME-Proxy: Feedback-ID: iccd040f4:Fastmail Received: by mail.messagingengine.com (Postfix) with ESMTPA; Tue, 19 Mar 2024 21:45:22 -0400 (EDT) From: Zi Yan To: linux-mm@kvack.org Cc: Zi Yan , Andrew Morton , "Matthew Wilcox (Oracle)" , Yang Shi , Huang Ying , "Kirill A . Shutemov" , Ryan Roberts , Baolin Wang , "Yin, Fengwei" , SeongJae Park , linux-kernel@vger.kernel.org Subject: [PATCH v4] mm/migrate: split source folio if it is on deferred split list Date: Tue, 19 Mar 2024 21:45:11 -0400 Message-ID: <20240320014511.306128-1-zi.yan@sent.com> X-Mailer: git-send-email 2.43.0 Reply-To: Zi Yan Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: 8bit From: Zi Yan If the source folio is on deferred split list, it is likely some subpages are not used. Split it before migration to avoid migrating unused subpages. Commit 616b8371539a6 ("mm: thp: enable thp migration in generic path") did not check if a THP is on deferred split list before migration, thus, the destination THP is never put on deferred split list even if the source THP might be. The opportunity of reclaiming free pages in a partially mapped THP during deferred list scanning is lost, but no other harmful consequence is present[1]. From v3: 1. Guarded deferred list code behind CONFIG_TRANSPARENT_HUGEPAGE to avoid compilation error (per SeongJae Park). From v2: 1. Split the source folio instead of migrating it (per Matthew Wilcox)[2]. From v1: 1. Used dst to get correct deferred split list after migration (per Ryan Roberts). [1]: https://lore.kernel.org/linux-mm/03CE3A00-917C-48CC-8E1C-6A98713C817C@nvidia.com/ [2]: https://lore.kernel.org/linux-mm/Ze_P6xagdTbcu1Kz@casper.infradead.org/ Fixes: 616b8371539a ("mm: thp: enable thp migration in generic path") Signed-off-by: Zi Yan --- mm/huge_memory.c | 22 ----------------- mm/internal.h | 25 +++++++++++++++++++ mm/migrate.c | 62 +++++++++++++++++++++++++++++++++++++++--------- 3 files changed, 76 insertions(+), 33 deletions(-) diff --git a/mm/huge_memory.c b/mm/huge_memory.c index b40bd9f3ead5..c77cedf45f3a 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -766,28 +766,6 @@ pmd_t maybe_pmd_mkwrite(pmd_t pmd, struct vm_area_struct *vma) return pmd; } -#ifdef CONFIG_MEMCG -static inline -struct deferred_split *get_deferred_split_queue(struct folio *folio) -{ - struct mem_cgroup *memcg = folio_memcg(folio); - struct pglist_data *pgdat = NODE_DATA(folio_nid(folio)); - - if (memcg) - return &memcg->deferred_split_queue; - else - return &pgdat->deferred_split_queue; -} -#else -static inline -struct deferred_split *get_deferred_split_queue(struct folio *folio) -{ - struct pglist_data *pgdat = NODE_DATA(folio_nid(folio)); - - return &pgdat->deferred_split_queue; -} -#endif - void folio_prep_large_rmappable(struct folio *folio) { if (!folio || !folio_test_large(folio)) diff --git a/mm/internal.h b/mm/internal.h index 85c3db43454d..56cf2051cb88 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -1106,6 +1106,31 @@ struct page *follow_trans_huge_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmd, unsigned int flags); +#ifdef CONFIG_TRANSPARENT_HUGEPAGE +#ifdef CONFIG_MEMCG +static inline +struct deferred_split *get_deferred_split_queue(struct folio *folio) +{ + struct mem_cgroup *memcg = folio_memcg(folio); + struct pglist_data *pgdat = NODE_DATA(folio_nid(folio)); + + if (memcg) + return &memcg->deferred_split_queue; + else + return &pgdat->deferred_split_queue; +} +#else +static inline +struct deferred_split *get_deferred_split_queue(struct folio *folio) +{ + struct pglist_data *pgdat = NODE_DATA(folio_nid(folio)); + + return &pgdat->deferred_split_queue; +} +#endif +#endif + + /* * mm/mmap.c */ diff --git a/mm/migrate.c b/mm/migrate.c index 73a052a382f1..e80cb0f46342 100644 --- a/mm/migrate.c +++ b/mm/migrate.c @@ -1654,25 +1654,65 @@ static int migrate_pages_batch(struct list_head *from, /* * Large folio migration might be unsupported or - * the allocation might be failed so we should retry - * on the same folio with the large folio split + * the folio is on deferred split list so we should + * retry on the same folio with the large folio split * to normal folios. * * Split folios are put in split_folios, and * we will migrate them after the rest of the * list is processed. */ - if (!thp_migration_supported() && is_thp) { - nr_failed++; - stats->nr_thp_failed++; - if (!try_split_folio(folio, split_folios)) { - stats->nr_thp_split++; - stats->nr_split++; + if (is_thp) { + bool is_on_deferred_list = false; + +#ifdef CONFIG_TRANSPARENT_HUGEPAGE + /* + * Check without taking split_queue_lock to + * reduce locking overheads. The worst case is + * that if the folio is put on the deferred + * split list after the check, it will be + * migrated and not put back on the list. + * The migrated folio will not be split + * via shrinker during memory pressure. + */ + if (!data_race(list_empty(&folio->_deferred_list))) { + struct deferred_split *ds_queue; + unsigned long flags; + + ds_queue = + get_deferred_split_queue(folio); + spin_lock_irqsave(&ds_queue->split_queue_lock, + flags); + /* + * Only check if the folio is on + * deferred split list without removing + * it. Since the folio can be on + * deferred_split_scan() local list and + * removing it can cause the local list + * corruption. Folio split process + * below can handle it with the help of + * folio_ref_freeze(). + */ + is_on_deferred_list = + !list_empty(&folio->_deferred_list); + spin_unlock_irqrestore(&ds_queue->split_queue_lock, + flags); + } +#endif + if (!thp_migration_supported() || + is_on_deferred_list) { + nr_failed++; + stats->nr_thp_failed++; + if (!try_split_folio(folio, + split_folios)) { + stats->nr_thp_split++; + stats->nr_split++; + continue; + } + stats->nr_failed_pages += nr_pages; + list_move_tail(&folio->lru, ret_folios); continue; } - stats->nr_failed_pages += nr_pages; - list_move_tail(&folio->lru, ret_folios); - continue; } rc = migrate_folio_unmap(get_new_folio, put_new_folio, -- 2.43.0