Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 62C49C7EE31 for ; Wed, 1 Mar 2023 11:03:26 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229907AbjCALDZ (ORCPT ); Wed, 1 Mar 2023 06:03:25 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:45330 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229898AbjCALDW (ORCPT ); Wed, 1 Mar 2023 06:03:22 -0500 Received: from out30-99.freemail.mail.aliyun.com (out30-99.freemail.mail.aliyun.com [115.124.30.99]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D3830A24B for ; Wed, 1 Mar 2023 03:03:18 -0800 (PST) X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R191e4;CH=green;DM=||false|;DS=||;FP=0|-1|-1|-1|0|-1|-1|-1;HT=ay29a033018045170;MF=baolin.wang@linux.alibaba.com;NM=1;PH=DS;RN=14;SR=0;TI=SMTPD_---0Vct2H0z_1677668593; Received: from 30.97.48.59(mailfrom:baolin.wang@linux.alibaba.com fp:SMTPD_---0Vct2H0z_1677668593) by smtp.aliyun-inc.com; Wed, 01 Mar 2023 19:03:14 +0800 Message-ID: Date: Wed, 1 Mar 2023 19:03:20 +0800 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:102.0) Gecko/20100101 Thunderbird/102.7.2 Subject: Re: [PATCH 3/3] migrate_pages: try migrate in batch asynchronously firstly To: "Huang, Ying" Cc: Andrew Morton , linux-mm@kvack.org, linux-kernel@vger.kernel.org, Hugh Dickins , "Xu, Pengfei" , Christoph Hellwig , Stefan Roesch , Tejun Heo , Xin Hao , Zi Yan , Yang Shi , Matthew Wilcox , Mike Kravetz References: <20230224141145.96814-1-ying.huang@intel.com> <20230224141145.96814-4-ying.huang@intel.com> <87zg8x9epg.fsf@yhuang6-desk2.ccr.corp.intel.com> From: Baolin Wang In-Reply-To: <87zg8x9epg.fsf@yhuang6-desk2.ccr.corp.intel.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 3/1/2023 2:18 PM, Huang, Ying wrote: > Baolin Wang writes: > >> On 2/24/2023 10:11 PM, Huang Ying wrote: >>> When we have locked more than one folios, we cannot wait the lock or >>> bit (e.g., page lock, buffer head lock, writeback bit) synchronously. >>> Otherwise deadlock may be triggered. This make it hard to batch the >>> synchronous migration directly. >>> This patch re-enables batching synchronous migration via trying to >>> migrate in batch asynchronously firstly. And any folios that are >>> failed to be migrated asynchronously will be migrated synchronously >>> one by one. >>> Test shows that this can restore the TLB flushing batching >>> performance >>> for synchronous migration effectively. >>> Signed-off-by: "Huang, Ying" >>> Cc: Hugh Dickins >>> Cc: "Xu, Pengfei" >>> Cc: Christoph Hellwig >>> Cc: Stefan Roesch >>> Cc: Tejun Heo >>> Cc: Xin Hao >>> Cc: Zi Yan >>> Cc: Yang Shi >>> Cc: Baolin Wang >>> Cc: Matthew Wilcox >>> Cc: Mike Kravetz >>> --- >>> mm/migrate.c | 65 ++++++++++++++++++++++++++++++++++++++++++++-------- >>> 1 file changed, 55 insertions(+), 10 deletions(-) >>> diff --git a/mm/migrate.c b/mm/migrate.c >>> index 91198b487e49..c17ce5ee8d92 100644 >>> --- a/mm/migrate.c >>> +++ b/mm/migrate.c >>> @@ -1843,6 +1843,51 @@ static int migrate_pages_batch(struct list_head *from, new_page_t get_new_page, >>> return rc; >>> } >>> +static int migrate_pages_sync(struct list_head *from, new_page_t >>> get_new_page, >>> + free_page_t put_new_page, unsigned long private, >>> + enum migrate_mode mode, int reason, struct list_head *ret_folios, >>> + struct list_head *split_folios, struct migrate_pages_stats *stats) >>> +{ >>> + int rc, nr_failed = 0; >>> + LIST_HEAD(folios); >>> + struct migrate_pages_stats astats; >>> + >>> + memset(&astats, 0, sizeof(astats)); >>> + /* Try to migrate in batch with MIGRATE_ASYNC mode firstly */ >>> + rc = migrate_pages_batch(from, get_new_page, put_new_page, private, MIGRATE_ASYNC, >>> + reason, &folios, split_folios, &astats, >>> + NR_MAX_MIGRATE_PAGES_RETRY); >>> + stats->nr_succeeded += astats.nr_succeeded; >>> + stats->nr_thp_succeeded += astats.nr_thp_succeeded; >>> + stats->nr_thp_split += astats.nr_thp_split; >>> + if (rc < 0) { >>> + stats->nr_failed_pages += astats.nr_failed_pages; >>> + stats->nr_thp_failed += astats.nr_thp_failed; >>> + list_splice_tail(&folios, ret_folios); >>> + return rc; >>> + } >>> + stats->nr_thp_failed += astats.nr_thp_split; >>> + nr_failed += astats.nr_thp_split; >>> + /* >>> + * Fall back to migrate all failed folios one by one synchronously. All >>> + * failed folios except split THPs will be retried, so their failure >>> + * isn't counted >>> + */ >>> + list_splice_tail_init(&folios, from); >>> + while (!list_empty(from)) { >>> + list_move(from->next, &folios); >>> + rc = migrate_pages_batch(&folios, get_new_page, put_new_page, >>> + private, mode, reason, ret_folios, >>> + split_folios, stats, NR_MAX_MIGRATE_PAGES_RETRY); >>> + list_splice_tail_init(&folios, ret_folios); >>> + if (rc < 0) >>> + return rc; >>> + nr_failed += rc; >>> + } >>> + >>> + return nr_failed; >>> +} >>> + >>> /* >>> * migrate_pages - migrate the folios specified in a list, to the free folios >>> * supplied as the target for the page migration >>> @@ -1874,7 +1919,7 @@ int migrate_pages(struct list_head *from, new_page_t get_new_page, >>> enum migrate_mode mode, int reason, unsigned int *ret_succeeded) >>> { >>> int rc, rc_gather; >>> - int nr_pages, batch; >>> + int nr_pages; >>> struct folio *folio, *folio2; >>> LIST_HEAD(folios); >>> LIST_HEAD(ret_folios); >>> @@ -1890,10 +1935,6 @@ int migrate_pages(struct list_head *from, new_page_t get_new_page, >>> if (rc_gather < 0) >>> goto out; >>> - if (mode == MIGRATE_ASYNC) >>> - batch = NR_MAX_BATCHED_MIGRATION; >>> - else >>> - batch = 1; >>> again: >>> nr_pages = 0; >>> list_for_each_entry_safe(folio, folio2, from, lru) { >>> @@ -1904,16 +1945,20 @@ int migrate_pages(struct list_head *from, new_page_t get_new_page, >>> } >>> nr_pages += folio_nr_pages(folio); >>> - if (nr_pages >= batch) >>> + if (nr_pages >= NR_MAX_BATCHED_MIGRATION) >>> break; >>> } >>> - if (nr_pages >= batch) >>> + if (nr_pages >= NR_MAX_BATCHED_MIGRATION) >>> list_cut_before(&folios, from, &folio2->lru); >>> else >>> list_splice_init(from, &folios); >>> - rc = migrate_pages_batch(&folios, get_new_page, put_new_page, private, >>> - mode, reason, &ret_folios, &split_folios, &stats, >>> - NR_MAX_MIGRATE_PAGES_RETRY); >>> + if (mode == MIGRATE_ASYNC) >>> + rc = migrate_pages_batch(&folios, get_new_page, put_new_page, private, >>> + mode, reason, &ret_folios, &split_folios, &stats, >>> + NR_MAX_MIGRATE_PAGES_RETRY); >>> + else >>> + rc = migrate_pages_sync(&folios, get_new_page, put_new_page, private, >>> + mode, reason, &ret_folios, &split_folios, &stats); >> >> For split folios, it seems also reasonable to use migrate_pages_sync() >> instead of always using fixed MIGRATE_ASYNC mode? > > For split folios, we only try to migrate them with minimal effort. > Previously, we decrease the retry number from 10 to 1. Now, I think > that it's reasonable to change the migration mode to MIGRATE_ASYNC to > reduce latency. They have been counted as failure anyway. Sounds reasonable. Thanks for explanation. Please feel free to add: Reviewed-by: Baolin Wang