Received: by 2002:a05:7412:b130:b0:e2:908c:2ebd with SMTP id az48csp2089924rdb; Mon, 20 Nov 2023 01:20:10 -0800 (PST) X-Google-Smtp-Source: AGHT+IFbRO3Orfe2pfKY6v+01b4Gwf6Vv3dDMSs0hut6nr9oJtbmFCp2MqXaOw6yd877wHuYMCPd X-Received: by 2002:a05:6870:c1d2:b0:1e9:8b78:899c with SMTP id i18-20020a056870c1d200b001e98b78899cmr9094330oad.55.1700472010681; Mon, 20 Nov 2023 01:20:10 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1700472010; cv=none; d=google.com; s=arc-20160816; b=xY9HRjmBpkoO5QjKvS6LtBMKRCqKJc6h2MVno3ON8tDC1N4SZL1johnIpzjWOo3k9w 5uYnVgiZf6zZexENxmeIaW/AM0K4vS0/1Tq04VRb3rd2uQ9Y+4kFqi+SZUIydNNXC0Mh 2MUyUzDIDCfU0iEcyDaj5UFPjuZCO6RXDgSJLRj4AybumJrmtKt9fGuKjqyDNdDOHFiR xrZ0uL/SeUq2YKr/FvPm4FiANQdNOA+HOTz+JaGW925qh4DdNx7kSF/MTlrkCB5K+4P0 k+UdfaafZrVAjoXhxKbcXOq96CeXwvyCrKLqv8rb01CqbeI7q29kfnjRiHi9Np9vZM1b 0i4Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:from :references:cc:to:subject:user-agent:mime-version:date:message-id; bh=2Z+Uk4JNIskL2AsgPhyZCAXwRbIwvsVXAJHPDbd7Smc=; fh=sYKSPbAuqRqYl0ekM99qCJlJrgrNE+GOWwc0V0zqqnI=; b=Mbpeoq5Pd7RF7swZ9QLiZCTx97QVIxtgUWmdJwU4RcBFvyAODBD/fI45qO30ORWdQ2 NgegDrsu6oRhyCkkbYvvFKNljC5nhJOITm/C5bqznwvr/wpw8DK2aZlcic/F+gu+cFRb +IGro0naL4YBPpUUBvb8bMXer+OFVO/6Mzy5sDgT+yETpYYtPsQZxcuG3lkgNV9c/QI7 u6qhJ9enN41JnaplJddvGlk4LQad3Io3/CxMG1OB6rqUuY4z2NBRx3k4d8RFnUGtFz1S N0dlyksBVMBvMXI7HNl43cWKBiT8qWwiu3hhSCuRi6ZQbi/OS4ijozuALeORHl6pjqZM hZIg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.37 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Return-Path: Received: from snail.vger.email (snail.vger.email. [23.128.96.37]) by mx.google.com with ESMTPS id h3-20020a056a00170300b006cb8cf06613si2366827pfc.2.2023.11.20.01.20.09 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 20 Nov 2023 01:20:10 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.37 as permitted sender) client-ip=23.128.96.37; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.37 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by snail.vger.email (Postfix) with ESMTP id 9440E806290D; Mon, 20 Nov 2023 01:18:29 -0800 (PST) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.11 at snail.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232500AbjKTJS0 (ORCPT + 99 others); Mon, 20 Nov 2023 04:18:26 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:57976 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232554AbjKTJSZ (ORCPT ); Mon, 20 Nov 2023 04:18:25 -0500 Received: from out30-101.freemail.mail.aliyun.com (out30-101.freemail.mail.aliyun.com [115.124.30.101]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id A554B137 for ; Mon, 20 Nov 2023 01:18:20 -0800 (PST) X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R151e4;CH=green;DM=||false|;DS=||;FP=0|-1|-1|-1|0|-1|-1|-1;HT=ay29a033018046060;MF=baolin.wang@linux.alibaba.com;NM=1;PH=DS;RN=19;SR=0;TI=SMTPD_---0VwkgKKM_1700471895; Received: from 30.97.48.46(mailfrom:baolin.wang@linux.alibaba.com fp:SMTPD_---0VwkgKKM_1700471895) by smtp.aliyun-inc.com; Mon, 20 Nov 2023 17:18:17 +0800 Message-ID: Date: Mon, 20 Nov 2023 17:18:29 +0800 MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v1 1/4] mm/compaction: enable compacting >0 order folios. To: Zi Yan , linux-mm@kvack.org, linux-kernel@vger.kernel.org Cc: "Huang, Ying" , Ryan Roberts , Andrew Morton , "Matthew Wilcox (Oracle)" , David Hildenbrand , "Yin, Fengwei" , Yu Zhao , Vlastimil Babka , "Kirill A . Shutemov" , Johannes Weiner , Kemeng Shi , Mel Gorman , Rohan Puri , Mcgrof Chamberlain , Adam Manzanares , "Vishal Moola (Oracle)" References: <20231113170157.280181-1-zi.yan@sent.com> <20231113170157.280181-2-zi.yan@sent.com> From: Baolin Wang In-Reply-To: <20231113170157.280181-2-zi.yan@sent.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-Spam-Status: No, score=-9.9 required=5.0 tests=BAYES_00, ENV_AND_HDR_SPF_MATCH,RCVD_IN_DNSWL_NONE,SPF_HELO_NONE,SPF_PASS, T_SCC_BODY_TEXT_LINE,UNPARSEABLE_RELAY,USER_IN_DEF_SPF_WL autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (snail.vger.email [0.0.0.0]); Mon, 20 Nov 2023 01:18:29 -0800 (PST) On 11/14/2023 1:01 AM, Zi Yan wrote: > From: Zi Yan > > migrate_pages() supports >0 order folio migration and during compaction, > even if compaction_alloc() cannot provide >0 order free pages, > migrate_pages() can split the source page and try to migrate the base pages > from the split. It can be a baseline and start point for adding support for > compacting >0 order folios. > > Suggested-by: Huang Ying > Signed-off-by: Zi Yan > --- > mm/compaction.c | 57 ++++++++++++++++++++++++++++++++++++------------- > 1 file changed, 42 insertions(+), 15 deletions(-) > > diff --git a/mm/compaction.c b/mm/compaction.c > index 01ba298739dd..5217dd35b493 100644 > --- a/mm/compaction.c > +++ b/mm/compaction.c > @@ -816,6 +816,21 @@ static bool too_many_isolated(struct compact_control *cc) > return too_many; > } > > +/* > + * 1. if the page order is larger than or equal to target_order (i.e., > + * cc->order and when it is not -1 for global compaction), skip it since > + * target_order already indicates no free page with larger than target_order > + * exists and later migrating it will most likely fail; > + * > + * 2. compacting > pageblock_order pages does not improve memory fragmentation, > + * skip them; > + */ > +static bool skip_isolation_on_order(int order, int target_order) > +{ > + return (target_order != -1 && order >= target_order) || > + order >= pageblock_order; > +} > + > /** > * isolate_migratepages_block() - isolate all migrate-able pages within > * a single pageblock > @@ -1009,7 +1024,7 @@ isolate_migratepages_block(struct compact_control *cc, unsigned long low_pfn, > /* > * Regardless of being on LRU, compound pages such as THP and > * hugetlbfs are not to be compacted unless we are attempting > - * an allocation much larger than the huge page size (eg CMA). > + * an allocation larger than the compound page size. > * We can potentially save a lot of iterations if we skip them > * at once. The check is racy, but we can consider only valid > * values and the only danger is skipping too much. > @@ -1017,11 +1032,18 @@ isolate_migratepages_block(struct compact_control *cc, unsigned long low_pfn, > if (PageCompound(page) && !cc->alloc_contig) { > const unsigned int order = compound_order(page); > > - if (likely(order <= MAX_ORDER)) { > - low_pfn += (1UL << order) - 1; > - nr_scanned += (1UL << order) - 1; > + /* > + * Skip based on page order and compaction target order > + * and skip hugetlbfs pages. > + */ > + if (skip_isolation_on_order(order, cc->order) || > + PageHuge(page)) { > + if (order <= MAX_ORDER) { > + low_pfn += (1UL << order) - 1; > + nr_scanned += (1UL << order) - 1; > + } > + goto isolate_fail; > } > - goto isolate_fail; > } > > /* > @@ -1144,17 +1166,18 @@ isolate_migratepages_block(struct compact_control *cc, unsigned long low_pfn, > goto isolate_abort; > } > } > + } > > - /* > - * folio become large since the non-locked check, > - * and it's on LRU. > - */ > - if (unlikely(folio_test_large(folio) && !cc->alloc_contig)) { > - low_pfn += folio_nr_pages(folio) - 1; > - nr_scanned += folio_nr_pages(folio) - 1; > - folio_set_lru(folio); > - goto isolate_fail_put; > - } > + /* > + * Check LRU folio order under the lock > + */ > + if (unlikely(skip_isolation_on_order(folio_order(folio), > + cc->order) && > + !cc->alloc_contig)) { > + low_pfn += folio_nr_pages(folio) - 1; > + nr_scanned += folio_nr_pages(folio) - 1; > + folio_set_lru(folio); > + goto isolate_fail_put; > } Why was this part moved out of the 'if (lruvec != locked)' block? If we hold the lru lock, then we do not need to check again, right?