Received: by 2002:a05:7412:f690:b0:e2:908c:2ebd with SMTP id ej16csp302815rdb; Thu, 19 Oct 2023 05:13:18 -0700 (PDT) X-Google-Smtp-Source: AGHT+IHhtkfdtWvIPaqV4Ubu2FziVAclu7lk1+ZMPIErnhn6yUVbw7Q85LFZUzbO2AKCMfKIGym/ X-Received: by 2002:a05:6870:120f:b0:1ea:7ec9:3dcd with SMTP id 15-20020a056870120f00b001ea7ec93dcdmr2217026oan.1.1697717598153; Thu, 19 Oct 2023 05:13:18 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1697717598; cv=none; d=google.com; s=arc-20160816; b=OPMpkXEJYohASDLUtyZB3lUSwTOCm5xZZF01l0gY3NhlQhYkhXYYTR2xY4jEeZfLD3 p7cF4dwR1hfdICBmexQFIkYKrJmq1hcDfdPZdLNGE5sv4dylAgNyEQNiABSKxL/US30/ h7kZmRc1c7rQuqpzZVzPW7uAEHFNjPCmjztwQYx/psopXoTE/9weyHFNqhv6fW47AlqP ov+N4x26RnVVBFXAz6ytRYWM1X4/qw6jJeu/EofSCOw5/9BTApzcfmpRjMokYGV84Ylx 8YmY98/mcsJ35QI7GJoI76P8bGgohbJZlBtgiGiU2Vj7drhKepHGoFP3kslYM3wN2EpA s1wA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date; bh=bJ4K/u4PLM7GC+IUYpRGVBX6AaneyJ2KnY/OZ2mnKew=; fh=HR4DBHbif49FM8LxiYZcqJ7T0XBSkbhO9SOtYE6LAJA=; b=1GFV3btk+UICPJkjxRlYRla1JoELJHuXln7UMRVxHqIQR8p2wsz214cgJsAJ1FhWLu O0JjFCb9G2oxYfxd7Emj1N4UHT7TamWsdA1Q1juS1pMUgQzQm9RQQKOeipxBrFovQGEa eYcJa5rOTDweOH+AWXdtcdD6FcqKsSEveYKZqW0EwLnnPGrbGKn1jFZdOlX3XJ8gKjMV OGUVsJRDb4txZZ114+P/pnh5KTPPXgcBUG4bI0HuBQvZ+op0OhGxow7M5RwgyEoIVXPf o4tUMGvKLDASJuD8Xjl8mzHyUFiM1Tb7FImMYThDSqZ6MjthTGy7JmQ5khdcT3MZLHpZ YisA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:6 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from pete.vger.email (pete.vger.email. [2620:137:e000::3:6]) by mx.google.com with ESMTPS id ca42-20020a056a0206aa00b005897813624fsi4966587pgb.476.2023.10.19.05.13.17 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 19 Oct 2023 05:13:18 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:6 as permitted sender) client-ip=2620:137:e000::3:6; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:6 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by pete.vger.email (Postfix) with ESMTP id E43F481A7D26; Thu, 19 Oct 2023 05:13:14 -0700 (PDT) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.10 at pete.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1345479AbjJSMNH (ORCPT + 99 others); Thu, 19 Oct 2023 08:13:07 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33668 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S233294AbjJSMNG (ORCPT ); Thu, 19 Oct 2023 08:13:06 -0400 Received: from outbound-smtp43.blacknight.com (outbound-smtp43.blacknight.com [46.22.139.229]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 5D8AEBE for ; Thu, 19 Oct 2023 05:13:03 -0700 (PDT) Received: from mail.blacknight.com (pemlinmail04.blacknight.ie [81.17.254.17]) by outbound-smtp43.blacknight.com (Postfix) with ESMTPS id 38B1A20BC for ; Thu, 19 Oct 2023 13:13:02 +0100 (IST) Received: (qmail 26744 invoked from network); 19 Oct 2023 12:13:02 -0000 Received: from unknown (HELO techsingularity.net) (mgorman@techsingularity.net@[84.203.199.31]) by 81.17.254.9 with ESMTPSA (AES256-SHA encrypted, authenticated); 19 Oct 2023 12:13:01 -0000 Date: Thu, 19 Oct 2023 13:12:58 +0100 From: Mel Gorman To: Huang Ying Cc: Andrew Morton , linux-mm@kvack.org, linux-kernel@vger.kernel.org, Arjan Van De Ven , Vlastimil Babka , David Hildenbrand , Johannes Weiner , Dave Hansen , Michal Hocko , Pavel Tatashin , Matthew Wilcox , Christoph Lameter Subject: Re: [PATCH -V3 4/9] mm: restrict the pcp batch scale factor to avoid too long latency Message-ID: <20231019121258.52y5o7aaivyq2ex7@techsingularity.net> References: <20231016053002.756205-1-ying.huang@intel.com> <20231016053002.756205-5-ying.huang@intel.com> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-15 Content-Disposition: inline In-Reply-To: <20231016053002.756205-5-ying.huang@intel.com> X-Spam-Status: No, score=-0.8 required=5.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on pete.vger.email Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (pete.vger.email [0.0.0.0]); Thu, 19 Oct 2023 05:13:15 -0700 (PDT) On Mon, Oct 16, 2023 at 01:29:57PM +0800, Huang Ying wrote: > In page allocator, PCP (Per-CPU Pageset) is refilled and drained in > batches to increase page allocation throughput, reduce page > allocation/freeing latency per page, and reduce zone lock contention. > But too large batch size will cause too long maximal > allocation/freeing latency, which may punish arbitrary users. So the > default batch size is chosen carefully (in zone_batchsize(), the value > is 63 for zone > 1GB) to avoid that. > > In commit 3b12e7e97938 ("mm/page_alloc: scale the number of pages that > are batch freed"), the batch size will be scaled for large number of > page freeing to improve page freeing performance and reduce zone lock > contention. Similar optimization can be used for large number of > pages allocation too. > > To find out a suitable max batch scale factor (that is, max effective > batch size), some tests and measurement on some machines were done as > follows. > > A set of debug patches are implemented as follows, > > - Set PCP high to be 2 * batch to reduce the effect of PCP high > > - Disable free batch size scaling to get the raw performance. > > - The code with zone lock held is extracted from rmqueue_bulk() and > free_pcppages_bulk() to 2 separate functions to make it easy to > measure the function run time with ftrace function_graph tracer. > > - The batch size is hard coded to be 63 (default), 127, 255, 511, > 1023, 2047, 4095. > > Then will-it-scale/page_fault1 is used to generate the page > allocation/freeing workload. The page allocation/freeing throughput > (page/s) is measured via will-it-scale. The page allocation/freeing > average latency (alloc/free latency avg, in us) and allocation/freeing > latency at 99 percentile (alloc/free latency 99%, in us) are measured > with ftrace function_graph tracer. > > The test results are as follows, > > Sapphire Rapids Server > ====================== > Batch throughput free latency free latency alloc latency alloc latency > page/s avg / us 99% / us avg / us 99% / us > ----- ---------- ------------ ------------ ------------- ------------- > 63 513633.4 2.33 3.57 2.67 6.83 > 127 517616.7 4.35 6.65 4.22 13.03 > 255 520822.8 8.29 13.32 7.52 25.24 > 511 524122.0 15.79 23.42 14.02 49.35 > 1023 525980.5 30.25 44.19 25.36 94.88 > 2047 526793.6 59.39 84.50 45.22 140.81 > > Ice Lake Server > =============== > Batch throughput free latency free latency alloc latency alloc latency > page/s avg / us 99% / us avg / us 99% / us > ----- ---------- ------------ ------------ ------------- ------------- > 63 620210.3 2.21 3.68 2.02 4.35 > 127 627003.0 4.09 6.86 3.51 8.28 > 255 630777.5 7.70 13.50 6.17 15.97 > 511 633651.5 14.85 22.62 11.66 31.08 > 1023 637071.1 28.55 42.02 20.81 54.36 > 2047 638089.7 56.54 84.06 39.28 91.68 > > Cascade Lake Server > =================== > Batch throughput free latency free latency alloc latency alloc latency > page/s avg / us 99% / us avg / us 99% / us > ----- ---------- ------------ ------------ ------------- ------------- > 63 404706.7 3.29 5.03 3.53 4.75 > 127 422475.2 6.12 9.09 6.36 8.76 > 255 411522.2 11.68 16.97 10.90 16.39 > 511 428124.1 22.54 31.28 19.86 32.25 > 1023 414718.4 43.39 62.52 40.00 66.33 > 2047 429848.7 86.64 120.34 71.14 106.08 > > Commet Lake Desktop > =================== > Batch throughput free latency free latency alloc latency alloc latency > page/s avg / us 99% / us avg / us 99% / us > ----- ---------- ------------ ------------ ------------- ------------- > > 63 795183.13 2.18 3.55 2.03 3.05 > 127 803067.85 3.91 6.56 3.85 5.52 > 255 812771.10 7.35 10.80 7.14 10.20 > 511 817723.48 14.17 27.54 13.43 30.31 > 1023 818870.19 27.72 40.10 27.89 46.28 > > Coffee Lake Desktop > =================== > Batch throughput free latency free latency alloc latency alloc latency > page/s avg / us 99% / us avg / us 99% / us > ----- ---------- ------------ ------------ ------------- ------------- > 63 510542.8 3.13 4.40 2.48 3.43 > 127 514288.6 5.97 7.89 4.65 6.04 > 255 516889.7 11.86 15.58 8.96 12.55 > 511 519802.4 23.10 28.81 16.95 26.19 > 1023 520802.7 45.30 52.51 33.19 45.95 > 2047 519997.1 90.63 104.00 65.26 81.74 > > From the above data, to restrict the allocation/freeing latency to be > less than 100 us in most times, the max batch scale factor needs to be > less than or equal to 5. > > Although it is reasonable to use 5 as max batch scale factor for the > systems tested, there are also slower systems. Where smaller value > should be used to constrain the page allocation/freeing latency. > > So, in this patch, a new kconfig option (PCP_BATCH_SCALE_MAX) is added > to set the max batch scale factor. Whose default value is 5, and > users can reduce it when necessary. > > Signed-off-by: "Huang, Ying" > Acked-by: Andrew Morton Acked-by: Mel Gorman -- Mel Gorman SUSE Labs