Received: by 2002:a05:7412:d1aa:b0:fc:a2b0:25d7 with SMTP id ba42csp1602951rdb; Wed, 31 Jan 2024 03:56:44 -0800 (PST) X-Google-Smtp-Source: AGHT+IHgNl596D12OExhU+i+cWUA8VTzeJmXg2zePSdTKcdsv20QqXtXUcqV0q5jqy8jkI6IwaIL X-Received: by 2002:a17:906:5a8a:b0:a36:600d:a7ca with SMTP id l10-20020a1709065a8a00b00a36600da7camr1028337ejq.18.1706702204197; Wed, 31 Jan 2024 03:56:44 -0800 (PST) ARC-Seal: i=2; a=rsa-sha256; t=1706702204; cv=pass; d=google.com; s=arc-20160816; b=AeNGCJHj+FAnEG157HNtRIAKTGWCmg3L3YkyURKlluMBQsJjVKAwI19zASGzlEZv3r oTRDSEdjdBZ0eRIn8KlNe9IfTVgnh0F4QHuv3OKWA8LZaeavtu3EbMwixrzGvGtbALQR DH2bHEXqCtJJT+jsCCe0kEFvh4C4+TGRv9Wo4pvppQU6CX1LaN9vadleJ6Y1vr4/Cv9k Fj2RaHv7AqVtg4m1RcB8XDlLi6gXR9OEa83XFdkGbqdltrTd4OTObpQIwiYwiMTYq5pb QlhE5QRnx5KO7Oup6L5tBID1pcLV74nl+3EEOnV+P2XF8H5YahPmFkOH8ya0xuA076s7 liLg== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:in-reply-to:from:content-language :references:cc:to:subject:user-agent:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:date:message-id; bh=ygaOgyEV+z8eAfnO2G8+JFFGpraVR3aGrA5RLncnQXw=; fh=sr/zZjOwDF5bMQ/htr0iZtN0X/zJIzyaAy6fyrwvSZM=; b=iZYkhzb5qcyFf5R5HB/N+09hr9O0CciEF6PUb5xgLMywZCpjLk1Ek2hGSt5erjelJd 2+SfHpGKFL+s91fcHQRkI6bxOg4D2MWqOKh6IFdHR7kHnRYpR+IX+RHR9Fd9zsBe4VUQ feBzijw0P6Q9jScTPI8SVzIvAwbxDDSs0nmJK3Y5W3niBTVBE3TNeK3VAcvrcZFmMfot Oh10WlefdHB/BZoLjlLHwxtugucXc5QtU5eD9mqipNdjxUbVh1r8x7ydV5/i4KAWUfX8 nlnQT6tEMQTzpjPlnAtupCgGXbu9Gk25B/D+g48dIW8GZdFYKcXSIefyhNy7xgTW/zYu ax+Q==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; arc=pass (i=1 spf=pass spfdomain=arm.com dmarc=pass fromdomain=arm.com); spf=pass (google.com: domain of linux-kernel+bounces-46380-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.80.249 as permitted sender) smtp.mailfrom="linux-kernel+bounces-46380-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com X-Forwarded-Encrypted: i=1; AJvYcCX44ZgqVQIhxs2tWHuegrWMTEj617i439hX7jq2q2OgKHCu36K+4P2ESCyB9RSh8H2I8B1TcaVZi7drKroUF1nuscXuE9dJczua3WH3ZQ== Return-Path: Received: from am.mirrors.kernel.org (am.mirrors.kernel.org. [147.75.80.249]) by mx.google.com with ESMTPS id c12-20020a17090654cc00b00a353ca3d908si4521504ejp.556.2024.01.31.03.56.44 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 31 Jan 2024 03:56:44 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-46380-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.80.249 as permitted sender) client-ip=147.75.80.249; Authentication-Results: mx.google.com; arc=pass (i=1 spf=pass spfdomain=arm.com dmarc=pass fromdomain=arm.com); spf=pass (google.com: domain of linux-kernel+bounces-46380-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.80.249 as permitted sender) smtp.mailfrom="linux-kernel+bounces-46380-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by am.mirrors.kernel.org (Postfix) with ESMTPS id ECF3A1F2DDF9 for ; Wed, 31 Jan 2024 11:49:42 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 7C7B6768F8; Wed, 31 Jan 2024 11:49:34 +0000 (UTC) Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 624B8762DC; Wed, 31 Jan 2024 11:49:31 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=217.140.110.172 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1706701773; cv=none; b=ttrs5gUKe6cHaP5/P39pUdUgH2RA45IqX9dVyzDuyC13FqP13vjZrPO/az981TeD6ewe2GXEEq+DuDNMR97Q4aReO/Xr1Ma4qdQA1D602fs+1LTY2t824gr9fpsXTDF5kmwK0g9RVVK2Qari417OrLNOyY9XWBPZYbNIM+gXfi8= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1706701773; c=relaxed/simple; bh=kPGAT+VjhwpGiz/PhiDeYTLyEcw/DmxhiOP0SJU7hzc=; h=Message-ID:Date:MIME-Version:Subject:To:Cc:References:From: In-Reply-To:Content-Type; b=NhPCcHrVrliqxSkYhE3LzS44JSIkxP6+G1Df8oL/2XyQ6N2mCioZOOXPNAi5NkZIzKlrUCJjnT3Fa09A1GZtWI1PrOPFM8EWIxGXDf90v2v6SBwJnqhNHSVmso8UmHBg+XKjE/YZrKF/T1JidCnmpf20ho45Z+G8nMobWtFsWcg= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=arm.com; spf=pass smtp.mailfrom=arm.com; arc=none smtp.client-ip=217.140.110.172 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=arm.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=arm.com Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 2681F11FB; Wed, 31 Jan 2024 03:50:14 -0800 (PST) Received: from [10.57.79.60] (unknown [10.57.79.60]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id E085D3F738; Wed, 31 Jan 2024 03:49:26 -0800 (PST) Message-ID: <714d0930-2202-48b6-9728-d248f820325e@arm.com> Date: Wed, 31 Jan 2024 11:49:25 +0000 Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v3 00/15] mm/memory: optimize fork() with PTE-mapped THP To: David Hildenbrand , linux-kernel@vger.kernel.org Cc: linux-mm@kvack.org, Andrew Morton , Matthew Wilcox , Russell King , Catalin Marinas , Will Deacon , Dinh Nguyen , Michael Ellerman , Nicholas Piggin , Christophe Leroy , "Aneesh Kumar K.V" , "Naveen N. Rao" , Paul Walmsley , Palmer Dabbelt , Albert Ou , Alexander Gordeev , Gerald Schaefer , Heiko Carstens , Vasily Gorbik , Christian Borntraeger , Sven Schnelle , "David S. Miller" , linux-arm-kernel@lists.infradead.org, linuxppc-dev@lists.ozlabs.org, linux-riscv@lists.infradead.org, linux-s390@vger.kernel.org, sparclinux@vger.kernel.org References: <20240129124649.189745-1-david@redhat.com> <57eb82c7-4816-42a2-b5ab-cc221e289b21@arm.com> Content-Language: en-GB From: Ryan Roberts In-Reply-To: Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit On 31/01/2024 11:28, David Hildenbrand wrote: > On 31.01.24 12:16, Ryan Roberts wrote: >> On 31/01/2024 11:06, David Hildenbrand wrote: >>> On 31.01.24 11:43, Ryan Roberts wrote: >>>> On 29/01/2024 12:46, David Hildenbrand wrote: >>>>> Now that the rmap overhaul[1] is upstream that provides a clean interface >>>>> for rmap batching, let's implement PTE batching during fork when processing >>>>> PTE-mapped THPs. >>>>> >>>>> This series is partially based on Ryan's previous work[2] to implement >>>>> cont-pte support on arm64, but its a complete rewrite based on [1] to >>>>> optimize all architectures independent of any such PTE bits, and to >>>>> use the new rmap batching functions that simplify the code and prepare >>>>> for further rmap accounting changes. >>>>> >>>>> We collect consecutive PTEs that map consecutive pages of the same large >>>>> folio, making sure that the other PTE bits are compatible, and (a) adjust >>>>> the refcount only once per batch, (b) call rmap handling functions only >>>>> once per batch and (c) perform batch PTE setting/updates. >>>>> >>>>> While this series should be beneficial for adding cont-pte support on >>>>> ARM64[2], it's one of the requirements for maintaining a total mapcount[3] >>>>> for large folios with minimal added overhead and further changes[4] that >>>>> build up on top of the total mapcount. >>>>> >>>>> Independent of all that, this series results in a speedup during fork with >>>>> PTE-mapped THP, which is the default with THPs that are smaller than a PMD >>>>> (for example, 16KiB to 1024KiB mTHPs for anonymous memory[5]). >>>>> >>>>> On an Intel Xeon Silver 4210R CPU, fork'ing with 1GiB of PTE-mapped folios >>>>> of the same size (stddev < 1%) results in the following runtimes >>>>> for fork() (shorter is better): >>>>> >>>>> Folio Size | v6.8-rc1 |      New | Change >>>>> ------------------------------------------ >>>>>         4KiB | 0.014328 | 0.014035 |   - 2% >>>>>        16KiB | 0.014263 | 0.01196  |   -16% >>>>>        32KiB | 0.014334 | 0.01094  |   -24% >>>>>        64KiB | 0.014046 | 0.010444 |   -26% >>>>>       128KiB | 0.014011 | 0.010063 |   -28% >>>>>       256KiB | 0.013993 | 0.009938 |   -29% >>>>>       512KiB | 0.013983 | 0.00985  |   -30% >>>>>      1024KiB | 0.013986 | 0.00982  |   -30% >>>>>      2048KiB | 0.014305 | 0.010076 |   -30% >>>> >>>> Just a heads up that I'm seeing some strange results on Apple M2. Fork for >>>> order-0 is seemingly costing ~17% more. I'm using GCC 13.2 and was pretty >>>> sure I >>>> didn't see this problem with version 1; although that was on a different >>>> baseline and I've thrown the numbers away so will rerun and try to debug this. Numbers for v1 of the series, both on top of 6.8-rc1 and rebased to the same mm-unstable base as v3 of the series (first 2 rows are from what I just posted for context): | kernel | mean_rel | std_rel | |:-------------------|-----------:|----------:| | mm-unstabe (base) | 0.0% | 1.1% | | mm-unstable + v3 | 16.7% | 0.8% | | mm-unstable + v1 | -2.5% | 1.7% | | v6.8-rc1 + v1 | -6.6% | 1.1% | So all looks good with v1. And seems to suggest mm-unstable has regressed by ~4% vs v6.8-rc1. Is this really a useful benchmark? Does the raw performance of fork() syscall really matter? Evidence suggests its moving all over the place - breath on the code and it changes - not a great place to be when using the test for gating purposes! Still with the old tests - I'll move to the new ones now. >>>> >>> >>> So far, on my x86 tests (Intel, AMD EPYC), I was not able to observe this. >>> fork() for order-0 was consistently effectively unchanged. Do you observe that >>> on other ARM systems as well? >> >> Nope; running the exact same kernel binary and user space on Altra, I see >> sensible numbers; >> >> fork order-0: -1.3% >> fork order-9: -7.6% >> dontneed order-0: -0.5% >> dontneed order-9: 0.1% >> munmap order-0: 0.0% >> munmap order-9: -67.9% >> >> So I guess some pipelining issue that causes the M2 to stall more? > > With one effective added folio_test_large(), it could only be a code layout > problem? Or the compiler does something stupid, but you say that you run the > exact same kernel binary, so that doesn't make sense. Yup, same binary. We know this code is very sensitive - 1 cycle makes a big difference. So could easily be code layout, branch prediction, etc... > > I'm also surprised about the dontneed vs. munmap numbers. You mean the ones for Altra that I posted? (I didn't post any for M2). The altra numbers look ok to me; dontneed has no change, and munmap has no change for order-0 and is massively improved for order-9. Doesn't make any sense > (again, there was this VMA merging problem but it would still allow for batching > within a single VMA that spans exactly one large folio). > > What are you using as baseline? Really just mm-unstable vs. mm-unstable+patches? yes. except for "v6.8-rc1 + v1" above. > > Let's see if the new test changes the numbers you measure. >