Received: by 2002:a05:6359:6284:b0:131:369:b2a3 with SMTP id se4csp420282rwb; Fri, 4 Aug 2023 15:16:10 -0700 (PDT) X-Google-Smtp-Source: AGHT+IHUfW5zsr3RZoU5K6ZNXkrZ/9fA5x8bGWR+YBAQnJ7r0LJVYUwlIRnJM4Iuz+N5yElXYZaU X-Received: by 2002:a17:90a:1f06:b0:262:e49b:12d0 with SMTP id u6-20020a17090a1f0600b00262e49b12d0mr2199993pja.48.1691187369910; Fri, 04 Aug 2023 15:16:09 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1691187369; cv=none; d=google.com; s=arc-20160816; b=RmJbmchbU0NE3Nu/40jwtQi7DSPR9gdf2LY0BjwVCe7ftEYnmrbMPsnzxXs+aFVPB5 p02jCZEYjSpO2PE5jsh+fsvm3O+MukO7z2JbtA2EaA/2eMAIip6aH187qte73r+Q4/KL dA7RUT5bHDtiaXj6cNh1jhmNGtMuiRxATboyI+zHbzW8WE2z1QgKuGHtNDM7j6VR5cTn dJ1qqvmsoAhdElFBJIGgkSYs1q+cdLVPaqFNCbrRZblTDvap3vntYYup8ZECIG4zho/F OXRKua+kAJF0kVr+wV40CkprhVV2iTe7K9ym/kKvLjqP2go5d08oK+46lshR6UyWwAP3 dIbQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:subject :organization:from:references:cc:to:content-language:user-agent :mime-version:date:message-id:dkim-signature; bh=e/k7/FOb2397vAPUhkAxLXzvl/jl3TjsQOaaW7NC/DA=; fh=w4ihpAb6qOgfGy/CQ9HhxWu9r3tWAjuE3x2qITYqANE=; b=apNsM4bzC61Gk7ulQaqnFFJug09gHLvRPU7XHeLbr9lg4vehdZA1mAo8Z14qPyn8I8 5QYTPieyHEFHh35LRtc9zr1woW+DjO+vSRrdrHCssmQneZrR9QPQ7teULutmyuVNDryN C/WSKLnmUQyW27iGKOxbJ2PrptXPLaOO2k8lter82j570/4jRDL/wjnYHRCJwmpL86sk Ac819Id1MXn8jXPCatzApt9kenL6UcS8uoUxCera91s3nnY1UP/U8ZBBJT/dOqJweYSr O9cNqZGVa0O+HTNnT9mREa+KIyOeqk6E7DX1XzGbwHK0mqpQMoYQa9cpTsThK8QKeTIb q0lw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=Cg5B2t4S; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id t22-20020a17090ad51600b0025c0d113469si149229pju.58.2023.08.04.15.15.57; Fri, 04 Aug 2023 15:16:09 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=Cg5B2t4S; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229825AbjHDVPD (ORCPT + 99 others); Fri, 4 Aug 2023 17:15:03 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:40310 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230498AbjHDVPB (ORCPT ); Fri, 4 Aug 2023 17:15:01 -0400 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.129.124]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id F01C5524E for ; Fri, 4 Aug 2023 14:13:58 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1691183638; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=e/k7/FOb2397vAPUhkAxLXzvl/jl3TjsQOaaW7NC/DA=; b=Cg5B2t4SJjbifOlZBLGhSnUsntroYZ0y26FIHrZ3CQbIpfcpOmdEk4lGyMwzayilB6o6Yk DwVagw7mGi81zya1tAthV9YX4iJv6/KQUvhKDUvzH2dhLptoqdL1TPlYYRwMS4sz4a6UxN um5LTdGQeVlbBaOHbHfaHnwWaihRFX4= Received: from mail-wr1-f71.google.com (mail-wr1-f71.google.com [209.85.221.71]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.3, cipher=TLS_AES_256_GCM_SHA384) id us-mta-571-n1Z78ubuMUujDHhH7dUlVQ-1; Fri, 04 Aug 2023 17:13:56 -0400 X-MC-Unique: n1Z78ubuMUujDHhH7dUlVQ-1 Received: by mail-wr1-f71.google.com with SMTP id ffacd0b85a97d-2f2981b8364so1416238f8f.1 for ; Fri, 04 Aug 2023 14:13:55 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20221208; t=1691183634; x=1691788434; h=content-transfer-encoding:in-reply-to:subject:organization:from :references:cc:to:content-language:user-agent:mime-version:date :message-id:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=e/k7/FOb2397vAPUhkAxLXzvl/jl3TjsQOaaW7NC/DA=; b=NhS8eX1itJR5GxfKuhb5sUWXM4Lkm7MW25u6YR31nbHLhZEGLaFn9HPwLq1iwXyBW6 3YnYwRjgvG2tm/7CZxYheoBftcvrtJh62UT98kYIBZDybCgmVZoiTd9ZihQWMaT+3xTg SxjTBdJgiVcz0QJXZXnrxayVyC1FyFm8F36GzxMyqicWXrNIpqkol/1K5KQON2CB6cMJ OxKGZIMHwc5/1nnf3erfkTrAyZ9TEYQjjq9HZQMf7nPkUGgHYx0G6O1++Vlg+OK0L0ni khkzxvzVTo4cSWvuRKIyz3TKtwKH01ly66D3k74l6goobcCVhvfbZj6SorxuC6+LDAuf o9Lw== X-Gm-Message-State: AOJu0YxmfQlQwdJy2RXGSFxOrIhbc649d0bVyFH4VeXbWKFciGqo+DwH /sRtVvoUQPzQRyPAGkQhc3BAei4uleK+SGOV3dcMAHUuDTOyeiHwBSlzEGYgt4liiADnqfh9HwL JOyVrxBTXKt/K7aLR5Z6Gv/Jn X-Received: by 2002:adf:e7c3:0:b0:314:4473:5bad with SMTP id e3-20020adfe7c3000000b0031444735badmr2103664wrn.65.1691183633906; Fri, 04 Aug 2023 14:13:53 -0700 (PDT) X-Received: by 2002:adf:e7c3:0:b0:314:4473:5bad with SMTP id e3-20020adfe7c3000000b0031444735badmr2103647wrn.65.1691183633457; Fri, 04 Aug 2023 14:13:53 -0700 (PDT) Received: from ?IPV6:2003:d8:2f2d:8e00:a20e:59bc:3c13:4806? (p200300d82f2d8e00a20e59bc3c134806.dip0.t-ipconnect.de. [2003:d8:2f2d:8e00:a20e:59bc:3c13:4806]) by smtp.gmail.com with ESMTPSA id t6-20020a5d6a46000000b003142e438e8csm3374296wrw.26.2023.08.04.14.13.52 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Fri, 04 Aug 2023 14:13:53 -0700 (PDT) Message-ID: <259ad8fc-c12b-69b9-ba16-adb9e3e6d672@redhat.com> Date: Fri, 4 Aug 2023 23:13:51 +0200 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:102.0) Gecko/20100101 Thunderbird/102.13.0 Content-Language: en-US To: Yu Zhao Cc: Ryan Roberts , Andrew Morton , Matthew Wilcox , Yin Fengwei , Catalin Marinas , Will Deacon , Anshuman Khandual , Yang Shi , "Huang, Ying" , Zi Yan , Luis Chamberlain , Itaru Kitayama , "Kirill A. Shutemov" , linux-mm@kvack.org, linux-kernel@vger.kernel.org, linux-arm-kernel@lists.infradead.org References: <20230726095146.2826796-1-ryan.roberts@arm.com> <20230726095146.2826796-3-ryan.roberts@arm.com> <5e595904-3dca-0e15-0769-7ed10975fd0d@arm.com> From: David Hildenbrand Organization: Red Hat Subject: Re: [PATCH v4 2/5] mm: LARGE_ANON_FOLIO for improved performance In-Reply-To: Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit X-Spam-Status: No, score=-2.2 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,NICE_REPLY_A, RCVD_IN_DNSWL_BLOCKED,RCVD_IN_MSPIKE_H4,RCVD_IN_MSPIKE_WL, SPF_HELO_NONE,SPF_NONE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 04.08.23 23:00, Yu Zhao wrote: > On Fri, Aug 4, 2023 at 2:23 PM David Hildenbrand wrote: >> >> On 04.08.23 10:27, Ryan Roberts wrote: >>> On 04/08/2023 00:50, Yu Zhao wrote: >>>> On Thu, Aug 3, 2023 at 6:43 AM Ryan Roberts wrote: >>>>> >>>>> + Kirill >>>>> >>>>> On 26/07/2023 10:51, Ryan Roberts wrote: >>>>>> Introduce LARGE_ANON_FOLIO feature, which allows anonymous memory to be >>>>>> allocated in large folios of a determined order. All pages of the large >>>>>> folio are pte-mapped during the same page fault, significantly reducing >>>>>> the number of page faults. The number of per-page operations (e.g. ref >>>>>> counting, rmap management lru list management) are also significantly >>>>>> reduced since those ops now become per-folio. >>>>>> >>>>>> The new behaviour is hidden behind the new LARGE_ANON_FOLIO Kconfig, >>>>>> which defaults to disabled for now; The long term aim is for this to >>>>>> defaut to enabled, but there are some risks around internal >>>>>> fragmentation that need to be better understood first. >>>>>> >>>>>> When enabled, the folio order is determined as such: For a vma, process >>>>>> or system that has explicitly disabled THP, we continue to allocate >>>>>> order-0. THP is most likely disabled to avoid any possible internal >>>>>> fragmentation so we honour that request. >>>>>> >>>>>> Otherwise, the return value of arch_wants_pte_order() is used. For vmas >>>>>> that have not explicitly opted-in to use transparent hugepages (e.g. >>>>>> where thp=madvise and the vma does not have MADV_HUGEPAGE), then >>>>>> arch_wants_pte_order() is limited to 64K (or PAGE_SIZE, whichever is >>>>>> bigger). This allows for a performance boost without requiring any >>>>>> explicit opt-in from the workload while limitting internal >>>>>> fragmentation. >>>>>> >>>>>> If the preferred order can't be used (e.g. because the folio would >>>>>> breach the bounds of the vma, or because ptes in the region are already >>>>>> mapped) then we fall back to a suitable lower order; first >>>>>> PAGE_ALLOC_COSTLY_ORDER, then order-0. >>>>>> >>>>> >>>>> ... >>>>> >>>>>> +#define ANON_FOLIO_MAX_ORDER_UNHINTED \ >>>>>> + (ilog2(max_t(unsigned long, SZ_64K, PAGE_SIZE)) - PAGE_SHIFT) >>>>>> + >>>>>> +static int anon_folio_order(struct vm_area_struct *vma) >>>>>> +{ >>>>>> + int order; >>>>>> + >>>>>> + /* >>>>>> + * If THP is explicitly disabled for either the vma, the process or the >>>>>> + * system, then this is very likely intended to limit internal >>>>>> + * fragmentation; in this case, don't attempt to allocate a large >>>>>> + * anonymous folio. >>>>>> + * >>>>>> + * Else, if the vma is eligible for thp, allocate a large folio of the >>>>>> + * size preferred by the arch. Or if the arch requested a very small >>>>>> + * size or didn't request a size, then use PAGE_ALLOC_COSTLY_ORDER, >>>>>> + * which still meets the arch's requirements but means we still take >>>>>> + * advantage of SW optimizations (e.g. fewer page faults). >>>>>> + * >>>>>> + * Finally if thp is enabled but the vma isn't eligible, take the >>>>>> + * arch-preferred size and limit it to ANON_FOLIO_MAX_ORDER_UNHINTED. >>>>>> + * This ensures workloads that have not explicitly opted-in take benefit >>>>>> + * while capping the potential for internal fragmentation. >>>>>> + */ >>>>>> + >>>>>> + if ((vma->vm_flags & VM_NOHUGEPAGE) || >>>>>> + test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags) || >>>>>> + !hugepage_flags_enabled()) >>>>>> + order = 0; >>>>>> + else { >>>>>> + order = max(arch_wants_pte_order(), PAGE_ALLOC_COSTLY_ORDER); >>>>>> + >>>>>> + if (!hugepage_vma_check(vma, vma->vm_flags, false, true, true)) >>>>>> + order = min(order, ANON_FOLIO_MAX_ORDER_UNHINTED); >>>>>> + } >>>>>> + >>>>>> + return order; >>>>>> +} >>>>> >>>>> >>>>> Hi All, >>>>> >>>>> I'm writing up the conclusions that we arrived at during discussion in the THP >>>>> meeting yesterday, regarding linkage with exiting THP ABIs. It would be great if >>>>> I can get explicit "agree" or disagree + rationale from at least David, Yu and >>>>> Kirill. >>>>> >>>>> In summary; I think we are converging on the approach that is already coded, but >>>>> I'd like confirmation. >>>>> >>>>> >>>>> >>>>> The THP situation today >>>>> ----------------------- >>>>> >>>>> - At system level: THP can be set to "never", "madvise" or "always" >>>>> - At process level: THP can be "never" or "defer to system setting" >>>>> - At VMA level: no-hint, MADV_HUGEPAGE, MADV_NOHUGEPAGE >>>>> >>>>> That gives us this table to describe how a page fault is handled, according to >>>>> process state (columns) and vma flags (rows): >>>>> >>>>> | never | madvise | always >>>>> ----------------|-----------|-----------|----------- >>>>> no hint | S | S | THP>S >>>>> MADV_HUGEPAGE | S | THP>S | THP>S >>>>> MADV_NOHUGEPAGE | S | S | S >>>>> >>>>> Legend: >>>>> S allocate single page (PTE-mapped) >>>>> LAF allocate lage anon folio (PTE-mapped) >>>>> THP allocate THP-sized folio (PMD-mapped) >>>>>> fallback (usually because vma size/alignment insufficient for folio) >>>>> >>>>> >>>>> >>>>> Principles for Large Anon Folios (LAF) >>>>> -------------------------------------- >>>>> >>>>> David tells us there are use cases today (e.g. qemu live migration) which use >>>>> MADV_NOHUGEPAGE to mean "don't fill any PTEs that are not explicitly faulted" >>>>> and these use cases will break (i.e. functionally incorrect) if this request is >>>>> not honoured. >>>> >>>> I don't remember David saying this. I think he was referring to UFFD, >>>> not MADV_NOHUGEPAGE, when discussing what we need to absolutely >>>> respect. >>> >>> My understanding was that MADV_NOHUGEPAGE was being applied to regions *before* >>> UFFD was being registered, and the app relied on MADV_NOHUGEPAGE to not back any >>> unfaulted pages. It's not completely clear to me how not honouring >>> MADV_NOHUGEPAGE would break things though. David? >> >> Sorry, I'm still lagging behind on some threads. >> >> Imagine the following for VM postcopy live migration: >> >> (1) Set MADV_NOHUGEPAGE on guest memory and discard all memory (e.g., >> MADV_DONTNEED), to start with a clean slate. >> (2) Migrates some pages during precopy from the source and stores them >> into guest memory on the destination. Some of the memory locations >> will have pages populated. >> (3) At some point, decide to enable postcopy: enable userfaultfd on >> guest memory. >> (4) Discard *selected* pages again that have been dirtied in the >> meantime on the source. These are pages that have been migrated >> previously. >> (5) Start running the VM on the destination. >> (6) Anything that's not populated will trigger userfaultfd missing >> faults. Then, you can request them from the source and place them. >> >> Assume you would populate more than required during 2), you can end up >> not getting userfaultfd faults during 4) and corrupt your guest state. >> It works if during (2) you migrated all guest memory, or if during 4) >> you zap everything that still needs migr > > I see what you mean now. Thanks. > > Yes, in this case we have to interpret MADV_NOHUGEPAGE as nothing >4KB. Note that it's still even unclear to me why we want to *not* call these things THP. It would certainly make everything less confusing if we call them THP, but with additional attributes. I think that is one of the first things we should figure out because it also indirectly tells us what all these toggles mean and how/if we should redefine them (and if they even apply). Currently THP == PMD size In 2016, Hugh already envisioned PUD/PGD THP (see 49920d28781d ("mm: make transparent hugepage size public")) when he explicitly exposed "hpage_pmd_size". Not "hpage_size". For hugetlb on arm64 we already support various sizes that are < PMD size and *not* call them differently. It's a huge(tlb) page. Sometimes we refer to them as cont-PTE hugetlb pages. So, nowadays we do have "PMD-sized THP", someday we might have "PUD-sized THP". Can't we come up with a name to describe sub-PMD THP? Is it really of value if we invent a new term for them? Yes, I was not enjoying "Flexible THP". Once we figured that out, we should figure out if MADV_HUGEPAGE meant "only PMD-sized THP" or anything else? Also, we can then figure out if MADV_NOHUGEPAGE meant "only PMD-sized THP" or anything else? The simplest approach to me would be "they imply any THP, and once we need more tunables we might add some", similar to what Kirill also raised. Again, it's all unclear to me at this point and I'm happy to hear opinions, because I really don't know. -- Cheers, David / dhildenb