Received: by 2002:a05:6358:bb9e:b0:b9:5105:a5b4 with SMTP id df30csp105378rwb; Fri, 2 Sep 2022 10:55:35 -0700 (PDT) X-Google-Smtp-Source: AA6agR7q8x9ePUWrRo0edEX4pG0g3aRTecZGiIUSgJCmb58LhAygRdhje9vHSYBVUYGwtu6SJQxC X-Received: by 2002:a63:6403:0:b0:42b:f6bc:817f with SMTP id y3-20020a636403000000b0042bf6bc817fmr22148510pgb.283.1662141335124; Fri, 02 Sep 2022 10:55:35 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1662141335; cv=none; d=google.com; s=arc-20160816; b=cjAunsDG8vcBBKM5EeBKUAu4W7hh7gKi+H+Ji/0HK2ruUhNiP8i21n0gn/G7zHeIPN RJjsdl8cZG2hpVQxft83p9wN9Za60Aopk24RNUpoJEU7YC27WPxIHEuGmCLc19hrz8Ch wps+R6MVhQwUZtBvsVpHlC8gLxZzYtfNPdl3L76TrWpkTcTPBiJj1OGJEzPpsLS014G1 q6LVklAHCdEa9eCebeSNI5EAn3BnJYII1RWo7+PXjyCRDBe5B0fUP068H0NeHomBsC8V px5jl8YsNgY62XAe7t5FWzfeoMO/M9ftRVgp8uCVtFCZFovaSmLp0kbWJO0uWzFQkKdZ MNCw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=cOrUWtTrBL3zDaJGPAOPIuPhvPakueyffj9KJ2tlTT4=; b=KBdfKjGff5/AptjnTVVZluu32783Y04d+3KyLtUpkQGR9Mfli5kpeTz6qcnkR7Q6Gh 0L6jlAb+zNmfE8Bz5YXQA/X02u/C2/NsKqbgi1BrAYUXfQsKsb8HoJgOKXTgQF6De4EE PPJrNxeQQcPfgykdlGW6KpTMLNUMOzjf4vfMugZi90u71kxNcyIQoNXJxar4JKgv1xwz ciXw+hDndw8je//Hq3Vapk0dmg1NZzqYZVayS+V7OJa9jhD+q5Z9SLGg782x13WdNiHg ODgNvrJqCeujQGNvvwmDklFoKKcQ/LY7/IYMHMxxQXN9UxrNNYfMcYZ8W+4WBnsSxggV qVBA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20210112 header.b=BE8sVUGO; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id l2-20020a63da42000000b0042b95537587si2569763pgj.359.2022.09.02.10.55.23; Fri, 02 Sep 2022 10:55:35 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20210112 header.b=BE8sVUGO; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232470AbiIBRaj (ORCPT + 99 others); Fri, 2 Sep 2022 13:30:39 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:52946 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S237164AbiIBRaX (ORCPT ); Fri, 2 Sep 2022 13:30:23 -0400 Received: from mail-pj1-x1036.google.com (mail-pj1-x1036.google.com [IPv6:2607:f8b0:4864:20::1036]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id ADAF83AB32 for ; Fri, 2 Sep 2022 10:30:21 -0700 (PDT) Received: by mail-pj1-x1036.google.com with SMTP id u9-20020a17090a1f0900b001fde6477464so6128365pja.4 for ; Fri, 02 Sep 2022 10:30:21 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:from:to:cc:subject:date; bh=cOrUWtTrBL3zDaJGPAOPIuPhvPakueyffj9KJ2tlTT4=; b=BE8sVUGOaOsUdYhSmBT7vTpDvHDWvzQDGTHMIlQNsZnjHfGVnIuueCCDcTbhjvK65o 159BgFFdcgoaTt6YrUKubRK1pVYkLNO0nUwJsLvgfKXJRtiekHXLq7MP1peZUZRGSGqq XdbONBuOmMFNP6xau6YsxzPNEBugrC53ZSfnnKkTazkEcsj3fA7I0a7LQMfPTw8n2xT6 mD7PDJ0rvPjbjyppGxGYN32UNmrsV0Qyw68U7CnJh6GoSR7w8wuKo2tU33dOlsMNp7/K yWmdUpEHAcn+EGkQMrNG1XHARNimffDj2uzSlC6SZLkhV3ndW7KIaihGCgr5290pEuKG QInQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:x-gm-message-state:from:to:cc:subject:date; bh=cOrUWtTrBL3zDaJGPAOPIuPhvPakueyffj9KJ2tlTT4=; b=3FM4yetgNOqUDMt+azWztrdoEDGsbmw7lGDvdJ/wWWX9GlvDgQmQDXl7ypnuPj3sUh smrnxWdpMJfomeaC8TxL4Bhdc/mKC0gOaRMJZf2qbv3ARSipekHtPL6gzULEp7RHO3A6 VeSd0+1slZpwywGpPPADvXO6NnG651OEfrBLv6/MysZ6OOuQwYDX/yYtqCAczSxaDbgo 09vN1AXCFhVs3PnACKnkjmymHBlSbT5l5RhGtNomaGyL7+2Z/nKkvUfmNqaWoT5QChr6 L4IXxN/s0ASHpLUsv0VxnPliLmfMslixhy8Q8oufsRHB4aWC3Eb1eeie0ZvZ7aFlb/N6 Qgxw== X-Gm-Message-State: ACgBeo1j6LVQJYeniVCyEC1B3RPEUFPCMR9AWBDr/aWlV3Mv21jwlqYQ F+/+Aw6YjaXPc1nsaG+Db28JgJtlntkek8v116k= X-Received: by 2002:a17:90a:bc9:b0:1fb:c5bf:e9d with SMTP id x9-20020a17090a0bc900b001fbc5bf0e9dmr5769637pjd.21.1662139821211; Fri, 02 Sep 2022 10:30:21 -0700 (PDT) MIME-Version: 1.0 References: <20220901222707.477402-1-shy828301@gmail.com> In-Reply-To: From: Yang Shi Date: Fri, 2 Sep 2022 10:30:08 -0700 Message-ID: Subject: Re: [PATCH] mm: gup: fix the fast GUP race against THP collapse To: Peter Xu Cc: david@redhat.com, kirill.shutemov@linux.intel.com, jhubbard@nvidia.com, jgg@nvidia.com, hughd@google.com, akpm@linux-foundation.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org Content-Type: text/plain; charset="UTF-8" X-Spam-Status: No, score=-1.9 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,FREEMAIL_ENVFROM_END_DIGIT, FREEMAIL_FROM,RCVD_IN_DNSWL_NONE,SPF_HELO_NONE,SPF_PASS, T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Sep 2, 2022 at 9:00 AM Peter Xu wrote: > > On Thu, Sep 01, 2022 at 04:50:45PM -0700, Yang Shi wrote: > > On Thu, Sep 1, 2022 at 4:26 PM Peter Xu wrote: > > > > > > Hi, Yang, > > > > > > On Thu, Sep 01, 2022 at 03:27:07PM -0700, Yang Shi wrote: > > > > Since general RCU GUP fast was introduced in commit 2667f50e8b81 ("mm: > > > > introduce a general RCU get_user_pages_fast()"), a TLB flush is no longer > > > > sufficient to handle concurrent GUP-fast in all cases, it only handles > > > > traditional IPI-based GUP-fast correctly. > > > > > > If TLB flush (or, IPI broadcasts) used to work to protect against gup-fast, > > > I'm kind of confused why it's not sufficient even if with RCU gup? Isn't > > > that'll keep working as long as interrupt disabled (which current fast-gup > > > will still do)? > > > > Actually the wording was copied from David's commit log for his > > PageAnonExclusive fix. My understanding is the IPI broadcast still > > works, but it may not be supported by all architectures and not > > preferred anymore. So we should avoid depending on IPI broadcast IIUC. > > > > > > > > IIUC the issue is you suspect not all archs correctly implemented > > > pmdp_collapse_flush(), or am I wrong? > > > > This is a possible fix, please see below for details. > > > > > > > > > On architectures that send > > > > an IPI broadcast on TLB flush, it works as expected. But on the > > > > architectures that do not use IPI to broadcast TLB flush, it may have > > > > the below race: > > > > > > > > CPU A CPU B > > > > THP collapse fast GUP > > > > gup_pmd_range() <-- see valid pmd > > > > gup_pte_range() <-- work on pte > > > > pmdp_collapse_flush() <-- clear pmd and flush > > > > __collapse_huge_page_isolate() > > > > check page pinned <-- before GUP bump refcount > > > > pin the page > > > > check PTE <-- no change > > > > __collapse_huge_page_copy() > > > > copy data to huge page > > > > ptep_clear() > > > > install huge pmd for the huge page > > > > return the stale page > > > > discard the stale page > > > > > > > > The race could be fixed by checking whether PMD is changed or not after > > > > taking the page pin in fast GUP, just like what it does for PTE. If the > > > > PMD is changed it means there may be parallel THP collapse, so GUP > > > > should back off. > > > > > > Could the race also be fixed by impl pmdp_collapse_flush() correctly for > > > the archs that are missing? Do you know which arch(s) is broken with it? > > > > Yes, and this was suggested by me in the first place, but per the > > suggestion from John and David, this is not the preferred way. I think > > it is because: > > > > Firstly, using IPI to serialize against fast GUP is not recommended > > anymore since fast GUP does check PTE then back off so we should avoid > > it. > > Secondly, if checking PMD then backing off could solve the problem, > > why do we still need broadcast IPI? It doesn't sound performant. > > > > > > > > It's just not clear to me whether this patch is an optimization or a fix, > > > if it's a fix whether the IPI broadcast in ppc pmdp_collapse_flush() would > > > still be needed. > > > > It is a fix and the fix will make IPI broadcast not useful anymore. > > How about another patch to remove the ppc impl too? Then it can be a two > patches series. > > So that ppc developers can be copied and maybe it helps to have the ppc > people looking at current approach too. I was thinking about it before sending this patch, but I thought it may be better to wait for ppc developers to confirm some questions. I think Aneesh was copied in another thread. > > Then the last piece of it is the s390 pmdp_collapse_flush(). I'm wondering > whether generic pmdp_collapse_flush() would be good enough, since the only > addition comparing to the s390 one will be flush_tlb_range() (which is a > further __tlb_flush_mm_lazy). David may have some thoughts. > > The patch itself looks good to me, one trivial nit below. > > > > > > > > > Thanks, > > > > > > > > > > > Also update the stale comment about serializing against fast GUP in > > > > khugepaged. > > > > > > > > Fixes: 2667f50e8b81 ("mm: introduce a general RCU get_user_pages_fast()") > > > > Signed-off-by: Yang Shi > > > > --- > > > > mm/gup.c | 30 ++++++++++++++++++++++++------ > > > > mm/khugepaged.c | 10 ++++++---- > > > > 2 files changed, 30 insertions(+), 10 deletions(-) > > > > > > > > diff --git a/mm/gup.c b/mm/gup.c > > > > index f3fc1f08d90c..4365b2811269 100644 > > > > --- a/mm/gup.c > > > > +++ b/mm/gup.c > > > > @@ -2380,8 +2380,9 @@ static void __maybe_unused undo_dev_pagemap(int *nr, int nr_start, > > > > } > > > > > > > > #ifdef CONFIG_ARCH_HAS_PTE_SPECIAL > > > > -static int gup_pte_range(pmd_t pmd, unsigned long addr, unsigned long end, > > > > - unsigned int flags, struct page **pages, int *nr) > > > > +static int gup_pte_range(pmd_t pmd, pmd_t *pmdp, unsigned long addr, > > > > + unsigned long end, unsigned int flags, > > > > + struct page **pages, int *nr) > > > > { > > > > struct dev_pagemap *pgmap = NULL; > > > > int nr_start = *nr, ret = 0; > > > > @@ -2423,7 +2424,23 @@ static int gup_pte_range(pmd_t pmd, unsigned long addr, unsigned long end, > > > > goto pte_unmap; > > > > } > > > > > > > > - if (unlikely(pte_val(pte) != pte_val(*ptep))) { > > > > + /* > > > > + * THP collapse conceptually does: > > > > + * 1. Clear and flush PMD > > > > + * 2. Check the base page refcount > > > > + * 3. Copy data to huge page > > > > + * 4. Clear PTE > > > > + * 5. Discard the base page > > > > + * > > > > + * So fast GUP may race with THP collapse then pin and > > > > + * return an old page since TLB flush is no longer sufficient > > > > + * to serialize against fast GUP. > > > > + * > > > > + * Check PMD, if it is changed just back off since it > > > > + * means there may be parallel THP collapse. > > Would you mind rewording this comment a bit? I feel it a bit weird to > suddenly mention about thp collapse especially its details. > > Maybe some statement on the whole history of why check pte, and in what > case pmd check is needed (where the thp collapse example can be moved to, > imho)? > > One of my attempt for reference.. > > /* > * Fast-gup relies on pte change detection to avoid > * concurrent pgtable operations. > * > * To pin the page, fast-gup needs to do below in order: > * (1) pin the page (by prefetching pte), then (2) check > * pte not changed. > * > * For the rest of pgtable operations where pgtable updates > * can be racy with fast-gup, we need to do (1) clear pte, > * then (2) check whether page is pinned. > * > * Above will work for all pte-level operations, including > * thp split. > * > * For thp collapse, it's a bit more complicated because > * with RCU pgtable free fast-gup can be walking a pgtable > * page that is being freed (so pte is still valid but pmd > * can be cleared already). To avoid race in such > * condition, we need to also check pmd here to make sure > * pmd doesn't change (corresponds to pmdp_collapse_flush() > * in the thp collide code path). > */ > > If you agree with the comment change, feel free to add: Thanks a lot, it looks good to me, will update the patch to incorporate it. > > Acked-by: Peter Xu > > Thanks, > > -- > Peter Xu >