Received: by 2002:a05:6a10:413:0:0:0:0 with SMTP id 19csp806396pxp; Fri, 11 Mar 2022 15:39:30 -0800 (PST) X-Google-Smtp-Source: ABdhPJwkJUBz/XVWDERntsy6NA9rF+6I5sS8rup1s8TaLZGyJPbOSBMrOqNUeuZKkVxUHe2kOr9s X-Received: by 2002:a17:90b:4a07:b0:1bf:b159:185 with SMTP id kk7-20020a17090b4a0700b001bfb1590185mr13370968pjb.65.1647041970191; Fri, 11 Mar 2022 15:39:30 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1647041970; cv=none; d=google.com; s=arc-20160816; b=0oqjS2fwKb2JdjEf7l8VQBCEZjOxnCnMZOM3W89bBkOndABcoNt/u70H81VbqlWKu2 HFwOT27siskrgp2E0R/DOCGLZdosifn1MPbzCwxbkWM4+yZHpIpNmeW1Bh5qF8MZuQMX 7DTDYcEsLs6zL2vNt+F6B9c4JqP4eUsRtUAhJQ41BLjh4kIdI3SauSpEJOlVhZp0wz9A aMyneaYwED/UZGO9aJXb9gIIUFYDsDTedm14GHUA92rUK0lRiGfRGCHtokDYN09OlqCL Ij7Zq1qBpDGP/Dr+XG0wPSHwR8krhhUwzR3Pyk6PgAXhiLrzSZrPbePisj0d2H+6FETM eGNQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=SzKX99Cocy/lwgU3tnCH1Vt8ZkKR8jjaQaAMFI9xwxc=; b=N7/iOY00+p92E0DNEtIGOt7UOGa6JQvJIhe2+b1M80SgI9/TFlO0NI2I4lXrmnNq4z XVXA810UkRJYUGRP+ypUusRViOZSFEoL/aqeBb1w+S35d9ysJJNPag7p9usxji7SPkgY L/h7nYcrEO/ajFeEwdMELnpVgNm716WWcj/TXLF8kPC1R3trdHKwtPwi57Y+M7vyKX8f orsxJ/GBytiMgu4SjlIugDTFGNWuaDbC6lAGrtlaWJbSOMJcRt3DqjGNO3EVv5Cg43T4 4axZoI0Cm7A9ZU0Eq3yhVqCokJrnU7XaM5ERWwfJhGq7Hz6dfyOO09ZSCycbpkprrgHj hZ1A== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@infradead.org header.s=casper.20170209 header.b=nRWaKR4q; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from lindbergh.monkeyblade.net (lindbergh.monkeyblade.net. [2620:137:e000::1:18]) by mx.google.com with ESMTPS id u1-20020a17090341c100b00151ee000674si9773744ple.450.2022.03.11.15.39.26 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 11 Mar 2022 15:39:30 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:18 as permitted sender) client-ip=2620:137:e000::1:18; Authentication-Results: mx.google.com; dkim=pass header.i=@infradead.org header.s=casper.20170209 header.b=nRWaKR4q; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by lindbergh.monkeyblade.net (Postfix) with ESMTP id 45AA31945EE; Fri, 11 Mar 2022 14:54:45 -0800 (PST) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229809AbiCKWzG (ORCPT + 99 others); Fri, 11 Mar 2022 17:55:06 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:34718 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230156AbiCKWyk (ORCPT ); Fri, 11 Mar 2022 17:54:40 -0500 Received: from casper.infradead.org (casper.infradead.org [IPv6:2001:8b0:10b:1236::1]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 7F9442BE122 for ; Fri, 11 Mar 2022 14:28:23 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=casper.20170209; h=In-Reply-To:Content-Type:MIME-Version: References:Message-ID:Subject:Cc:To:From:Date:Sender:Reply-To: Content-Transfer-Encoding:Content-ID:Content-Description; bh=SzKX99Cocy/lwgU3tnCH1Vt8ZkKR8jjaQaAMFI9xwxc=; b=nRWaKR4qwPKfFozcskyUoAynnV jv/gIbaWr0G1z+JE77V8iiPRV4WKN1S/2xymJG5jcRHfRLh0T4A8ZwzHeWyOk2n35+GEh71diCwg5 izPd/42UPCFqErRfIqqh4r2DBmO+RZFwI3Aff8KbvJASbG2Nt/5RLYEkCtHUubuTyFL6iqDpqQWxa m/XbgA9zah0cwIIsuevRuaOp0n8DvPWmwaWpdPpyTfpfx6x/Rz1LahjBDgObRj+5KZdKS1EZYABGF peDL4T18VlLbtMm2XFEcPyDDKr9z5MwBNw2Kiu7AH1eX3vDM2iLu/s0aomnYK3GXY4qDeFA/YCtiY VjyL3oUA==; Received: from willy by casper.infradead.org with local (Exim 4.94.2 #2 (Red Hat Linux)) id 1nSmOp-001uVY-CC; Fri, 11 Mar 2022 21:02:11 +0000 Date: Fri, 11 Mar 2022 21:02:11 +0000 From: Matthew Wilcox To: David Hildenbrand Cc: linux-kernel@vger.kernel.org, Andrew Morton , Hugh Dickins , Linus Torvalds , David Rientjes , Shakeel Butt , John Hubbard , Jason Gunthorpe , Mike Kravetz , Mike Rapoport , Yang Shi , "Kirill A . Shutemov" , Vlastimil Babka , Jann Horn , Michal Hocko , Nadav Amit , Rik van Riel , Roman Gushchin , Andrea Arcangeli , Peter Xu , Donald Dutile , Christoph Hellwig , Oleg Nesterov , Jan Kara , Liang Zhang , Pedro Gomes , Oded Gabbay , linux-mm@kvack.org Subject: Re: [PATCH v1 10/15] mm/page-flags: reuse PG_slab as PG_anon_exclusive for PageAnon() pages Message-ID: References: <20220308141437.144919-1-david@redhat.com> <20220308141437.144919-11-david@redhat.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: X-Spam-Status: No, score=-2.0 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,RDNS_NONE,SPF_HELO_NONE,T_SCC_BODY_TEXT_LINE autolearn=no autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Mar 11, 2022 at 07:46:39PM +0100, David Hildenbrand wrote: > I'm currently testing with the following. My tests so far with a swapfile on > all different kinds of weird filesystems (excluding networking fs, though) > revealed no surprises so far: I like this a lot better than reusing PG_swap. Thanks! I'm somewhat reluctant to introduce a new flag that can be set on tail pages. Do we lose much if it's always set only on the head page? > +++ b/include/linux/page-flags.h > @@ -142,6 +142,60 @@ enum pageflags { > > PG_readahead = PG_reclaim, > > + /* > + * Depending on the way an anonymous folio can be mapped into a page > + * table (e.g., single PMD/PUD/CONT of the head page vs. PTE-mapped > + * THP), PG_anon_exclusive may be set only for the head page or for > + * subpages of an anonymous folio. > + * > + * PG_anon_exclusive is *usually* only expressive in combination with a > + * page table entry. Depending on the page table entry type it might > + * store the following information: > + * > + * Is what's mapped via this page table entry exclusive to the > + * single process and can be mapped writable without further > + * checks? If not, it might be shared and we might have to COW. > + * > + * For now, we only expect PTE-mapped THPs to make use of > + * PG_anon_exclusive in subpages. For other anonymous compound > + * folios (i.e., hugetlb), only the head page is logically mapped and > + * holds this information. > + * > + * For example, an exclusive, PMD-mapped THP only has PG_anon_exclusive > + * set on the head page. When replacing the PMD by a page table full > + * of PTEs, PG_anon_exclusive, if set on the head page, will be set on > + * all tail pages accordingly. Note that converting from a PTE-mapping > + * to a PMD mapping using the same compound page is currently not > + * possible and consequently doesn't require care. > + * > + * If GUP wants to take a reliable pin (FOLL_PIN) on an anonymous page, > + * it should only pin if the relevant PG_anon_bit is set. In that case, > + * the pin will be fully reliable and stay consistent with the pages > + * mapped into the page table, as the bit cannot get cleared (e.g., by > + * fork(), KSM) while the page is pinned. For anonymous pages that > + * are mapped R/W, PG_anon_exclusive can be assumed to always be set > + * because such pages cannot possibly be shared. > + * > + * The page table lock protecting the page table entry is the primary > + * synchronization mechanism for PG_anon_exclusive; GUP-fast that does > + * not take the PT lock needs special care when trying to clear the > + * flag. > + * > + * Page table entry types and PG_anon_exclusive: > + * * Present: PG_anon_exclusive applies. > + * * Swap: the information is lost. PG_anon_exclusive was cleared. > + * * Migration: the entry holds this information instead. > + * PG_anon_exclusive was cleared. > + * * Device private: PG_anon_exclusive applies. > + * * Device exclusive: PG_anon_exclusive applies. > + * * HW Poison: PG_anon_exclusive is stale and not changed. > + * > + * If the page may be pinned (FOLL_PIN), clearing PG_anon_exclusive is > + * not allowed and the flag will stick around until the page is freed > + * and folio->mapping is cleared. > + */ ... I also don't think this is the right place for this comment. Not sure where it should go. > +static __always_inline void SetPageAnonExclusive(struct page *page) > +{ > + VM_BUG_ON_PGFLAGS(!PageAnon(page) || PageKsm(page), page); hm. seems to me like we should have a PageAnonNotKsm which just does return ((unsigned long)page->mapping & PAGE_MAPPING_FLAGS) == PAGE_MAPPING_ANON; because that's "a bit" inefficient. OK, that's just a VM_BUG_ON, but we have other users in real code: mm/migrate.c: if (PageAnon(page) && !PageKsm(page)) mm/page_idle.c: need_lock = !PageAnon(page) || PageKsm(page); mm/rmap.c: if (!is_locked && (!PageAnon(page) || PageKsm(page))) {