Received: by 2002:ab2:3319:0:b0:1ef:7a0f:c32d with SMTP id i25csp548260lqc; Fri, 8 Mar 2024 05:32:53 -0800 (PST) X-Forwarded-Encrypted: i=3; AJvYcCVNod8yOq4y954wXOJHLF+492oyItNpGwP9DKXjQ3c5wjfytMgcp9pJKX8JDALZ+ht0mwZMlvKRpT7xNHtEPgLvMLE/IdMjKCraYyOraQ== X-Google-Smtp-Source: AGHT+IHsjFbgY+XETOyywS6H5b/cYcaBz6el6b2fBy0szyYITMzWUgRU79A2tg578Ty7hO4bX9aA X-Received: by 2002:a05:622a:349:b0:42e:8a12:f8c6 with SMTP id r9-20020a05622a034900b0042e8a12f8c6mr12420339qtw.19.1709904772911; Fri, 08 Mar 2024 05:32:52 -0800 (PST) ARC-Seal: i=2; a=rsa-sha256; t=1709904772; cv=pass; d=google.com; s=arc-20160816; b=vQcB3oWszp9dgWGqEJ+vLiXDNwu03sBuUIymzWvuscDwgOKY5zDaNv8kmi0sfG9hcJ StBuImNjdxeSf82CyNuf8GWmqYFINmEzAN42gCOU9WEZqchKANMBBe7gWeOe+zqu8xLs XheGy0zaiyfSEqFriL0ForafnXNdS7Ln86SJIhE2n5ou1utR6CtSs4pak9yCsqw9QZ8L Ok7XD6axfPx6U5z6wocuAyTN49K8+Ys0EcBeyHtlP2plvAaEHQRPbDI3R1Pwh9I0pXjj IvRXHs06Y+B82EBJHz8exMMx25dsK5Bx6SYXrE/PGlX6SAvmYJ0czD/F8p4Fcq869OHl wzvw== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:in-reply-to:organization:autocrypt:from :references:cc:to:content-language:subject:user-agent:mime-version :list-unsubscribe:list-subscribe:list-id:precedence:date:message-id :dkim-signature; bh=ro1n/1u64aO9dY5Vhpk3ABFekUnLq7v1DTG5WgNOC9I=; fh=g26/a92C7J/Xb5V+bX0YFzEo3OSk/MmLxYjfA0/LnP4=; b=QI8HJeqjIhrbAitiB0kPo2EjUr4Rl/Lbssoc1Di4fhaUjE19vfPp89V13I4FLFajnQ hjgi21hslkwWz9Ba9O8XkH6qT4mf0A6TPF0YwwAp1RZ5CZX3yzXEEAp2NhXy+ITaZT// fcSq6qpXTZ5rGJoFAIKtEBYij/iJPmDiDc/LVh7YHCRh8e0C5WF9v5skdeIzwHwbug/j ABSIB+5pKkeHb0o0nFQTem8rMrURCZCJYYnlSWaPPKnMtao7qhHEDyxer4WsYbN5ZR56 1LF33cxKNoLjfcnoujw4WAolnR9dBKFQoHKjgVKJ66qRBMKqYJpcK8NOyfaD+onfoZQI PO5Q==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=H6IaXjrF; arc=pass (i=1 spf=pass spfdomain=redhat.com dkim=pass dkdomain=redhat.com dmarc=pass fromdomain=redhat.com); spf=pass (google.com: domain of linux-kernel+bounces-97090-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) smtp.mailfrom="linux-kernel+bounces-97090-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from ny.mirrors.kernel.org (ny.mirrors.kernel.org. [147.75.199.223]) by mx.google.com with ESMTPS id z15-20020a05622a028f00b0042ece029236si2109507qtw.473.2024.03.08.05.32.52 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 08 Mar 2024 05:32:52 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-97090-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) client-ip=147.75.199.223; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=H6IaXjrF; arc=pass (i=1 spf=pass spfdomain=redhat.com dkim=pass dkdomain=redhat.com dmarc=pass fromdomain=redhat.com); spf=pass (google.com: domain of linux-kernel+bounces-97090-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) smtp.mailfrom="linux-kernel+bounces-97090-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ny.mirrors.kernel.org (Postfix) with ESMTPS id 949141C20ED4 for ; Fri, 8 Mar 2024 13:32:52 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id E478D36AF8; Fri, 8 Mar 2024 13:32:49 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=redhat.com header.i=@redhat.com header.b="H6IaXjrF" Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.129.124]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 26ED41EB3E for ; Fri, 8 Mar 2024 13:32:45 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=170.10.129.124 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1709904768; cv=none; b=Ye9M30/dNOpi1VBMscQVZjorCyNLjL70W6SQy6HYYAQhDiiFb0TPJkFUVxrajgEXP+jFGlj2CsUiBYvkkfTzli504ubmvHbequC3CYq88ub5GIBZMxUvH+GUAOnwGU1IwVbiX6eCuNNXczE8/BTmIjrNd9Z0KrEMvf2vkp28eig= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1709904768; c=relaxed/simple; bh=2kcrjdycob94G13NhjtZJYx57m/5RZQa8QoP2t3mKKk=; h=Message-ID:Date:MIME-Version:Subject:To:Cc:References:From: In-Reply-To:Content-Type; b=RmLxcSyOAase+t9jQpOURoVpPrbnTa70aUv98WtT2Eq7KWz1XDYNKk+ponWFx1X6kUk5wKg+6+h31QPNftZ4it8pBMznmsH9GkFToH5GP49ZgqI/zVC+7+bwxArq8Fk99IA31Qz+dyicachSih7ad2ApuJV4WipgL/WMmYK0bcQ= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=redhat.com; spf=pass smtp.mailfrom=redhat.com; dkim=pass (1024-bit key) header.d=redhat.com header.i=@redhat.com header.b=H6IaXjrF; arc=none smtp.client-ip=170.10.129.124 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=redhat.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=redhat.com DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1709904765; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:autocrypt:autocrypt; bh=ro1n/1u64aO9dY5Vhpk3ABFekUnLq7v1DTG5WgNOC9I=; b=H6IaXjrFNSMtqC4ZZGyQF5CFlNvb7z50fO2xB/odTl4vtroHumyUSnUPGZw6pjAVtrFovO a1SuBfsUmhgjLZH3r45DXmjW+Jx1TVLdkWHrbjrR/wQuqO/xQO5dC/Z5wTMg315psCdDTl HwtwKb0kvhvQERUkAT3l85UUbZJhxsY= Received: from mail-lf1-f72.google.com (mail-lf1-f72.google.com [209.85.167.72]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.3, cipher=TLS_AES_256_GCM_SHA384) id us-mta-587-_RdZOJeGMSu8q9sU1jFP1Q-1; Fri, 08 Mar 2024 08:32:41 -0500 X-MC-Unique: _RdZOJeGMSu8q9sU1jFP1Q-1 Received: by mail-lf1-f72.google.com with SMTP id 2adb3069b0e04-5134251ef6eso2413586e87.2 for ; Fri, 08 Mar 2024 05:32:41 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1709904449; x=1710509249; h=content-transfer-encoding:in-reply-to:organization:autocrypt:from :references:cc:to:content-language:subject:user-agent:mime-version :date:message-id:x-gm-message-state:from:to:cc:subject:date :message-id:reply-to; bh=ro1n/1u64aO9dY5Vhpk3ABFekUnLq7v1DTG5WgNOC9I=; b=czqVRtFwd09K16MgNOYqibM5BNWmgRRj2a6P4o2eMxlE+xduVXg877hmGH9gtnuFsb NgWJMzWvi36cYC7WhhgEht8LOguMuToPeebdl192PHqRuv0hTcDTRBFNYA6BwC3agJV4 T/IE22zoU8X5vCVEpkI2DxkSZacAIslApb9STm67GNfJZPPmBpdd4lextUSnCQiXMaQg bhCZDqPIExiVVhEIcwrdG8W+L5mFeVam0BhyBm9zoe2qLDFeETQCsaVZ5CVB786I7yXL NhVkguWeSGdXgJm/5016MBm0fG276loEMZMqvR6dmq5LvObQIUlmsYniRYyH4x7yTH5b Y4Hw== X-Forwarded-Encrypted: i=1; AJvYcCXAJIl+JvGttitAxJKD23dhfMO6XBonfUKDZft8fTA5bGworHnXUDUUWiroA/Y2oAao6P9ISV81x/jEeVpdCf3VxNGMmU4PhvkZJZf+ X-Gm-Message-State: AOJu0YxCqwgE4x4RHaFAo0yntfrn0ZdFxIBweaGDxEzbeJ6VWIcgxWto Ts3e0ZDOAUn7F3oYqH4X/bZarQ9ESVhdxEnoRfHul9+tknLUpEX2ed64WGwh3NZKGsb7+WFgTua rt6x2ekrYMCiHE2MZeOoVt9jpWtZLLJHiwjflKdxeksREFY57azBN61nBEr7FGQ== X-Received: by 2002:a19:8c11:0:b0:513:91c4:aa6 with SMTP id o17-20020a198c11000000b0051391c40aa6mr1492433lfd.63.1709904449028; Fri, 08 Mar 2024 05:27:29 -0800 (PST) X-Received: by 2002:a19:8c11:0:b0:513:91c4:aa6 with SMTP id o17-20020a198c11000000b0051391c40aa6mr1492409lfd.63.1709904448506; Fri, 08 Mar 2024 05:27:28 -0800 (PST) Received: from ?IPV6:2003:cb:c707:3500:5cd0:78d9:ca49:21f3? (p200300cbc70735005cd078d9ca4921f3.dip0.t-ipconnect.de. [2003:cb:c707:3500:5cd0:78d9:ca49:21f3]) by smtp.gmail.com with ESMTPSA id j6-20020a05600c488600b00412f195254csm5714585wmp.36.2024.03.08.05.27.27 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Fri, 08 Mar 2024 05:27:27 -0800 (PST) Message-ID: <2962269e-fb74-4da3-b7b7-b75a5c436e0e@redhat.com> Date: Fri, 8 Mar 2024 14:27:26 +0100 Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v2 1/1] mm/madvise: enhance lazyfreeing with mTHP in madvise_free Content-Language: en-US To: Ryan Roberts , Barry Song <21cnbao@gmail.com> Cc: Lance Yang , Vishal Moola , akpm@linux-foundation.org, zokeefe@google.com, shy828301@gmail.com, mhocko@suse.com, fengwei.yin@intel.com, xiehuan09@gmail.com, wangkefeng.wang@huawei.com, songmuchun@bytedance.com, peterx@redhat.com, minchan@kernel.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org References: <20240307061425.21013-1-ioworker0@gmail.com> <03458c20-5544-411b-9b8d-b4600a9b802f@arm.com> <501c9f77-1459-467a-8619-78e86b46d300@arm.com> <8f84c7d6-982a-4933-a7a7-3f640df64991@redhat.com> <60dc7309-cb38-45e3-b2c0-ff0119202a12@arm.com> From: David Hildenbrand Autocrypt: addr=david@redhat.com; keydata= xsFNBFXLn5EBEAC+zYvAFJxCBY9Tr1xZgcESmxVNI/0ffzE/ZQOiHJl6mGkmA1R7/uUpiCjJ dBrn+lhhOYjjNefFQou6478faXE6o2AhmebqT4KiQoUQFV4R7y1KMEKoSyy8hQaK1umALTdL QZLQMzNE74ap+GDK0wnacPQFpcG1AE9RMq3aeErY5tujekBS32jfC/7AnH7I0v1v1TbbK3Gp XNeiN4QroO+5qaSr0ID2sz5jtBLRb15RMre27E1ImpaIv2Jw8NJgW0k/D1RyKCwaTsgRdwuK Kx/Y91XuSBdz0uOyU/S8kM1+ag0wvsGlpBVxRR/xw/E8M7TEwuCZQArqqTCmkG6HGcXFT0V9 PXFNNgV5jXMQRwU0O/ztJIQqsE5LsUomE//bLwzj9IVsaQpKDqW6TAPjcdBDPLHvriq7kGjt WhVhdl0qEYB8lkBEU7V2Yb+SYhmhpDrti9Fq1EsmhiHSkxJcGREoMK/63r9WLZYI3+4W2rAc UucZa4OT27U5ZISjNg3Ev0rxU5UH2/pT4wJCfxwocmqaRr6UYmrtZmND89X0KigoFD/XSeVv jwBRNjPAubK9/k5NoRrYqztM9W6sJqrH8+UWZ1Idd/DdmogJh0gNC0+N42Za9yBRURfIdKSb B3JfpUqcWwE7vUaYrHG1nw54pLUoPG6sAA7Mehl3nd4pZUALHwARAQABzSREYXZpZCBIaWxk ZW5icmFuZCA8ZGF2aWRAcmVkaGF0LmNvbT7CwZgEEwEIAEICGwMGCwkIBwMCBhUIAgkKCwQW AgMBAh4BAheAAhkBFiEEG9nKrXNcTDpGDfzKTd4Q9wD/g1oFAl8Ox4kFCRKpKXgACgkQTd4Q 9wD/g1oHcA//a6Tj7SBNjFNM1iNhWUo1lxAja0lpSodSnB2g4FCZ4R61SBR4l/psBL73xktp rDHrx4aSpwkRP6Epu6mLvhlfjmkRG4OynJ5HG1gfv7RJJfnUdUM1z5kdS8JBrOhMJS2c/gPf wv1TGRq2XdMPnfY2o0CxRqpcLkx4vBODvJGl2mQyJF/gPepdDfcT8/PY9BJ7FL6Hrq1gnAo4 3Iv9qV0JiT2wmZciNyYQhmA1V6dyTRiQ4YAc31zOo2IM+xisPzeSHgw3ONY/XhYvfZ9r7W1l pNQdc2G+o4Di9NPFHQQhDw3YTRR1opJaTlRDzxYxzU6ZnUUBghxt9cwUWTpfCktkMZiPSDGd KgQBjnweV2jw9UOTxjb4LXqDjmSNkjDdQUOU69jGMUXgihvo4zhYcMX8F5gWdRtMR7DzW/YE BgVcyxNkMIXoY1aYj6npHYiNQesQlqjU6azjbH70/SXKM5tNRplgW8TNprMDuntdvV9wNkFs 9TyM02V5aWxFfI42+aivc4KEw69SE9KXwC7FSf5wXzuTot97N9Phj/Z3+jx443jo2NR34XgF 89cct7wJMjOF7bBefo0fPPZQuIma0Zym71cP61OP/i11ahNye6HGKfxGCOcs5wW9kRQEk8P9 M/k2wt3mt/fCQnuP/mWutNPt95w9wSsUyATLmtNrwccz63XOwU0EVcufkQEQAOfX3n0g0fZz Bgm/S2zF/kxQKCEKP8ID+Vz8sy2GpDvveBq4H2Y34XWsT1zLJdvqPI4af4ZSMxuerWjXbVWb T6d4odQIG0fKx4F8NccDqbgHeZRNajXeeJ3R7gAzvWvQNLz4piHrO/B4tf8svmRBL0ZB5P5A 2uhdwLU3NZuK22zpNn4is87BPWF8HhY0L5fafgDMOqnf4guJVJPYNPhUFzXUbPqOKOkL8ojk CXxkOFHAbjstSK5Ca3fKquY3rdX3DNo+EL7FvAiw1mUtS+5GeYE+RMnDCsVFm/C7kY8c2d0G NWkB9pJM5+mnIoFNxy7YBcldYATVeOHoY4LyaUWNnAvFYWp08dHWfZo9WCiJMuTfgtH9tc75 7QanMVdPt6fDK8UUXIBLQ2TWr/sQKE9xtFuEmoQGlE1l6bGaDnnMLcYu+Asp3kDT0w4zYGsx 5r6XQVRH4+5N6eHZiaeYtFOujp5n+pjBaQK7wUUjDilPQ5QMzIuCL4YjVoylWiBNknvQWBXS lQCWmavOT9sttGQXdPCC5ynI+1ymZC1ORZKANLnRAb0NH/UCzcsstw2TAkFnMEbo9Zu9w7Kv AxBQXWeXhJI9XQssfrf4Gusdqx8nPEpfOqCtbbwJMATbHyqLt7/oz/5deGuwxgb65pWIzufa N7eop7uh+6bezi+rugUI+w6DABEBAAHCwXwEGAEIACYCGwwWIQQb2cqtc1xMOkYN/MpN3hD3 AP+DWgUCXw7HsgUJEqkpoQAKCRBN3hD3AP+DWrrpD/4qS3dyVRxDcDHIlmguXjC1Q5tZTwNB boaBTPHSy/Nksu0eY7x6HfQJ3xajVH32Ms6t1trDQmPx2iP5+7iDsb7OKAb5eOS8h+BEBDeq 3ecsQDv0fFJOA9ag5O3LLNk+3x3q7e0uo06XMaY7UHS341ozXUUI7wC7iKfoUTv03iO9El5f XpNMx/YrIMduZ2+nd9Di7o5+KIwlb2mAB9sTNHdMrXesX8eBL6T9b+MZJk+mZuPxKNVfEQMQ a5SxUEADIPQTPNvBewdeI80yeOCrN+Zzwy/Mrx9EPeu59Y5vSJOx/z6OUImD/GhX7Xvkt3kq Er5KTrJz3++B6SH9pum9PuoE/k+nntJkNMmQpR4MCBaV/J9gIOPGodDKnjdng+mXliF3Ptu6 3oxc2RCyGzTlxyMwuc2U5Q7KtUNTdDe8T0uE+9b8BLMVQDDfJjqY0VVqSUwImzTDLX9S4g/8 kC4HRcclk8hpyhY2jKGluZO0awwTIMgVEzmTyBphDg/Gx7dZU1Xf8HFuE+UZ5UDHDTnwgv7E th6RC9+WrhDNspZ9fJjKWRbveQgUFCpe1sa77LAw+XFrKmBHXp9ZVIe90RMe2tRL06BGiRZr jPrnvUsUUsjRoRNJjKKA/REq+sAnhkNPPZ/NNMjaZ5b8Tovi8C0tmxiCHaQYqj7G2rgnT0kt WNyWQQ== Organization: Red Hat In-Reply-To: <60dc7309-cb38-45e3-b2c0-ff0119202a12@arm.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit On 08.03.24 14:05, Ryan Roberts wrote: > On 07/03/2024 18:54, Barry Song wrote: >> On Fri, Mar 8, 2024 at 12:31 AM Ryan Roberts wrote: >>> >>> On 07/03/2024 12:01, Barry Song wrote: >>>> On Thu, Mar 7, 2024 at 7:45 PM David Hildenbrand wrote: >>>>> >>>>> On 07.03.24 12:42, Ryan Roberts wrote: >>>>>> On 07/03/2024 11:31, David Hildenbrand wrote: >>>>>>> On 07.03.24 12:26, Barry Song wrote: >>>>>>>> On Thu, Mar 7, 2024 at 7:13 PM Ryan Roberts wrote: >>>>>>>>> >>>>>>>>> On 07/03/2024 10:54, David Hildenbrand wrote: >>>>>>>>>> On 07.03.24 11:54, David Hildenbrand wrote: >>>>>>>>>>> On 07.03.24 11:50, Ryan Roberts wrote: >>>>>>>>>>>> On 07/03/2024 09:33, Barry Song wrote: >>>>>>>>>>>>> On Thu, Mar 7, 2024 at 10:07 PM Ryan Roberts wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>> On 07/03/2024 08:10, Barry Song wrote: >>>>>>>>>>>>>>> On Thu, Mar 7, 2024 at 9:00 PM Lance Yang wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hey Barry, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks for taking time to review! >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Thu, Mar 7, 2024 at 3:00 PM Barry Song <21cnbao@gmail.com> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Thu, Mar 7, 2024 at 7:15 PM Lance Yang wrote: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> [...] >>>>>>>>>>>>>>>>>> +static inline bool can_mark_large_folio_lazyfree(unsigned long addr, >>>>>>>>>>>>>>>>>> + struct folio *folio, >>>>>>>>>>>>>>>>>> pte_t *start_pte) >>>>>>>>>>>>>>>>>> +{ >>>>>>>>>>>>>>>>>> + int nr_pages = folio_nr_pages(folio); >>>>>>>>>>>>>>>>>> + fpb_t flags = FPB_IGNORE_DIRTY | FPB_IGNORE_SOFT_DIRTY; >>>>>>>>>>>>>>>>>> + >>>>>>>>>>>>>>>>>> + for (int i = 0; i < nr_pages; i++) >>>>>>>>>>>>>>>>>> + if (page_mapcount(folio_page(folio, i)) != 1) >>>>>>>>>>>>>>>>>> + return false; >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> we have moved to folio_estimated_sharers though it is not precise, so >>>>>>>>>>>>>>>>> we don't do >>>>>>>>>>>>>>>>> this check with lots of loops and depending on the subpage's mapcount. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> If we don't check the subpage’s mapcount, and there is a cow folio >>>>>>>>>>>>>>>> associated >>>>>>>>>>>>>>>> with this folio and the cow folio has smaller size than this folio, >>>>>>>>>>>>>>>> should we still >>>>>>>>>>>>>>>> mark this folio as lazyfree? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I agree, this is true. However, we've somehow accepted the fact that >>>>>>>>>>>>>>> folio_likely_mapped_shared >>>>>>>>>>>>>>> can result in false negatives or false positives to balance the >>>>>>>>>>>>>>> overhead. So I really don't know :-) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Maybe David and Vishal can give some comments here. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> BTW, do we need to rebase our work against David's changes[1]? >>>>>>>>>>>>>>>>> [1] >>>>>>>>>>>>>>>>> https://lore.kernel.org/linux-mm/20240227201548.857831-1-david@redhat.com/ >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Yes, we should rebase our work against David’s changes. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> + >>>>>>>>>>>>>>>>>> + return nr_pages == folio_pte_batch(folio, addr, start_pte, >>>>>>>>>>>>>>>>>> + ptep_get(start_pte), nr_pages, >>>>>>>>>>>>>>>>>> flags, NULL); >>>>>>>>>>>>>>>>>> +} >>>>>>>>>>>>>>>>>> + >>>>>>>>>>>>>>>>>> static int madvise_free_pte_range(pmd_t *pmd, unsigned long addr, >>>>>>>>>>>>>>>>>> unsigned long end, struct mm_walk >>>>>>>>>>>>>>>>>> *walk) >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> @@ -676,11 +690,45 @@ static int madvise_free_pte_range(pmd_t *pmd, >>>>>>>>>>>>>>>>>> unsigned long addr, >>>>>>>>>>>>>>>>>> */ >>>>>>>>>>>>>>>>>> if (folio_test_large(folio)) { >>>>>>>>>>>>>>>>>> int err; >>>>>>>>>>>>>>>>>> + unsigned long next_addr, align; >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> - if (folio_estimated_sharers(folio) != 1) >>>>>>>>>>>>>>>>>> - break; >>>>>>>>>>>>>>>>>> - if (!folio_trylock(folio)) >>>>>>>>>>>>>>>>>> - break; >>>>>>>>>>>>>>>>>> + if (folio_estimated_sharers(folio) != 1 || >>>>>>>>>>>>>>>>>> + !folio_trylock(folio)) >>>>>>>>>>>>>>>>>> + goto skip_large_folio; >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I don't think we can skip all the PTEs for nr_pages, as some of them >>>>>>>>>>>>>>>>> might be >>>>>>>>>>>>>>>>> pointing to other folios. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> for example, for a large folio with 16PTEs, you do MADV_DONTNEED(15-16), >>>>>>>>>>>>>>>>> and write the memory of PTE15 and PTE16, you get page faults, thus PTE15 >>>>>>>>>>>>>>>>> and PTE16 will point to two different small folios. We can only skip >>>>>>>>>>>>>>>>> when we >>>>>>>>>>>>>>>>> are sure nr_pages == folio_pte_batch() is sure. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Agreed. Thanks for pointing that out. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> + >>>>>>>>>>>>>>>>>> + align = folio_nr_pages(folio) * PAGE_SIZE; >>>>>>>>>>>>>>>>>> + next_addr = ALIGN_DOWN(addr + align, align); >>>>>>>>>>>>>>>>>> + >>>>>>>>>>>>>>>>>> + /* >>>>>>>>>>>>>>>>>> + * If we mark only the subpages as lazyfree, or >>>>>>>>>>>>>>>>>> + * cannot mark the entire large folio as >>>>>>>>>>>>>>>>>> lazyfree, >>>>>>>>>>>>>>>>>> + * then just split it. >>>>>>>>>>>>>>>>>> + */ >>>>>>>>>>>>>>>>>> + if (next_addr > end || next_addr - addr != >>>>>>>>>>>>>>>>>> align || >>>>>>>>>>>>>>>>>> + !can_mark_large_folio_lazyfree(addr, folio, >>>>>>>>>>>>>>>>>> pte)) >>>>>>>>>>>>>>>>>> + goto split_large_folio; >>>>>>>>>>>>>>>>>> + >>>>>>>>>>>>>>>>>> + /* >>>>>>>>>>>>>>>>>> + * Avoid unnecessary folio splitting if the >>>>>>>>>>>>>>>>>> large >>>>>>>>>>>>>>>>>> + * folio is entirely within the given range. >>>>>>>>>>>>>>>>>> + */ >>>>>>>>>>>>>>>>>> + folio_clear_dirty(folio); >>>>>>>>>>>>>>>>>> + folio_unlock(folio); >>>>>>>>>>>>>>>>>> + for (; addr != next_addr; pte++, addr += >>>>>>>>>>>>>>>>>> PAGE_SIZE) { >>>>>>>>>>>>>>>>>> + ptent = ptep_get(pte); >>>>>>>>>>>>>>>>>> + if (pte_young(ptent) || >>>>>>>>>>>>>>>>>> pte_dirty(ptent)) { >>>>>>>>>>>>>>>>>> + ptent = >>>>>>>>>>>>>>>>>> ptep_get_and_clear_full( >>>>>>>>>>>>>>>>>> + mm, addr, pte, >>>>>>>>>>>>>>>>>> tlb->fullmm); >>>>>>>>>>>>>>>>>> + ptent = pte_mkold(ptent); >>>>>>>>>>>>>>>>>> + ptent = pte_mkclean(ptent); >>>>>>>>>>>>>>>>>> + set_pte_at(mm, addr, pte, >>>>>>>>>>>>>>>>>> ptent); >>>>>>>>>>>>>>>>>> + tlb_remove_tlb_entry(tlb, pte, >>>>>>>>>>>>>>>>>> addr); >>>>>>>>>>>>>>>>>> + } >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Can we do this in batches? for a CONT-PTE mapped large folio, you are >>>>>>>>>>>>>>>>> unfolding >>>>>>>>>>>>>>>>> and folding again. It seems quite expensive. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I'm not convinced we should be doing this in batches. We want the initial >>>>>>>>>>>>>> folio_pte_batch() to be as loose as possible regarding permissions so >>>>>>>>>>>>>> that we >>>>>>>>>>>>>> reduce our chances of splitting folios to the min. (e.g. ignore SW bits >>>>>>>>>>>>>> like >>>>>>>>>>>>>> soft dirty, etc). I think it might be possible that some PTEs are RO and >>>>>>>>>>>>>> other >>>>>>>>>>>>>> RW too (e.g. due to cow - although with the current cow impl, probably not. >>>>>>>>>>>>>> But >>>>>>>>>>>>>> its fragile to assume that). Anyway, if we do an initial batch that ignores >>>>>>>>>>>>>> all >>>>>>>>>>>>> >>>>>>>>>>>>> You are correct. I believe this scenario could indeed occur. For instance, >>>>>>>>>>>>> if process A forks process B and then unmaps itself, leaving B as the >>>>>>>>>>>>> sole process owning the large folio. The current wp_page_reuse() function >>>>>>>>>>>>> will reuse PTE one by one while the specific subpage is written. >>>>>>>>>>>> >>>>>>>>>>>> Hmm - I thought it would only reuse if the total mapcount for the folio >>>>>>>>>>>> was 1. >>>>>>>>>>>> And since it is a large folio with each page mapped once in proc B, I thought >>>>>>>>>>>> every subpage write would cause a copy except the last one? I haven't >>>>>>>>>>>> looked at >>>>>>>>>>>> the code for a while. But I had it in my head that this is an area we need to >>>>>>>>>>>> improve for mTHP. >>>>>>>> >>>>>>>> So sad I am wrong again ???? >>>>>>>> >>>>>>>>>>> >>>>>>>>>>> wp_page_reuse() will currently reuse a PTE part of a large folio only if >>>>>>>>>>> a single PTE remains mapped (refcount == 0). >>>>>>>>>> >>>>>>>>>> ^ == 1 >>>>>>>> >>>>>>>> seems this needs improvement. it is a waste the last subpage can >>>>>>> >>>>>>> My take that is WIP: >>>>>>> >>>>>>> https://lore.kernel.org/all/20231124132626.235350-1-david@redhat.com/T/#u >>>>>>> >>>>>>>> reuse the whole large folio. i was doing it in a quite different way, >>>>>>>> if the large folio had only one subpage left, i would do copy and >>>>>>>> released the large folio[1]. and if i could reuse the whole large folio >>>>>>>> with CONT-PTE, i would reuse the whole large folio[2]. in mainline, >>>>>>>> we don't have this cont-pte luxury exposed to mm, so i guess we can >>>>>>>> not do [2] easily, but [1] seems to be an optimization. >>>>>>> >>>>>>> Yeah, I had essentially the same idea: just free up the large folio if most of >>>>>>> the stuff is unmapped. But that's rather a corner-case optimization, so I did >>>>>>> not proceed with that. >>>>>>> >>>>>> >>>>>> I'm not sure it's a corner case, really? - process forks, then both parent and >>>>>> child and write to all pages in what was previously a fully & contiguously >>>>>> mapped large folio? >>>>> >>>>> Well, with 2 MiB my assumption was that while it can happen, it's rather >>>>> rare. With smaller THP it might get more likely, agreed. >>>>> >>>>>> >>>>>> Reggardless, why is it an optimization to do the copy for the last subpage and >>>>>> syncrhonously free the large folio? It's already partially mapped so is on the >>>>>> deferred split list and can be split if memory is tight. >>>> >>>> we don't want reclamation overhead later. and we want memories immediately >>>> available to others. >>> >>> But by that logic, you also don't want to leave the large folio partially mapped >>> all the way until the last subpage is CoWed. Surely you would want to reclaim it >>> when you reach partial map status? >> >> To some extent, I agree. But then we will have two many copies. The last >> subpage is small, and a safe place to copy instead. >> >> We actually had to tune userspace to decrease partial map as too much >> partial map both unfolded CONT-PTE and wasted too much memory. if a >> vma had too much partial map, we disabled mTHP on this VMA. > > I actually had a whacky idea around introducing selectable page size ABI > per-process that might help here. I know Android is doing work to make the > system 16K page compatible. You could run most of the system processes with 16K > ABI on top of 4K kernel. Then those processes don't even have the ability to > madvise/munmap/mprotect/mremap anything less than 16K alignment so that acts as > an anti-fragmentation mechanism while allowing non-16K capable processes to run > side-by-side. Just a passing thought... It sounds interesting, but and also like a lot of work. -- Cheers, David / dhildenb