Received: by 2002:ab2:3350:0:b0:1f4:6588:b3a7 with SMTP id o16csp1295423lqe; Mon, 8 Apr 2024 05:11:13 -0700 (PDT) X-Forwarded-Encrypted: i=3; AJvYcCVxhjtYzgYxr2Y++21KtOv9nsOGrpp8T3QTwzsU7tBy+AKhAJBhoboO2ujXsvP1a5QINTw0TPNy4M11xFjul+/FvJEz/6G5gPc+BiVB3w== X-Google-Smtp-Source: AGHT+IHXlbIGwwD7fdXu5ANjeg84mdQIRMmlzLb7xpA13a733XTjAJyPlJ4fbD7CH5IAapcXSQjv X-Received: by 2002:a17:902:f68c:b0:1e2:23b9:ead3 with SMTP id l12-20020a170902f68c00b001e223b9ead3mr8393632plg.24.1712578273184; Mon, 08 Apr 2024 05:11:13 -0700 (PDT) ARC-Seal: i=2; a=rsa-sha256; t=1712578273; cv=pass; d=google.com; s=arc-20160816; b=DhX7WzoWnCEMDulbtZHYNm9K59YD6hyIzIFRXtldftKIuCS94v7UtEHdRUDZvInTIp wvQihqh652eh6WEV0ZhUgwyKmh0+SQm3QGJPxVUVRJqLQHBA56eiWIsu0X7fdL0CCCZ/ YNpoAS8ffNNVCsUAS3eDaEUGYTn5Y0L0Upbjv/3R68oUgG7yER6HK6IvPOMCof5IRGv+ E5U/3Xm1WorRiihYwkauUeWUZ6i0EwIbV65y/pLJcnI2AfJcX3VkapIiTZslGsP/TrX5 gJaN1WNtwP+L0fhH240lBogZ+1y67SsiYijELh7FFFK9lhr3k1+mn6qEXT00/jvXbKS7 /fxg== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:in-reply-to:from:references:cc:to :content-language:subject:user-agent:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:date:message-id; bh=DUjsFzxFXfC2KgiKHcWQcBklKwFuFt3vSaMX+WKKn2I=; fh=X8Ye0zNbQ3DGHGIuZ73PsDHIYx4AqcRuTr5uMr01yYM=; b=iuDlLepDDIQPbVrd3LdIbt+squzegJguS0ubSM4lzmQl2au+JFyX2MQ6w86npaQ09w soN/IxmYRoS4O4iu0lnPzuTtfJh6zvLaRDTr+2hwdOX05wkmyOnnp7FO9ZL23E2dzfz4 KbzMxWMKOVoclJiSh03ew6uTFsX5KXlbaCzdM4cqZQAJykYgct4DGjom5m04OVHFeRZq WrSEW1N6UP/jEjDxyit+NGLZioOeZe8p/CXqpE4wUJw78WglQzV1YIgYrjpoKUgD4ITq /+0VqT/BQSgoZsiFQ0r35xpmZHwvfq9LOKhvxwp9fs+uPi/LdPWRwQGNJFWILFa3LN6a +nLQ==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; arc=pass (i=1 spf=pass spfdomain=arm.com dmarc=pass fromdomain=arm.com); spf=pass (google.com: domain of linux-kernel+bounces-135305-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:40f1:3f00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-135305-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Return-Path: Received: from sy.mirrors.kernel.org (sy.mirrors.kernel.org. [2604:1380:40f1:3f00::1]) by mx.google.com with ESMTPS id q5-20020a170902c9c500b001e463053710si523022pld.45.2024.04.08.05.11.12 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 08 Apr 2024 05:11:13 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel+bounces-135305-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:40f1:3f00::1 as permitted sender) client-ip=2604:1380:40f1:3f00::1; Authentication-Results: mx.google.com; arc=pass (i=1 spf=pass spfdomain=arm.com dmarc=pass fromdomain=arm.com); spf=pass (google.com: domain of linux-kernel+bounces-135305-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:40f1:3f00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-135305-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by sy.mirrors.kernel.org (Postfix) with ESMTPS id 9292AB23EC8 for ; Mon, 8 Apr 2024 12:08:25 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 9F8EA6CDA8; Mon, 8 Apr 2024 12:07:49 +0000 (UTC) Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 4ACC86A333 for ; Mon, 8 Apr 2024 12:07:46 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=217.140.110.172 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1712578069; cv=none; b=UHSm81FcImiGx6xtajl/c8WKo+k0aNcnMOeU9e5DzOISG+1k0WS0DOlG+K2P4baMg12ikfi77oFNUh8pmcGrzUUMqu9HMt9ulGAzPnG7i9IZzmfvd165dx/xDjqm+enSJW4eFGofFG1wqCCajzWn+nGgpxRg7Fzx681WA0p9tzA= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1712578069; c=relaxed/simple; bh=b8pWrBbsgVx0W4YG0AiioljQVukONKMHKcEisvgp6Ig=; h=Message-ID:Date:MIME-Version:Subject:To:Cc:References:From: In-Reply-To:Content-Type; b=oQUE+EreTL0v4qmSC39XkxZALYBAGvA0lg4s7aW2wRq26k1MdkikmNlsEVT4ikXnH5MBNGz8imVSeFJ7Om6gM+BV+Koy0VBaL4YrchMKXOUkcXHoZySxBdE6s/8O6Qv44LfIhgzh7LUd9jpuiWV6JwAgVdzi36GaY4R8hFVWlUc= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=arm.com; spf=pass smtp.mailfrom=arm.com; arc=none smtp.client-ip=217.140.110.172 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=arm.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=arm.com Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id ED007DA7; Mon, 8 Apr 2024 05:08:15 -0700 (PDT) Received: from [10.57.73.169] (unknown [10.57.73.169]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id A01823F64C; Mon, 8 Apr 2024 05:07:43 -0700 (PDT) Message-ID: <4110bb1d-65e5-4cf0-91ad-62749975829d@arm.com> Date: Mon, 8 Apr 2024 13:07:42 +0100 Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v6 2/6] mm: swap: free_swap_and_cache_nr() as batched free_swap_and_cache() Content-Language: en-GB To: David Hildenbrand , Andrew Morton , Matthew Wilcox , Huang Ying , Gao Xiang , Yu Zhao , Yang Shi , Michal Hocko , Kefeng Wang , Barry Song <21cnbao@gmail.com>, Chris Li , Lance Yang Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org References: <20240403114032.1162100-1-ryan.roberts@arm.com> <20240403114032.1162100-3-ryan.roberts@arm.com> <051052af-3b56-4290-98d3-fd5a1eb11ce1@redhat.com> From: Ryan Roberts In-Reply-To: <051052af-3b56-4290-98d3-fd5a1eb11ce1@redhat.com> Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit [...] > > [...] > >> + >> +/** >> + * swap_pte_batch - detect a PTE batch for a set of contiguous swap entries >> + * @start_ptep: Page table pointer for the first entry. >> + * @max_nr: The maximum number of table entries to consider. >> + * @entry: Swap entry recovered from the first table entry. >> + * >> + * Detect a batch of contiguous swap entries: consecutive (non-present) PTEs >> + * containing swap entries all with consecutive offsets and targeting the same >> + * swap type. >> + * > > Likely you should document that any swp pte bits are ignored? () Now that I understand what swp pte bits are, I think the simplest thing is to just make this function always consider the pte bits by using pte_same() as you suggest below? I don't think there is ever a case for ignoring the swp pte bits? And then I don't need to do anything special for uffd-wp either (below you suggested not doing batching when the VMA has uffd enabled). Any concerns? > >> + * max_nr must be at least one and must be limited by the caller so scanning >> + * cannot exceed a single page table. >> + * >> + * Return: the number of table entries in the batch. >> + */ >> +static inline int swap_pte_batch(pte_t *start_ptep, int max_nr, >> +                 swp_entry_t entry) >> +{ >> +    const pte_t *end_ptep = start_ptep + max_nr; >> +    unsigned long expected_offset = swp_offset(entry) + 1; >> +    unsigned int expected_type = swp_type(entry); >> +    pte_t *ptep = start_ptep + 1; >> + >> +    VM_WARN_ON(max_nr < 1); >> +    VM_WARN_ON(non_swap_entry(entry)); >> + >> +    while (ptep < end_ptep) { >> +        pte_t pte = ptep_get(ptep); >> + >> +        if (pte_none(pte) || pte_present(pte)) >> +            break; >> + >> +        entry = pte_to_swp_entry(pte); >> + >> +        if (non_swap_entry(entry) || >> +            swp_type(entry) != expected_type || >> +            swp_offset(entry) != expected_offset) >> +            break; >> + >> +        expected_offset++; >> +        ptep++; >> +    } >> + >> +    return ptep - start_ptep; >> +} > > Looks very clean :) > > I was wondering whether we could similarly construct the expected swp PTE and > only check pte_same. > > expected_pte = __swp_entry_to_pte(__swp_entry(expected_type, expected_offset)); So planning to do this. > > ... or have a variant to increase only the swp offset for an existing pte. But > non-trivial due to the arch-dependent format. not this - I agree this will be difficult due to per-arch changes. I'd rather just do the generic version and leave the compiler to do the best it can to simplify and optimize. > > But then, we'd fail on mismatch of other swp pte bits. > > > On swapin, when reusing this function (likely!), we'll might to make sure that > the PTE bits match as well. > > See below regarding uffd-wp. > > >>   #endif /* CONFIG_MMU */ >>     void __acct_reclaim_writeback(pg_data_t *pgdat, struct folio *folio, >> diff --git a/mm/madvise.c b/mm/madvise.c >> index 1f77a51baaac..070bedb4996e 100644 >> --- a/mm/madvise.c >> +++ b/mm/madvise.c >> @@ -628,6 +628,7 @@ static int madvise_free_pte_range(pmd_t *pmd, unsigned >> long addr, >>       struct folio *folio; >>       int nr_swap = 0; >>       unsigned long next; >> +    int nr, max_nr; >>         next = pmd_addr_end(addr, end); >>       if (pmd_trans_huge(*pmd)) >> @@ -640,7 +641,8 @@ static int madvise_free_pte_range(pmd_t *pmd, unsigned >> long addr, >>           return 0; >>       flush_tlb_batched_pending(mm); >>       arch_enter_lazy_mmu_mode(); >> -    for (; addr != end; pte++, addr += PAGE_SIZE) { >> +    for (; addr != end; pte += nr, addr += PAGE_SIZE * nr) { >> +        nr = 1; >>           ptent = ptep_get(pte); >>             if (pte_none(ptent)) >> @@ -655,9 +657,11 @@ static int madvise_free_pte_range(pmd_t *pmd, unsigned >> long addr, >>                 entry = pte_to_swp_entry(ptent); >>               if (!non_swap_entry(entry)) { >> -                nr_swap--; >> -                free_swap_and_cache(entry); >> -                pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); >> +                max_nr = (end - addr) / PAGE_SIZE; >> +                nr = swap_pte_batch(pte, max_nr, entry); >> +                nr_swap -= nr; >> +                free_swap_and_cache_nr(entry, nr); >> +                clear_not_present_full_ptes(mm, addr, pte, nr, tlb->fullmm); >>               } else if (is_hwpoison_entry(entry) || >>                      is_poisoned_swp_entry(entry)) { >>                   pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); >> diff --git a/mm/memory.c b/mm/memory.c >> index 7dc6c3d9fa83..ef2968894718 100644 >> --- a/mm/memory.c >> +++ b/mm/memory.c >> @@ -1637,12 +1637,13 @@ static unsigned long zap_pte_range(struct mmu_gather >> *tlb, >>                   folio_remove_rmap_pte(folio, page, vma); >>               folio_put(folio); >>           } else if (!non_swap_entry(entry)) { >> -            /* Genuine swap entry, hence a private anon page */ >> +            max_nr = (end - addr) / PAGE_SIZE; >> +            nr = swap_pte_batch(pte, max_nr, entry); >> +            /* Genuine swap entries, hence a private anon pages */ >>               if (!should_zap_cows(details)) >>                   continue; >> -            rss[MM_SWAPENTS]--; >> -            if (unlikely(!free_swap_and_cache(entry))) >> -                print_bad_pte(vma, addr, ptent, NULL); >> +            rss[MM_SWAPENTS] -= nr; >> +            free_swap_and_cache_nr(entry, nr); >>           } else if (is_migration_entry(entry)) { >>               folio = pfn_swap_entry_folio(entry); >>               if (!should_zap_folio(details, folio)) >> @@ -1665,8 +1666,8 @@ static unsigned long zap_pte_range(struct mmu_gather *tlb, >>               pr_alert("unrecognized swap entry 0x%lx\n", entry.val); >>               WARN_ON_ONCE(1); >>           } >> -        pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); >> -        zap_install_uffd_wp_if_needed(vma, addr, pte, 1, details, ptent); >> +        clear_not_present_full_ptes(mm, addr, pte, nr, tlb->fullmm); > > For zap_install_uffd_wp_if_needed(), the uffd-wp bit has to match. > > zap_install_uffd_wp_if_needed() will use the uffd-wp information in > ptent->pteval to make a decision whether to place PTE_MARKER_UFFD_WP markers. > > On mixture, you either lose some or place too many markers. > > A simple workaround would be to disable any such batching if the VMA does have > uffd-wp enabled. Rather than this, I'll just consider all the swp pte bits when batching. > >> +        zap_install_uffd_wp_if_needed(vma, addr, pte, nr, details, ptent); >>       } while (pte += nr, addr += PAGE_SIZE * nr, addr != end); [...]