Received: by 2002:a05:6a10:8c0a:0:0:0:0 with SMTP id go10csp1325721pxb; Fri, 26 Feb 2021 08:04:42 -0800 (PST) X-Google-Smtp-Source: ABdhPJzj/fVV/DI1pbsOemTAzIwhffOuztYrQZ1tAAZVPOO414+IIKiGs/wXcMY/M9oxrprO8t+h X-Received: by 2002:a17:906:7687:: with SMTP id o7mr4221257ejm.209.1614355482125; Fri, 26 Feb 2021 08:04:42 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1614355482; cv=none; d=google.com; s=arc-20160816; b=Rj5a3UOlZwvpgUsOVlVKO9qwNNhOD9qwt2wUHrTTeTOOYc2tj5PoC3eWJCdn41bRmG xsNFugt96FYxnwWQB5ppHxTQZw2jqRBhN8x0WwaqA0x8zYhHBn+OiDkPquUA0OxNp4gq 5S9fxAcfhx1y7xbqRA0ENTMDmXxTC7ZBx+k17P3yYuPacH/jwUm/zOJ3SJdYHiBS2pl3 F3o0vygTA7iTmvtiKOAjT77ghtwW0/8VriIR0/rzkk+Xv5WhvDpYPzjDm+bPYijMVLtU c6EA4lVNXAoaDOtPuG99aimZNAqnsge33IrzbUa7IHRq3yh7jHaSII5fBXZM1LqxxaAd DUdw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=LBaKHm661e2/pCw2/WvK+r7G9CV6irEmIaQaUTUrM/I=; b=uMzF0ppnSPZaKQZRlqXbMKi8QJ2mM3xZFL1rjMnkMOhnsaAy3a1lIZ4yHYhAqxWa/m zqu7dMNo7bIwXEbX6xSIdL82Fk+IvxWFHyH61cjELIrmWESakgk0qprWqP2sZkDQZJWr jv4uftn1rGqn+FCAHTHvgdXhmH349Ez9Bvvw2jDUGeDMjXSb8+6JP+aK5mYxPs5DnJEY 3OCSr1sZW8GzYhZH9oJk/7W26mzUNage0ZnPrlxkiZs/QJOXfpLeg7wpfuzDfy9BUiD6 /9wkdX0rLiF6tvSjgioPjT33hKvnvEtLVqtfn28cPNrpiUcgNrinZCHyJfDpKwT/PVN0 ALew== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@infradead.org header.s=casper.20170209 header.b="sW/OjqIp"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id l24si6048128ejk.270.2021.02.26.08.04.18; Fri, 26 Feb 2021 08:04:42 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@infradead.org header.s=casper.20170209 header.b="sW/OjqIp"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230147AbhBZQAj (ORCPT + 99 others); Fri, 26 Feb 2021 11:00:39 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:42776 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229849AbhBZQAW (ORCPT ); Fri, 26 Feb 2021 11:00:22 -0500 Received: from casper.infradead.org (casper.infradead.org [IPv6:2001:8b0:10b:1236::1]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 3B974C06174A; Fri, 26 Feb 2021 07:59:41 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=casper.20170209; h=In-Reply-To:Content-Type:MIME-Version: References:Message-ID:Subject:Cc:To:From:Date:Sender:Reply-To: Content-Transfer-Encoding:Content-ID:Content-Description; bh=LBaKHm661e2/pCw2/WvK+r7G9CV6irEmIaQaUTUrM/I=; b=sW/OjqIpj6duvHj3cvoK44nFoC Bbj/3yA8GY9h9BMJAKOOsa5sW+pBKuHPgCK5xKV5CxRPmDmphYR8ev591rOD/YKAznDD1X97olu6c JqboHG2zspmXFBui0uTbvpSPO3gWFX/hBaAd3ovQ/+mnW7ZYkfHh6Yzw3clipK+bmYjnenJBrNncl EVwfdz5giJvu80Ps9BWVccGfx72XsL1x7uaKseu4ViUxtPcy92u9OLoA8RiBo1yzcbbiZES8i4h+w rkil1V1rVJNY20RGNZhZ2F9vLWiPqyCS7yIrh4N+HToimxqsr/RB254l0GAlWbiSjJ2eTpKg8QpRR xL62pMxg==; Received: from hch by casper.infradead.org with local (Exim 4.94 #2 (Red Hat Linux)) id 1lFfWH-00CD0S-J5; Fri, 26 Feb 2021 15:59:11 +0000 Date: Fri, 26 Feb 2021 15:59:09 +0000 From: Christoph Hellwig To: Alistair Popple Cc: linux-mm@kvack.org, nouveau@lists.freedesktop.org, bskeggs@redhat.com, akpm@linux-foundation.org, linux-doc@vger.kernel.org, linux-kernel@vger.kernel.org, dri-devel@lists.freedesktop.org, jhubbard@nvidia.com, rcampbell@nvidia.com, jglisse@redhat.com, jgg@nvidia.com, hch@infradead.org, daniel@ffwll.ch Subject: Re: [PATCH v3 1/8] mm: Remove special swap entry functions Message-ID: <20210226155909.GA2907711@infradead.org> References: <20210226071832.31547-1-apopple@nvidia.com> <20210226071832.31547-2-apopple@nvidia.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20210226071832.31547-2-apopple@nvidia.com> X-SRS-Rewrite: SMTP reverse-path rewritten from by casper.infradead.org. See http://www.infradead.org/rpr.html Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org > - struct page *page = migration_entry_to_page(entry); > + struct page *page = pfn_to_page(swp_offset(entry)); I wonder if keeping a single special_entry_to_page() helper would still me a useful. But I'm not entirely sure. There are also two more open coded copies of this in the THP migration code. > -#define free_swap_and_cache(e) ({(is_migration_entry(e) || is_device_private_entry(e));}) > -#define swapcache_prepare(e) ({(is_migration_entry(e) || is_device_private_entry(e));}) > +#define free_swap_and_cache(e) is_special_entry(e) > +#define swapcache_prepare(e) is_special_entry(e) Staring at this I'm really, really confused at what this is doing. Looking a little closer these are the !CONFIG_SWAP stubs, but it could probably use a comment or two. > } else if (is_migration_entry(entry)) { > - page = migration_entry_to_page(entry); > + page = pfn_to_page(swp_offset(entry)); > > rss[mm_counter(page)]++; > > @@ -737,7 +737,7 @@ copy_nonpresent_pte(struct mm_struct *dst_mm, struct mm_struct *src_mm, > set_pte_at(src_mm, addr, src_pte, pte); > } > } else if (is_device_private_entry(entry)) { > - page = device_private_entry_to_page(entry); > + page = pfn_to_page(swp_offset(entry)); > > /* > * Update rss count even for unaddressable pages, as > @@ -1274,7 +1274,7 @@ static unsigned long zap_pte_range(struct mmu_gather *tlb, > > entry = pte_to_swp_entry(ptent); > if (is_device_private_entry(entry)) { > - struct page *page = device_private_entry_to_page(entry); > + struct page *page = pfn_to_page(swp_offset(entry)); > > if (unlikely(details && details->check_mapping)) { > /* > @@ -1303,7 +1303,7 @@ static unsigned long zap_pte_range(struct mmu_gather *tlb, > else if (is_migration_entry(entry)) { > struct page *page; > > - page = migration_entry_to_page(entry); > + page = pfn_to_page(swp_offset(entry)); > rss[mm_counter(page)]--; > } > if (unlikely(!free_swap_and_cache(entry))) > @@ -3271,7 +3271,7 @@ vm_fault_t do_swap_page(struct vm_fault *vmf) > migration_entry_wait(vma->vm_mm, vmf->pmd, > vmf->address); > } else if (is_device_private_entry(entry)) { > - vmf->page = device_private_entry_to_page(entry); > + vmf->page = pfn_to_page(swp_offset(entry)); > ret = vmf->page->pgmap->ops->migrate_to_ram(vmf); > } else if (is_hwpoison_entry(entry)) { > ret = VM_FAULT_HWPOISON; > diff --git a/mm/migrate.c b/mm/migrate.c > index 20ca887ea769..72adcc3d8f5b 100644 > --- a/mm/migrate.c > +++ b/mm/migrate.c > @@ -321,7 +321,7 @@ void __migration_entry_wait(struct mm_struct *mm, pte_t *ptep, > if (!is_migration_entry(entry)) > goto out; > > - page = migration_entry_to_page(entry); > + page = pfn_to_page(swp_offset(entry)); > > /* > * Once page cache replacement of page migration started, page_count > @@ -361,7 +361,7 @@ void pmd_migration_entry_wait(struct mm_struct *mm, pmd_t *pmd) > ptl = pmd_lock(mm, pmd); > if (!is_pmd_migration_entry(*pmd)) > goto unlock; > - page = migration_entry_to_page(pmd_to_swp_entry(*pmd)); > + page = pfn_to_page(swp_offset(pmd_to_swp_entry(*pmd))); > if (!get_page_unless_zero(page)) > goto unlock; > spin_unlock(ptl); > @@ -2437,7 +2437,7 @@ static int migrate_vma_collect_pmd(pmd_t *pmdp, > if (!is_device_private_entry(entry)) > goto next; > > - page = device_private_entry_to_page(entry); > + page = pfn_to_page(swp_offset(entry)); > if (!(migrate->flags & > MIGRATE_VMA_SELECT_DEVICE_PRIVATE) || > page->pgmap->owner != migrate->pgmap_owner) > diff --git a/mm/page_vma_mapped.c b/mm/page_vma_mapped.c > index 86e3a3688d59..34230d08556a 100644 > --- a/mm/page_vma_mapped.c > +++ b/mm/page_vma_mapped.c > @@ -96,7 +96,7 @@ static bool check_pte(struct page_vma_mapped_walk *pvmw) > if (!is_migration_entry(entry)) > return false; > > - pfn = migration_entry_to_pfn(entry); > + pfn = swp_offset(entry); > } else if (is_swap_pte(*pvmw->pte)) { > swp_entry_t entry; > > @@ -105,7 +105,7 @@ static bool check_pte(struct page_vma_mapped_walk *pvmw) > if (!is_device_private_entry(entry)) > return false; > > - pfn = device_private_entry_to_pfn(entry); > + pfn = swp_offset(entry); > } else { > if (!pte_present(*pvmw->pte)) > return false; > @@ -200,7 +200,7 @@ bool page_vma_mapped_walk(struct page_vma_mapped_walk *pvmw) > if (is_migration_entry(pmd_to_swp_entry(*pvmw->pmd))) { > swp_entry_t entry = pmd_to_swp_entry(*pvmw->pmd); > > - if (migration_entry_to_page(entry) != page) > + if (pfn_to_page(swp_offset(entry)) != page) > return not_found(pvmw); > return true; > } > -- > 2.20.1 > ---end quoted text---