Received: by 2002:a25:1506:0:0:0:0:0 with SMTP id 6csp53108ybv; Tue, 18 Feb 2020 17:33:40 -0800 (PST) X-Google-Smtp-Source: APXvYqxKg4dfVF9hUtQU+958BbTViVSiY5pappwL65dcrnRWBcABsnBTxRpml2mY/blFvsiQobCG X-Received: by 2002:a9d:63d6:: with SMTP id e22mr18237319otl.185.1582076020001; Tue, 18 Feb 2020 17:33:40 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1582076019; cv=none; d=google.com; s=arc-20160816; b=TtFSCS27VAZP5YiSWjnvbaiSmna/1OFFivoskHWYsVzkIYvHXuz4ruuX6jaSLaGVqe 71FdgObo3p1IcRtsYRzs0h+Ix6kmOuQpf0AWtL0B+SblhAM48yFVTH7nMiPmHtqMflLr J2bEvWPzM8I8bTWNHyaDQ98VWLxmkpadqM3SlLMpRRTT3VlShiGEkjrO/zY6x5kb2qKz zYb2pmk7asPDscBUEZklLPEUP7IXgg/QowVptQCglZPEqfMNlIJHWh1VRuF44EjqpXbM J7I6G7oYoBjbtg6KvbOVprf+aqY3BuQt67XdSZox4Fyox0lZHdLdjECJ8paziu5fMS3z ZDtQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:dkim-signature:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=a8ZomntyDZQd4T0ETmKn7kyy8sPKQZmX4XFar/4v0t8=; b=Hfj2CN52OpanLSHaGu8kTBCVhmvy+WDDzxTpSoXNJcsZEgMQSeUrgKlUXNSpvua+dE SNf/6PpOLeTs8sEco7pYCFVqPL+CqcuhoxyERKnBHTrUQ5C6chio1RcUG0j9HeBKrd8t fbU3jSjZHlp3sHWw+3vFb7zn1ujkL6ubGYGTHicBNJcnpQhvwA5+278PCRlkSaZ4aXXA DKmXNWFBX+oy4Z926MLWrxtimwloO63THYUiQ/7jltHq+UNRXHDC23r3HlGw27Xhz34i tJ6jSGkSfJXzQXbdrTrZHOK4CMOrHchsQ+ymipi7e6I1juOsJgu7cIIEjstEHpQdk3zI ce+A== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b=D1CWU0QJ; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id w135si9285069oiw.44.2020.02.18.17.33.27; Tue, 18 Feb 2020 17:33:39 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b=D1CWU0QJ; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727906AbgBSBcd (ORCPT + 99 others); Tue, 18 Feb 2020 20:32:33 -0500 Received: from hqnvemgate25.nvidia.com ([216.228.121.64]:3956 "EHLO hqnvemgate25.nvidia.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726482AbgBSBcd (ORCPT ); Tue, 18 Feb 2020 20:32:33 -0500 Received: from hqpgpgate102.nvidia.com (Not Verified[216.228.121.13]) by hqnvemgate25.nvidia.com (using TLS: TLSv1.2, DES-CBC3-SHA) id ; Tue, 18 Feb 2020 17:31:59 -0800 Received: from hqmail.nvidia.com ([172.20.161.6]) by hqpgpgate102.nvidia.com (PGP Universal service); Tue, 18 Feb 2020 17:32:31 -0800 X-PGP-Universal: processed; by hqpgpgate102.nvidia.com on Tue, 18 Feb 2020 17:32:31 -0800 Received: from [10.110.48.28] (10.124.1.5) by HQMAIL107.nvidia.com (172.20.187.13) with Microsoft SMTP Server (TLS) id 15.0.1473.3; Wed, 19 Feb 2020 01:32:31 +0000 Subject: Re: [PATCH v6 09/19] mm: Add page_cache_readahead_limit To: Matthew Wilcox , CC: , , , , , , , , References: <20200217184613.19668-1-willy@infradead.org> <20200217184613.19668-16-willy@infradead.org> X-Nvconfidentiality: public From: John Hubbard Message-ID: <1263603d-f446-c447-2eac-697d105fa76c@nvidia.com> Date: Tue, 18 Feb 2020 17:32:31 -0800 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.5.0 MIME-Version: 1.0 In-Reply-To: <20200217184613.19668-16-willy@infradead.org> X-Originating-IP: [10.124.1.5] X-ClientProxiedBy: HQMAIL111.nvidia.com (172.20.187.18) To HQMAIL107.nvidia.com (172.20.187.13) Content-Type: text/plain; charset="utf-8" Content-Language: en-US Content-Transfer-Encoding: 7bit DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=nvidia.com; s=n1; t=1582075919; bh=a8ZomntyDZQd4T0ETmKn7kyy8sPKQZmX4XFar/4v0t8=; h=X-PGP-Universal:Subject:To:CC:References:X-Nvconfidentiality:From: Message-ID:Date:User-Agent:MIME-Version:In-Reply-To: X-Originating-IP:X-ClientProxiedBy:Content-Type:Content-Language: Content-Transfer-Encoding; b=D1CWU0QJZ/vJUMktO/sgLa0NnHUxAzBi1kB3pVWooMjgO2jXs6q5zixf/wAqu/IOj hr6pMYPnaPYV3Ou3y08ljYOt3n9jfzxdK6636RtEz2ipjrYs3b4hz65TVfyMrWizE8 BxPE/Qh4U52+XVl1w9Y1JqrHWmAs8zmb1n5IGtB0T/8XlxD4CpXKwc33AZc13ZXKhr j3FFEWf0HWoIz+2Nn/3nFVgupnlhaFia1orj1/9DKmtEAFVKf0ZH9VZvLLwMWQNW7A KAP7iN3AiTFahALL3seMXUw0/Tn+K2lQR+L66VGAqWlcnZtVglzmkuRCBxmUoAyDSr JDYpNY+TxVunw== Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2/17/20 10:45 AM, Matthew Wilcox wrote: > From: "Matthew Wilcox (Oracle)" > > ext4 and f2fs have duplicated the guts of the readahead code so > they can read past i_size. Instead, separate out the guts of the > readahead code so they can call it directly. > > Signed-off-by: Matthew Wilcox (Oracle) > --- > fs/ext4/verity.c | 35 ++--------------------- > fs/f2fs/verity.c | 35 ++--------------------- > include/linux/pagemap.h | 4 +++ > mm/readahead.c | 61 +++++++++++++++++++++++++++++------------ > 4 files changed, 52 insertions(+), 83 deletions(-) Just some minor ideas below, mostly documentation, so: Reviewed-by: John Hubbard > > diff --git a/fs/ext4/verity.c b/fs/ext4/verity.c > index dc5ec724d889..f6e0bf05933e 100644 > --- a/fs/ext4/verity.c > +++ b/fs/ext4/verity.c > @@ -342,37 +342,6 @@ static int ext4_get_verity_descriptor(struct inode *inode, void *buf, > return desc_size; > } > > -/* > - * Prefetch some pages from the file's Merkle tree. > - * > - * This is basically a stripped-down version of __do_page_cache_readahead() > - * which works on pages past i_size. > - */ > -static void ext4_merkle_tree_readahead(struct address_space *mapping, > - pgoff_t start_index, unsigned long count) > -{ > - LIST_HEAD(pages); > - unsigned int nr_pages = 0; > - struct page *page; > - pgoff_t index; > - struct blk_plug plug; > - > - for (index = start_index; index < start_index + count; index++) { > - page = xa_load(&mapping->i_pages, index); > - if (!page || xa_is_value(page)) { > - page = __page_cache_alloc(readahead_gfp_mask(mapping)); > - if (!page) > - break; > - page->index = index; > - list_add(&page->lru, &pages); > - nr_pages++; > - } > - } > - blk_start_plug(&plug); > - ext4_mpage_readpages(mapping, &pages, NULL, nr_pages, true); > - blk_finish_plug(&plug); > -} > - > static struct page *ext4_read_merkle_tree_page(struct inode *inode, > pgoff_t index, > unsigned long num_ra_pages) > @@ -386,8 +355,8 @@ static struct page *ext4_read_merkle_tree_page(struct inode *inode, > if (page) > put_page(page); > else if (num_ra_pages > 1) > - ext4_merkle_tree_readahead(inode->i_mapping, index, > - num_ra_pages); > + page_cache_readahead_limit(inode->i_mapping, NULL, > + index, LONG_MAX, num_ra_pages, 0); LONG_MAX seems bold at first, but then again I can't think of anything smaller that makes any sense, and the previous code didn't have a limit either...OK. I also wondered about the NULL file parameter, and wonder if we're stripping out information that is needed for authentication, given that that's what the newly written kerneldoc says the "file" arg is for. But it seems that if we're this deep in the fs code's read routines, file system authentication has long since been addressed. Any actually I don't yet (still working through the patches) see any authentication, so maybe that parameter will turn out to be unnecessary. Anyway, It's nice to see this factored out into a single routine. > page = read_mapping_page(inode->i_mapping, index, NULL); > } > return page; > diff --git a/fs/f2fs/verity.c b/fs/f2fs/verity.c > index d7d430a6f130..71a3e36721fa 100644 > --- a/fs/f2fs/verity.c > +++ b/fs/f2fs/verity.c > @@ -222,37 +222,6 @@ static int f2fs_get_verity_descriptor(struct inode *inode, void *buf, > return size; > } > > -/* > - * Prefetch some pages from the file's Merkle tree. > - * > - * This is basically a stripped-down version of __do_page_cache_readahead() > - * which works on pages past i_size. > - */ > -static void f2fs_merkle_tree_readahead(struct address_space *mapping, > - pgoff_t start_index, unsigned long count) > -{ > - LIST_HEAD(pages); > - unsigned int nr_pages = 0; > - struct page *page; > - pgoff_t index; > - struct blk_plug plug; > - > - for (index = start_index; index < start_index + count; index++) { > - page = xa_load(&mapping->i_pages, index); > - if (!page || xa_is_value(page)) { > - page = __page_cache_alloc(readahead_gfp_mask(mapping)); > - if (!page) > - break; > - page->index = index; > - list_add(&page->lru, &pages); > - nr_pages++; > - } > - } > - blk_start_plug(&plug); > - f2fs_mpage_readpages(mapping, &pages, NULL, nr_pages, true); > - blk_finish_plug(&plug); > -} > - > static struct page *f2fs_read_merkle_tree_page(struct inode *inode, > pgoff_t index, > unsigned long num_ra_pages) > @@ -266,8 +235,8 @@ static struct page *f2fs_read_merkle_tree_page(struct inode *inode, > if (page) > put_page(page); > else if (num_ra_pages > 1) > - f2fs_merkle_tree_readahead(inode->i_mapping, index, > - num_ra_pages); > + page_cache_readahead_limit(inode->i_mapping, NULL, > + index, LONG_MAX, num_ra_pages, 0); > page = read_mapping_page(inode->i_mapping, index, NULL); > } > return page; > diff --git a/include/linux/pagemap.h b/include/linux/pagemap.h > index bd4291f78f41..4f36c06d064d 100644 > --- a/include/linux/pagemap.h > +++ b/include/linux/pagemap.h > @@ -389,6 +389,10 @@ extern struct page * read_cache_page_gfp(struct address_space *mapping, > pgoff_t index, gfp_t gfp_mask); > extern int read_cache_pages(struct address_space *mapping, > struct list_head *pages, filler_t *filler, void *data); > +void page_cache_readahead_limit(struct address_space *mapping, > + struct file *file, pgoff_t offset, pgoff_t end_index, > + unsigned long nr_to_read, unsigned long lookahead_size); > + > > static inline struct page *read_mapping_page(struct address_space *mapping, > pgoff_t index, void *data) > diff --git a/mm/readahead.c b/mm/readahead.c > index 975ff5e387be..94d499cfb657 100644 > --- a/mm/readahead.c > +++ b/mm/readahead.c > @@ -142,35 +142,38 @@ static void read_pages(struct readahead_control *rac, struct list_head *pages) > blk_finish_plug(&plug); > } > > -/* > - * __do_page_cache_readahead() actually reads a chunk of disk. It allocates > - * the pages first, then submits them for I/O. This avoids the very bad > - * behaviour which would occur if page allocations are causing VM writeback. > - * We really don't want to intermingle reads and writes like that. > +/** > + * page_cache_readahead_limit - Start readahead beyond a file's i_size. Maybe: "Start readahead to a caller-specified end point" ? (It's only *potentially* beyond files's i_size.) > + * @mapping: File address space. > + * @file: This instance of the open file; used for authentication. > + * @offset: First page index to read. > + * @end_index: The maximum page index to read. > + * @nr_to_read: The number of pages to read. How about: "The number of pages to read, as long as end_index is not exceeded." > + * @lookahead_size: Where to start the next readahead. Pre-existing, but...it's hard to understand how a size is "where to start". Should we rename this arg? > + * > + * This function is for filesystems to call when they want to start > + * readahead potentially beyond a file's stated i_size. If you want > + * to start readahead on a normal file, you probably want to call > + * page_cache_async_readahead() or page_cache_sync_readahead() instead. > + * > + * Context: File is referenced by caller. Mutexes may be held by caller. > + * May sleep, but will not reenter filesystem to reclaim memory. In fact, can we say "must not reenter filesystem"? > */ > -void __do_page_cache_readahead(struct address_space *mapping, > - struct file *filp, pgoff_t offset, unsigned long nr_to_read, > - unsigned long lookahead_size) > +void page_cache_readahead_limit(struct address_space *mapping, > + struct file *file, pgoff_t offset, pgoff_t end_index, > + unsigned long nr_to_read, unsigned long lookahead_size) > { > - struct inode *inode = mapping->host; > - unsigned long end_index; /* The last page we want to read */ > LIST_HEAD(page_pool); > unsigned long i; > - loff_t isize = i_size_read(inode); > gfp_t gfp_mask = readahead_gfp_mask(mapping); > bool use_list = mapping->a_ops->readpages; > struct readahead_control rac = { > .mapping = mapping, > - .file = filp, > + .file = file, > ._start = offset, > ._nr_pages = 0, > }; > > - if (isize == 0) > - return; > - > - end_index = ((isize - 1) >> PAGE_SHIFT); > - > /* > * Preallocate as many pages as we will need. > */ > @@ -225,6 +228,30 @@ void __do_page_cache_readahead(struct address_space *mapping, > read_pages(&rac, &page_pool); > BUG_ON(!list_empty(&page_pool)); > } > +EXPORT_SYMBOL_GPL(page_cache_readahead_limit); > + > +/* > + * __do_page_cache_readahead() actually reads a chunk of disk. It allocates > + * the pages first, then submits them for I/O. This avoids the very bad > + * behaviour which would occur if page allocations are causing VM writeback. > + * We really don't want to intermingle reads and writes like that. > + */ > +void __do_page_cache_readahead(struct address_space *mapping, > + struct file *file, pgoff_t offset, unsigned long nr_to_read, > + unsigned long lookahead_size) > +{ > + struct inode *inode = mapping->host; > + unsigned long end_index; /* The last page we want to read */ > + loff_t isize = i_size_read(inode); > + > + if (isize == 0) > + return; > + > + end_index = ((isize - 1) >> PAGE_SHIFT); > + > + page_cache_readahead_limit(mapping, file, offset, end_index, > + nr_to_read, lookahead_size); > +} > > /* > * Chunk the readahead into 2 megabyte units, so that we don't pin too much > thanks, -- John Hubbard NVIDIA