Received: by 2002:a25:1506:0:0:0:0:0 with SMTP id 6csp1437521ybv; Thu, 20 Feb 2020 20:30:53 -0800 (PST) X-Google-Smtp-Source: APXvYqzkKUpbZ2RAqg3ZOLu7kh6+om4n+wAqC2vukcEcBsQjpC3eFi3CTlaeYrIbdeNjcc/X8sE2 X-Received: by 2002:a05:6808:3b2:: with SMTP id n18mr446566oie.146.1582259453573; Thu, 20 Feb 2020 20:30:53 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1582259453; cv=none; d=google.com; s=arc-20160816; b=Dt4nLYSALmDp04uhCEmVZe2vC7bV0KFsRoilJA3kX2Mir4oCJQT2Vh5jWJ0hK93xLz V0PXug10JUC2l8bQ4t23B9/WOFIaPFFpVfbbGdHT1s6HQ9toMBzMVwqBl2hYA1w6Pbu6 yIsWITkYqny6VLgQjEpnOj6Ny8cxWsF9jz+UB0bgtoEjXmm2wTHCUEHoQBCf1aEvJ/66 ogArPm/4mWXPHezJGhofXp+ZUSnO56YqkvBatZt2AwzYTzL447dfv6aJGf5W03UojHyU 7fjzyuJcGPHH7ROdLt7tfRYgNnZ+p5jf4s0+ae9KsoK2lbS/TLu8vEs5KretJDpRPDtL fang== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:dkim-signature:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=+Wm3044wXz/CoDYYTB4qibyZMJiORNPZJOuYKY+0G8w=; b=IY8JXP/qf/LWDPa+31YljZxgTkwjHey82RAkAR452KNXt9PHNp1yZPD4/aqcFDHdm8 0nS/Ard9UV4VPJorj9/lGCXFR6RT6BAdjZ35P0CEcNvPn0WC/PsZj/GLqRmGY6ONjHNs VHNg4RLmqTf/qvddO7j1+25EUsUdFiJZGUgKcwc2YqECH/UP6HLfmA4DzndrXUgLupA4 Q+dHKqS6Ots+yMsrVFvbsuI8LN5Z834N3id04WpY4LVEBFj54zygImOK+/SecO8mnUfX 4KhzJP9aLU++hZjsvSp2GvD5AoLUGf+xuIvf6rr8pfR+wA/h9sDQM/Y91I8+PZu+y6U5 8DLA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b=UitvsuI1; spf=pass (google.com: best guess record for domain of linux-ext4-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-ext4-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id m20si852514otf.143.2020.02.20.20.30.42; Thu, 20 Feb 2020 20:30:53 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-ext4-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b=UitvsuI1; spf=pass (google.com: best guess record for domain of linux-ext4-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-ext4-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1729595AbgBUEak (ORCPT + 99 others); Thu, 20 Feb 2020 23:30:40 -0500 Received: from hqnvemgate24.nvidia.com ([216.228.121.143]:2874 "EHLO hqnvemgate24.nvidia.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1729562AbgBUEak (ORCPT ); Thu, 20 Feb 2020 23:30:40 -0500 Received: from hqpgpgate102.nvidia.com (Not Verified[216.228.121.13]) by hqnvemgate24.nvidia.com (using TLS: TLSv1.2, DES-CBC3-SHA) id ; Thu, 20 Feb 2020 20:29:25 -0800 Received: from hqmail.nvidia.com ([172.20.161.6]) by hqpgpgate102.nvidia.com (PGP Universal service); Thu, 20 Feb 2020 20:30:38 -0800 X-PGP-Universal: processed; by hqpgpgate102.nvidia.com on Thu, 20 Feb 2020 20:30:38 -0800 Received: from [10.110.48.28] (10.124.1.5) by HQMAIL107.nvidia.com (172.20.187.13) with Microsoft SMTP Server (TLS) id 15.0.1473.3; Fri, 21 Feb 2020 04:30:38 +0000 Subject: Re: [PATCH v7 10/24] mm: Add readahead address space operation To: Matthew Wilcox , CC: , , , , , , , , References: <20200219210103.32400-1-willy@infradead.org> <20200219210103.32400-11-willy@infradead.org> X-Nvconfidentiality: public From: John Hubbard Message-ID: <45d67cb1-9ca3-e2b8-d71f-1f0c2e085155@nvidia.com> Date: Thu, 20 Feb 2020 20:30:37 -0800 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.5.0 MIME-Version: 1.0 In-Reply-To: <20200219210103.32400-11-willy@infradead.org> X-Originating-IP: [10.124.1.5] X-ClientProxiedBy: HQMAIL105.nvidia.com (172.20.187.12) To HQMAIL107.nvidia.com (172.20.187.13) Content-Type: text/plain; charset="utf-8" Content-Language: en-US Content-Transfer-Encoding: 7bit DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=nvidia.com; s=n1; t=1582259366; bh=+Wm3044wXz/CoDYYTB4qibyZMJiORNPZJOuYKY+0G8w=; h=X-PGP-Universal:Subject:To:CC:References:X-Nvconfidentiality:From: Message-ID:Date:User-Agent:MIME-Version:In-Reply-To: X-Originating-IP:X-ClientProxiedBy:Content-Type:Content-Language: Content-Transfer-Encoding; b=UitvsuI1yoXnzml76+cSz7F+oTZFS/iyT7dOV5LH/5L9YZnf2JDmcICsW5QQsDOWw JEmxfGjJhpm6Mw0DpkQ4I97JHP2YxMVlsg4TdE5MBveHV+wfxOmkMx2E9X1fQGLWfG f8znlDQrWpx92I4FW23hWS3bxNs0j+rkOmq+Rdu1ptzaIP6nxGv9ByYUjCJJs79YE7 seWhQosGZtfe20jPzoLRnPNfNQlt0kzqk9SCpCzrG7HeKnuVPkQFIfGbTqOGCn3/Pw /XB5cMIwzFk+62xO1plJYkkUcmyRC6IeEYi54p0+giupM67/DiXtRdSSVw1QBxLyH7 Qd6o1BBujjHhg== Sender: linux-ext4-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-ext4@vger.kernel.org On 2/19/20 1:00 PM, Matthew Wilcox wrote: > From: "Matthew Wilcox (Oracle)" > > This replaces ->readpages with a saner interface: > - Return void instead of an ignored error code. > - Page cache is already populated with locked pages when ->readahead > is called. > - New arguments can be passed to the implementation without changing > all the filesystems that use a common helper function like > mpage_readahead(). > > Signed-off-by: Matthew Wilcox (Oracle) > --- > Documentation/filesystems/locking.rst | 6 +++++- > Documentation/filesystems/vfs.rst | 15 +++++++++++++++ > include/linux/fs.h | 2 ++ > include/linux/pagemap.h | 18 ++++++++++++++++++ > mm/readahead.c | 12 ++++++++++-- > 5 files changed, 50 insertions(+), 3 deletions(-) This all looks correct to me, so: Reviewed-by: John Hubbard thanks, -- John Hubbard NVIDIA > > diff --git a/Documentation/filesystems/locking.rst b/Documentation/filesystems/locking.rst > index 5057e4d9dcd1..0af2e0e11461 100644 > --- a/Documentation/filesystems/locking.rst > +++ b/Documentation/filesystems/locking.rst > @@ -239,6 +239,7 @@ prototypes:: > int (*readpage)(struct file *, struct page *); > int (*writepages)(struct address_space *, struct writeback_control *); > int (*set_page_dirty)(struct page *page); > + void (*readahead)(struct readahead_control *); > int (*readpages)(struct file *filp, struct address_space *mapping, > struct list_head *pages, unsigned nr_pages); > int (*write_begin)(struct file *, struct address_space *mapping, > @@ -271,7 +272,8 @@ writepage: yes, unlocks (see below) > readpage: yes, unlocks > writepages: > set_page_dirty no > -readpages: > +readahead: yes, unlocks > +readpages: no > write_begin: locks the page exclusive > write_end: yes, unlocks exclusive > bmap: > @@ -295,6 +297,8 @@ the request handler (/dev/loop). > ->readpage() unlocks the page, either synchronously or via I/O > completion. > > +->readahead() unlocks the pages that I/O is attempted on like ->readpage(). > + > ->readpages() populates the pagecache with the passed pages and starts > I/O against them. They come unlocked upon I/O completion. > > diff --git a/Documentation/filesystems/vfs.rst b/Documentation/filesystems/vfs.rst > index 7d4d09dd5e6d..ed17771c212b 100644 > --- a/Documentation/filesystems/vfs.rst > +++ b/Documentation/filesystems/vfs.rst > @@ -706,6 +706,7 @@ cache in your filesystem. The following members are defined: > int (*readpage)(struct file *, struct page *); > int (*writepages)(struct address_space *, struct writeback_control *); > int (*set_page_dirty)(struct page *page); > + void (*readahead)(struct readahead_control *); > int (*readpages)(struct file *filp, struct address_space *mapping, > struct list_head *pages, unsigned nr_pages); > int (*write_begin)(struct file *, struct address_space *mapping, > @@ -781,12 +782,26 @@ cache in your filesystem. The following members are defined: > If defined, it should set the PageDirty flag, and the > PAGECACHE_TAG_DIRTY tag in the radix tree. > > +``readahead`` > + Called by the VM to read pages associated with the address_space > + object. The pages are consecutive in the page cache and are > + locked. The implementation should decrement the page refcount > + after starting I/O on each page. Usually the page will be > + unlocked by the I/O completion handler. If the filesystem decides > + to stop attempting I/O before reaching the end of the readahead > + window, it can simply return. The caller will decrement the page > + refcount and unlock the remaining pages for you. Set PageUptodate > + if the I/O completes successfully. Setting PageError on any page > + will be ignored; simply unlock the page if an I/O error occurs. > + > ``readpages`` > called by the VM to read pages associated with the address_space > object. This is essentially just a vector version of readpage. > Instead of just one page, several pages are requested. > readpages is only used for read-ahead, so read errors are > ignored. If anything goes wrong, feel free to give up. > + This interface is deprecated and will be removed by the end of > + 2020; implement readahead instead. > > ``write_begin`` > Called by the generic buffered write code to ask the filesystem > diff --git a/include/linux/fs.h b/include/linux/fs.h > index 3cd4fe6b845e..d4e2d2964346 100644 > --- a/include/linux/fs.h > +++ b/include/linux/fs.h > @@ -292,6 +292,7 @@ enum positive_aop_returns { > struct page; > struct address_space; > struct writeback_control; > +struct readahead_control; > > /* > * Write life time hint values. > @@ -375,6 +376,7 @@ struct address_space_operations { > */ > int (*readpages)(struct file *filp, struct address_space *mapping, > struct list_head *pages, unsigned nr_pages); > + void (*readahead)(struct readahead_control *); > > int (*write_begin)(struct file *, struct address_space *mapping, > loff_t pos, unsigned len, unsigned flags, > diff --git a/include/linux/pagemap.h b/include/linux/pagemap.h > index 4989d330fada..b3008605fd1b 100644 > --- a/include/linux/pagemap.h > +++ b/include/linux/pagemap.h > @@ -669,6 +669,24 @@ static inline struct page *readahead_page(struct readahead_control *rac) > return page; > } > > +/* The byte offset into the file of this readahead block */ > +static inline loff_t readahead_pos(struct readahead_control *rac) > +{ > + return (loff_t)rac->_index * PAGE_SIZE; > +} > + > +/* The number of bytes in this readahead block */ > +static inline loff_t readahead_length(struct readahead_control *rac) > +{ > + return (loff_t)rac->_nr_pages * PAGE_SIZE; > +} > + > +/* The index of the first page in this readahead block */ > +static inline unsigned int readahead_index(struct readahead_control *rac) > +{ > + return rac->_index; > +} > + > /* The number of pages in this readahead block */ > static inline unsigned int readahead_count(struct readahead_control *rac) > { > diff --git a/mm/readahead.c b/mm/readahead.c > index aaa209559ba2..07cdfbf00f4b 100644 > --- a/mm/readahead.c > +++ b/mm/readahead.c > @@ -124,7 +124,14 @@ static void read_pages(struct readahead_control *rac, struct list_head *pages) > > blk_start_plug(&plug); > > - if (aops->readpages) { > + if (aops->readahead) { > + aops->readahead(rac); > + /* Clean up the remaining pages */ > + while ((page = readahead_page(rac))) { > + unlock_page(page); > + put_page(page); > + } > + } else if (aops->readpages) { > aops->readpages(rac->file, rac->mapping, pages, > readahead_count(rac)); > /* Clean up the remaining pages */ > @@ -234,7 +241,8 @@ void force_page_cache_readahead(struct address_space *mapping, > struct file_ra_state *ra = &filp->f_ra; > unsigned long max_pages; > > - if (unlikely(!mapping->a_ops->readpage && !mapping->a_ops->readpages)) > + if (unlikely(!mapping->a_ops->readpage && !mapping->a_ops->readpages && > + !mapping->a_ops->readahead)) > return; > > /* >