Return-Path: Received: from ipmail07.adl2.internode.on.net ([150.101.137.131]:36569 "EHLO ipmail07.adl2.internode.on.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1750803AbcA1VLG (ORCPT ); Thu, 28 Jan 2016 16:11:06 -0500 Date: Fri, 29 Jan 2016 08:10:49 +1100 From: Dave Chinner To: Chuck Lever Cc: Jan Kara , lsf-pc@lists.linux-foundation.org, Linux RDMA Mailing List , linux-fsdevel , Linux NFS Mailing List Subject: Re: [Lsf-pc] [LSF/MM TOPIC] Remote access to pmem on storage targets Message-ID: <20160128211049.GP6033@dastard> References: <06414D5A-0632-4C74-B76C-038093E8AED3@oracle.com> <20160126082533.GR24938@quack.suse.cz> <20160127000404.GN6033@dastard> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii In-Reply-To: Sender: linux-nfs-owner@vger.kernel.org List-ID: On Wed, Jan 27, 2016 at 10:55:36AM -0500, Chuck Lever wrote: > > > On Jan 26, 2016, at 7:04 PM, Dave Chinner wrote: > > > > On Tue, Jan 26, 2016 at 10:58:44AM -0500, Chuck Lever wrote: > >> It is not going to be like the well-worn paradigm that > >> involves a page cache on the storage target backed by > >> slow I/O operations. The protocol layers on storage > >> targets need a way to discover memory addresses of > >> persistent memory that will be used as source/sink > >> buffers for RDMA operations. > >> > >> And making data durable after a write is going to need > >> some thought. So I believe some new plumbing will be > >> necessary. > > > > Haven't we already solve this for the pNFS file driver that XFS > > implements? i.e. these export operations: > > > > int (*get_uuid)(struct super_block *sb, u8 *buf, u32 *len, u64 *offset); > > int (*map_blocks)(struct inode *inode, loff_t offset, > > u64 len, struct iomap *iomap, > > bool write, u32 *device_generation); > > int (*commit_blocks)(struct inode *inode, struct iomap *iomaps, > > int nr_iomaps, struct iattr *iattr); > > > > so mapping/allocation of file offset to sector mappings, which can > > then trivially be used to grab the memory address through the bdev > > ->direct_access method, yes? > > Thanks, that makes sense. How would such addresses be > utilized? That's a different problem, and you need to talk to the IO guys about that. > I'll speak about the NFS/RDMA server for this example, as > I am more familiar with that than with block targets. When > I say "NFS server" here I mean the software service on the > storage target that speaks the NFS protocol. > > In today's RDMA-enabled storage protocols, an initiator > exposes its memory (in small segments) to storage targets, > sends a request, and the target's network transport performs > RDMA Read and Write operations to move the payload data in > that request. > > Assuming the NFS server is somehow aware that what it is > getting from ->direct_access is a persistent memory address > and not an LBA, it would then have to pass it down to the > transport layer (svcrdma) so that the address can be used > as a source or sink buffer for RDMA operations. > > For an NFS READ, this should be straightforward. An RPC > request comes in, the NFS server identifies the memory that > is to source the READ reply and passes the address of that > memory to the transport, which then pushes the data in > that memory via an RDMA Write to the client. Right, it's no different from using the page cache, except for however the memory adress is then mapped by the IO subsystem for the DMA transfer... > NFS WRITES are more difficult. An RPC request comes in, > and today the transport layer gathers incoming payload data > in anonymous pages before the NFS server even knows there > is an incoming RPC. We'd have to add some kind of hook to > enable the NFS server and the underlying filesystem to > provide appropriate sink buffers to the transport. ->map_blocks needs to be called to allocate/map the file offset and return a memory address before the data is sent from the client. > After the NFS WRITE request has been wholly received, the > NFS server today uses vfs_writev to put that data into the > target file. We'd probably want something more efficient > for pmem-backed filesystems. We want something more > efficient for traditional page cache-based filesystems > anyway. Yup. see above. > Every NFS WRITE larger than a page would be essentially > CoW, since the filesystem would need to provide "anonymous" > blocks to sink incoming WRITE data and then transition > those blocks into the target file? Not sure how this works > for pNFS with block devices. No, ->map_blocks can return blocks that are already allocated to the file at the given offset, hence overwrite in place works just fine. > Finally a client needs to perform an NFS COMMIT to ensure > that the written data is at rest on durable storage. We > could insist that all NFS WRITE operations to pmem will > be DATA_SYNC or better (in other words, abandon UNSTABLE > mode). You could, but you'd still need the two map/commit calls into the filesystem to get the memory and mark the write done... > If not, then a separate NFS COMMIT/LAYOUTCOMMIT > is necessary to flush memory caches and ensure data > durability. An extra RPC round trip is likely not a good > idea when the cost structure of NFS WRITE is so much > different than it is for traditional block devices. IIRC, ->commit_blocks is called from the LAYOUTCOMMIT operation. You'll need to call this to pair the ->map_blocks call above that provided the memory as the data sink for the write. This is because ->map_blocks allocates unwritten extents so that stale data will not be exposed before the write is complete and ->commit_blocks is called to remove the unwritten extent flag. > I imagine that the issues are similar for block targets, if > they assume block devices are fronted by a memory cache. Yup, hence the "three phase" write operation - map blocks, write data, commit blocks. Cheers, Dave. -- Dave Chinner david@fromorbit.com