Received: by 2002:a05:6a10:22f:0:0:0:0 with SMTP id 15csp467555pxk; Wed, 9 Sep 2020 09:56:08 -0700 (PDT) X-Google-Smtp-Source: ABdhPJw6WmLv5+MjVgfjaErUzCWSdOnT7zUK3LHwhmM8Cq6rQ4frJDsE+MKIRJ0Zx9ewUTLlneid X-Received: by 2002:a17:907:118c:: with SMTP id uz12mr4452918ejb.321.1599670567873; Wed, 09 Sep 2020 09:56:07 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1599670567; cv=none; d=google.com; s=arc-20160816; b=lndZnPj5lLoamgKfcyBKq1KjGPeCXj/p8dozv7uw9759zY61DgkeH7XjYlSx+rGoVs K5ESFKXripb5r8VyVaoVXJvMFMigdaxxlvI34RoSHz6qoElfxiIk75PCe2fLMoGeJ50y CI+FCs/xvg+0YQqjG/CZ5ItJccZrU3w7nRcK+zkG1U30aCqCO+YdUsuUM6Q5QHgKIcTf vPkxD4mJR/92o20LiPMvCN8P15X54/ya2E8bMIhkMfXcBsj1Y4AKps8aQnLLGPJbrFzn 4ku7DakdfcMVXkI9ee3neO6MjKxv/eoclzlWyCBQQacJDudGz9Lruc/YodYmgsH7tt4Y xftw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=snpdJa8IBy01ZPM2bsTR1z2qbOmy44Bnyd/h1RTnFXQ=; b=cDG3aIlSsZXJ7nR8SfwctSSwWfOBoIR66ru3JbWpZl80Fn1hYlPItXMO923UGunT1Q ab1ikWeRBUFv80kANsGmWyVQyZ6+/U650acpXkny6i3vJyHpMfeeU1Vgq4P5hW+xnlSK 7QIU+zl/BWBaY55ounOHox/8PsnLDtxXD9eW8CIRyVhMmyPOze99uoEVtBkfIQm7ccn5 VZDEmfvBbMU5QHVRfJn3FRwq8PZgYWfb9Llh66mH6rsf6Em/jYGBWfkkYM00YNDMX1C1 fQcthXUHIT9DZauAk5BjwQEdZ2v7+6v1lWzfvYapLmv9S+xC23ZxhkMLlIlU4DwB9HWy 3O8g== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=QXMhBtJB; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id gl17si1838133ejb.311.2020.09.09.09.55.43; Wed, 09 Sep 2020 09:56:07 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=QXMhBtJB; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1730996AbgIIQxk (ORCPT + 99 others); Wed, 9 Sep 2020 12:53:40 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:60566 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1730939AbgIIQxf (ORCPT ); Wed, 9 Sep 2020 12:53:35 -0400 Received: from mail-ej1-x644.google.com (mail-ej1-x644.google.com [IPv6:2a00:1450:4864:20::644]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 4C3A0C061573 for ; Wed, 9 Sep 2020 09:53:35 -0700 (PDT) Received: by mail-ej1-x644.google.com with SMTP id z22so4556604ejl.7 for ; Wed, 09 Sep 2020 09:53:35 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=snpdJa8IBy01ZPM2bsTR1z2qbOmy44Bnyd/h1RTnFXQ=; b=QXMhBtJByRcyCYV372I9aBbU4ualy4wX3nIJFbe1nxJDloQ76NBJGpmhVUGmSuAsi0 dPXufn0dk9S4Nlq49VgQer8DAJjR+feE7eOTBW+sAVAGoM5QQEt/0Fv+TfhjsODBP0dI eqm/AWN14Yx3mSuDKzPwiGN42csrnPBh69nhpm17dtQcO+/5NR5sun3HtBr+GYu8lsVO 2m0gtiHPhMItZpmdM8nCNHXa1Teu9AIt2GPEVMjPwy/PnPhON65U3sQxq1dWut7FWei9 LSsPDFXIbgpId8TmmOR6zJYkt6c49m2CZoor9L66h/uFteVcj60JmQ2g/A5NSs+aebZH U1Xw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=snpdJa8IBy01ZPM2bsTR1z2qbOmy44Bnyd/h1RTnFXQ=; b=HPBqJIbpeEBY+yS2PqHsFX82wC8QViOicVY9JoQkBUMmhNSn+I/aTkm6In7rRwFQ3z xisn0GbaSejUSzHoopuod4tXBswc+hSh7lqb6huLEGr5HlYhi7AJXYIM/L7nMTdjvwNs aIB0LIOjVc9g3PKD6R2XmEqnSC38yFlhLeAtTiif7lf53p0xMTp02RMu/bcAXJASIb6H 3qFF2CF9TZkhhGDkD8BZDhOuM7saHqoPXs78K4HRYQJDvuZ9oCBH2ZlM159erQKssFHy 3dKSaMu1HDbse0AGmv4X3J2kmEwjPzv0OHjIUbA5bjGK7RF5Ror+Admsc24KPJsiW7Gc 1QsQ== X-Gm-Message-State: AOAM531gr0B2aZEROQbC5gkODR/mSfQiTmaZgYIRHYl/VjCLc1/lKPTK LEB8dmuJjuHIjLMOHGDkwwGSH5jF/r5Tze61VBNp24V1 X-Received: by 2002:a17:906:b090:: with SMTP id x16mr4573651ejy.46.1599670413892; Wed, 09 Sep 2020 09:53:33 -0700 (PDT) MIME-Version: 1.0 References: <20200908162559.509113-1-Anna.Schumaker@Netapp.com> <20200908162559.509113-3-Anna.Schumaker@Netapp.com> <20200908194245.GB6256@pick.fieldses.org> In-Reply-To: <20200908194245.GB6256@pick.fieldses.org> From: Anna Schumaker Date: Wed, 9 Sep 2020 12:53:18 -0400 Message-ID: Subject: Re: [PATCH v5 2/5] NFSD: Add READ_PLUS data support To: "J. Bruce Fields" Cc: Chuck Lever , Linux NFS Mailing List Content-Type: text/plain; charset="UTF-8" Sender: linux-nfs-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-nfs@vger.kernel.org On Tue, Sep 8, 2020 at 3:42 PM J. Bruce Fields wrote: > > On Tue, Sep 08, 2020 at 12:25:56PM -0400, schumaker.anna@gmail.com wrote: > > From: Anna Schumaker > > > > This patch adds READ_PLUS support for returning a single > > NFS4_CONTENT_DATA segment to the client. This is basically the same as > > the READ operation, only with the extra information about data segments. > > > > Signed-off-by: Anna Schumaker > > > > --- > > v5: Fix up nfsd4_read_plus_rsize() calculation > > --- > > fs/nfsd/nfs4proc.c | 20 +++++++++++ > > fs/nfsd/nfs4xdr.c | 83 ++++++++++++++++++++++++++++++++++++++++++++-- > > 2 files changed, 101 insertions(+), 2 deletions(-) > > > > diff --git a/fs/nfsd/nfs4proc.c b/fs/nfsd/nfs4proc.c > > index eaf50eafa935..0a3df5f10501 100644 > > --- a/fs/nfsd/nfs4proc.c > > +++ b/fs/nfsd/nfs4proc.c > > @@ -2591,6 +2591,19 @@ static inline u32 nfsd4_read_rsize(struct svc_rqst *rqstp, struct nfsd4_op *op) > > return (op_encode_hdr_size + 2 + XDR_QUADLEN(rlen)) * sizeof(__be32); > > } > > > > +static inline u32 nfsd4_read_plus_rsize(struct svc_rqst *rqstp, struct nfsd4_op *op) > > +{ > > + u32 maxcount = svc_max_payload(rqstp); > > + u32 rlen = min(op->u.read.rd_length, maxcount); > > + /* > > + * Worst case is a single large data segment, so make > > + * sure we have enough room to encode that > > + */ > > After the last patch we allow an unlimited number of segments. So a > zillion 1-byte segments is also possible, and is a worse case. > > Possible ways to avoid that kind of thing: > > - when encoding, stop and return a short read if the xdr-encoded > result would exceed the limit calculated here. Doesn't this happen automatically through calls to xdr_reserve_space()? > - round SEEK_HOLE results up to the nearest (say) 512 bytes, and > SEEK_DATA result down to the nearest 512 bytes. May also need > some logic to avoid encoding 0-length segments. > - talk to linux-fsdevel, see if we can get a minimum hole > alignment guarantee from filesystems. > > I'm thinking #1 is simplest. > > --b. > > > + u32 seg_len = 1 + 2 + 1; > > + > > + return (op_encode_hdr_size + 2 + seg_len + XDR_QUADLEN(rlen)) * sizeof(__be32); > > +} > > + > > static inline u32 nfsd4_readdir_rsize(struct svc_rqst *rqstp, struct nfsd4_op *op) > > { > > u32 maxcount = 0, rlen = 0; > > @@ -3163,6 +3176,13 @@ static const struct nfsd4_operation nfsd4_ops[] = { > > .op_name = "OP_COPY", > > .op_rsize_bop = nfsd4_copy_rsize, > > }, > > + [OP_READ_PLUS] = { > > + .op_func = nfsd4_read, > > + .op_release = nfsd4_read_release, > > + .op_name = "OP_READ_PLUS", > > + .op_rsize_bop = nfsd4_read_plus_rsize, > > + .op_get_currentstateid = nfsd4_get_readstateid, > > + }, > > [OP_SEEK] = { > > .op_func = nfsd4_seek, > > .op_name = "OP_SEEK", > > diff --git a/fs/nfsd/nfs4xdr.c b/fs/nfsd/nfs4xdr.c > > index 0be194de4888..26d12e3edf33 100644 > > --- a/fs/nfsd/nfs4xdr.c > > +++ b/fs/nfsd/nfs4xdr.c > > @@ -2173,7 +2173,7 @@ static const nfsd4_dec nfsd4_dec_ops[] = { > > [OP_LAYOUTSTATS] = (nfsd4_dec)nfsd4_decode_notsupp, > > [OP_OFFLOAD_CANCEL] = (nfsd4_dec)nfsd4_decode_offload_status, > > [OP_OFFLOAD_STATUS] = (nfsd4_dec)nfsd4_decode_offload_status, > > - [OP_READ_PLUS] = (nfsd4_dec)nfsd4_decode_notsupp, > > + [OP_READ_PLUS] = (nfsd4_dec)nfsd4_decode_read, > > [OP_SEEK] = (nfsd4_dec)nfsd4_decode_seek, > > [OP_WRITE_SAME] = (nfsd4_dec)nfsd4_decode_notsupp, > > [OP_CLONE] = (nfsd4_dec)nfsd4_decode_clone, > > @@ -4597,6 +4597,85 @@ nfsd4_encode_offload_status(struct nfsd4_compoundres *resp, __be32 nfserr, > > return nfserr_resource; > > p = xdr_encode_hyper(p, os->count); > > *p++ = cpu_to_be32(0); > > + return nfserr; > > +} > > + > > +static __be32 > > +nfsd4_encode_read_plus_data(struct nfsd4_compoundres *resp, > > + struct nfsd4_read *read, > > + unsigned long maxcount, u32 *eof) > > +{ > > + struct xdr_stream *xdr = &resp->xdr; > > + struct file *file = read->rd_nf->nf_file; > > + int starting_len = xdr->buf->len; > > + __be32 nfserr; > > + __be32 *p, tmp; > > + __be64 tmp64; > > + > > + /* Content type, offset, byte count */ > > + p = xdr_reserve_space(xdr, 4 + 8 + 4); > > + if (!p) > > + return nfserr_resource; > > + > > + read->rd_vlen = xdr_reserve_space_vec(xdr, resp->rqstp->rq_vec, maxcount); > > + if (read->rd_vlen < 0) > > + return nfserr_resource; > > + > > + nfserr = nfsd_readv(resp->rqstp, read->rd_fhp, file, read->rd_offset, > > + resp->rqstp->rq_vec, read->rd_vlen, &maxcount, eof); > > + if (nfserr) > > + return nfserr; > > + > > + tmp = htonl(NFS4_CONTENT_DATA); > > + write_bytes_to_xdr_buf(xdr->buf, starting_len, &tmp, 4); > > + tmp64 = cpu_to_be64(read->rd_offset); > > + write_bytes_to_xdr_buf(xdr->buf, starting_len + 4, &tmp64, 8); > > + tmp = htonl(maxcount); > > + write_bytes_to_xdr_buf(xdr->buf, starting_len + 12, &tmp, 4); > > + return nfs_ok; > > +} > > + > > +static __be32 > > +nfsd4_encode_read_plus(struct nfsd4_compoundres *resp, __be32 nfserr, > > + struct nfsd4_read *read) > > +{ > > + unsigned long maxcount; > > + struct xdr_stream *xdr = &resp->xdr; > > + struct file *file; > > + int starting_len = xdr->buf->len; > > + int segments = 0; > > + __be32 *p, tmp; > > + u32 eof; > > + > > + if (nfserr) > > + return nfserr; > > + file = read->rd_nf->nf_file; > > + > > + /* eof flag, segment count */ > > + p = xdr_reserve_space(xdr, 4 + 4); > > + if (!p) > > + return nfserr_resource; > > + xdr_commit_encode(xdr); > > + > > + maxcount = svc_max_payload(resp->rqstp); > > + maxcount = min_t(unsigned long, maxcount, > > + (xdr->buf->buflen - xdr->buf->len)); > > + maxcount = min_t(unsigned long, maxcount, read->rd_length); > > + > > + eof = read->rd_offset >= i_size_read(file_inode(file)); > > + if (!eof) { > > + nfserr = nfsd4_encode_read_plus_data(resp, read, maxcount, &eof); > > + segments++; > > + } > > + > > + if (nfserr) > > + xdr_truncate_encode(xdr, starting_len); > > + else { > > + tmp = htonl(eof); > > + write_bytes_to_xdr_buf(xdr->buf, starting_len, &tmp, 4); > > + tmp = htonl(segments); > > + write_bytes_to_xdr_buf(xdr->buf, starting_len + 4, &tmp, 4); > > + } > > > > return nfserr; > > } > > @@ -4974,7 +5053,7 @@ static const nfsd4_enc nfsd4_enc_ops[] = { > > [OP_LAYOUTSTATS] = (nfsd4_enc)nfsd4_encode_noop, > > [OP_OFFLOAD_CANCEL] = (nfsd4_enc)nfsd4_encode_noop, > > [OP_OFFLOAD_STATUS] = (nfsd4_enc)nfsd4_encode_offload_status, > > - [OP_READ_PLUS] = (nfsd4_enc)nfsd4_encode_noop, > > + [OP_READ_PLUS] = (nfsd4_enc)nfsd4_encode_read_plus, > > [OP_SEEK] = (nfsd4_enc)nfsd4_encode_seek, > > [OP_WRITE_SAME] = (nfsd4_enc)nfsd4_encode_noop, > > [OP_CLONE] = (nfsd4_enc)nfsd4_encode_noop, > > -- > > 2.28.0 > > >