Received: by 2002:a25:1506:0:0:0:0:0 with SMTP id 6csp967639ybv; Thu, 20 Feb 2020 10:31:17 -0800 (PST) X-Google-Smtp-Source: APXvYqxdj90AFZNC0vrTYfFGesD99vyW0uDlA0Fyb7zkDKMfcMiVJtVmexwotPgetnRYILnNbMyy X-Received: by 2002:aca:f10:: with SMTP id 16mr2983688oip.117.1582223477165; Thu, 20 Feb 2020 10:31:17 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1582223477; cv=none; d=google.com; s=arc-20160816; b=ENcVbOhaUxSTZCvq1Jbu7OEswxOchCVzh08E+Ff9bxeJQCq/sSB+nvUWpurF+A0Xuj dOwqQvqtdaGZsTumRPxobFFnKd2zXPcKJ2curIXmz9Uq5RsEN0N9+lSCYvWQ0ccQh6vI 7wqxJAFBTSpdNrr5GoweniCShxsFHhb8hxr2g7lo7zQtgry5RfRgT1WKht/jBdtCnqZH o+GbhiMDXKaX7WmG3T6gR8kStuB1HOP0lJbB8XsE+AQVGpmbyTL0P3DRoK5M/F4pbWX4 PABHvqg2aHu16NNDoHh2E4OPK4WacuC5z6nMBPQY72VHun1ykqS+bOAJYylbfg+w7t9s feAA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:to:references:message-id :content-transfer-encoding:cc:date:in-reply-to:from:subject :mime-version:dkim-signature; bh=HPMTPrlzxuePQNHkJApc24LTsV8QV+cJm9y+1ag6Kkk=; b=xeFbBGiielp2ZAPrD8IsJiQrRw1TI9CicA8MVb4ukxL3T5XspfqiTjus8fJ9iZD2xz mn08fjR2Eg6KkdRz4DyE6hK/KK3IuwjSNwCxHcvJzwXqGNNxe2lReZcAbQkRPKGiiHau eCeJSBSoTJwf2ovhshZvclSN07psq5VmcRI6ZV2lLmOpB6hvvCk93no2DSECenYlJn93 q7v/3h2l1xfZHoGVeR69vKnzM5O14JhEAYe+/nOcJNBGNR4mRs5gM2KOSTTzO7xZbehW 87W8dykkWLDZl5AiDDXy0GxPat7M8xuFILFrcq6bDnZPoHm/YfS8G7EX5UODGwiG6ptX AfCQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@oracle.com header.s=corp-2020-01-29 header.b=BT7CSxXN; spf=pass (google.com: best guess record for domain of linux-nfs-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=oracle.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id i23si98236oto.206.2020.02.20.10.31.04; Thu, 20 Feb 2020 10:31:17 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-nfs-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@oracle.com header.s=corp-2020-01-29 header.b=BT7CSxXN; spf=pass (google.com: best guess record for domain of linux-nfs-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=oracle.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728111AbgBTSbE (ORCPT + 99 others); Thu, 20 Feb 2020 13:31:04 -0500 Received: from aserp2120.oracle.com ([141.146.126.78]:32854 "EHLO aserp2120.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727535AbgBTSbD (ORCPT ); Thu, 20 Feb 2020 13:31:03 -0500 Received: from pps.filterd (aserp2120.oracle.com [127.0.0.1]) by aserp2120.oracle.com (8.16.0.42/8.16.0.42) with SMTP id 01KIRNOh171519; Thu, 20 Feb 2020 18:30:58 GMT DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=oracle.com; h=content-type : mime-version : subject : from : in-reply-to : date : cc : content-transfer-encoding : message-id : references : to; s=corp-2020-01-29; bh=HPMTPrlzxuePQNHkJApc24LTsV8QV+cJm9y+1ag6Kkk=; b=BT7CSxXN6EvJJCMlOn1UnpNIVRtkuzKuvLDXHRd/IqID0q/K6jPfeKk3KROHSwCS/JEi X+tnMHlE8Cl9KeUrLqsg1kFXmHR9HMg2DXWpkcdqqR/76RW83+uLlOSlKtjvxBWE7KgI MRb6fYYRElBz5cpH2fGUbPgOXyIURqtZfRXEtDUk1a5TRIFGjbln/YCYx5IbX7arfOp9 uAUFCB60ODSG+zxS5AsjeMME7EiIffpTPS9g4tWpQ3B6rZI38Uc4pb6urjLLNqkem7k5 mBV50Njh42HI70IOkdIHsQ4HpYpRwjH/b+lthMaeElwLf+4BVBRxqE5xkzolYRcCnqQr nA== Received: from aserp3020.oracle.com (aserp3020.oracle.com [141.146.126.70]) by aserp2120.oracle.com with ESMTP id 2y8udkknaw-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Thu, 20 Feb 2020 18:30:57 +0000 Received: from pps.filterd (aserp3020.oracle.com [127.0.0.1]) by aserp3020.oracle.com (8.16.0.42/8.16.0.42) with SMTP id 01KIS8fY110699; Thu, 20 Feb 2020 18:30:57 GMT Received: from aserv0122.oracle.com (aserv0122.oracle.com [141.146.126.236]) by aserp3020.oracle.com with ESMTP id 2y8udenrc1-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Thu, 20 Feb 2020 18:30:57 +0000 Received: from abhmp0011.oracle.com (abhmp0011.oracle.com [141.146.116.17]) by aserv0122.oracle.com (8.14.4/8.14.4) with ESMTP id 01KIUuKT031009; Thu, 20 Feb 2020 18:30:56 GMT Received: from anon-dhcp-153.1015granger.net (/68.61.232.219) by default (Oracle Beehive Gateway v4.0) with ESMTP ; Thu, 20 Feb 2020 10:30:56 -0800 Content-Type: text/plain; charset=us-ascii Mime-Version: 1.0 (Mac OS X Mail 12.4 \(3445.104.11\)) Subject: Re: [PATCH v2 4/6] NFS: Add READ_PLUS data segment support From: Chuck Lever In-Reply-To: Date: Thu, 20 Feb 2020 13:30:55 -0500 Cc: Trond.Myklebust@hammerspace.com, Linux NFS Mailing List Content-Transfer-Encoding: quoted-printable Message-Id: <93E71BBA-029B-44AE-B580-0332E157D0A2@oracle.com> References: <20200214211227.407836-1-Anna.Schumaker@Netapp.com> <20200214211227.407836-5-Anna.Schumaker@Netapp.com> <7621b7d84295dd3086e2036f8cb389ceb47cbbc2.camel@gmail.com> To: Anna Schumaker X-Mailer: Apple Mail (2.3445.104.11) X-Proofpoint-Virus-Version: vendor=nai engine=6000 definitions=9537 signatures=668685 X-Proofpoint-Spam-Details: rule=notspam policy=default score=0 spamscore=0 mlxlogscore=999 phishscore=0 suspectscore=0 mlxscore=0 malwarescore=0 adultscore=0 bulkscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2001150001 definitions=main-2002200136 X-Proofpoint-Virus-Version: vendor=nai engine=6000 definitions=9537 signatures=668685 X-Proofpoint-Spam-Details: rule=notspam policy=default score=0 lowpriorityscore=0 suspectscore=0 spamscore=0 priorityscore=1501 adultscore=0 mlxscore=0 clxscore=1015 malwarescore=0 mlxlogscore=999 phishscore=0 impostorscore=0 bulkscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2001150001 definitions=main-2002200136 Sender: linux-nfs-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-nfs@vger.kernel.org > On Feb 20, 2020, at 1:28 PM, Anna Schumaker = wrote: >=20 > On Thu, 2020-02-20 at 09:55 -0500, Chuck Lever wrote: >>> On Feb 20, 2020, at 9:42 AM, Anna Schumaker = >>> wrote: >>>=20 >>> On Fri, 2020-02-14 at 17:28 -0500, Chuck Lever wrote: >>>>> On Feb 14, 2020, at 4:12 PM, schumaker.anna@gmail.com wrote: >>>>>=20 >>>>> From: Anna Schumaker >>>>>=20 >>>>> This patch adds client support for decoding a single = NFS4_CONTENT_DATA >>>>> segment returned by the server. This is the simplest = implementation >>>>> possible, since it does not account for any hole segments in the = reply. >>>>>=20 >>>>> Signed-off-by: Anna Schumaker >>>>> --- >>>>> fs/nfs/nfs42xdr.c | 138 = ++++++++++++++++++++++++++++++++++++++ >>>>> fs/nfs/nfs4proc.c | 43 +++++++++++- >>>>> fs/nfs/nfs4xdr.c | 1 + >>>>> include/linux/nfs4.h | 2 +- >>>>> include/linux/nfs_fs_sb.h | 1 + >>>>> include/linux/nfs_xdr.h | 2 +- >>>>> 6 files changed, 182 insertions(+), 5 deletions(-) >>>>>=20 >>>>> diff --git a/fs/nfs/nfs42xdr.c b/fs/nfs/nfs42xdr.c >>>>> index c03f3246d6c5..bf118ecabe2c 100644 >>>>> --- a/fs/nfs/nfs42xdr.c >>>>> +++ b/fs/nfs/nfs42xdr.c >>>>> @@ -45,6 +45,15 @@ >>>>> #define encode_deallocate_maxsz (op_encode_hdr_maxsz + \ >>>>> encode_fallocate_maxsz) >>>>> #define decode_deallocate_maxsz (op_decode_hdr_maxsz) >>>>> +#define encode_read_plus_maxsz (op_encode_hdr_maxsz + \ >>>>> + encode_stateid_maxsz + 3) >>>>> +#define NFS42_READ_PLUS_SEGMENT_SIZE (1 /* data_content4 */ + = \ >>>>> + 2 /* data_info4.di_offset */ + >>>>> \ >>>>> + 2 /* data_info4.di_length */) >>>>> +#define decode_read_plus_maxsz (op_decode_hdr_maxsz + \ >>>>> + 1 /* rpr_eof */ + \ >>>>> + 1 /* rpr_contents count */ + \ >>>>> + NFS42_READ_PLUS_SEGMENT_SIZE) >>>>> #define encode_seek_maxsz (op_encode_hdr_maxsz + \ >>>>> encode_stateid_maxsz + \ >>>>> 2 /* offset */ + \ >>>>> @@ -128,6 +137,14 @@ >>>>> decode_putfh_maxsz + \ >>>>> decode_deallocate_maxsz + \ >>>>> decode_getattr_maxsz) >>>>> +#define NFS4_enc_read_plus_sz = (compound_encode_hdr_maxsz + \ >>>>> + encode_sequence_maxsz + \ >>>>> + encode_putfh_maxsz + \ >>>>> + encode_read_plus_maxsz) >>>>> +#define NFS4_dec_read_plus_sz = (compound_decode_hdr_maxsz + \ >>>>> + decode_sequence_maxsz + \ >>>>> + decode_putfh_maxsz + \ >>>>> + decode_read_plus_maxsz) >>>>> #define NFS4_enc_seek_sz (compound_encode_hdr_maxsz + \ >>>>> encode_sequence_maxsz + \ >>>>> encode_putfh_maxsz + \ >>>>> @@ -252,6 +269,16 @@ static void encode_deallocate(struct = xdr_stream >>>>> *xdr, >>>>> encode_fallocate(xdr, args); >>>>> } >>>>>=20 >>>>> +static void encode_read_plus(struct xdr_stream *xdr, >>>>> + const struct nfs_pgio_args *args, >>>>> + struct compound_hdr *hdr) >>>>> +{ >>>>> + encode_op_hdr(xdr, OP_READ_PLUS, decode_read_plus_maxsz, hdr); >>>>> + encode_nfs4_stateid(xdr, &args->stateid); >>>>> + encode_uint64(xdr, args->offset); >>>>> + encode_uint32(xdr, args->count); >>>>> +} >>>>> + >>>>> static void encode_seek(struct xdr_stream *xdr, >>>>> const struct nfs42_seek_args *args, >>>>> struct compound_hdr *hdr) >>>>> @@ -446,6 +473,29 @@ static void nfs4_xdr_enc_deallocate(struct = rpc_rqst >>>>> *req, >>>>> encode_nops(&hdr); >>>>> } >>>>>=20 >>>>> +/* >>>>> + * Encode READ_PLUS request >>>>> + */ >>>>> +static void nfs4_xdr_enc_read_plus(struct rpc_rqst *req, >>>>> + struct xdr_stream *xdr, >>>>> + const void *data) >>>>> +{ >>>>> + const struct nfs_pgio_args *args =3D data; >>>>> + struct compound_hdr hdr =3D { >>>>> + .minorversion =3D = nfs4_xdr_minorversion(&args->seq_args), >>>>> + }; >>>>> + >>>>> + encode_compound_hdr(xdr, req, &hdr); >>>>> + encode_sequence(xdr, &args->seq_args, &hdr); >>>>> + encode_putfh(xdr, args->fh, &hdr); >>>>> + encode_read_plus(xdr, args, &hdr); >>>>> + >>>>> + rpc_prepare_reply_pages(req, args->pages, args->pgbase, >>>>> + args->count, hdr.replen); >>>>> + req->rq_rcv_buf.flags |=3D XDRBUF_READ; >>>>=20 >>>> IMO this line is incorrect. >>>=20 >>> You're right, this line causes problems for RDMA with READ_PLUS. I = added it >>> to >>> match how the other xdr read encoders were set up >>=20 >> Ja, I think just removing that line should be sufficient. >> Better would be replacing it with a comment explaining >> why this encoder does not set XDRBUF_READ. :-) >>=20 >>=20 >>>> RFC 8267 Section 6.1 does not list any part of the result of = READ_PLUS >>>> as DDP-eligible. There's no way for a client to know how to set up >>>> Write chunks, unless it knows exactly where the file's holes are in >>>> advance. Even then... racy. >>>>=20 >>>> Just curious, have you tried READ_PLUS with proto=3Drdma ? >>>=20 >>> I haven't done in-depth performance testing, but I have been able to = run it. >>=20 >> We should figure out whether that will have a regressive >> impact on NFS/RDMA workloads. I expect that it will, but >> the client can always set up the Reply chunk so that the >> READ payload fits precisely in an RDMA segment that lines >> up with page cache pages. That mitigates some impact. >>=20 >> If your patch set already changes NFSv4.2 mounts to always >> use READ_PLUS in place of READ, it might be prudent for the >> "proto=3Drdma" mount option to also set "noreadplus", at least >> for the time being. >=20 > I can make this change. >=20 >>=20 >> The down-side here is that would make NFSv4.2 on RDMA >> unable to recognize holes in files the same way as it >> does on TCP, and that's a pretty significant variation >> in behavior. Does "noreadplus" even deal with that? >=20 > Setting "noreadplus" just causes the client to use the READ operation = instead, > so there should be no difference between v4.1 and v4.2 if the option = is set. My concern is the difference between NFSv4.2 with noreadplus and NFSv4.2 with readplus. The former is not able to detect holes in files on the server, but the latter is. Is that worth mentioning in the man page, or in release notes when NFSv4.2 becomes the default? > Anna >=20 >>=20 >>=20 >>> Anna >>>=20 >>>>=20 >>>>> + encode_nops(&hdr); >>>>> +} >>>>> + >>>>> /* >>>>> * Encode SEEK request >>>>> */ >>>>> @@ -694,6 +744,67 @@ static int decode_deallocate(struct = xdr_stream >>>>> *xdr, >>>>> struct nfs42_falloc_res *re >>>>> return decode_op_hdr(xdr, OP_DEALLOCATE); >>>>> } >>>>>=20 >>>>> +static uint32_t decode_read_plus_data(struct xdr_stream *xdr, = struct >>>>> nfs_pgio_res *res, >>>>> + uint32_t *eof) >>>>> +{ >>>>> + __be32 *p; >>>>> + uint32_t count, recvd; >>>>> + uint64_t offset; >>>>> + >>>>> + p =3D xdr_inline_decode(xdr, 8 + 4); >>>>> + if (unlikely(!p)) >>>>> + return -EIO; >>>>> + >>>>> + p =3D xdr_decode_hyper(p, &offset); >>>>> + count =3D be32_to_cpup(p); >>>>> + if (count =3D=3D 0) >>>>> + return 0; >>>>> + >>>>> + recvd =3D xdr_read_pages(xdr, count); >>>>> + if (count > recvd) { >>>>> + dprintk("NFS: server cheating in read reply: " >>>>> + "count %u > recvd %u\n", count, recvd); >>>>> + count =3D recvd; >>>>> + *eof =3D 0; >>>>> + } >>>>> + >>>>> + return count; >>>>> +} >>>>> + >>>>> +static int decode_read_plus(struct xdr_stream *xdr, struct = nfs_pgio_res >>>>> *res) >>>>> +{ >>>>> + __be32 *p; >>>>> + uint32_t count, eof, segments, type; >>>>> + int status; >>>>> + >>>>> + status =3D decode_op_hdr(xdr, OP_READ_PLUS); >>>>> + if (status) >>>>> + return status; >>>>> + >>>>> + p =3D xdr_inline_decode(xdr, 4 + 4); >>>>> + if (unlikely(!p)) >>>>> + return -EIO; >>>>> + >>>>> + eof =3D be32_to_cpup(p++); >>>>> + segments =3D be32_to_cpup(p++); >>>>> + if (segments =3D=3D 0) >>>>> + return 0; >>>>> + >>>>> + p =3D xdr_inline_decode(xdr, 4); >>>>> + if (unlikely(!p)) >>>>> + return -EIO; >>>>> + >>>>> + type =3D be32_to_cpup(p++); >>>>> + if (type =3D=3D NFS4_CONTENT_DATA) >>>>> + count =3D decode_read_plus_data(xdr, res, &eof); >>>>> + else >>>>> + return -EINVAL; >>>>> + >>>>> + res->eof =3D eof; >>>>> + res->count =3D count; >>>>> + return 0; >>>>> +} >>>>> + >>>>> static int decode_seek(struct xdr_stream *xdr, struct = nfs42_seek_res >>>>> *res) >>>>> { >>>>> int status; >>>>> @@ -870,6 +981,33 @@ static int nfs4_xdr_dec_deallocate(struct = rpc_rqst >>>>> *rqstp, >>>>> return status; >>>>> } >>>>>=20 >>>>> +/* >>>>> + * Decode READ_PLUS request >>>>> + */ >>>>> +static int nfs4_xdr_dec_read_plus(struct rpc_rqst *rqstp, >>>>> + struct xdr_stream *xdr, >>>>> + void *data) >>>>> +{ >>>>> + struct nfs_pgio_res *res =3D data; >>>>> + struct compound_hdr hdr; >>>>> + int status; >>>>> + >>>>> + status =3D decode_compound_hdr(xdr, &hdr); >>>>> + if (status) >>>>> + goto out; >>>>> + status =3D decode_sequence(xdr, &res->seq_res, rqstp); >>>>> + if (status) >>>>> + goto out; >>>>> + status =3D decode_putfh(xdr); >>>>> + if (status) >>>>> + goto out; >>>>> + status =3D decode_read_plus(xdr, res); >>>>> + if (!status) >>>>> + status =3D res->count; >>>>> +out: >>>>> + return status; >>>>> +} >>>>> + >>>>> /* >>>>> * Decode SEEK request >>>>> */ >>>>> diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c >>>>> index 95d07a3dc5d1..ed3ec8c36273 100644 >>>>> --- a/fs/nfs/nfs4proc.c >>>>> +++ b/fs/nfs/nfs4proc.c >>>>> @@ -69,6 +69,10 @@ >>>>>=20 >>>>> #include "nfs4trace.h" >>>>>=20 >>>>> +#ifdef CONFIG_NFS_V4_2 >>>>> +#include "nfs42.h" >>>>> +#endif /* CONFIG_NFS_V4_2 */ >>>>> + >>>>> #define NFSDBG_FACILITY NFSDBG_PROC >>>>>=20 >>>>> #define NFS4_BITMASK_SZ 3 >>>>> @@ -5199,28 +5203,60 @@ static bool = nfs4_read_stateid_changed(struct >>>>> rpc_task *task, >>>>> return true; >>>>> } >>>>>=20 >>>>> +static bool nfs4_read_plus_not_supported(struct rpc_task *task, >>>>> + struct nfs_pgio_header *hdr) >>>>> +{ >>>>> + struct nfs_server *server =3D NFS_SERVER(hdr->inode); >>>>> + struct rpc_message *msg =3D &task->tk_msg; >>>>> + >>>>> + if (msg->rpc_proc =3D=3D = &nfs4_procedures[NFSPROC4_CLNT_READ_PLUS] >>>>> && >>>>> + server->caps & NFS_CAP_READ_PLUS && task->tk_status =3D=3D >>>>> -ENOTSUPP) { >>>>> + server->caps &=3D ~NFS_CAP_READ_PLUS; >>>>> + msg->rpc_proc =3D &nfs4_procedures[NFSPROC4_CLNT_READ]; >>>>> + rpc_restart_call_prepare(task); >>>>> + return true; >>>>> + } >>>>> + return false; >>>>> +} >>>>> + >>>>> static int nfs4_read_done(struct rpc_task *task, struct = nfs_pgio_header >>>>> *hdr) >>>>> { >>>>> - >>>>> dprintk("--> %s\n", __func__); >>>>>=20 >>>>> if (!nfs4_sequence_done(task, &hdr->res.seq_res)) >>>>> return -EAGAIN; >>>>> if (nfs4_read_stateid_changed(task, &hdr->args)) >>>>> return -EAGAIN; >>>>> + if (nfs4_read_plus_not_supported(task, hdr)) >>>>> + return -EAGAIN; >>>>> if (task->tk_status > 0) >>>>> nfs_invalidate_atime(hdr->inode); >>>>> return hdr->pgio_done_cb ? hdr->pgio_done_cb(task, hdr) : >>>>> nfs4_read_done_cb(task, hdr); >>>>> } >>>>>=20 >>>>> +#ifdef CONFIG_NFS_V4_2 >>>>> +static void nfs42_read_plus_support(struct nfs_server *server, = struct >>>>> rpc_message *msg) >>>>> +{ >>>>> + if (server->caps & NFS_CAP_READ_PLUS) >>>>> + msg->rpc_proc =3D >>>>> &nfs4_procedures[NFSPROC4_CLNT_READ_PLUS]; >>>>> + else >>>>> + msg->rpc_proc =3D &nfs4_procedures[NFSPROC4_CLNT_READ]; >>>>> +} >>>>> +#else >>>>> +static void nfs42_read_plus_support(struct nfs_server *server, = struct >>>>> rpc_message *msg) >>>>> +{ >>>>> + msg->rpc_proc =3D &nfs4_procedures[NFSPROC4_CLNT_READ]; >>>>> +} >>>>> +#endif /* CONFIG_NFS_V4_2 */ >>>>> + >>>>> static void nfs4_proc_read_setup(struct nfs_pgio_header *hdr, >>>>> struct rpc_message *msg) >>>>> { >>>>> hdr->timestamp =3D jiffies; >>>>> if (!hdr->pgio_done_cb) >>>>> hdr->pgio_done_cb =3D nfs4_read_done_cb; >>>>> - msg->rpc_proc =3D &nfs4_procedures[NFSPROC4_CLNT_READ]; >>>>> + nfs42_read_plus_support(NFS_SERVER(hdr->inode), msg); >>>>> nfs4_init_sequence(&hdr->args.seq_args, &hdr->res.seq_res, 0, >>>>> 0); >>>>> } >>>>>=20 >>>>> @@ -9970,7 +10006,8 @@ static const struct nfs4_minor_version_ops >>>>> nfs_v4_2_minor_ops =3D { >>>>> | NFS_CAP_SEEK >>>>> | NFS_CAP_LAYOUTSTATS >>>>> | NFS_CAP_CLONE >>>>> - | NFS_CAP_LAYOUTERROR, >>>>> + | NFS_CAP_LAYOUTERROR >>>>> + | NFS_CAP_READ_PLUS, >>>>> .init_client =3D nfs41_init_client, >>>>> .shutdown_client =3D nfs41_shutdown_client, >>>>> .match_stateid =3D nfs41_match_stateid, >>>>> diff --git a/fs/nfs/nfs4xdr.c b/fs/nfs/nfs4xdr.c >>>>> index 47817ef0aadb..68b2917d0537 100644 >>>>> --- a/fs/nfs/nfs4xdr.c >>>>> +++ b/fs/nfs/nfs4xdr.c >>>>> @@ -7584,6 +7584,7 @@ const struct rpc_procinfo nfs4_procedures[] = =3D { >>>>> PROC42(COPY_NOTIFY, enc_copy_notify, dec_copy_notify) >>>>> , >>>>> PROC(LOOKUPP, enc_lookupp, dec_lookupp), >>>>> PROC42(LAYOUTERROR, enc_layouterror, dec_layouterror) >>>>> , >>>>> + PROC42(READ_PLUS, enc_read_plus, dec_read_plus), >>>>> }; >>>>>=20 >>>>> static unsigned int = nfs_version4_counts[ARRAY_SIZE(nfs4_procedures)]; >>>>> diff --git a/include/linux/nfs4.h b/include/linux/nfs4.h >>>>> index 82d8fb422092..c1eeef52545c 100644 >>>>> --- a/include/linux/nfs4.h >>>>> +++ b/include/linux/nfs4.h >>>>> @@ -540,8 +540,8 @@ enum { >>>>>=20 >>>>> NFSPROC4_CLNT_LOOKUPP, >>>>> NFSPROC4_CLNT_LAYOUTERROR, >>>>> - >>>>> NFSPROC4_CLNT_COPY_NOTIFY, >>>>> + NFSPROC4_CLNT_READ_PLUS, >>>>> }; >>>>>=20 >>>>> /* nfs41 types */ >>>>> diff --git a/include/linux/nfs_fs_sb.h b/include/linux/nfs_fs_sb.h >>>>> index 465fa98258a3..11248c5a7b24 100644 >>>>> --- a/include/linux/nfs_fs_sb.h >>>>> +++ b/include/linux/nfs_fs_sb.h >>>>> @@ -281,5 +281,6 @@ struct nfs_server { >>>>> #define NFS_CAP_OFFLOAD_CANCEL (1U << 25) >>>>> #define NFS_CAP_LAYOUTERROR (1U << 26) >>>>> #define NFS_CAP_COPY_NOTIFY (1U << 27) >>>>> +#define NFS_CAP_READ_PLUS (1U << 28) >>>>>=20 >>>>> #endif >>>>> diff --git a/include/linux/nfs_xdr.h b/include/linux/nfs_xdr.h >>>>> index 94c77ed55ce1..8efbf3d8b263 100644 >>>>> --- a/include/linux/nfs_xdr.h >>>>> +++ b/include/linux/nfs_xdr.h >>>>> @@ -655,7 +655,7 @@ struct nfs_pgio_args { >>>>> struct nfs_pgio_res { >>>>> struct nfs4_sequence_res seq_res; >>>>> struct nfs_fattr * fattr; >>>>> - __u32 count; >>>>> + __u64 count; >>>>> __u32 op_status; >>>>> union { >>>>> struct { >>>>> --=20 >>>>> 2.25.0 >>>>>=20 >>>>=20 >>>> -- >>>> Chuck Lever >>=20 >> -- >> Chuck Lever -- Chuck Lever