Received: by 2002:a05:6a11:4021:0:0:0:0 with SMTP id ky33csp151888pxb; Tue, 28 Sep 2021 17:59:56 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwQ5KhT+iRIPEVTsgoaYdf0S4k9SgLXyQwUe1uSe2m+UbD0a5Bfki7oCMnTmmcnRQoVQhAH X-Received: by 2002:aa7:9e9a:0:b0:43e:d9b:cd93 with SMTP id p26-20020aa79e9a000000b0043e0d9bcd93mr8235054pfq.50.1632877195958; Tue, 28 Sep 2021 17:59:55 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1632877195; cv=none; d=google.com; s=arc-20160816; b=1DATNxh4Lw/vjc+efyaysVgqivHlzMs/IdYfNb98W39kShi/ZX2a8QahKc/8KYorWD rF2l8DZo3tE2BRoOEDu/D+zg5NABCKsNFSPxiqe5YnKKndLVmUw8vFlOMWxqxPwHoaVI g7xKKlfzqQnqhqlo9LwB3307qnMLErqjPaMwfui+2pB41V2llDUwXx4SM3z3wQ/hMAyG LEp4NKR9S51pZmkTMuihuo0ncJ1dKsDROwXzFLz2kPij9hdvfjC2NIDcib7WB9rW6pTn +DfwXYlL3bvTXULcJHATolGtxHWCJiJbeQV5Hs2ucTB4r3ugpCjPHwvWlmKqjoHsTu/c GKPg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :dkim-signature; bh=Rlf9+AHVeG2NbrCRR3lZJydSlIZG82P50DkfosDID6E=; b=ttgotGTWJk4I9WAMx78UTXXlfvBrCqEZHqygZpbWEIyEE/FgNkznkemcQ5tcN1V8Ng NmeTDF4H0pQsEvPZxp4Pv2bajyMnmlbCPNV82BqwqK7A6/2cFQQ/m6ggQ2cve15Sm9/w b0HUKoMV2PVKzdQ5bwaJB+wcNLDmcLqG4RU0JBxzrl1XZho9F6CYYxR24wxBVT7UpjQY gzlcb/zlKgAdSK2jkv5BT3why6dq0BM2xRbMwGh8Pfh6k6blUSy5mAaOhfuuOYr/h57e Exz4OPNKFidifGHyqmoKsxp7bXp9/zaXbhayYp4DPx1qAyVMviL/iDQWJlgZe+ova8A2 3tww== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@oracle.com header.s=corp-2021-07-09 header.b=LNZFRsRc; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=oracle.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id q34si5069327pja.107.2021.09.28.17.59.25; Tue, 28 Sep 2021 17:59:55 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@oracle.com header.s=corp-2021-07-09 header.b=LNZFRsRc; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=oracle.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S243559AbhI2A6a (ORCPT + 99 others); Tue, 28 Sep 2021 20:58:30 -0400 Received: from mx0a-00069f02.pphosted.com ([205.220.165.32]:43038 "EHLO mx0a-00069f02.pphosted.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S243534AbhI2A63 (ORCPT ); Tue, 28 Sep 2021 20:58:29 -0400 Received: from pps.filterd (m0246629.ppops.net [127.0.0.1]) by mx0b-00069f02.pphosted.com (8.16.1.2/8.16.1.2) with SMTP id 18T0uGE3027809; Wed, 29 Sep 2021 00:56:48 GMT DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=oracle.com; h=from : to : cc : subject : date : message-id : in-reply-to : references : mime-version : content-transfer-encoding; s=corp-2021-07-09; bh=Rlf9+AHVeG2NbrCRR3lZJydSlIZG82P50DkfosDID6E=; b=LNZFRsRcnB81YFQeuUd1a1B0uI/mcQTRWe0Y1svVXvCYrdn4AoBSHy7PCcP1JcLT2zfC NgnGR6Wvv4SOzbk0lpmkuWpKR8BT2fjOY471aTAJEqqUU7PLkh99frn59qdgX4/Pjwwg m+cyBPZSURjGNb/VfWDNbcbkbWAV346EybLqC37AEVH3nauRJAcmoEFGZtQ6ttt84K6f E9SgXdlnlpM9lu+bR1/8uqmU1+zcKPWYdUEqDIxc7ZXpr5ynAYJBMNUkRkyJMRXCwxz3 /Py6j0MDD/sVlBQQRlWoFb/BMnhrocfEVWU6bSBA50GmCiBd6QX4fOzb50YSi5sUL21O QA== Received: from userp3030.oracle.com (userp3030.oracle.com [156.151.31.80]) by mx0b-00069f02.pphosted.com with ESMTP id 3bbj90t613-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Wed, 29 Sep 2021 00:56:48 +0000 Received: from pps.filterd (userp3030.oracle.com [127.0.0.1]) by userp3030.oracle.com (8.16.0.42/8.16.0.42) with SMTP id 18T0uPvk052328; Wed, 29 Sep 2021 00:56:47 GMT Received: from pps.reinject (localhost [127.0.0.1]) by userp3030.oracle.com with ESMTP id 3bc3bj5mc3-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Wed, 29 Sep 2021 00:56:47 +0000 Received: from userp3030.oracle.com (userp3030.oracle.com [127.0.0.1]) by pps.reinject (8.16.0.36/8.16.0.36) with SMTP id 18T0ujKX053426; Wed, 29 Sep 2021 00:56:46 GMT Received: from userp3020.oracle.com (ksplice-shell2.us.oracle.com [10.152.118.36]) by userp3030.oracle.com with ESMTP id 3bc3bj5mau-3; Wed, 29 Sep 2021 00:56:46 +0000 From: Dai Ngo To: bfields@fieldses.org Cc: chuck.lever@oracle.com, linux-nfs@vger.kernel.org, linux-fsdevel@vger.kernel.org Subject: [PATCH RFC v5 2/2] nfsd: Initial implementation of NFSv4 Courteous Server Date: Tue, 28 Sep 2021 20:56:41 -0400 Message-Id: <20210929005641.60861-3-dai.ngo@oracle.com> X-Mailer: git-send-email 2.20.1.1226.g1595ea5.dirty In-Reply-To: <20210929005641.60861-1-dai.ngo@oracle.com> References: <20210929005641.60861-1-dai.ngo@oracle.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Proofpoint-GUID: LK-kkEp3VW9vbbiB1jOxlHj2CRPeRfBm X-Proofpoint-ORIG-GUID: LK-kkEp3VW9vbbiB1jOxlHj2CRPeRfBm Precedence: bulk List-ID: X-Mailing-List: linux-nfs@vger.kernel.org Currently an NFSv4 client must maintain its lease by using the at least one of the state tokens or if nothing else, by issuing a RENEW (4.0), or a singleton SEQUENCE (4.1) at least once during each lease period. If the client fails to renew the lease, for any reason, the Linux server expunges the state tokens immediately upon detection of the "failure to renew the lease" condition and begins returning NFS4ERR_EXPIRED if the client should reconnect and attempt to use the (now) expired state. The default lease period for the Linux server is 90 seconds. The typical client cuts that in half and will issue a lease renewing operation every 45 seconds. The 90 second lease period is very short considering the potential for moderately long term network partitions. A network partition refers to any loss of network connectivity between the NFS client and the NFS server, regardless of its root cause. This includes NIC failures, NIC driver bugs, network misconfigurations & administrative errors, routers & switches crashing and/or having software updates applied, even down to cables being physically pulled. In most cases, these network failures are transient, although the duration is unknown. A server which does not immediately expunge the state on lease expiration is known as a Courteous Server. A Courteous Server continues to recognize previously generated state tokens as valid until conflict arises between the expired state and the requests from another client, or the server reboots. The initial implementation of the Courteous Server will do the following: . when the laundromat thread detects an expired client and if that client still has established states on the Linux server and there is no waiters for the client's locks then mark the client as a COURTESY_CLIENT and skip destroying the client and all its states, otherwise destroy the client as usual. . detects conflict of OPEN request with COURTESY_CLIENT, destroys the expired client and all its states, skips the delegation recall then allows the conflicting request to succeed. . detects conflict of LOCK/LOCKT, NLM LOCK and TEST, and local locks requests with COURTESY_CLIENT, destroys the expired client and all its states then allows the conflicting request to succeed. Signed-off-by: Dai Ngo --- fs/nfsd/nfs4state.c | 269 +++++++++++++++++++++++++++++++++++++++++++++++++++- fs/nfsd/state.h | 3 + 2 files changed, 269 insertions(+), 3 deletions(-) diff --git a/fs/nfsd/nfs4state.c b/fs/nfsd/nfs4state.c index 3f4027a5de88..f6d835d6e99b 100644 --- a/fs/nfsd/nfs4state.c +++ b/fs/nfsd/nfs4state.c @@ -125,6 +125,11 @@ static void free_session(struct nfsd4_session *); static const struct nfsd4_callback_ops nfsd4_cb_recall_ops; static const struct nfsd4_callback_ops nfsd4_cb_notify_lock_ops; +static struct workqueue_struct *laundry_wq; +static void laundromat_main(struct work_struct *); + +static int courtesy_client_expiry = (24 * 60 * 60); /* in secs */ + static bool is_session_dead(struct nfsd4_session *ses) { return ses->se_flags & NFS4_SESSION_DEAD; @@ -172,6 +177,7 @@ renew_client_locked(struct nfs4_client *clp) list_move_tail(&clp->cl_lru, &nn->client_lru); clp->cl_time = ktime_get_boottime_seconds(); + clear_bit(NFSD4_COURTESY_CLIENT, &clp->cl_flags); } static void put_client_renew_locked(struct nfs4_client *clp) @@ -2389,6 +2395,10 @@ static int client_info_show(struct seq_file *m, void *v) seq_puts(m, "status: confirmed\n"); else seq_puts(m, "status: unconfirmed\n"); + seq_printf(m, "courtesy client: %s\n", + test_bit(NFSD4_COURTESY_CLIENT, &clp->cl_flags) ? "yes" : "no"); + seq_printf(m, "seconds from last renew: %lld\n", + ktime_get_boottime_seconds() - clp->cl_time); seq_printf(m, "name: "); seq_quote_mem(m, clp->cl_name.data, clp->cl_name.len); seq_printf(m, "\nminor version: %d\n", clp->cl_minorversion); @@ -4662,6 +4672,33 @@ static void nfsd_break_one_deleg(struct nfs4_delegation *dp) nfsd4_run_cb(&dp->dl_recall); } +/* + * This function is called when a file is opened and there is a + * delegation conflict with another client. If the other client + * is a courtesy client then kick start the laundromat to destroy + * it. + */ +static bool +nfsd_check_courtesy_client(struct nfs4_delegation *dp) +{ + struct svc_rqst *rqst; + struct nfs4_client *clp = dp->dl_recall.cb_clp; + struct nfsd_net *nn = net_generic(clp->net, nfsd_net_id); + + if (!i_am_nfsd()) + goto out; + rqst = kthread_data(current); + if (rqst->rq_prog != NFS_PROGRAM || rqst->rq_vers < 4) + return false; +out: + if (test_bit(NFSD4_COURTESY_CLIENT, &clp->cl_flags)) { + set_bit(NFSD4_DESTROY_COURTESY_CLIENT, &clp->cl_flags); + mod_delayed_work(laundry_wq, &nn->laundromat_work, 0); + return true; + } + return false; +} + /* Called from break_lease() with i_lock held. */ static bool nfsd_break_deleg_cb(struct file_lock *fl) @@ -4670,6 +4707,8 @@ nfsd_break_deleg_cb(struct file_lock *fl) struct nfs4_delegation *dp = (struct nfs4_delegation *)fl->fl_owner; struct nfs4_file *fp = dp->dl_stid.sc_file; + if (nfsd_check_courtesy_client(dp)) + return false; trace_nfsd_cb_recall(&dp->dl_stid); /* @@ -4912,6 +4951,104 @@ nfsd4_truncate(struct svc_rqst *rqstp, struct svc_fh *fh, return nfsd_setattr(rqstp, fh, &iattr, 0, (time64_t)0); } +static bool +__nfs4_check_deny_bmap(struct nfs4_file *fp, struct nfs4_ol_stateid *stp, + u32 access, bool share_access) +{ + if (share_access) { + if (!stp->st_deny_bmap) + return false; + + if ((stp->st_deny_bmap & (1 << NFS4_SHARE_DENY_BOTH)) || + (access & NFS4_SHARE_ACCESS_READ && + stp->st_deny_bmap & (1 << NFS4_SHARE_DENY_READ)) || + (access & NFS4_SHARE_ACCESS_WRITE && + stp->st_deny_bmap & (1 << NFS4_SHARE_DENY_WRITE))) { + return true; + } + return false; + } + if ((access & NFS4_SHARE_DENY_BOTH) || + (access & NFS4_SHARE_DENY_READ && + stp->st_access_bmap & (1 << NFS4_SHARE_ACCESS_READ)) || + (access & NFS4_SHARE_DENY_WRITE && + stp->st_access_bmap & (1 << NFS4_SHARE_ACCESS_WRITE))) { + return true; + } + return false; +} + +/* + * access: access mode if share_access is true else is deny mode + */ +static bool +nfs4_check_deny_bmap(struct nfs4_client *clp, struct nfs4_file *fp, + u32 access, bool share_access) +{ + int i; + struct nfs4_openowner *oo; + struct nfs4_stateowner *so, *tmp; + struct nfs4_ol_stateid *stp, *stmp; + + spin_lock(&clp->cl_lock); + for (i = 0; i < OWNER_HASH_SIZE; i++) { + list_for_each_entry_safe(so, tmp, &clp->cl_ownerstr_hashtbl[i], + so_strhash) { + if (!so->so_is_open_owner) + continue; + oo = openowner(so); + list_for_each_entry_safe(stp, stmp, + &oo->oo_owner.so_stateids, st_perstateowner) { + if (__nfs4_check_deny_bmap(fp, stp, access, + share_access)) { + spin_unlock(&clp->cl_lock); + return true; + } + } + } + } + spin_unlock(&clp->cl_lock); + return false; +} + +static int +nfs4_destroy_clnts_with_sresv_conflict(struct svc_rqst *rqstp, + struct nfs4_client *clp, struct nfs4_file *fp, + u32 access, bool share_access) +{ + int cnt = 0; + struct nfs4_client *cl; + struct list_head *pos, *next, reaplist; + struct nfsd_net *nn = net_generic(SVC_NET(rqstp), nfsd_net_id); + + INIT_LIST_HEAD(&reaplist); + spin_lock(&nn->client_lock); + list_for_each_safe(pos, next, &nn->client_lru) { + cl = list_entry(pos, struct nfs4_client, cl_lru); + if (cl == clp || + (!test_bit(NFSD4_COURTESY_CLIENT, &cl->cl_flags) && + !test_bit(NFSD4_DESTROY_COURTESY_CLIENT, &cl->cl_flags))) + continue; + /* + * check CS client for deny_bmap that + * matches nfs4_file and access mode + */ + if (nfs4_check_deny_bmap(cl, fp, access, share_access)) { + if (mark_client_expired_locked(cl)) + continue; + list_add(&cl->cl_lru, &reaplist); + } + } + spin_unlock(&nn->client_lock); + list_for_each_safe(pos, next, &reaplist) { + cl = list_entry(pos, struct nfs4_client, cl_lru); + list_del_init(&cl->cl_lru); + expire_client(cl); + cnt++; + } + return cnt; +} + static __be32 nfs4_get_vfs_file(struct svc_rqst *rqstp, struct nfs4_file *fp, struct svc_fh *cur_fh, struct nfs4_ol_stateid *stp, struct nfsd4_open *open) @@ -4921,6 +5058,7 @@ static __be32 nfs4_get_vfs_file(struct svc_rqst *rqstp, struct nfs4_file *fp, int oflag = nfs4_access_to_omode(open->op_share_access); int access = nfs4_access_to_access(open->op_share_access); unsigned char old_access_bmap, old_deny_bmap; + int cnt = 0; spin_lock(&fp->fi_lock); @@ -4928,16 +5066,46 @@ static __be32 nfs4_get_vfs_file(struct svc_rqst *rqstp, struct nfs4_file *fp, * Are we trying to set a deny mode that would conflict with * current access? */ +chk_deny: status = nfs4_file_check_deny(fp, open->op_share_deny); if (status != nfs_ok) { spin_unlock(&fp->fi_lock); + if (status != nfserr_share_denied) + goto out; + /* + * search and destroy all courtesy clients that own + * openstate of nfs4_file, fp, that has st_access_bmap + * that matches open->op_share_deny + */ + cnt = nfs4_destroy_clnts_with_sresv_conflict(rqstp, + stp->st_stateowner->so_client, fp, + open->op_share_deny, false); + if (cnt) { + spin_lock(&fp->fi_lock); + goto chk_deny; + } goto out; } /* set access to the file */ +get_access: status = nfs4_file_get_access(fp, open->op_share_access); if (status != nfs_ok) { spin_unlock(&fp->fi_lock); + if (status != nfserr_share_denied) + goto out; + /* + * search and destroy all courtesy clients that own + * openstate of nfs4_file, fp, that has st_deny_bmap + * that matches open->op_share_access. + */ + cnt = nfs4_destroy_clnts_with_sresv_conflict(rqstp, + stp->st_stateowner->so_client, fp, + open->op_share_access, true); + if (cnt) { + spin_lock(&fp->fi_lock); + goto get_access; + } goto out; } @@ -5289,6 +5457,22 @@ static void nfsd4_deleg_xgrade_none_ext(struct nfsd4_open *open, */ } +static bool +nfs4_destroy_courtesy_client(struct nfs4_client *clp) +{ + struct nfsd_net *nn = net_generic(clp->net, nfsd_net_id); + + spin_lock(&nn->client_lock); + if (!test_bit(NFSD4_COURTESY_CLIENT, &clp->cl_flags) || + mark_client_expired_locked(clp)) { + spin_unlock(&nn->client_lock); + return false; + } + spin_unlock(&nn->client_lock); + expire_client(clp); + return true; +} + __be32 nfsd4_process_open2(struct svc_rqst *rqstp, struct svc_fh *current_fh, struct nfsd4_open *open) { @@ -5572,6 +5756,47 @@ static void nfsd4_ssc_expire_umount(struct nfsd_net *nn) } #endif +static +bool nfs4_anylock_conflict(struct nfs4_client *clp) +{ + int i; + struct nfs4_stateowner *so, *tmp; + struct nfs4_lockowner *lo; + struct nfs4_ol_stateid *stp; + struct nfs4_file *nf; + struct inode *ino; + struct file_lock_context *ctx; + struct file_lock *fl; + + for (i = 0; i < OWNER_HASH_SIZE; i++) { + /* scan each lock owner */ + list_for_each_entry_safe(so, tmp, &clp->cl_ownerstr_hashtbl[i], + so_strhash) { + if (so->so_is_open_owner) + continue; + + /* scan lock states of this lock owner */ + lo = lockowner(so); + list_for_each_entry(stp, &lo->lo_owner.so_stateids, + st_perstateowner) { + nf = stp->st_stid.sc_file; + ino = nf->fi_inode; + ctx = ino->i_flctx; + if (!ctx) + continue; + /* check each lock belongs to this lock state */ + list_for_each_entry(fl, &ctx->flc_posix, fl_list) { + if (fl->fl_owner != lo) + continue; + if (!list_empty(&fl->fl_blocked_requests)) + return true; + } + } + } + } + return false; +} + static time64_t nfs4_laundromat(struct nfsd_net *nn) { @@ -5587,7 +5812,9 @@ nfs4_laundromat(struct nfsd_net *nn) }; struct nfs4_cpntf_state *cps; copy_stateid_t *cps_t; + struct nfs4_stid *stid; int i; + int id = 0; if (clients_still_reclaiming(nn)) { lt.new_timeo = 0; @@ -5608,8 +5835,33 @@ nfs4_laundromat(struct nfsd_net *nn) spin_lock(&nn->client_lock); list_for_each_safe(pos, next, &nn->client_lru) { clp = list_entry(pos, struct nfs4_client, cl_lru); + if (test_bit(NFSD4_DESTROY_COURTESY_CLIENT, &clp->cl_flags)) { + clear_bit(NFSD4_COURTESY_CLIENT, &clp->cl_flags); + goto exp_client; + } + if (test_bit(NFSD4_COURTESY_CLIENT, &clp->cl_flags)) { + if (ktime_get_boottime_seconds() >= clp->courtesy_client_expiry) + goto exp_client; + /* + * after umount, v4.0 client is still + * around waiting to be expired + */ + if (clp->cl_minorversion) + continue; + } if (!state_expired(<, clp->cl_time)) break; + spin_lock(&clp->cl_lock); + stid = idr_get_next(&clp->cl_stateids, &id); + spin_unlock(&clp->cl_lock); + if (stid && !nfs4_anylock_conflict(clp)) { + /* client still has states */ + clp->courtesy_client_expiry = + ktime_get_boottime_seconds() + courtesy_client_expiry; + set_bit(NFSD4_COURTESY_CLIENT, &clp->cl_flags); + continue; + } +exp_client: if (mark_client_expired_locked(clp)) continue; list_add(&clp->cl_lru, &reaplist); @@ -5689,9 +5941,6 @@ nfs4_laundromat(struct nfsd_net *nn) return max_t(time64_t, lt.new_timeo, NFSD_LAUNDROMAT_MINTIMEOUT); } -static struct workqueue_struct *laundry_wq; -static void laundromat_main(struct work_struct *); - static void laundromat_main(struct work_struct *laundry) { @@ -6496,6 +6745,19 @@ nfs4_transform_lock_offset(struct file_lock *lock) lock->fl_end = OFFSET_MAX; } +/* return true if lock was expired else return false */ +static bool +nfsd4_fl_expire_lock(struct file_lock *fl, bool testonly) +{ + struct nfs4_lockowner *lo = (struct nfs4_lockowner *)fl->fl_owner; + struct nfs4_client *clp = lo->lo_owner.so_client; + + if (testonly) + return test_bit(NFSD4_COURTESY_CLIENT, &clp->cl_flags) ? + true : false; + return nfs4_destroy_courtesy_client(clp); +} + static fl_owner_t nfsd4_fl_get_owner(fl_owner_t owner) { @@ -6543,6 +6805,7 @@ static const struct lock_manager_operations nfsd_posix_mng_ops = { .lm_notify = nfsd4_lm_notify, .lm_get_owner = nfsd4_fl_get_owner, .lm_put_owner = nfsd4_fl_put_owner, + .lm_expire_lock = nfsd4_fl_expire_lock, }; static inline void diff --git a/fs/nfsd/state.h b/fs/nfsd/state.h index e73bdbb1634a..93e30b101578 100644 --- a/fs/nfsd/state.h +++ b/fs/nfsd/state.h @@ -345,6 +345,8 @@ struct nfs4_client { #define NFSD4_CLIENT_UPCALL_LOCK (5) /* upcall serialization */ #define NFSD4_CLIENT_CB_FLAG_MASK (1 << NFSD4_CLIENT_CB_UPDATE | \ 1 << NFSD4_CLIENT_CB_KILL) +#define NFSD4_COURTESY_CLIENT (6) /* be nice to expired client */ +#define NFSD4_DESTROY_COURTESY_CLIENT (7) unsigned long cl_flags; const struct cred *cl_cb_cred; struct rpc_clnt *cl_cb_client; @@ -385,6 +387,7 @@ struct nfs4_client { struct list_head async_copies; /* list of async copies */ spinlock_t async_lock; /* lock for async copies */ atomic_t cl_cb_inflight; /* Outstanding callbacks */ + int courtesy_client_expiry; }; /* struct nfs4_client_reset -- 2.9.5