Received: by 2002:a05:6358:111d:b0:dc:6189:e246 with SMTP id f29csp3020094rwi; Tue, 1 Nov 2022 14:57:27 -0700 (PDT) X-Google-Smtp-Source: AMsMyM4Ql32rymsLPkbUswdcoSbG2WtDE3gXuxus03nqGrcMIdaBGXPS21bc0loDrqKNVp4MMNoc X-Received: by 2002:a17:906:5a63:b0:7ad:ac42:150 with SMTP id my35-20020a1709065a6300b007adac420150mr20360984ejc.583.1667339847113; Tue, 01 Nov 2022 14:57:27 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1667339847; cv=none; d=google.com; s=arc-20160816; b=uR8qTdynCwW8A1CPBmniT61QZWvIzOCYJHPIqUZp6nxJYJoI1EWxsqMxpVrRcRJ6qA RPSKcK0GvfKuARFmBn9+wVYR1LDHT9doktvJxQPZaVHRw0zSQerOjJ0LoHRzhkGCT6p4 mW6yCjpAtQ1O92OH2g13zK+hNTgC2u115bbaebvM3RGoqX1b7y3UA/A3mYhxBrSnZTjY SaRt7X6vwm8n8ulnsSQhDEy+EpOrPHbl1TDtIqB/thYNLGqYDZzuVlXJ6gXCNvP31XRw Q4QMrNGa4wP/3S6I72viXVMW5bttNUlOlKftQCR4QZUk52yOgk7BrL87Xx67m2C1FCf7 x6lg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:mime-version:user-agent :content-transfer-encoding:references:in-reply-to:date:cc:to:from :subject:message-id:dkim-signature; bh=utpp2y8TLiQredu9gslk7c5q+j49TDxJOCVmo0hCYnY=; b=IAmFfHUzWprQwtmzuzQBVHmtIgljYlg2EqhRgMC7kACZk/0ZaQOLv0MFn3zjQ4ERLn 60YUq4GBSWbE3Zuwzn0l3MUYlE8f3Ih9QSIgiU5pLhiFoDG4eo/qJgxlOmEingiTRlIO Uzx0NZVqTu/ydJOkCHRYlUJYeeeB9zjHZXW2CxfwzoY2vEygv/f702qTS+9Zttj84LDm XVzrZMeQms5iZIK30AWjRPDWsl1f0utZciuX2a9ikG+TqK1rQ166AIEgpCSXrRfs7LYd fcRgim9iPigoaXfn+SeZtSe21DdiX5QvFDIIwkMaRe4lAqu/6OXQt3wNnQzLx6wTTxjs rNLg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=MfgMUMk1; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id du20-20020a17090772d400b00781a47397b1si14646896ejc.502.2022.11.01.14.57.02; Tue, 01 Nov 2022 14:57:27 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=MfgMUMk1; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229770AbiKAVjL (ORCPT + 99 others); Tue, 1 Nov 2022 17:39:11 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:52212 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229591AbiKAVjK (ORCPT ); Tue, 1 Nov 2022 17:39:10 -0400 Received: from ams.source.kernel.org (ams.source.kernel.org [145.40.68.75]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id E0A091CB13 for ; Tue, 1 Nov 2022 14:39:08 -0700 (PDT) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ams.source.kernel.org (Postfix) with ESMTPS id 7BDE8B81136 for ; Tue, 1 Nov 2022 21:39:07 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id CB146C433D6; Tue, 1 Nov 2022 21:39:05 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1667338746; bh=IMgmiNYGLzNmwHO3z416BJGVbGCQLf+0gzL81GIViFQ=; h=Subject:From:To:Cc:Date:In-Reply-To:References:From; b=MfgMUMk10CyhpvTCPcKrIpMvptRLMa9MUDaken4gt+m6B4Bd0aQOLGYuWlzl7S6C0 rYlLYL1Gu0kCraV6GwwAF20JGSj+jNqq81Y6/fU9NGnH9n+tLds9R+4W0Q/XqeoiT2 mBAvwoZ3G5IO41Z93XnbE6l9GehdE1ZM5YK3QFxnWDRmiB7TmoT66YnN1toqc4CgSy VFuxZTEMhRj+7uPf8tkisqca3c2/YDVks07fCVzJITyU2XXnVYL28hoxKyBedcGcqu Hk1hODrHjP1S9iP+qBYk7RuWcTMvfF2bSXyn/pVKPM8Yv1ROS4HWzsQyE0g4Q3GkVk XaPDsKeaNpbFw== Message-ID: Subject: Re: [PATCH v5 3/5] nfsd: rework refcounting in filecache From: Jeff Layton To: NeilBrown Cc: chuck.lever@oracle.com, linux-nfs@vger.kernel.org Date: Tue, 01 Nov 2022 17:39:04 -0400 In-Reply-To: <166733783854.19313.2332783814411405159@noble.neil.brown.name> References: <20221101144647.136696-1-jlayton@kernel.org> , <20221101144647.136696-4-jlayton@kernel.org> <166733783854.19313.2332783814411405159@noble.neil.brown.name> Content-Type: text/plain; charset="ISO-8859-15" Content-Transfer-Encoding: quoted-printable User-Agent: Evolution 3.44.4 (3.44.4-2.fc36) MIME-Version: 1.0 X-Spam-Status: No, score=-8.2 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_HI, SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-nfs@vger.kernel.org On Wed, 2022-11-02 at 08:23 +1100, NeilBrown wrote: > On Wed, 02 Nov 2022, Jeff Layton wrote: > > The filecache refcounting is a bit non-standard for something searchabl= e > > by RCU, in that we maintain a sentinel reference while it's hashed. Thi= s > > in turn requires that we have to do things differently in the "put" > > depending on whether its hashed, which we believe to have led to races. > >=20 > > There are other problems in here too. nfsd_file_close_inode_sync can en= d > > up freeing an nfsd_file while there are still outstanding references to > > it, and there are a number of subtle ToC/ToU races. > >=20 > > Rework the code so that the refcount is what drives the lifecycle. When > > the refcount goes to zero, then unhash and rcu free the object. > >=20 > > With this change, the LRU carries a reference. Take special care to > > deal with it when removing an entry from the list. >=20 > The refcounting and lru management all look sane here. >=20 > You need to have moved the final put (and corresponding fsync) to > different threads. I think I see you and Chuck discussing that and I > have no sense of what is "right".=A0 >=20 Yeah, this is a tough call. I get Chuck's reticence. One thing we could consider is offloading the SYNC_NONE writeback submission to a workqueue. I'm not sure though whether that's a win -- it might just add needless context switches. OTOH, that would make it fairly simple to kick off writeback when the REFERENCED flag is cleared, which would probably be the best time to do it. An entry that ends up being harvested by the LRU scanner is going to be touched by it at least twice: once to clear the REFERENCED flag, and again ~2s later to reap it. If we schedule writeback when we clear the flag then we have a pretty good indication that nothing else is going to be using it (though I think we need to clear REFERENCED even when nfsd_file_check_writeback returns true -- I'll fix that in the coming series). In any case, I'd probably like to do that sort of change in a separate series after we get the first part sorted. > But it would be nice to explain in > the comment what is being moved and why, so I could then confirm that > the code matches the intent. >=20 I'm happy to add comments, but I'm a little unclear on what you're confused by here. It's a bit too big of a patch for me to give a full play-by-play description. Can you elaborate on what you'd like to see? > >=20 > > Signed-off-by: Jeff Layton > > --- > > fs/nfsd/filecache.c | 247 ++++++++++++++++++++++---------------------- > > fs/nfsd/trace.h | 1 + > > 2 files changed, 123 insertions(+), 125 deletions(-) > >=20 > > diff --git a/fs/nfsd/filecache.c b/fs/nfsd/filecache.c > > index 0bf3727455e2..e67297ad12bf 100644 > > --- a/fs/nfsd/filecache.c > > +++ b/fs/nfsd/filecache.c > > @@ -303,8 +303,7 @@ nfsd_file_alloc(struct nfsd_file_lookup_key *key, u= nsigned int may) > > if (key->gc) > > __set_bit(NFSD_FILE_GC, &nf->nf_flags); > > nf->nf_inode =3D key->inode; > > - /* nf_ref is pre-incremented for hash table */ > > - refcount_set(&nf->nf_ref, 2); > > + refcount_set(&nf->nf_ref, 1); > > nf->nf_may =3D key->need; > > nf->nf_mark =3D NULL; > > } > > @@ -353,24 +352,35 @@ nfsd_file_unhash(struct nfsd_file *nf) > > return false; > > } > > =20 > > -static bool > > +static void > > nfsd_file_free(struct nfsd_file *nf) > > { > > s64 age =3D ktime_to_ms(ktime_sub(ktime_get(), nf->nf_birthtime)); > > - bool flush =3D false; > > =20 > > trace_nfsd_file_free(nf); > > =20 > > this_cpu_inc(nfsd_file_releases); > > this_cpu_add(nfsd_file_total_age, age); > > =20 > > + nfsd_file_unhash(nf); > > + > > + /* > > + * We call fsync here in order to catch writeback errors. It's not > > + * strictly required by the protocol, but an nfsd_file coule get > > + * evicted from the cache before a COMMIT comes in. If another > > + * task were to open that file in the interim and scrape the error, > > + * then the client may never see it. By calling fsync here, we ensure > > + * that writeback happens before the entry is freed, and that any > > + * errors reported result in the write verifier changing. > > + */ > > + nfsd_file_fsync(nf); > > + > > if (nf->nf_mark) > > nfsd_file_mark_put(nf->nf_mark); > > if (nf->nf_file) { > > get_file(nf->nf_file); > > filp_close(nf->nf_file, NULL); > > fput(nf->nf_file); > > - flush =3D true; > > } > > =20 > > /* > > @@ -378,10 +388,9 @@ nfsd_file_free(struct nfsd_file *nf) > > * WARN and leak it to preserve system stability. > > */ > > if (WARN_ON_ONCE(!list_empty(&nf->nf_lru))) > > - return flush; > > + return; > > =20 > > call_rcu(&nf->nf_rcu, nfsd_file_slab_free); > > - return flush; > > } > > =20 > > static bool > > @@ -397,17 +406,23 @@ nfsd_file_check_writeback(struct nfsd_file *nf) > > mapping_tagged(mapping, PAGECACHE_TAG_WRITEBACK); > > } > > =20 > > -static void nfsd_file_lru_add(struct nfsd_file *nf) > > +static bool nfsd_file_lru_add(struct nfsd_file *nf) > > { > > set_bit(NFSD_FILE_REFERENCED, &nf->nf_flags); > > - if (list_lru_add(&nfsd_file_lru, &nf->nf_lru)) > > + if (list_lru_add(&nfsd_file_lru, &nf->nf_lru)) { > > trace_nfsd_file_lru_add(nf); > > + return true; > > + } > > + return false; > > } > > =20 > > -static void nfsd_file_lru_remove(struct nfsd_file *nf) > > +static bool nfsd_file_lru_remove(struct nfsd_file *nf) > > { > > - if (list_lru_del(&nfsd_file_lru, &nf->nf_lru)) > > + if (list_lru_del(&nfsd_file_lru, &nf->nf_lru)) { > > trace_nfsd_file_lru_del(nf); > > + return true; > > + } > > + return false; > > } > > =20 > > struct nfsd_file * > > @@ -418,86 +433,80 @@ nfsd_file_get(struct nfsd_file *nf) > > return NULL; > > } > > =20 > > -static void > > +/** > > + * nfsd_file_unhash_and_queue - unhash a file and queue it to the disp= ose list > > + * @nf: nfsd_file to be unhashed and queued > > + * @dispose: list to which it should be queued > > + * > > + * Attempt to unhash a nfsd_file and queue it to the given list. Each = file > > + * will have a reference held on behalf of the list. That reference ma= y come > > + * from the LRU, or we may need to take one. If we can't get a referen= ce, > > + * ignore it altogether. > > + */ > > +static bool > > nfsd_file_unhash_and_queue(struct nfsd_file *nf, struct list_head *dis= pose) > > { > > trace_nfsd_file_unhash_and_queue(nf); > > if (nfsd_file_unhash(nf)) { > > - /* caller must call nfsd_file_dispose_list() later */ > > - nfsd_file_lru_remove(nf); > > + /* > > + * If we remove it from the LRU, then just use that > > + * reference for the dispose list. Otherwise, we need > > + * to take a reference. If that fails, just ignore > > + * the file altogether. > > + */ > > + if (!nfsd_file_lru_remove(nf) && !nfsd_file_get(nf)) > > + return false; > > list_add(&nf->nf_lru, dispose); > > + return true; > > } > > + return false; > > } > > =20 > > -static void > > -nfsd_file_put_noref(struct nfsd_file *nf) > > -{ > > - trace_nfsd_file_put(nf); > > - > > - if (refcount_dec_and_test(&nf->nf_ref)) { > > - WARN_ON(test_bit(NFSD_FILE_HASHED, &nf->nf_flags)); > > - nfsd_file_lru_remove(nf); > > - nfsd_file_free(nf); > > - } > > -} > > - > > -static void > > -nfsd_file_unhash_and_put(struct nfsd_file *nf) > > -{ > > - if (nfsd_file_unhash(nf)) > > - nfsd_file_put_noref(nf); > > -} > > - > > +/** > > + * nfsd_file_put - put the reference to a nfsd_file > > + * @nf: nfsd_file of which to put the reference > > + * > > + * Put a reference to a nfsd_file. In the v4 case, we just put the > > + * reference immediately. In the GC case, if the reference would be > > + * the last one, the put it on the LRU instead to be cleaned up later. > > + */ > > void > > nfsd_file_put(struct nfsd_file *nf) > > { > > might_sleep(); > > + trace_nfsd_file_put(nf); > > =20 > > - if (test_bit(NFSD_FILE_GC, &nf->nf_flags)) > > - nfsd_file_lru_add(nf); > > - else if (refcount_read(&nf->nf_ref) =3D=3D 2) > > - nfsd_file_unhash_and_put(nf); > > - > > - if (!test_bit(NFSD_FILE_HASHED, &nf->nf_flags)) { > > - nfsd_file_fsync(nf); > > - nfsd_file_put_noref(nf); > > - } else if (nf->nf_file && test_bit(NFSD_FILE_GC, &nf->nf_flags)) { > > - nfsd_file_put_noref(nf); > > - nfsd_file_schedule_laundrette(); > > - } else > > - nfsd_file_put_noref(nf); > > -} > > - > > -static void > > -nfsd_file_dispose_list(struct list_head *dispose) > > -{ > > - struct nfsd_file *nf; > > - > > - while(!list_empty(dispose)) { > > - nf =3D list_first_entry(dispose, struct nfsd_file, nf_lru); > > - list_del_init(&nf->nf_lru); > > - nfsd_file_fsync(nf); > > - nfsd_file_put_noref(nf); > > + /* > > + * The HASHED check is racy. We may end up with the occasional > > + * unhashed entry on the LRU, but they should get cleaned up > > + * like any other. > > + */ > > + if (test_bit(NFSD_FILE_GC, &nf->nf_flags) && > > + test_bit(NFSD_FILE_HASHED, &nf->nf_flags)) { > > + /* > > + * If this is the last reference (nf_ref =3D=3D 1), then transfer > > + * it to the LRU. If the add to the LRU fails, just put it as > > + * usual. > > + */ > > + if (refcount_dec_not_one(&nf->nf_ref) || nfsd_file_lru_add(nf)) { > > + nfsd_file_schedule_laundrette(); > > + return; > > + } > > } > > + if (refcount_dec_and_test(&nf->nf_ref)) > > + nfsd_file_free(nf); > > } > > =20 > > static void > > -nfsd_file_dispose_list_sync(struct list_head *dispose) > > +nfsd_file_dispose_list(struct list_head *dispose) > > { > > - bool flush =3D false; > > struct nfsd_file *nf; > > =20 > > while(!list_empty(dispose)) { > > nf =3D list_first_entry(dispose, struct nfsd_file, nf_lru); > > list_del_init(&nf->nf_lru); > > - nfsd_file_fsync(nf); > > - if (!refcount_dec_and_test(&nf->nf_ref)) > > - continue; > > - if (nfsd_file_free(nf)) > > - flush =3D true; > > + nfsd_file_free(nf); > > } > > - if (flush) > > - flush_delayed_fput(); > > } > > =20 > > static void > > @@ -567,21 +576,8 @@ nfsd_file_lru_cb(struct list_head *item, struct li= st_lru_one *lru, > > struct list_head *head =3D arg; > > struct nfsd_file *nf =3D list_entry(item, struct nfsd_file, nf_lru); > > =20 > > - /* > > - * Do a lockless refcount check. The hashtable holds one reference, s= o > > - * we look to see if anything else has a reference, or if any have > > - * been put since the shrinker last ran. Those don't get unhashed and > > - * released. > > - * > > - * Note that in the put path, we set the flag and then decrement the > > - * counter. Here we check the counter and then test and clear the fla= g. > > - * That order is deliberate to ensure that we can do this locklessly. > > - */ > > - if (refcount_read(&nf->nf_ref) > 1) { > > - list_lru_isolate(lru, &nf->nf_lru); > > - trace_nfsd_file_gc_in_use(nf); > > - return LRU_REMOVED; > > - } > > + /* We should only be dealing with GC entries here */ > > + WARN_ON_ONCE(!test_bit(NFSD_FILE_GC, &nf->nf_flags)); > > =20 > > /* > > * Don't throw out files that are still undergoing I/O or > > @@ -592,40 +588,30 @@ nfsd_file_lru_cb(struct list_head *item, struct l= ist_lru_one *lru, > > return LRU_SKIP; > > } > > =20 > > + /* If it was recently added to the list, skip it */ > > if (test_and_clear_bit(NFSD_FILE_REFERENCED, &nf->nf_flags)) { > > trace_nfsd_file_gc_referenced(nf); > > return LRU_ROTATE; > > } > > =20 > > - if (!test_and_clear_bit(NFSD_FILE_HASHED, &nf->nf_flags)) { > > - trace_nfsd_file_gc_hashed(nf); > > - return LRU_SKIP; > > + /* > > + * Put the reference held on behalf of the LRU. If it wasn't the last > > + * one, then just remove it from the LRU and ignore it. > > + */ > > + if (!refcount_dec_and_test(&nf->nf_ref)) { > > + trace_nfsd_file_gc_in_use(nf); > > + list_lru_isolate(lru, &nf->nf_lru); > > + return LRU_REMOVED; > > } > > =20 > > + /* Refcount went to zero. Unhash it and queue it to the dispose list = */ > > + nfsd_file_unhash(nf); > > list_lru_isolate_move(lru, &nf->nf_lru, head); > > this_cpu_inc(nfsd_file_evictions); > > trace_nfsd_file_gc_disposed(nf); > > return LRU_REMOVED; > > } > > =20 > > -/* > > - * Unhash items on @dispose immediately, then queue them on the > > - * disposal workqueue to finish releasing them in the background. > > - * > > - * cel: Note that between the time list_lru_shrink_walk runs and > > - * now, these items are in the hash table but marked unhashed. > > - * Why release these outside of lru_cb ? There's no lock ordering > > - * problem since lru_cb currently takes no lock. > > - */ > > -static void nfsd_file_gc_dispose_list(struct list_head *dispose) > > -{ > > - struct nfsd_file *nf; > > - > > - list_for_each_entry(nf, dispose, nf_lru) > > - nfsd_file_hash_remove(nf); > > - nfsd_file_dispose_list_delayed(dispose); > > -} > > - > > static void > > nfsd_file_gc(void) > > { > > @@ -635,7 +621,7 @@ nfsd_file_gc(void) > > ret =3D list_lru_walk(&nfsd_file_lru, nfsd_file_lru_cb, > > &dispose, list_lru_count(&nfsd_file_lru)); > > trace_nfsd_file_gc_removed(ret, list_lru_count(&nfsd_file_lru)); > > - nfsd_file_gc_dispose_list(&dispose); > > + nfsd_file_dispose_list_delayed(&dispose); > > } > > =20 > > static void > > @@ -660,7 +646,7 @@ nfsd_file_lru_scan(struct shrinker *s, struct shrin= k_control *sc) > > ret =3D list_lru_shrink_walk(&nfsd_file_lru, sc, > > nfsd_file_lru_cb, &dispose); > > trace_nfsd_file_shrinker_removed(ret, list_lru_count(&nfsd_file_lru))= ; > > - nfsd_file_gc_dispose_list(&dispose); > > + nfsd_file_dispose_list_delayed(&dispose); > > return ret; > > } > > =20 > > @@ -671,8 +657,11 @@ static struct shrinker nfsd_file_shrinker =3D { > > }; > > =20 > > /* > > - * Find all cache items across all net namespaces that match @inode an= d > > - * move them to @dispose. The lookup is atomic wrt nfsd_file_acquire()= . > > + * Find all cache items across all net namespaces that match @inode, u= nhash > > + * them, take references and then put them on @dispose if that was suc= cessful. > > + * > > + * The nfsd_file objects on the list will be unhashed, and each will h= ave a > > + * reference taken. > > */ > > static unsigned int > > __nfsd_file_close_inode(struct inode *inode, struct list_head *dispose= ) > > @@ -690,8 +679,9 @@ __nfsd_file_close_inode(struct inode *inode, struct= list_head *dispose) > > nfsd_file_rhash_params); > > if (!nf) > > break; > > - nfsd_file_unhash_and_queue(nf, dispose); > > - count++; > > + > > + if (nfsd_file_unhash_and_queue(nf, dispose)) > > + count++; > > } while (1); > > rcu_read_unlock(); > > return count; > > @@ -703,15 +693,23 @@ __nfsd_file_close_inode(struct inode *inode, stru= ct list_head *dispose) > > * > > * Unhash and put all cache item associated with @inode. > > */ > > -static void > > +static unsigned int > > nfsd_file_close_inode(struct inode *inode) > > { > > - LIST_HEAD(dispose); > > + struct nfsd_file *nf; > > unsigned int count; > > + LIST_HEAD(dispose); > > =20 > > count =3D __nfsd_file_close_inode(inode, &dispose); > > trace_nfsd_file_close_inode(inode, count); > > - nfsd_file_dispose_list_delayed(&dispose); > > + while(!list_empty(&dispose)) { > > + nf =3D list_first_entry(&dispose, struct nfsd_file, nf_lru); > > + list_del_init(&nf->nf_lru); > > + trace_nfsd_file_closing(nf); > > + if (refcount_dec_and_test(&nf->nf_ref)) > > + nfsd_file_free(nf); > > + } > > + return count; > > } > > =20 > > /** > > @@ -723,19 +721,15 @@ nfsd_file_close_inode(struct inode *inode) > > void > > nfsd_file_close_inode_sync(struct inode *inode) > > { > > - LIST_HEAD(dispose); > > - unsigned int count; > > - > > - count =3D __nfsd_file_close_inode(inode, &dispose); > > - trace_nfsd_file_close_inode_sync(inode, count); > > - nfsd_file_dispose_list_sync(&dispose); > > + if (nfsd_file_close_inode(inode)) > > + flush_delayed_fput(); > > } > > =20 > > /** > > * nfsd_file_delayed_close - close unused nfsd_files > > * @work: dummy > > * > > - * Walk the LRU list and close any entries that have not been used sin= ce > > + * Walk the LRU list and destroy any entries that have not been used s= ince > > * the last scan. > > */ > > static void > > @@ -1056,8 +1050,10 @@ nfsd_file_do_acquire(struct svc_rqst *rqstp, str= uct svc_fh *fhp, > > rcu_read_lock(); > > nf =3D rhashtable_lookup(&nfsd_file_rhash_tbl, &key, > > nfsd_file_rhash_params); > > - if (nf) > > - nf =3D nfsd_file_get(nf); > > + if (nf) { > > + if (!nfsd_file_lru_remove(nf)) > > + nf =3D nfsd_file_get(nf); > > + } > > rcu_read_unlock(); > > if (nf) > > goto wait_for_construction; > > @@ -1092,11 +1088,11 @@ nfsd_file_do_acquire(struct svc_rqst *rqstp, st= ruct svc_fh *fhp, > > goto out; > > } > > open_retry =3D false; > > - nfsd_file_put_noref(nf); > > + if (refcount_dec_and_test(&nf->nf_ref)) > > + nfsd_file_free(nf); > > goto retry; > > } > > =20 > > - nfsd_file_lru_remove(nf); > > this_cpu_inc(nfsd_file_cache_hits); > > =20 > > status =3D nfserrno(nfsd_open_break_lease(file_inode(nf->nf_file), ma= y_flags)); > > @@ -1106,7 +1102,8 @@ nfsd_file_do_acquire(struct svc_rqst *rqstp, stru= ct svc_fh *fhp, > > this_cpu_inc(nfsd_file_acquisitions); > > *pnf =3D nf; > > } else { > > - nfsd_file_put(nf); > > + if (refcount_dec_and_test(&nf->nf_ref)) > > + nfsd_file_free(nf); > > nf =3D NULL; > > } > > =20 > > @@ -1133,7 +1130,7 @@ nfsd_file_do_acquire(struct svc_rqst *rqstp, stru= ct svc_fh *fhp, > > * then unhash. > > */ > > if (status !=3D nfs_ok || key.inode->i_nlink =3D=3D 0) > > - nfsd_file_unhash_and_put(nf); > > + nfsd_file_unhash(nf); > > clear_bit_unlock(NFSD_FILE_PENDING, &nf->nf_flags); > > smp_mb__after_atomic(); > > wake_up_bit(&nf->nf_flags, NFSD_FILE_PENDING); > > diff --git a/fs/nfsd/trace.h b/fs/nfsd/trace.h > > index 940252482fd4..a44ded06af87 100644 > > --- a/fs/nfsd/trace.h > > +++ b/fs/nfsd/trace.h > > @@ -906,6 +906,7 @@ DEFINE_EVENT(nfsd_file_class, name, \ > > DEFINE_NFSD_FILE_EVENT(nfsd_file_free); > > DEFINE_NFSD_FILE_EVENT(nfsd_file_unhash); > > DEFINE_NFSD_FILE_EVENT(nfsd_file_put); > > +DEFINE_NFSD_FILE_EVENT(nfsd_file_closing); > > DEFINE_NFSD_FILE_EVENT(nfsd_file_unhash_and_queue); > > =20 > > TRACE_EVENT(nfsd_file_alloc, > > --=20 > > 2.38.1 > >=20 > >=20 --=20 Jeff Layton