Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp701841pxj; Thu, 10 Jun 2021 10:33:03 -0700 (PDT) X-Google-Smtp-Source: ABdhPJyxVTP0nGJ04tbRQTzf09rsqQD0q+ian5WZxanBQanmG03RjbolVn8nnIy3fJ5VP3ekvm2n X-Received: by 2002:a17:906:d967:: with SMTP id rp7mr686257ejb.424.1623346383357; Thu, 10 Jun 2021 10:33:03 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1623346383; cv=none; d=google.com; s=arc-20160816; b=t+nVZA7Y5tzE5s7MQmAj35D9ViWZCUtKY9KWQxD7ZmB/0goujZPKDKJeSXGyHvHBSj 31fkJJAXvRxthp/vNGq43P5qPOVE7QdY3iL0APt/iErh3i2XgFKb7yxi2QQJslLZMdeH 1PSuwuTkMuziHW15mDaOEALqRCy9VB7141p80ydh88xyRNeKsYL8lpZ2X47BkK4/QegA dBmXnMa+kWoDCz4BI3j3G8jANHwgOwvKKZqDl7JdTDYHC80JfrMWXTiZa3VZ2Br+vih7 pZgGWB93TxnVacZCJ7dwzNt6SYuPtFTR8bGk6d672xfIhL4DuMBxRgWzJ7l01nhLLO3K X1Xw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=pTBLxJFKTdwzcDe+JrfxQOJonn6Ld8Apq22DTEd0Oug=; b=SLifBhlBmp85ywgbVZ1Npcib0F8734OHOEkkGni8ho6pHZiM1MoY3CshrXI3hpCtmO oI7hRpTt9+gpIenAKT1NnNv9JdGbf6WIEELycDIQttwMl+8cB+Y3fi2NEYwdKlOIKgjb CY3sefRPZrFjEqD3q3Q2DjRmGbDUARzTLUsuTHdQhHAz5LZF4Vr1O05BVMJP0/UHZ76U 7XHk13gAVQmGOAnmrR0jpabTmJ7XAed0I83iLTcJ7AvVkSDdirM5kqLm3Eqz5kl0X8Cz mSgjcNo/We6+BGn2hZg8zdoQ/UKawoqy3WYl+uit3rTGcXwsvNHPZaez72xbZJdfPY1q ly2A== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=PIXq2Gcg; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id n25si3107244ejl.314.2021.06.10.10.32.30; Thu, 10 Jun 2021 10:33:03 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=PIXq2Gcg; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229937AbhFJRdm (ORCPT + 99 others); Thu, 10 Jun 2021 13:33:42 -0400 Received: from mail-ed1-f47.google.com ([209.85.208.47]:41712 "EHLO mail-ed1-f47.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229802AbhFJRdl (ORCPT ); Thu, 10 Jun 2021 13:33:41 -0400 Received: by mail-ed1-f47.google.com with SMTP id g18so31985124edq.8 for ; Thu, 10 Jun 2021 10:31:28 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=pTBLxJFKTdwzcDe+JrfxQOJonn6Ld8Apq22DTEd0Oug=; b=PIXq2GcgwBpAJfRt9rM/leAaHji82TFrMSuUfWuIUOHWZ/2zQhn4X/3sOr1YXTmLkU x3RX/U8Nks1yoT/U98PGEU8ioMr8EoiIISSmT2cjLJiSMjXlThRB3ge67yHWVXHRRYMM fYRV67eoekoMjIE/381L551ww01KXSFpar9K/fhAw7msw+bzAlpXhPv2Jo3tmtbmzD3d IIQvERPEazVWOq2yyoWuz9Gforb4KMiUFW17XbbmeSX/Vtoj6vW/5W1hMblcfhcQZ4oK NBB5H+h5YKF1hmJVFXVGa9IGXjWgIDeHHJImXmnL2e7MIYM0fzTVrBVIM+knrmH2YQ8k jHKg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=pTBLxJFKTdwzcDe+JrfxQOJonn6Ld8Apq22DTEd0Oug=; b=HnmH7aDyDTFHfJn2SnyaXwkjQm/F1dz42xWz58WkEjMrkyX7g6cHDdNVf+YEHH1ag4 ZN5vKjCy3j0zEv8l+RCOO2FvrYXgAAkODzEInQWsR9znS87fqE5VcIgRVYGKmhW1ZTs4 7tkUbYWkutrJG1THZTJ35Ww2afKlfieXtyMnQDA/liQ15iZBt8AUmJ90mHP4oRmXwZSX JDzCFan8l4v71V4MnbZDBC3HO06ArE8/P5MGl0/RAwBNHtrdjsHZZoHkhdPhYQkkLIbJ eS5bxz+jlIQ5441rK1x/cX20xf6fzcDhqL4AcD0a8IGSle2NBxO+UKGJT/HJC1jxbuLY 0hiA== X-Gm-Message-State: AOAM5338UN5xj/ruBvVFwNkaEr4JbRJgk5IOhq6gCJ0AKX4XY2RgCepc vKJNhtwuUTgREtNuCGc2nRCCZXXiC/0f/hFYgEY= X-Received: by 2002:aa7:cb5a:: with SMTP id w26mr595888edt.139.1623346227961; Thu, 10 Jun 2021 10:30:27 -0700 (PDT) MIME-Version: 1.0 References: <20210609215319.5518-1-olga.kornievskaia@gmail.com> <20210609215319.5518-3-olga.kornievskaia@gmail.com> <6C64456A-931F-4CAD-A559-412A12F0F741@oracle.com> <6bca80de292b5aa36734e7d942d0e9f53430903b.camel@hammerspace.com> <83D9342E-FDAF-4B2C-A518-20BF0A9AD073@oracle.com> <3658c226b43fb190de38c00e5199ccf35ccc4369.camel@hammerspace.com> In-Reply-To: From: Olga Kornievskaia Date: Thu, 10 Jun 2021 13:30:16 -0400 Message-ID: Subject: Re: [PATCH v2 2/3] NFSv4 introduce max_connect mount options To: Trond Myklebust Cc: "linux-nfs@vger.kernel.org" , "anna.schumaker@netapp.com" , "chuck.lever@oracle.com" Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-nfs@vger.kernel.org On Thu, Jun 10, 2021 at 12:36 PM Trond Myklebust wrote: > > On Thu, 2021-06-10 at 12:14 -0400, Olga Kornievskaia wrote: > > On Thu, Jun 10, 2021 at 10:56 AM Trond Myklebust > > wrote: > > > > > > On Thu, 2021-06-10 at 10:31 -0400, Olga Kornievskaia wrote: > > > > On Thu, Jun 10, 2021 at 10:13 AM Trond Myklebust > > > > wrote: > > > > > > > > > > On Thu, 2021-06-10 at 13:56 +0000, Chuck Lever III wrote: > > > > > > > > > > > > > > > > > > > On Jun 10, 2021, at 9:34 AM, Trond Myklebust < > > > > > > > trondmy@hammerspace.com> wrote: > > > > > > > > > > > > > > On Thu, 2021-06-10 at 13:30 +0000, Chuck Lever III wrote: > > > > > > > > > > > > > > > > > > > > > > > > > On Jun 9, 2021, at 5:53 PM, Olga Kornievskaia < > > > > > > > > > olga.kornievskaia@gmail.com> wrote: > > > > > > > > > > > > > > > > > > From: Olga Kornievskaia > > > > > > > > > > > > > > > > > > This option will control up to how many xprts can the > > > > > > > > > client > > > > > > > > > establish to the server. This patch parses the value > > > > > > > > > and > > > > > > > > > sets > > > > > > > > > up structures that keep track of max_connect. > > > > > > > > > > > > > > > > > > Signed-off-by: Olga Kornievskaia > > > > > > > > > --- > > > > > > > > > fs/nfs/client.c | 1 + > > > > > > > > > fs/nfs/fs_context.c | 8 ++++++++ > > > > > > > > > fs/nfs/internal.h | 2 ++ > > > > > > > > > fs/nfs/nfs4client.c | 12 ++++++++++-- > > > > > > > > > fs/nfs/super.c | 2 ++ > > > > > > > > > include/linux/nfs_fs_sb.h | 1 + > > > > > > > > > 6 files changed, 24 insertions(+), 2 deletions(-) > > > > > > > > > > > > > > > > > > diff --git a/fs/nfs/client.c b/fs/nfs/client.c > > > > > > > > > index 330f65727c45..486dec59972b 100644 > > > > > > > > > --- a/fs/nfs/client.c > > > > > > > > > +++ b/fs/nfs/client.c > > > > > > > > > @@ -179,6 +179,7 @@ struct nfs_client > > > > > > > > > *nfs_alloc_client(const > > > > > > > > > struct nfs_client_initdata *cl_init) > > > > > > > > > > > > > > > > > > clp->cl_proto = cl_init->proto; > > > > > > > > > clp->cl_nconnect = cl_init->nconnect; > > > > > > > > > + clp->cl_max_connect = cl_init->max_connect ? > > > > > > > > > cl_init- > > > > > > > > > > max_connect : 1; > > > > > > > > > > > > > > > > So, 1 is the default setting, meaning the "add another > > > > > > > > transport" > > > > > > > > facility is disabled by default. Would it be less > > > > > > > > surprising > > > > > > > > for > > > > > > > > an admin to allow some extra connections by default? > > > > > > > > > > > > > > > > > > > > > > > > > clp->cl_net = get_net(cl_init->net); > > > > > > > > > > > > > > > > > > clp->cl_principal = "*"; > > > > > > > > > diff --git a/fs/nfs/fs_context.c b/fs/nfs/fs_context.c > > > > > > > > > index d95c9a39bc70..cfbff7098f8e 100644 > > > > > > > > > --- a/fs/nfs/fs_context.c > > > > > > > > > +++ b/fs/nfs/fs_context.c > > > > > > > > > @@ -29,6 +29,7 @@ > > > > > > > > > #endif > > > > > > > > > > > > > > > > > > #define NFS_MAX_CONNECTIONS 16 > > > > > > > > > +#define NFS_MAX_TRANSPORTS 128 > > > > > > > > > > > > > > > > This maximum seems excessive... again, there are > > > > > > > > diminishing > > > > > > > > returns to adding more connections to the same server. > > > > > > > > what's > > > > > > > > wrong with re-using NFS_MAX_CONNECTIONS for the maximum? > > > > > > > > > > > > > > > > As always, I'm a little queasy about adding yet another > > > > > > > > mount > > > > > > > > option. Are there real use cases where a whole-client > > > > > > > > setting > > > > > > > > (like a sysfs attribute) would be inadequate? Is there a > > > > > > > > way > > > > > > > > the client could figure out a reasonable maximum without > > > > > > > > a > > > > > > > > human intervention, say, by counting the number of NICs > > > > > > > > on > > > > > > > > the system? > > > > > > > > > > > > > > Oh, hell no! We're not tying anything to the number of > > > > > > > NICs... > > > > > > > > > > > > That's a bit of an over-reaction. :-) A little more > > > > > > explanation > > > > > > would be welcome. I mean, don't you expect someone to ask > > > > > > "How > > > > > > do I pick a good value?" and someone might reasonably answer > > > > > > "Well, start with the number of NICs on your client times 3" > > > > > > or > > > > > > something like that. > > > > > > > > > > > > IMO we're about to add another admin setting without > > > > > > understanding > > > > > > how it will be used, how to select a good maximum value, or > > > > > > even > > > > > > whether this maximum needs to be adjustable. In a previous e- > > > > > > mail > > > > > > Olga has already demonstrated that it will be difficult to > > > > > > explain > > > > > > how to use this setting with nconnect=. > > > > > > > > > > > > Thus I would favor a (moderate) soldered-in maximum to start > > > > > > with, > > > > > > and then as real world use cases arise, consider adding a > > > > > > tuning > > > > > > mechanism based on actual requirements. > > > > > > > > > > It's not an overreaction. It's insane to think that counting > > > > > NICs > > > > > gives > > > > > you any notion whatsoever about the network topology and > > > > > connectivity > > > > > between the client and server. It doesn't even tell you how > > > > > many of > > > > > those NICs might potentially be available to your application. > > > > > > > > > > We're not doing any automation based on that kind of layering > > > > > violation. > > > > > > > > I'm not suggesting to programmatically determine the number of > > > > NIC to > > > > determine the value of max_connect. > > > > > > > > > > > No, but that's what Chuck appeared to be suggesting in order to > > > avoid > > > the need for the mount option. > > > > > > To me, the main reason for the mount option is to allow the user to > > > limit the number of new IP addresses being added so that if the DNS > > > server is configured to hand out lots of different addresses for > > > the > > > same servername, the user can basically say 'no, I just want to use > > > the > > > one IP address that I'm already connected to' (i.e. max_connect=1). > > > I > > > can imagine that some clustered setups might need that ability in > > > order > > > to work efficiently. > > > > > > I'm fine with the idea of nconnect setting the number of > > > connections > > > per IP address, but that would need some plumbing in > > > rpc_clnt_test_and_add_xprt() to allow us to add up to 'nconnect' > > > copies > > > of a given transport. > > > Presumably rpc_xprt_switch_has_addr() would need to return a count > > > of > > > the number of copies of the transport that are already present so > > > that > > > we can decide whether or not we should add a new one. > > > > I think the last paragraph is what I'm asking for. But I would like > > to > > again confirm if you still mean "max_connect" to be the total number > > of connections since you say we could/will allow for nconnect number > > of connections per IP address. Would max_connect need to be a > > multiple > > of nconnect (max_connect = X *nconnect)? > > No. Your suggestion to make the two independent is growing on me, > however in that case we do want to ensure that if nconnect=X, then we > always add X transports when we add a new IP address. Ok. I'm glad to hear independ idea still has life. Are you still thinking "max_connect" is the right name for it? I guess if we explain the feature in the man pages the name doesn't matter so much. I would have still liked it to be something like "max_session_xprts". > > Actually when I said supporting (or rather allowing for) nconnect * > > max_connect transport, is that correct? Given how the code works now > > this is going to be nconnect + max_connect (only if 1st mount had > > nconnect option). We can't "add" nconnect connections to the new > > mounts (but with my patch we can add a single trunk connection). By > > that I mean: say the first was "mount IP1:/vol1 /mnt1" (1 connection > > to IP2). Now the client is doing "mount IP2:/vol2 /mnt2". IP1 and IP2 > > are trunkable addresses of the same server so we add a trunk. We > > currently don't allow for doing "mount -o nconnec=2 IP2:vol2 /mnt2" > > and then also add "nconnect" connections to IP2 along with a trunk. > > In > > the 2nd example, we'd have 1 connections to IP1, then 2 connections > > to > > IP2. Can we allow for that (with needed code change)? If not, then > > we > > really need to commit to only support nconnect (16) connections + > > some > > number of trunkable connections. > > > I think we want to have nconnect be server-global. i.e. nconnect > entries of each IP address. Thank you both, Trond and Chuck. I'll work on v3. > > -- > Trond Myklebust > Linux NFS client maintainer, Hammerspace > trond.myklebust@hammerspace.com > >