Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp645418pxj; Thu, 10 Jun 2021 09:16:46 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxbn81nG131WKjxIMWC0HTAZYsOVKd4agCL26C8auquaxhZryCS2bt1CMu6mRuIKYHUdfe8 X-Received: by 2002:a05:6402:61a:: with SMTP id n26mr260928edv.220.1623341805817; Thu, 10 Jun 2021 09:16:45 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1623341805; cv=none; d=google.com; s=arc-20160816; b=VUMJO3f02DJcYO+UxqZ1ugPJGrRd+aDmoFl4pAw8h8WOZN4AG+EovpAOll7b3b9I6l Hc3GpIWJQKo90VnKv/V6gcQ5+UBzVg1lvf4pYAz6jrCbN4FbyZ4ZjTLSZhr6M8PVyPSf eKxMeo9FJxwPvVh66T7VRilgaNzpRrt6z70aJ0t5QJDsdKlubvLg+Fp/iRIsfP/FNYRK nXS59COzcWu8L9LgcOQE6EZeE2TzD0crP7pu/i6SEbq9uIGXcxBUz6dEnppHBxaNqQwN OsGZNrUoUAuV2BQCHParDq0apl20ZQdCsNHGXeKN0hSzvd+DBzvr/yDRuxG+6Jj/6+f2 BhIQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=WiAmtG22QSgpfH+MBUOUlX4nB+pCVCd2TjW8Kj93Kuw=; b=eeBw3a22wFy3SKQtKIH4eI2OLk7GfiCPte5356OczeQz3SOA7t04Bn9wEg1QPx6ajB 4dwjGRzf1NvBEBpQOeIYiuEgQsNCZIfzrGNoF8QJ8lzg2JoQgrBnBFAjBtohbbh0yuXZ I2GEKRnE6gd4hdXu8nczgEsRpLmc2dB/pB8L/16s+GNE1SLd3sF/c7X67I84dUZ4gd7Z xbMFkmEGswVTBJAxqjwuoP0hqObYF+uWCH/tYFiZNLe47Y9WUUMJrtReKA6ifJXZzC/s qe2po7Hnx8ZlZ1yjhFzjdv/xj7V+6Fu3CQLAEq2yiVtdmowcAXM89ZbvwwoffZmjor5d AoEQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=nnrGRYHT; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id l18si2519923ejr.283.2021.06.10.09.16.21; Thu, 10 Jun 2021 09:16:45 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=nnrGRYHT; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229802AbhFJQQm (ORCPT + 99 others); Thu, 10 Jun 2021 12:16:42 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:40982 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230134AbhFJQQm (ORCPT ); Thu, 10 Jun 2021 12:16:42 -0400 Received: from mail-ej1-x62b.google.com (mail-ej1-x62b.google.com [IPv6:2a00:1450:4864:20::62b]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 2ABC2C061574 for ; Thu, 10 Jun 2021 09:14:36 -0700 (PDT) Received: by mail-ej1-x62b.google.com with SMTP id ce15so96098ejb.4 for ; Thu, 10 Jun 2021 09:14:36 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=WiAmtG22QSgpfH+MBUOUlX4nB+pCVCd2TjW8Kj93Kuw=; b=nnrGRYHTautUY0U8MbHLsxA87F577gtZvkpbaa2Nb3xEbgsS7O/yn9qMR28wTiZaWo 7frK4gmDjkZCMyMvgoU+wDvWwr3/JUs6xkn3LoL3EF/0NrOB6Uhx+2JtsLEdWDy1OG5l DJ8BRqXhUtnVAznqmnvDWkRYwIDW4DctAkvQIj1odVZ8XcpAFB0gPSRVhISaC+2JV/h7 Qf+Mq2uZJYBovdDw8EgPr5u4olSQ014gtudV4GYnaeh82W7ObZ/kPUydk3sadKaVYcUI i8zdVHAtzcLnGa1f63VaKuQNkZOBsn85BOvxLpAeOuy9AltIg/LPPr898nezI6LWfE3+ fJSg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=WiAmtG22QSgpfH+MBUOUlX4nB+pCVCd2TjW8Kj93Kuw=; b=akY5yV3+FM3tnk842usCv6j5ONSevD10IwnrvUjU1cbL5FdMHZ2zwJrphAeXdreVh9 FvMCq50Scwc03wh68TMzl5eK0BmLVk3wbOyKklqgWL+e70O3RviCyaYCGWKiiWsshWxC BrkKN6iR3AKr0W107sEdnFX1Tc6gZ1SsFvrDT3whhdSdDhwgH86A1p8tO5cO25q9nGQN Kmj1mYJbI80G2sVcW0qNwpdB0t8f8c+CL7ucGjH1SODTGEkrRwY513uLtBrMhp0gdRPt NUL/7fWTTx7Cq3Z17PHua2dp/K35+JhVB/KRJsTIL/4uhU+XcOjbqnCrtp+metZ2At22 +VZA== X-Gm-Message-State: AOAM5303HRu8/7ElZUjsaygwinTm44MW2R7CFyQ96PB0sg+eNf8YSj/J qHzYhHdmeHy1S0x+F9XmtyKxne+rGr0AuGsk3FI= X-Received: by 2002:a17:907:3e26:: with SMTP id hp38mr329326ejc.451.1623341674615; Thu, 10 Jun 2021 09:14:34 -0700 (PDT) MIME-Version: 1.0 References: <20210609215319.5518-1-olga.kornievskaia@gmail.com> <20210609215319.5518-3-olga.kornievskaia@gmail.com> <6C64456A-931F-4CAD-A559-412A12F0F741@oracle.com> <6bca80de292b5aa36734e7d942d0e9f53430903b.camel@hammerspace.com> <83D9342E-FDAF-4B2C-A518-20BF0A9AD073@oracle.com> <3658c226b43fb190de38c00e5199ccf35ccc4369.camel@hammerspace.com> In-Reply-To: From: Olga Kornievskaia Date: Thu, 10 Jun 2021 12:14:23 -0400 Message-ID: Subject: Re: [PATCH v2 2/3] NFSv4 introduce max_connect mount options To: Trond Myklebust Cc: "linux-nfs@vger.kernel.org" , "anna.schumaker@netapp.com" , "chuck.lever@oracle.com" Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-nfs@vger.kernel.org On Thu, Jun 10, 2021 at 10:56 AM Trond Myklebust wrote: > > On Thu, 2021-06-10 at 10:31 -0400, Olga Kornievskaia wrote: > > On Thu, Jun 10, 2021 at 10:13 AM Trond Myklebust > > wrote: > > > > > > On Thu, 2021-06-10 at 13:56 +0000, Chuck Lever III wrote: > > > > > > > > > > > > > On Jun 10, 2021, at 9:34 AM, Trond Myklebust < > > > > > trondmy@hammerspace.com> wrote: > > > > > > > > > > On Thu, 2021-06-10 at 13:30 +0000, Chuck Lever III wrote: > > > > > > > > > > > > > > > > > > > On Jun 9, 2021, at 5:53 PM, Olga Kornievskaia < > > > > > > > olga.kornievskaia@gmail.com> wrote: > > > > > > > > > > > > > > From: Olga Kornievskaia > > > > > > > > > > > > > > This option will control up to how many xprts can the > > > > > > > client > > > > > > > establish to the server. This patch parses the value and > > > > > > > sets > > > > > > > up structures that keep track of max_connect. > > > > > > > > > > > > > > Signed-off-by: Olga Kornievskaia > > > > > > > --- > > > > > > > fs/nfs/client.c | 1 + > > > > > > > fs/nfs/fs_context.c | 8 ++++++++ > > > > > > > fs/nfs/internal.h | 2 ++ > > > > > > > fs/nfs/nfs4client.c | 12 ++++++++++-- > > > > > > > fs/nfs/super.c | 2 ++ > > > > > > > include/linux/nfs_fs_sb.h | 1 + > > > > > > > 6 files changed, 24 insertions(+), 2 deletions(-) > > > > > > > > > > > > > > diff --git a/fs/nfs/client.c b/fs/nfs/client.c > > > > > > > index 330f65727c45..486dec59972b 100644 > > > > > > > --- a/fs/nfs/client.c > > > > > > > +++ b/fs/nfs/client.c > > > > > > > @@ -179,6 +179,7 @@ struct nfs_client > > > > > > > *nfs_alloc_client(const > > > > > > > struct nfs_client_initdata *cl_init) > > > > > > > > > > > > > > clp->cl_proto = cl_init->proto; > > > > > > > clp->cl_nconnect = cl_init->nconnect; > > > > > > > + clp->cl_max_connect = cl_init->max_connect ? > > > > > > > cl_init- > > > > > > > > max_connect : 1; > > > > > > > > > > > > So, 1 is the default setting, meaning the "add another > > > > > > transport" > > > > > > facility is disabled by default. Would it be less surprising > > > > > > for > > > > > > an admin to allow some extra connections by default? > > > > > > > > > > > > > > > > > > > clp->cl_net = get_net(cl_init->net); > > > > > > > > > > > > > > clp->cl_principal = "*"; > > > > > > > diff --git a/fs/nfs/fs_context.c b/fs/nfs/fs_context.c > > > > > > > index d95c9a39bc70..cfbff7098f8e 100644 > > > > > > > --- a/fs/nfs/fs_context.c > > > > > > > +++ b/fs/nfs/fs_context.c > > > > > > > @@ -29,6 +29,7 @@ > > > > > > > #endif > > > > > > > > > > > > > > #define NFS_MAX_CONNECTIONS 16 > > > > > > > +#define NFS_MAX_TRANSPORTS 128 > > > > > > > > > > > > This maximum seems excessive... again, there are diminishing > > > > > > returns to adding more connections to the same server. what's > > > > > > wrong with re-using NFS_MAX_CONNECTIONS for the maximum? > > > > > > > > > > > > As always, I'm a little queasy about adding yet another mount > > > > > > option. Are there real use cases where a whole-client setting > > > > > > (like a sysfs attribute) would be inadequate? Is there a way > > > > > > the client could figure out a reasonable maximum without a > > > > > > human intervention, say, by counting the number of NICs on > > > > > > the system? > > > > > > > > > > Oh, hell no! We're not tying anything to the number of NICs... > > > > > > > > That's a bit of an over-reaction. :-) A little more explanation > > > > would be welcome. I mean, don't you expect someone to ask "How > > > > do I pick a good value?" and someone might reasonably answer > > > > "Well, start with the number of NICs on your client times 3" or > > > > something like that. > > > > > > > > IMO we're about to add another admin setting without > > > > understanding > > > > how it will be used, how to select a good maximum value, or even > > > > whether this maximum needs to be adjustable. In a previous e-mail > > > > Olga has already demonstrated that it will be difficult to > > > > explain > > > > how to use this setting with nconnect=. > > > > > > > > Thus I would favor a (moderate) soldered-in maximum to start > > > > with, > > > > and then as real world use cases arise, consider adding a tuning > > > > mechanism based on actual requirements. > > > > > > It's not an overreaction. It's insane to think that counting NICs > > > gives > > > you any notion whatsoever about the network topology and > > > connectivity > > > between the client and server. It doesn't even tell you how many of > > > those NICs might potentially be available to your application. > > > > > > We're not doing any automation based on that kind of layering > > > violation. > > > > I'm not suggesting to programmatically determine the number of NIC to > > determine the value of max_connect. > > > > > No, but that's what Chuck appeared to be suggesting in order to avoid > the need for the mount option. > > To me, the main reason for the mount option is to allow the user to > limit the number of new IP addresses being added so that if the DNS > server is configured to hand out lots of different addresses for the > same servername, the user can basically say 'no, I just want to use the > one IP address that I'm already connected to' (i.e. max_connect=1). I > can imagine that some clustered setups might need that ability in order > to work efficiently. > > I'm fine with the idea of nconnect setting the number of connections > per IP address, but that would need some plumbing in > rpc_clnt_test_and_add_xprt() to allow us to add up to 'nconnect' copies > of a given transport. > Presumably rpc_xprt_switch_has_addr() would need to return a count of > the number of copies of the transport that are already present so that > we can decide whether or not we should add a new one. I think the last paragraph is what I'm asking for. But I would like to again confirm if you still mean "max_connect" to be the total number of connections since you say we could/will allow for nconnect number of connections per IP address. Would max_connect need to be a multiple of nconnect (max_connect = X *nconnect)? Actually when I said supporting (or rather allowing for) nconnect * max_connect transport, is that correct? Given how the code works now this is going to be nconnect + max_connect (only if 1st mount had nconnect option). We can't "add" nconnect connections to the new mounts (but with my patch we can add a single trunk connection). By that I mean: say the first was "mount IP1:/vol1 /mnt1" (1 connection to IP2). Now the client is doing "mount IP2:/vol2 /mnt2". IP1 and IP2 are trunkable addresses of the same server so we add a trunk. We currently don't allow for doing "mount -o nconnec=2 IP2:vol2 /mnt2" and then also add "nconnect" connections to IP2 along with a trunk. In the 2nd example, we'd have 1 connections to IP1, then 2 connections to IP2. Can we allow for that (with needed code change)? If not, then we really need to commit to only support nconnect (16) connections + some number of trunkable connections. > > -- > Trond Myklebust > Linux NFS client maintainer, Hammerspace > trond.myklebust@hammerspace.com > >