Received: by 2002:ac0:a5b6:0:0:0:0:0 with SMTP id m51-v6csp166698imm; Thu, 31 May 2018 21:25:38 -0700 (PDT) X-Google-Smtp-Source: ADUXVKKCewAaHMGLsxgHz3brCK2/jTQeL+nqMLkujBmYrlcfRU53ILMseWIAI/FmnMv6VQFU46Kc X-Received: by 2002:a63:ac57:: with SMTP id z23-v6mr7386697pgn.394.1527827138385; Thu, 31 May 2018 21:25:38 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1527827138; cv=none; d=google.com; s=arc-20160816; b=QTTzEVg4bUUoxJ4+M4O2wIQlQHcGprmb2aSFvX/HEC/9Njhb2eGk7vBJ1tyRegYa+H YknnWdfkyXJblwt+TRT344Rs1sLElNzcyy32SCrWWhkmya9pmJRpfg91N4Jx7ZkNmhit pNCjDGXHpnulEPozJkGx/HWjw7UZ2a0SHfougFn43eRoisHrE05Pt8SKXuhDLCV5zNnH b1hFMpFP8kMvFUxfv6syGzRgQhanpdU+zd3qAJnVevro41yqf+1xpmA0hvESdEEv+u/E gwKyLqN/DIEwPfB0/PCrQnsGgnAW0vQph7uEvzPjirxNOBjk0jDtO7yXNcMDaRRF95wl 2pOA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:arc-authentication-results; bh=xxpLiG4uAkKcxZQa0N0p6gjMsRpyOSouocnjBLWYEU8=; b=yfw8BM1xeFmojTBE5WmHNTRZLGVP0Bq11MwwwqsTEupzq9hU+B6gX/G9/mDMhXMejM WJ5Tv/1n6P/kn31serkkVQ/6Bb3Fx9p638BaX0bafOj8QKh+hzXjLnpPKEKLo2xrVa+p QZ31Dwe5jTZ030cHxreUnw5+unaiNxkvIK0eQOb78JctRcOy0d+PYY2Ckc7nr5TjejAF T/BqO+m6L4vApCZTB4EK0zQmwJeR+3TQ1lIrlM1H2GkcRwoks2tiSw0z3D0WL/Kbq6uI iUVezyYyUDWokdbFxTxNMV4XEVXExuXbRrBCAJdzseiewB6ylNYaACkvHpLp0SLdxPLp s/2Q== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id 90-v6si3965741pla.38.2018.05.31.21.25.23; Thu, 31 May 2018 21:25:38 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1750881AbeFAEYq (ORCPT + 99 others); Fri, 1 Jun 2018 00:24:46 -0400 Received: from mx3-rdu2.redhat.com ([66.187.233.73]:50708 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1750771AbeFAEYn (ORCPT ); Fri, 1 Jun 2018 00:24:43 -0400 Received: from smtp.corp.redhat.com (int-mx06.intmail.prod.int.rdu2.redhat.com [10.11.54.6]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id EE4EE402290A; Fri, 1 Jun 2018 04:24:42 +0000 (UTC) Received: from localhost (unknown [10.18.25.149]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 74F0E201E820; Fri, 1 Jun 2018 04:24:42 +0000 (UTC) Date: Fri, 1 Jun 2018 00:24:41 -0400 From: Mike Snitzer To: "Martin K. Petersen" Cc: Christoph Hellwig , Sagi Grimberg , Johannes Thumshirn , Keith Busch , Hannes Reinecke , Laurence Oberman , Ewan Milne , James Smart , Linux Kernel Mailinglist , Linux NVMe Mailinglist , Martin George , John Meneghini , axboe@kernel.dk Subject: Re: [PATCH 0/3] Provide more fine grained control over multipathing Message-ID: <20180601042441.GB14244@redhat.com> References: <20180525125322.15398-1-jthumshirn@suse.de> <20180525130535.GA24239@lst.de> <20180525135813.GB9591@redhat.com> <20180530220206.GA7037@redhat.com> <20180531163311.GA30954@lst.de> <20180531181757.GB11848@redhat.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.5.21 (2010-09-15) X-Scanned-By: MIMEDefang 2.78 on 10.11.54.6 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.11.55.6]); Fri, 01 Jun 2018 04:24:43 +0000 (UTC) X-Greylist: inspected by milter-greylist-4.5.16 (mx1.redhat.com [10.11.55.6]); Fri, 01 Jun 2018 04:24:43 +0000 (UTC) for IP:'10.11.54.6' DOMAIN:'int-mx06.intmail.prod.int.rdu2.redhat.com' HELO:'smtp.corp.redhat.com' FROM:'msnitzer@redhat.com' RCPT:'' Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, May 31 2018 at 10:40pm -0400, Martin K. Petersen wrote: > > Mike, > > > 1) container A is tasked with managing some dedicated NVMe technology > > that absolutely needs native NVMe multipath. > > > 2) container B is tasked with offering some canned layered product > > that was developed ontop of dm-multipath with its own multipath-tools > > oriented APIs, etc. And it is to manage some other NVMe technology on > > the same host as container A. > > This assumes there is something to manage. And that the administrative > model currently employed by DM multipath will be easily applicable to > ANA devices. I don't believe that's the case. The configuration happens > on the storage side, not on the host. Fair point. > With ALUA (and the proprietary implementations that predated the spec), > it was very fuzzy whether it was the host or the target that owned > responsibility for this or that. Part of the reason was that ALUA was > deliberately vague to accommodate everybody's existing, non-standards > compliant multipath storage implementations. > > With ANA the heavy burden falls entirely on the storage. Most of the > things you would currently configure in multipath.conf have no meaning > in the context of ANA. Things that are currently the domain of > dm-multipath or multipathd are inextricably living either in the storage > device or in the NVMe ANA "device handler". And I think you are > significantly underestimating the effort required to expose that > information up the stack and to make use of it. That's not just a > multipath personality toggle switch. I'm aware that most everything in multipath.conf is SCSI/FC specific. That isn't the point. dm-multipath and multipathd are an existing framework for managing multipath storage. It could be made to work with NVMe. But yes it would not be easy. Especially not with the native NVMe multipath crew being so damn hostile. > If you want to make multipath -ll show something meaningful for ANA > devices, then by all means go ahead. I don't have any problem with > that. Thanks so much for your permission ;) But I'm actually not very involved with multipathd development anyway. It is likely a better use of time in the near-term though. Making the multipath tools and libraries be able to understand native NVMe multipath in all its glory might be a means to an end from a compatibility with existing monitoring applications perspective. Though NVMe just doesn't have per-device accounting at all. Also not yet aware how nvme cli conveys paths being down vs up, etc. Glad that isn't my problem ;) > But I don't think the burden of allowing multipathd/DM to inject > themselves into the path transition state machine has any benefit > whatsoever to the user. It's only complicating things and therefore we'd > be doing people a disservice rather than a favor. This notion that only native NVMe multipath can be successful is utter bullshit. And the mere fact that I've gotten such a reaction from a select few speaks to some serious control issues. Imagine if XFS developers just one day imposed that it is the _only_ filesystem that can be used on persistent memory. Just please dial it back.. seriously tiresome.