Received: by 2002:a25:ad19:0:0:0:0:0 with SMTP id y25csp7157991ybi; Mon, 22 Jul 2019 08:03:58 -0700 (PDT) X-Google-Smtp-Source: APXvYqzM+O4M59kZ1OD96fsnPv13byOvKpl0jVuOm/JCgMI4qll0zcAvOqbwETDCTgW1uSly8bnt X-Received: by 2002:a65:64ce:: with SMTP id t14mr6312523pgv.137.1563807838630; Mon, 22 Jul 2019 08:03:58 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1563807838; cv=none; d=google.com; s=arc-20160816; b=tlqfexeHGxZ8MeU3nHW/ldxtFotlLMIPTL7RO0hTb4VjoZNZwzuGaj7VRCWiEYzWac AvzsFCzTW96XEADnEvRYj+jGcY5vhAgO+9jDg36yk9EQHKWw/kxrjl1sOMGnwWIWoOl7 e2O5yjb4O7q2ZETTHK+vTVWa31uyx5A1L3bsjK00P2T0P0IaBjgkBIu6Bk767xrs7G/W fi0jk23GrZbBc9TTbqA860iDegE2aI5gHHM5r34EfetOUFD3r9rzeAf389p9PLhwdFev aeOcRgJdDHY/iPWDh7Qfob3MaX5IMZFM9bEysQt2vYU/nS2LNRmxIy/F923yWoIflYId sGZA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :organization:references:in-reply-to:message-id:subject:cc:to:from :date; bh=WiEo/8a5XfurZ5Rvf9iuWGbwrL3TguHNr1UyJvTzaGo=; b=pLDkVQ4yBWk6vjYeMzqhP8Fty5Aw5/Qy5eGpOFeW1j8yzNItTetVO0aJZFkECI1och rAdXU1CTuIy5eAcqrU+41f3VWN5lKIX4cMvoAaYz0z/1MVGWt9cvcliSPINFi0gpsgF3 552U01KJN0wHx/5G4Pk2Y5Xacwf2DsXC4LTpaMPPY3lbDloorDdrmBPOXccd2E7YbkgY QDHofdSnlICdVNPE+mRQAw9LG/G6klcdVG9JVpIhEiq4/uXPyb4I+ZlrUdnqVFWpZxVO QkJ5J8JCB6cSwmdKNBBbcI/R3xWZiR6u+JYMQSuwEmMFL82n59A5WrtPlAroeX1prQZE +LuA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id t20si8664991pjr.107.2019.07.22.08.03.42; Mon, 22 Jul 2019 08:03:58 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1729971AbfGVLn2 (ORCPT + 99 others); Mon, 22 Jul 2019 07:43:28 -0400 Received: from mx1.redhat.com ([209.132.183.28]:35202 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726339AbfGVLn1 (ORCPT ); Mon, 22 Jul 2019 07:43:27 -0400 Received: from smtp.corp.redhat.com (int-mx02.intmail.prod.int.phx2.redhat.com [10.5.11.12]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id C75C530832C6; Mon, 22 Jul 2019 11:43:26 +0000 (UTC) Received: from gondolin (dhcp-192-181.str.redhat.com [10.33.192.181]) by smtp.corp.redhat.com (Postfix) with ESMTP id AB1B160BF1; Mon, 22 Jul 2019 11:43:19 +0000 (UTC) Date: Mon, 22 Jul 2019 13:43:17 +0200 From: Cornelia Huck To: Christian Borntraeger Cc: "Dr. David Alan Gilbert" , Dan Williams , Vivek Goyal , Halil Pasic , Collin Walling , Sebastian Ott , KVM list , Miklos Szeredi , linux-nvdimm , Linux Kernel Mailing List , Stefan Hajnoczi , linux-fsdevel , Steven Whitehouse , Heiko Carstens , David Hildenbrand Subject: Re: [PATCH v2 18/30] virtio_fs, dax: Set up virtio_fs dax_device Message-ID: <20190722134317.39b148ce.cohuck@redhat.com> In-Reply-To: References: <20190515192715.18000-1-vgoyal@redhat.com> <20190515192715.18000-19-vgoyal@redhat.com> <20190717192725.25c3d146.pasic@linux.ibm.com> <20190718131532.GA13883@redhat.com> <20190722105630.GC3035@work-vm> Organization: Red Hat GmbH MIME-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit X-Scanned-By: MIMEDefang 2.79 on 10.5.11.12 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.44]); Mon, 22 Jul 2019 11:43:26 +0000 (UTC) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, 22 Jul 2019 13:20:18 +0200 Christian Borntraeger wrote: > On 22.07.19 12:56, Dr. David Alan Gilbert wrote: > > * Christian Borntraeger (borntraeger@de.ibm.com) wrote: > >> > >> > >> On 18.07.19 16:30, Dan Williams wrote: > >>> On Thu, Jul 18, 2019 at 6:15 AM Vivek Goyal wrote: > >>>> > >>>> On Wed, Jul 17, 2019 at 07:27:25PM +0200, Halil Pasic wrote: > >>>>> On Wed, 15 May 2019 15:27:03 -0400 > >>>>> Vivek Goyal wrote: > >>>>> > >>>>>> From: Stefan Hajnoczi > >>>>>> > >>>>>> Setup a dax device. > >>>>>> > >>>>>> Use the shm capability to find the cache entry and map it. > >>>>>> > >>>>>> The DAX window is accessed by the fs/dax.c infrastructure and must have > >>>>>> struct pages (at least on x86). Use devm_memremap_pages() to map the > >>>>>> DAX window PCI BAR and allocate struct page. > >>>>>> > >>>>> > >>>>> Sorry for being this late. I don't see any more recent version so I will > >>>>> comment here. > >>>>> > >>>>> I'm trying to figure out how is this supposed to work on s390. My concern > >>>>> is, that on s390 PCI memory needs to be accessed by special > >>>>> instructions. This is taken care of by the stuff defined in > >>>>> arch/s390/include/asm/io.h. E.g. we 'override' __raw_writew so it uses > >>>>> the appropriate s390 instruction. However if the code does not use the > >>>>> linux abstractions for accessing PCI memory, but assumes it can be > >>>>> accessed like RAM, we have a problem. > >>>>> > >>>>> Looking at this patch, it seems to me, that we might end up with exactly > >>>>> the case described. For example AFAICT copy_to_iter() (3) resolves to > >>>>> the function in lib/iov_iter.c which does not seem to cater for s390 > >>>>> oddities. > >>>>> > >>>>> I didn't have the time to investigate this properly, and since virtio-fs > >>>>> is virtual, we may be able to get around what is otherwise a > >>>>> limitation on s390. My understanding of these areas is admittedly > >>>>> shallow, and since I'm not sure I'll have much more time to > >>>>> invest in the near future I decided to raise concern. > >>>>> > >>>>> Any opinions? > >>>> > >>>> Hi Halil, > >>>> > >>>> I don't understand s390 and how PCI works there as well. Is there any > >>>> other transport we can use there to map IO memory directly and access > >>>> using DAX? > >>>> > >>>> BTW, is DAX supported for s390. > >>>> > >>>> I am also hoping somebody who knows better can chip in. Till that time, > >>>> we could still use virtio-fs on s390 without DAX. > >>> > >>> s390 has so-called "limited" dax support, see CONFIG_FS_DAX_LIMITED. > >>> In practice that means that support for PTE_DEVMAP is missing which > >>> means no get_user_pages() support for dax mappings. Effectively it's > >>> only useful for execute-in-place as operations like fork() and ptrace > >>> of dax mappings will fail. > >> > >> > >> This is only true for the dcssblk device driver (drivers/s390/block/dcssblk.c > >> and arch/s390/mm/extmem.c). > >> > >> For what its worth, the dcssblk looks to Linux like normal memory (just above the > >> previously detected memory) that can be used like normal memory. In previous time > >> we even had struct pages for this memory - this was removed long ago (when it was > >> still xip) to reduce the memory footprint for large dcss blocks and small memory > >> guests. > >> Can the CONFIG_FS_DAX_LIMITED go away if we have struct pages for that memory? > >> > >> Now some observations: > >> - dcssblk is z/VM only (not KVM) > >> - Setting CONFIG_FS_DAX_LIMITED globally as a Kconfig option depending on wether > >> a device driver is compiled in or not seems not flexible enough in case if you > >> have device driver that does have struct pages and another one that doesn't > >> - I do not see a reason why we should not be able to map anything from QEMU > >> into the guest real memory via an additional KVM memory slot. > >> We would need to handle that in the guest somehow (and not as a PCI bar), > >> register this with struct pages etc. You mean for ccw, right? I don't think we want pci to behave differently than everywhere else. > >> - we must then look how we can create the link between the guest memory and the > >> virtio-fs driver. For virtio-ccw we might be able to add a new ccw command or > >> whatever. Maybe we could also piggy-back on some memory hotplug work from David > >> Hildenbrand (add cc). > >> > >> Regarding limitations on the platform: > >> - while we do have PCI, the virtio devices are usually plugged via the ccw bus. > >> That implies no PCI bars. I assume you use those PCI bars only to implicitely > >> have the location of the shared memory > >> Correct? > > > > Right. > > So in essence we just have to provide a vm_get_shm_region callback in the virtio-ccw > guest code? > > How many regions do we have to support? One region per device? Or many? > Even if we need more, this should be possible with a 2 new CCWs, e.g READ_SHM_BASE(id) > and READ_SHM_SIZE(id) I'd just add a single CCW with a control block containing id and size. The main issue is where we put those regions, and what happens if we use both virtio-pci and virtio-ccw on the same machine. > > > > > >> - no real memory mapped I/O. Instead there are instructions that work on the mmio. > >> As I understand things, this is of no concern regarding virtio-fs as you do not > >> need mmio in the sense that a memory access of the guest to such an address > >> triggers an exit. You just need the shared memory as a mean to have the data > >> inside the guest. Any notification is done via normal virtqueue mechanisms > >> Correct? > > > > Yep. >