Received: by 2002:a89:2c3:0:b0:1ed:23cc:44d1 with SMTP id d3csp360533lqs; Tue, 5 Mar 2024 04:31:29 -0800 (PST) X-Forwarded-Encrypted: i=3; AJvYcCUvYIzmBEpesC7o13Cv+zKITj2aBETuY3Zf76nkoFji3HYFGv2Nofs/whC1cxM/AYIVnCzVfubgv93A1irhN1lN0SUV3Pri5oTfg8734Q== X-Google-Smtp-Source: AGHT+IHmpviqCYewj41iXhpVzXUKTPUeVWPaI9nj9dBcobRXGUgiG5apwU3RA8wddsaVaEZn8YK6 X-Received: by 2002:a05:6808:b:b0:3c1:aadd:e397 with SMTP id u11-20020a056808000b00b003c1aadde397mr1556717oic.53.1709641889389; Tue, 05 Mar 2024 04:31:29 -0800 (PST) ARC-Seal: i=2; a=rsa-sha256; t=1709641889; cv=pass; d=google.com; s=arc-20160816; b=j8xauZfoj9sCxctFDTEUhGQ0ajLFtrduF1CHsKeLD1/tWH6McxxCGZ+y7n4btxnoVs hba4RWVXpuNaKyjBCdXvQ+5bNtoXf/RuNiruTaa9x9gXOJiJIXBbqsybuRMMOzivBT0O 9nwsUqAzuam6dr6zlI0X2RWgeGDA9otyKI2jMfVroFIDeliF5aKCSa8ZQc0ukVPLx+7l OQrGP41ccxBgBZRMUMlIwihAZcf4fQer2TSb8n2HlCtaI9zo9lneoKBKZ8XWZGBEUczA 2v3KTWjBe8aAawy+n3axLSpHg0+xeSLssptWM/AcnJfGBSg9NIHOGU2aqf/2yki92Fj7 6qTw== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=in-reply-to:content-disposition:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:references:message-id:subject:cc :to:from:date:dkim-signature; bh=aJ5iE+LWxcrer1v/M54isoTyRWcTu5SpNDLz6j/UZx0=; fh=GhEVavzl/DXM+eozDuJ7Ib/qVrbBUMI01LnBNDhE8VM=; b=StwdUzYfW3jeqjW317UQQAMxSKUnohwvOcM1bqN5wk2cwraQK1ZFFgUXyUlw0MlMqz xzKEX7jhWJn09QG1BYefnf7qjqz9kZNOF90VbovVFnWx9qD564d0jUuv1eB2oDFL0203 CUFnAyXDZCUGH39/1DC5yT05MLAuizFQHhYIcHmjsKwkC23w1o7NwXtsD2S1s8Z8nmHs Vq8dl7+nymNkeoaf+0vH+4kHsppoZYUOCPY4jp6r4tlVfE9smFtRjqduyXAkLWcwR15x nZG1VoXaJhQ1D6srWB2VZsNVa4EPP3nZFi8MjJ7kNMjXKcWYVO2df2xBenX41ORBgeL+ uT/Q==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=edE9W4ex; arc=pass (i=1 dkim=pass dkdomain=kernel.org); spf=pass (google.com: domain of linux-kernel+bounces-92377-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-92377-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from ny.mirrors.kernel.org (ny.mirrors.kernel.org. [2604:1380:45d1:ec00::1]) by mx.google.com with ESMTPS id i6-20020ac85c06000000b0042f04df6326si607659qti.603.2024.03.05.04.31.29 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 05 Mar 2024 04:31:29 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-92377-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) client-ip=2604:1380:45d1:ec00::1; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=edE9W4ex; arc=pass (i=1 dkim=pass dkdomain=kernel.org); spf=pass (google.com: domain of linux-kernel+bounces-92377-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-92377-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ny.mirrors.kernel.org (Postfix) with ESMTPS id C2D351C22B07 for ; Tue, 5 Mar 2024 12:30:43 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 050495B666; Tue, 5 Mar 2024 12:29:42 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b="edE9W4ex" Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id C5ABA5A4C0; Tue, 5 Mar 2024 12:29:40 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1709641780; cv=none; b=hfYNNUReNlZMXRZxX7n+9ff5ePUYIFOeRNJ8GMdEs3H9zzy+kyaZ3isQJxiQMRtBUHfCrhvUjQEX/uLDEdb0tMg8UJdHEWs0sKGCEOnAnFdjRJz2/EjNtMYAOt50jNj8GPDKNIne5VPU/POin1oodQSoLmnx7XedopnZKt6fwbU= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1709641780; c=relaxed/simple; bh=P93hlccw+xdxdbFmx+6Lr6WNHOyANq8FCYqtCUP40hE=; h=Date:From:To:Cc:Subject:Message-ID:References:MIME-Version: Content-Type:Content-Disposition:In-Reply-To; b=ZrRZobJVkqLI6tkAw6zoFQ1Er+LgswJYKbECr2TeydSJK4T3yADjX+fOdGgUbgsUGWQpIKoOocd9PCW9Moq+d8RRtt9DkjIRMU5vLZzfoulBJmm5/p4OF7U1ECFXd6RaWwlP1qdVefOxy6qb4zEexBOJzjkF+9dukApA8zwXY4w= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b=edE9W4ex; arc=none smtp.client-ip=10.30.226.201 Received: by smtp.kernel.org (Postfix) with ESMTPSA id 7AE6CC43394; Tue, 5 Mar 2024 12:29:39 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1709641780; bh=P93hlccw+xdxdbFmx+6Lr6WNHOyANq8FCYqtCUP40hE=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=edE9W4exQ/mLyQoFZ0ecW6HlLDr9Z099WMeUuaFujQg2FZTZLAQnGDjK498JWa0ni hqFMO5hIGroOgoA/TTNGplduxItE+LvgFwOBWjD33xVKfmTFkJ5U+9CtOOpdfL7udr qhAc/itUetIkKuPj6Jv0+bNurWlOja/yk+vF05wSc8ytz4RcHmc2Kw2jU+9t5qdqtK 2MmLJphfTq56uO+Iq2YwNgSXMY5Hs5aBypVWlDYoEFPocQlvJgPL/m9HUD+GamQVTE DFv6scJ4w1oMC0aHDm5eYXuKeZW4m+eUbOnkHeJ/XxbIduCTrPWiF//VLFhaMH4KXv NmAjytaBBw2hw== Date: Tue, 5 Mar 2024 14:29:35 +0200 From: Leon Romanovsky To: Robin Murphy Cc: Christoph Hellwig , Marek Szyprowski , Joerg Roedel , Will Deacon , Jason Gunthorpe , Chaitanya Kulkarni , Jonathan Corbet , Jens Axboe , Keith Busch , Sagi Grimberg , Yishai Hadas , Shameer Kolothum , Kevin Tian , Alex Williamson , =?iso-8859-1?B?Suly9G1l?= Glisse , Andrew Morton , linux-doc@vger.kernel.org, linux-kernel@vger.kernel.org, linux-block@vger.kernel.org, linux-rdma@vger.kernel.org, iommu@lists.linux.dev, linux-nvme@lists.infradead.org, kvm@vger.kernel.org, linux-mm@kvack.org, Bart Van Assche , Damien Le Moal , Amir Goldstein , "josef@toxicpanda.com" , "Martin K. Petersen" , "daniel@iogearbox.net" , Dan Williams , "jack@suse.com" , Zhu Yanjun Subject: Re: [RFC RESEND 00/16] Split IOMMU DMA mapping operation to two steps Message-ID: <20240305122935.GB36868@unreal> References: <47afacda-3023-4eb7-b227-5f725c3187c2@arm.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <47afacda-3023-4eb7-b227-5f725c3187c2@arm.com> On Tue, Mar 05, 2024 at 12:05:23PM +0000, Robin Murphy wrote: > On 2024-03-05 11:18 am, Leon Romanovsky wrote: > > This is complimentary part to the proposed LSF/MM topic. > > https://lore.kernel.org/linux-rdma/22df55f8-cf64-4aa8-8c0b-b556c867b926@linux.dev/T/#m85672c860539fdbbc8fe0f5ccabdc05b40269057 > > > > This is posted as RFC to get a feedback on proposed split, but RDMA, VFIO and > > DMA patches are ready for review and inclusion, the NVMe patches are still in > > progress as they require agreement on API first. > > > > Thanks > > > > ------------------------------------------------------------------------------- > > The DMA mapping operation performs two steps at one same time: allocates > > IOVA space and actually maps DMA pages to that space. This one shot > > operation works perfectly for non-complex scenarios, where callers use > > that DMA API in control path when they setup hardware. > > > > However in more complex scenarios, when DMA mapping is needed in data > > path and especially when some sort of specific datatype is involved, > > such one shot approach has its drawbacks. > > > > That approach pushes developers to introduce new DMA APIs for specific > > datatype. For example existing scatter-gather mapping functions, or > > latest Chuck's RFC series to add biovec related DMA mapping [1] and > > probably struct folio will need it too. > > > > These advanced DMA mapping APIs are needed to calculate IOVA size to > > allocate it as one chunk and some sort of offset calculations to know > > which part of IOVA to map. > > I don't follow this part at all - at *some* point, something must know a > range of memory addresses involved in a DMA transfer, so that's where it > should map that range for DMA. In all presented cases in this series, the overall DMA size is known in advance. In RDMA case, it is known when user registers the memory, in VFIO, when live migration is happening and in NVMe, when BIO is created. So once we allocated IOVA, we will need to link ranges, which si the same as map but without IOVA allocation. > Even in a badly-designed system where the > point it's most practical to make the mapping is further out and only knows > that DMA will touch some subset of a buffer, but doesn't know exactly what > subset yet, you'd usually just map the whole buffer. I don't see why the DMA > API would ever need to know about anything other than pages/PFNs and > dma_addr_ts (yes, it does also accept them being wrapped together in > scatterlists; yes, scatterlists are awful and it would be nice to replace > them with a better general DMA descriptor; that is a whole other subject of > its own). This is exactly what was done here, we got rid of scatterlists. > > > Instead of teaching DMA to know these specific datatypes, let's separate > > existing DMA mapping routine to two steps and give an option to advanced > > callers (subsystems) perform all calculations internally in advance and > > map pages later when it is needed. > > From a brief look, this is clearly an awkward reinvention of the IOMMU API. > If IOMMU-aware drivers/subsystems want to explicitly manage IOMMU address > spaces then they can and should use the IOMMU API. Perhaps there's room for > some quality-of-life additions to the IOMMU API to help with common usage > patterns, but the generic DMA mapping API is absolutely not the place for > it. DMA mapping gives nice abstraction from IOMMU, and allows us to have same flow for IOMMU and non-IOMMU flows without duplicating code, while you suggest to teach almost every part in the kernel to know about IOMMU. In this series, we changed RDMA, VFIO and NVMe, and in all cases we removed more code than added. From what I saw, VDPA and virito-blk will benefit from proposed API too. Even in this RFC, where Chaitanya did partial job and didn't convert whole driver, the gain is pretty obvious: https://lore.kernel.org/linux-rdma/016fc02cbfa9be3c156a6f74df38def1e09c08f1.1709635535.git.leon@kernel.org/T/#u drivers/nvme/host/pci.c | 220 ++++++++++++++++++++++++++++++++++++++++++++++---------------------------------------------------------------------------------------------------------------------------------------------------------------- 1 file changed, 49 insertions(+), 171 deletions(-) Thanks > > Thanks, > Robin. > > > In this series, three users are converted and each of such conversion > > presents different positive gain: > > 1. RDMA simplifies and speeds up its pagefault handling for > > on-demand-paging (ODP) mode. > > 2. VFIO PCI live migration code saves huge chunk of memory. > > 3. NVMe PCI avoids intermediate SG table manipulation and operates > > directly on BIOs. > > > > Thanks > > > > [1] https://lore.kernel.org/all/169772852492.5232.17148564580779995849.stgit@klimt.1015granger.net > > > > Chaitanya Kulkarni (2): > > block: add dma_link_range() based API > > nvme-pci: use blk_rq_dma_map() for NVMe SGL > > > > Leon Romanovsky (14): > > mm/hmm: let users to tag specific PFNs > > dma-mapping: provide an interface to allocate IOVA > > dma-mapping: provide callbacks to link/unlink pages to specific IOVA > > iommu/dma: Provide an interface to allow preallocate IOVA > > iommu/dma: Prepare map/unmap page functions to receive IOVA > > iommu/dma: Implement link/unlink page callbacks > > RDMA/umem: Preallocate and cache IOVA for UMEM ODP > > RDMA/umem: Store ODP access mask information in PFN > > RDMA/core: Separate DMA mapping to caching IOVA and page linkage > > RDMA/umem: Prevent UMEM ODP creation with SWIOTLB > > vfio/mlx5: Explicitly use number of pages instead of allocated length > > vfio/mlx5: Rewrite create mkey flow to allow better code reuse > > vfio/mlx5: Explicitly store page list > > vfio/mlx5: Convert vfio to use DMA link API > > > > Documentation/core-api/dma-attributes.rst | 7 + > > block/blk-merge.c | 156 ++++++++++++++ > > drivers/infiniband/core/umem_odp.c | 219 +++++++------------ > > drivers/infiniband/hw/mlx5/mlx5_ib.h | 1 + > > drivers/infiniband/hw/mlx5/odp.c | 59 +++-- > > drivers/iommu/dma-iommu.c | 129 ++++++++--- > > drivers/nvme/host/pci.c | 220 +++++-------------- > > drivers/vfio/pci/mlx5/cmd.c | 252 ++++++++++++---------- > > drivers/vfio/pci/mlx5/cmd.h | 22 +- > > drivers/vfio/pci/mlx5/main.c | 136 +++++------- > > include/linux/blk-mq.h | 9 + > > include/linux/dma-map-ops.h | 13 ++ > > include/linux/dma-mapping.h | 39 ++++ > > include/linux/hmm.h | 3 + > > include/rdma/ib_umem_odp.h | 22 +- > > include/rdma/ib_verbs.h | 54 +++++ > > kernel/dma/debug.h | 2 + > > kernel/dma/direct.h | 7 +- > > kernel/dma/mapping.c | 91 ++++++++ > > mm/hmm.c | 34 +-- > > 20 files changed, 870 insertions(+), 605 deletions(-) > >