Received: by 2002:a05:7412:b995:b0:f9:9502:5bb8 with SMTP id it21csp317465rdb; Thu, 21 Dec 2023 09:57:15 -0800 (PST) X-Google-Smtp-Source: AGHT+IE+uzQ2PL5biJpPSKbtT8yQWQvMXHXe3eyqG6rgJscRNna0lC0t9/ikaOPilzJ34oWxKhII X-Received: by 2002:a50:9983:0:b0:553:3a99:789f with SMTP id m3-20020a509983000000b005533a99789fmr3075038edb.29.1703181434920; Thu, 21 Dec 2023 09:57:14 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1703181434; cv=none; d=google.com; s=arc-20160816; b=iPnjFaOUIvlE1Otu3Ki0HvQHpm12vl52FAXj/g2qxadjOP8I2ble5HjASk84sU9XX7 bNecl0oQLWKplRIAKT7XzQh7YTFiY6//MLUOJQroYPJZr47MFRJjtecRJUXW0VfQfQRh gJnIV9m1xKuEnf2YRKAwRNcqLengttIPUBftRFwu+xtWstYCarAbuHc0eYYlqk/owcT2 S1N+vv+3adE3u64Xl2epoxo6cH8M2KD3yCs14LjIH046ok37AzFrfnDx6KaWzLPCyLXe hekDdUw+jiX0bHoLWfSw/yO8fcByW0rg7DapWBVW2BF8uj1R74ekbpiQ95gaJmJ53sgf /paQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=mime-version:list-unsubscribe:list-subscribe:list-id:precedence :content-transfer-encoding:autocrypt:references:in-reply-to:date:cc :to:from:subject:message-id:dkim-signature; bh=tW16GHCoszQpqU8oQ+WYfaEUemfkUyVapaAXPlGWyZc=; fh=r2m5yWJAWH4Y7NIIq14FbHhSv8/L+ICP6/Nsgp3jkqg=; b=HxAq0TMKROBMLhV+9EkhIYKVttD32QvEBHl+kyIaRw9pF0+eKOJe+LHDmTQ1hjtxGr mCBtuxVyjeLG48aV4x5IWBpjMwEOAhFKWsEF6n+m5VN40dP9ttF8iqzQzqy0Ym4EN/go 1VAZsfE7GQXwtc8mpf4lVI+8RCeZMeUANHXahZRGW69tTAS+m1i79Gr/rYr4ZA4URCde 0WObNDa9XOJz3cZe+dvD5opVOFif+sLEoLx5hQ7t4jww7dfel5EMfM01PAdQN+xQm5YY 1KDXbBlj9Vg2Qr64R5WrA9Y0uFD3ptgf+07V78ERElVjRpXDeZ7atRG+MbM2ohOmm1zd qJPQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@crapouillou.net header.s=mail header.b=S0+YXFQN; spf=pass (google.com: domain of linux-kernel+bounces-8882-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:4601:e00::3 as permitted sender) smtp.mailfrom="linux-kernel+bounces-8882-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=crapouillou.net Return-Path: Received: from am.mirrors.kernel.org (am.mirrors.kernel.org. [2604:1380:4601:e00::3]) by mx.google.com with ESMTPS id s19-20020a056402521300b00553f8be9748si1104025edd.396.2023.12.21.09.57.14 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 21 Dec 2023 09:57:14 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-8882-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:4601:e00::3 as permitted sender) client-ip=2604:1380:4601:e00::3; Authentication-Results: mx.google.com; dkim=pass header.i=@crapouillou.net header.s=mail header.b=S0+YXFQN; spf=pass (google.com: domain of linux-kernel+bounces-8882-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:4601:e00::3 as permitted sender) smtp.mailfrom="linux-kernel+bounces-8882-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=crapouillou.net Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by am.mirrors.kernel.org (Postfix) with ESMTPS id 823041F25BB8 for ; Thu, 21 Dec 2023 17:57:14 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 4CBA863503; Thu, 21 Dec 2023 17:57:00 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=crapouillou.net header.i=@crapouillou.net header.b="S0+YXFQN" X-Original-To: linux-kernel@vger.kernel.org Received: from aposti.net (aposti.net [89.234.176.197]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id C396555E7B; Thu, 21 Dec 2023 17:56:57 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=crapouillou.net Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=crapouillou.net DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=crapouillou.net; s=mail; t=1703181415; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:autocrypt:autocrypt; bh=tW16GHCoszQpqU8oQ+WYfaEUemfkUyVapaAXPlGWyZc=; b=S0+YXFQNlZOmKQI/q412vuw+c+sS1iiaBsZiv7j2MysLfOJWqD/9l2AVS+cPr0YJm9IJrP RZwcd4vioTZVcY5dDpY8q4GoCUoGVCsjx3pJ2Zg3XF6NYeTGdBtjCv9TKnGaUJlCOI0sDB iHwLhQGbXPaZnW2H3b7FNZxS7FAuKeE= Message-ID: Subject: Re: [PATCH v5 0/8] iio: new DMABUF based API, v5 From: Paul Cercueil To: Jonathan Cameron Cc: Lars-Peter Clausen , Sumit Semwal , Christian =?ISO-8859-1?Q?K=F6nig?= , Vinod Koul , Jonathan Corbet , linux-doc@vger.kernel.org, linux-kernel@vger.kernel.org, dmaengine@vger.kernel.org, linux-iio@vger.kernel.org, linux-media@vger.kernel.org, dri-devel@lists.freedesktop.org, linaro-mm-sig@lists.linaro.org, Nuno =?ISO-8859-1?Q?S=E1?= , Michael Hennerich Date: Thu, 21 Dec 2023 18:56:52 +0100 In-Reply-To: <20231221163031.1a410905@jic23-huawei> References: <20231219175009.65482-1-paul@crapouillou.net> <20231221163031.1a410905@jic23-huawei> Autocrypt: addr=paul@crapouillou.net; prefer-encrypt=mutual; keydata=mQENBF0KhcEBCADkfmrzdTOp/gFOMQX0QwKE2WgeCJiHPWkpEuPH81/HB2dpjPZNW03ZMLQfECbbaEkdbN4YnPfXgcc1uBe5mwOAPV1MBlaZcEt4M67iYQwSNrP7maPS3IaQJ18ES8JJ5Uf5UzFZaUawgH+oipYGW+v31cX6L3k+dGsPRM0Pyo0sQt52fsopNPZ9iag0iY7dGNuKenaEqkYNjwEgTtNz8dt6s3hMpHIKZFL3OhAGi88wF/21isv0zkF4J0wlf9gYUTEEY3Eulx80PTVqGIcHZzfavlWIdzhe+rxHTDGVwseR2Y1WjgFGQ2F+vXetAB8NEeygXee+i9nY5qt9c07m8mzjABEBAAG0JFBhdWwgQ2VyY3VlaWwgPHBhdWxAY3JhcG91aWxsb3UubmV0PokBTgQTAQoAOBYhBNdHYd8OeCBwpMuVxnPua9InSr1BBQJdCoXBAhsDBQsJCAcCBhUKCQgLAgQWAgMBAh4BAheAAAoJEHPua9InSr1BgvIH/0kLyrI3V0f33a6D3BJwc1grbygPVYGuC5l5eMnAI+rDmLR19E2yvibRpgUc87NmPEQPpbbtAZt8On/2WZoE5OIPdlId/AHNpdgAtGXo0ZX4LGeVPjxjdkbrKVHxbcdcnY+zzaFglpbVSvp76pxqgVg8PgxkAAeeJV+ET4t0823Gz2HzCL/6JZhvKAEtHVulOWoBh368SYdolp1TSfORWmHzvQiCCCA+j0cMkYVGzIQzEQhX7Urf9N/nhU5/SGLFEi9DcBfXoGzhyQyLXflhJtKm3XGB1K/pPulbKaPcKAl6rIDWPuFpHkSbmZ9r4KFlBwgAhlGy6nqP7O3u7q23hRW5AQ0EXQqFwQEIAMo+MgvYHsyjX3Ja4Oolg1Txzm8woj30ch2nACFCqaO0R/1kLj2VVeLrDyQUOlXx9PD6IQI4M8wy8m0sR4wV2p/g/paw7k65cjzYYLh+FdLNyO7IW YXndJO+wDPi3aK/YKUYepqlP+QsmaHNYNdXEQDRKqNfJg8t0f5rfzp9ryxd1tCnbV+tG8VHQWiZXNqN7062DygSNXFUfQ0vZ3J2D4oAcIAEXTymRQ2+hr3Hf7I61KMHWeSkCvCG2decTYsHlw5Erix/jYWqVOtX0roOOLqWkqpQQJWtU+biWrAksmFmCp5fXIg1Nlg39v21xCXBGxJkxyTYuhdWyu1yDQ+LSIUAEQEAAYkBNgQYAQoAIBYhBNdHYd8OeCBwpMuVxnPua9InSr1BBQJdCoXBAhsMAAoJEHPua9InSr1B4wsH/Az767YCT0FSsMNt1jkkdLCBi7nY0GTW+PLP1a4zvVqFMo/vD6uz1ZflVTUAEvcTi3VHYZrlgjcxmcGu239oruqUS8Qy/xgZBp9KF0NTWQSl1iBfVbIU5VV1vHS6r77W5x0qXgfvAUWOH4gmN3MnF01SH2zMcLiaUGF+mcwl15rHbjnT3Nu2399aSE6cep86igfCAyFUOXjYEGlJy+c6UyT+DUylpjQg0nl8MlZ/7Whg2fAU9+FALIbQYQzGlT4c71SibR9T741jnegHhlmV4WXXUD6roFt54t0MSAFSVxzG8mLcSjR2cLUJ3NIPXixYUSEn3tQhfZj07xIIjWxAYZo= Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Hi Jonathan, Le jeudi 21 d=C3=A9cembre 2023 =C3=A0 16:30 +0000, Jonathan Cameron a =C3= =A9crit=C2=A0: > On Tue, 19 Dec 2023 18:50:01 +0100 > Paul Cercueil wrote: >=20 > > [V4 was: "iio: Add buffer write() support"][1] > >=20 > > Hi Jonathan, > >=20 > Hi Paul, >=20 > > This is a respin of the V3 of my patchset that introduced a new > > interface based on DMABUF objects [2]. >=20 > Great to see this moving forwards. >=20 > >=20 > > The V4 was a split of the patchset, to attempt to upstream buffer > > write() support first. But since there is no current user upstream, > > it > > was not merged. This V5 is about doing the opposite, and contains > > the > > new DMABUF interface, without adding the buffer write() support. It > > can > > already be used with the upstream adi-axi-adc driver. >=20 > Seems like a sensible path in the short term. >=20 > >=20 > > In user-space, Libiio uses it to transfer back and forth blocks of > > samples between the hardware and the applications, without having > > to > > copy the data. > >=20 > > On a ZCU102 with a FMComms3 daughter board, running Libiio from the > > pcercuei/dev-new-dmabuf-api branch [3], compiled with > > WITH_LOCAL_DMABUF_API=3DOFF (so that it uses fileio): > > =C2=A0 sudo utils/iio_rwdev -b 4096 -B cf-ad9361-lpc > > =C2=A0 Throughput: 116 MiB/s > >=20 > > Same hardware, with the DMABUF API (WITH_LOCAL_DMABUF_API=3DON): > > =C2=A0 sudo utils/iio_rwdev -b 4096 -B cf-ad9361-lpc > > =C2=A0 Throughput: 475 MiB/s > >=20 > > This benchmark only measures the speed at which the data can be > > fetched > > to iio_rwdev's internal buffers, and does not actually try to read > > the > > data (e.g. to pipe it to stdout). It shows that fetching the data > > is > > more than 4x faster using the new interface. > >=20 > > When actually reading the data, the performance difference isn't > > that > > impressive (maybe because in case of DMABUF the data is not in > > cache): >=20 > This needs a bit more investigation ideally. Perhaps perf counters > can be > used to establish that cache misses are the main different between > dropping it on the floor and actually reading the data. Yes, we'll work on it. The other big difference is that fileio uses dma_alloc_coherent() while the DMABUFs use non-coherent mappings. I guess coherent memory is faster for the typical access pattern (which is "read/write everything sequentially once"). > >=20 > > WITH_LOCAL_DMABUF_API=3DOFF (so that it uses fileio): > > =C2=A0 sudo utils/iio_rwdev -b 4096 cf-ad9361-lpc | dd of=3D/dev/zero > > status=3Dprogress > > =C2=A0 2446422528 bytes (2.4 GB, 2.3 GiB) copied, 22 s, 111 MB/s > >=20 > > WITH_LOCAL_DMABUF_API=3DON: > > =C2=A0 sudo utils/iio_rwdev -b 4096 cf-ad9361-lpc | dd of=3D/dev/zero > > status=3Dprogress > > =C2=A0 2334388736 bytes (2.3 GB, 2.2 GiB) copied, 21 s, 114 MB/s > >=20 > > One interesting thing to note is that fileio is (currently) > > actually > > faster than the DMABUF interface if you increase a lot the buffer > > size. > > My explanation is that the cache invalidation routine takes more > > and > > more time the bigger the DMABUF gets. This is because the DMABUF is > > backed by small-size pages, so a (e.g.) 64 MiB DMABUF is backed by > > up > > to 16 thousands pages, that have to be invalidated one by one. This > > can > > be addressed by using huge pages, but the udmabuf driver does not > > (yet) > > support creating DMABUFs backed by huge pages. >=20 > I'd imagine folios of reasonable size will help sort of a huge page > as then hopefully it will use the flush by va range instructions if > available. >=20 > >=20 > > Anyway, the real benefits happen when the DMABUFs are either shared > > between IIO devices, or between the IIO subsystem and another > > filesystem. In that case, the DMABUFs are simply passed around > > drivers, > > without the data being copied at any moment. > >=20 > > We use that feature to transfer samples from our transceivers to > > USB, > > using a DMABUF interface to FunctionFS [4]. > >=20 > > This drastically increases the throughput, to about 274 MiB/s over > > a > > USB3 link, vs. 127 MiB/s using IIO's fileio interface + write() to > > the > > FunctionFS endpoints, for a lower CPU usage (0.85 vs. 0.65 load > > avg.). >=20 > This is a nice example.=C2=A0 Where are you with getting the patch merged= ? I'll send a new version (mostly a [RESEND]...) in the coming days. As you can see from the review on my last attempt, the main blocker is that nobody wants to merge a new interface if the rest of the kernel bits aren't upstream yet. Kind of a chicken-and-egg problem :) > Overall, this code looks fine to me, though there are some parts that > need review by other maintainers (e.g. Vinod for the dmaengine > callback) > and I'd like a 'looks fine' at least form those who know a lot more > about dmabuf than I do. >=20 > To actually make this useful sounds like either udmabuf needs some > perf > improvements, or there has to be an upstream case of sharing it > without > something else (e.g your functionfs patches).=C2=A0 So what do we need to > get in before the positive benefit becomes worth carrying this extra > complexity? (which isn't too bad so I'm fine with a small benefit and > promises of riches :) I think the FunctionFS DMABUF interface can be pushed as well for 5.9, in parallel of this one, as the feedback on the V1 was good. I might just need some help pushing it forward (kind of a "I merge it if you merge it" guarantee). Cheers, -Paul >=20 > Jonathan >=20 > >=20 > > Based on linux-next/next-20231219. > >=20 > > Cheers, > > -Paul > >=20 > > [1] > > https://lore.kernel.org/all/20230807112113.47157-1-paul@crapouillou.net= / > > [2] > > https://lore.kernel.org/all/20230403154800.215924-1-paul@crapouillou.ne= t/ > > [3] > > https://github.com/analogdevicesinc/libiio/tree/pcercuei/dev-new-dmabuf= -api > > [4] > > https://lore.kernel.org/all/20230322092118.9213-1-paul@crapouillou.net/ > >=20 > > --- > > Changelog: > > - [3/8]: Replace V3's dmaengine_prep_slave_dma_array() with a new > > =C2=A0 dmaengine_prep_slave_dma_vec(), which uses a new 'dma_vec' > > struct. > > =C2=A0 Note that at some point we will need to support cyclic transfers > > =C2=A0 using dmaengine_prep_slave_dma_vec(). Maybe with a new "flags" > > =C2=A0 parameter to the function? > >=20 > > - [4/8]: Implement .device_prep_slave_dma_vec() instead of V3's > > =C2=A0 .device_prep_slave_dma_array(). > >=20 > > =C2=A0 @Vinod: this patch will cause a small conflict with my other > > =C2=A0 patchset adding scatter-gather support to the axi-dmac driver. > > =C2=A0 This patch adds a call to axi_dmac_alloc_desc(num_sgs), but the > > =C2=A0 prototype of this function changed in my other patchset - it > > would > > =C2=A0 have to be passed the "chan" variable. I don't know how you > > prefer it > > =C2=A0 to be resolved. Worst case scenario (and if @Jonathan is okay > > with > > =C2=A0 that) this one patch can be re-sent later, but it would make thi= s > > =C2=A0 patchset less "atomic". > >=20 > > - [5/8]: > > =C2=A0 - Use dev_err() instead of pr_err() > > =C2=A0 - Inline to_iio_dma_fence() > > =C2=A0 - Add comment to explain why we unref twice when detaching dmabu= f > > =C2=A0 - Remove TODO comment. It is actually safe to free the file's > > =C2=A0=C2=A0=C2=A0 private data even when transfers are still pending b= ecause it > > =C2=A0=C2=A0=C2=A0 won't be accessed. > > =C2=A0 - Fix documentation of new fields in struct > > iio_buffer_access_funcs > > =C2=A0 - iio_dma_resv_lock() does not need to be exported, make it > > static > >=20 > > - [7/8]: > > =C2=A0 - Use the new dmaengine_prep_slave_dma_vec(). > > =C2=A0 - Restrict to input buffers, since output buffers are not yet > > =C2=A0=C2=A0=C2=A0 supported by IIO buffers. > >=20 > > - [8/8]: > > =C2=A0 Use description lists for the documentation of the three new > > IOCTLs > > =C2=A0 instead of abusing subsections. > >=20 > > --- > > Alexandru Ardelean (1): > > =C2=A0 iio: buffer-dma: split iio_dma_buffer_fileio_free() function > >=20 > > Paul Cercueil (7): > > =C2=A0 iio: buffer-dma: Get rid of outgoing queue > > =C2=A0 dmaengine: Add API function dmaengine_prep_slave_dma_vec() > > =C2=A0 dmaengine: dma-axi-dmac: Implement device_prep_slave_dma_vec > > =C2=A0 iio: core: Add new DMABUF interface infrastructure > > =C2=A0 iio: buffer-dma: Enable support for DMABUFs > > =C2=A0 iio: buffer-dmaengine: Support new DMABUF based userspace API > > =C2=A0 Documentation: iio: Document high-speed DMABUF based API > >=20 > > =C2=A0Documentation/iio/dmabuf_api.rst=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2= =A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 |=C2=A0 54 +++ > > =C2=A0Documentation/iio/index.rst=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0= =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 |= =C2=A0=C2=A0 2 + > > =C2=A0drivers/dma/dma-axi-dmac.c=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2= =A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0= |=C2=A0 40 ++ > > =C2=A0drivers/iio/buffer/industrialio-buffer-dma.c=C2=A0 | 242 ++++++++= --- > > =C2=A0.../buffer/industrialio-buffer-dmaengine.c=C2=A0=C2=A0=C2=A0 |=C2= =A0 52 ++- > > =C2=A0drivers/iio/industrialio-buffer.c=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0= =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 | 402 > > ++++++++++++++++++ > > =C2=A0include/linux/dmaengine.h=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2= =A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0= =C2=A0 |=C2=A0 25 ++ > > =C2=A0include/linux/iio/buffer-dma.h=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2= =A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 |=C2=A0 33 +- > > =C2=A0include/linux/iio/buffer_impl.h=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2= =A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 |=C2=A0 26 ++ > > =C2=A0include/uapi/linux/iio/buffer.h=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2= =A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 |=C2=A0 22 + > > =C2=A010 files changed, 836 insertions(+), 62 deletions(-) > > =C2=A0create mode 100644 Documentation/iio/dmabuf_api.rst > >=20 >=20