Received: by 2002:a05:7412:8d10:b0:f3:1519:9f41 with SMTP id bj16csp4229600rdb; Mon, 11 Dec 2023 12:37:25 -0800 (PST) X-Google-Smtp-Source: AGHT+IGkjOHtcw+iyhmHXwcU+bHdniYhI9UibYvPUlB27SzMB+SICaItMcE6a1E9hIXkSUZADAUj X-Received: by 2002:a17:902:b706:b0:1cf:9790:f244 with SMTP id d6-20020a170902b70600b001cf9790f244mr2505780pls.47.1702327044968; Mon, 11 Dec 2023 12:37:24 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1702327044; cv=none; d=google.com; s=arc-20160816; b=YooJjeEMD1BB5BKcDpsD2Ji8eNoLcj3UQ6nBs4XoL/WeHEqgmA1St3sUA58n/v1r0r 3VXLkjKFDFsPq77yAg3to2Zs3IT15YhMaXIX8Z/GKp16mfBt74w0NpMp/0TCpSgLs1Pj NZNCbx1Hqar0zJNQi2ScvSzezTYQmxs5Ze6XT8VXeHx8AaYSwski/QWOPTcdKQJzD+ET dahSkRfLrlZWUw4AFAj1YUkJkp3Bs2ithGwFcEuKYanXqKuMS3EokN2KBPStqL1/97za zN/vZzDswqvpLPBBZA07L/Hpao9mCpKYNRrKfeXVoKxgh64RCb1vf3qxUqRfqxmpoSOO cLKw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:from :content-language:references:cc:to:subject:user-agent:mime-version :date:message-id:dkim-signature; bh=5mREG4zmhlrjDE364pycILMuHT3yybZyzypq/2nes4o=; fh=g6R8eOJ4BHfVwsSKujzw1tI9AAeFHxXwUAvPpPn0m0M=; b=BXxBUYfFNHrGtZYdxT2tXdORDowfgOeSBBZ6Mdm2z+uyhYSYrC1YyGfhsisftww+Nr VJ96A+r1/3z6Kcf549Ks0XAQEpsxZz86e/9Ln7V+yeC7Mnlqvyy7uSKTL98SQAs9EGQx TCL8rS2jlArsD7jexD4VSrizzakUhnJ/dZ8imYS4qHqQ6VXXKiahlFb5mVaTVL3+30gr DFodtRNpdZa4ZFiRnhygS0nsLlJ+P/RoxXBb0T1QScm17N4DRzz9511wAIL5G4u/QyiC DHxO4w7V2bMyeYI8xTGqlBUHKeuTXh99GQKrmDGnvD87iQSxMVAQpjFuaukfRrsKYV0l ZQ/Q== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20230601 header.b=V6q1zdkV; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:7 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from snail.vger.email (snail.vger.email. [2620:137:e000::3:7]) by mx.google.com with ESMTPS id k17-20020a170902c41100b001cfe100a99fsi6822088plk.487.2023.12.11.12.37.24 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 11 Dec 2023 12:37:24 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:7 as permitted sender) client-ip=2620:137:e000::3:7; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20230601 header.b=V6q1zdkV; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:7 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by snail.vger.email (Postfix) with ESMTP id 638208051913; Mon, 11 Dec 2023 12:37:23 -0800 (PST) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.11 at snail.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1344409AbjLKUhO (ORCPT + 99 others); Mon, 11 Dec 2023 15:37:14 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54200 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229587AbjLKUhN (ORCPT ); Mon, 11 Dec 2023 15:37:13 -0500 Received: from mail-wr1-x436.google.com (mail-wr1-x436.google.com [IPv6:2a00:1450:4864:20::436]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id B7DBDD6; Mon, 11 Dec 2023 12:37:18 -0800 (PST) Received: by mail-wr1-x436.google.com with SMTP id ffacd0b85a97d-3332e351670so4516544f8f.0; Mon, 11 Dec 2023 12:37:18 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20230601; t=1702327037; x=1702931837; darn=vger.kernel.org; h=content-transfer-encoding:in-reply-to:from:content-language :references:cc:to:subject:user-agent:mime-version:date:message-id :from:to:cc:subject:date:message-id:reply-to; bh=5mREG4zmhlrjDE364pycILMuHT3yybZyzypq/2nes4o=; b=V6q1zdkV+wxIKXpM7DWq66B2NPzZBUPK4hr3WRtG+MTVSfExKywo7PK4Bt3nluEqDT MkYrCuuDgvHFyqtkeTL5EVEmHFlAqoCG4wzA1E0AOVgzV7GgxwkQ9vs2R6aGYX15ri2/ sp1/AuQvVGRbOCriyt32YoOicKW7+0NZ50d5dTXoKn41lk//umXIHVlwVFUrW8yo+vYp lrn9Cd0RaMPdIaG8HHIVCsPOhnrT3YGAomYLF9MeJLVuBcHmmXArVou5P2kL/Ae82RHR ZnSLIq6BPwppSyaCaayBA0PoehLyk2aJOEhKI7Uz3FPmnyjCKCcFU5bviWtGyL9hN/ou 07XA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1702327037; x=1702931837; h=content-transfer-encoding:in-reply-to:from:content-language :references:cc:to:subject:user-agent:mime-version:date:message-id :x-gm-message-state:from:to:cc:subject:date:message-id:reply-to; bh=5mREG4zmhlrjDE364pycILMuHT3yybZyzypq/2nes4o=; b=XUGOs0D1C5EKC/b3Z0mxu1ogVHxrWSHJxBA65PBIYaZiH9yi8GafUYjsyV8T4DwD50 iI/eYdSCMb2iXzA3PXMw5XKi5yiYfZAet5sASmkxXIgBAHe7dc7e+81MnVu4zONtwbgJ e9xqoQTLS1VsUyMbW8DnbrP/FQ/3S73ZrSodhoH31n3CPmfICPsRdUav5uIZ0y+l2p2f BseSMW0UlBock5k7uzKK3pN9UsOmxf/FSvZP268P+JIQRei1jaSBheA1/uU9tAhth7Dp fvX9tXjqjGZbrnAh1rP5Lge2WvBvNqA5x/tTQ4UEppdSvJroTP+V//2x/hUYGgTp/CUf WKGg== X-Gm-Message-State: AOJu0YxI/1+RqepqcdtaXs8ZPwTXd9Ha46KBp8URvAL6Hv6fQJGeJM1S HOj5ktQZ5YXRgkSxBIL4Nfo= X-Received: by 2002:adf:a198:0:b0:333:38eb:8947 with SMTP id u24-20020adfa198000000b0033338eb8947mr1043696wru.275.1702327036762; Mon, 11 Dec 2023 12:37:16 -0800 (PST) Received: from [192.168.8.100] ([85.255.234.108]) by smtp.gmail.com with ESMTPSA id o4-20020a5d58c4000000b0033333bee379sm9328312wrf.107.2023.12.11.12.37.14 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Mon, 11 Dec 2023 12:37:16 -0800 (PST) Message-ID: <661c1bae-d7d3-457e-b545-5f67b9ef4197@gmail.com> Date: Mon, 11 Dec 2023 20:35:54 +0000 MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [net-next v1 08/16] memory-provider: dmabuf devmem memory provider To: Mina Almasry Cc: Shailend Chand , netdev@vger.kernel.org, linux-kernel@vger.kernel.org, linux-doc@vger.kernel.org, linux-arch@vger.kernel.org, linux-kselftest@vger.kernel.org, bpf@vger.kernel.org, linux-media@vger.kernel.org, dri-devel@lists.freedesktop.org, "David S. Miller" , Eric Dumazet , Jakub Kicinski , Paolo Abeni , Jonathan Corbet , Jeroen de Borst , Praveen Kaligineedi , Jesper Dangaard Brouer , Ilias Apalodimas , Arnd Bergmann , David Ahern , Willem de Bruijn , Shuah Khan , Sumit Semwal , =?UTF-8?Q?Christian_K=C3=B6nig?= , Yunsheng Lin , Harshitha Ramamurthy , Shakeel Butt , Willem de Bruijn , Kaiyuan Zhang References: <20231208005250.2910004-1-almasrymina@google.com> <20231208005250.2910004-9-almasrymina@google.com> Content-Language: en-US From: Pavel Begunkov In-Reply-To: Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,FREEMAIL_FROM, RCVD_IN_DNSWL_BLOCKED,SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (snail.vger.email [0.0.0.0]); Mon, 11 Dec 2023 12:37:23 -0800 (PST) On 12/11/23 02:30, Mina Almasry wrote: > On Sat, Dec 9, 2023 at 7:05 PM Pavel Begunkov wrote: >> >> On 12/8/23 23:25, Mina Almasry wrote: >>> On Fri, Dec 8, 2023 at 2:56 PM Pavel Begunkov wrote: >>>> >>>> On 12/8/23 00:52, Mina Almasry wrote: >>> ... >>>>> + if (pool->p.queue) >>>>> + binding = READ_ONCE(pool->p.queue->binding); >>>>> + >>>>> + if (binding) { >>>>> + pool->mp_ops = &dmabuf_devmem_ops; >>>>> + pool->mp_priv = binding; >>>>> + } >>>> >>>> Hmm, I don't understand why would we replace a nice transparent >>>> api with page pool relying on a queue having devmem specific >>>> pointer? It seemed more flexible and cleaner in the last RFC. >>>> >>> >>> Jakub requested this change and may chime in, but I suspect it's to >>> further abstract the devmem changes from driver. In this iteration, >>> the driver grabs the netdev_rx_queue and passes it to the page_pool, >>> and any future configurations between the net stack and page_pool can >>> be passed this way with the driver unbothered. >> >> Ok, that makes sense, but even if passed via an rx queue I'd >> at least hope it keeping abstract provider parameters, e.g. >> ops, but not hard coded with devmem specific code. >> >> It might even be better done with a helper like >> create_page_pool_from_queue(), unless there is some deeper >> interaction b/w pp and rx queues is predicted. >> > > Off hand I don't see the need for a new create_page_pool_from_queue(). > page_pool_create() already takes in a param arg that lets us pass in > the queue as well as any other params. > >>>>> + >>>>> if (pool->mp_ops) { >>>>> err = pool->mp_ops->init(pool); >>>>> if (err) { >>>>> @@ -1020,3 +1033,77 @@ void page_pool_update_nid(struct page_pool *pool, int new_nid) >>>>> } >>>>> } >>>>> EXPORT_SYMBOL(page_pool_update_nid); >>>>> + >>>>> +void __page_pool_iov_free(struct page_pool_iov *ppiov) >>>>> +{ >>>>> + if (WARN_ON(ppiov->pp->mp_ops != &dmabuf_devmem_ops)) >>>>> + return; >>>>> + >>>>> + netdev_free_dmabuf(ppiov); >>>>> +} >>>>> +EXPORT_SYMBOL_GPL(__page_pool_iov_free); >>>> >>>> I didn't look too deep but I don't think I immediately follow >>>> the pp refcounting. It increments pages_state_hold_cnt on >>>> allocation, but IIUC doesn't mark skbs for recycle? Then, they all >>>> will be put down via page_pool_iov_put_many() bypassing >>>> page_pool_return_page() and friends. That will call >>>> netdev_free_dmabuf(), which doesn't bump pages_state_release_cnt. >>>> >>>> At least I couldn't make it work with io_uring, and for my purposes, >>>> I forced all puts to go through page_pool_return_page(), which calls >>>> the ->release_page callback. The callback will put the reference and >>>> ask its page pool to account release_cnt. It also gets rid of >>>> __page_pool_iov_free(), as we'd need to add a hook there for >>>> customization otherwise. >>>> >>>> I didn't care about overhead because the hot path for me is getting >>>> buffers from a ring, which is somewhat analogous to sock_devmem_dontneed(), >>>> but done on pp allocations under napi, and it's done separately. >>>> >>>> Completely untested with TCP devmem: >>>> >>>> https://github.com/isilence/linux/commit/14bd56605183dc80b540999e8058c79ac92ae2d8 >>>> >>> >>> This was a mistake in the last RFC, which should be fixed in v1. In >>> the RFC I was not marking the skbs as skb_mark_for_recycle(), so the >>> unreffing path wasn't as expected. >>> >>> In this iteration, that should be completely fixed. I suspect since I >>> just posted this you're actually referring to the issue tested on the >>> last RFC? Correct me if wrong. >> >> Right, it was with RFCv3 >> >>> In this iteration, the reffing story: >>> >>> - memory provider allocs ppiov and returns it to the page pool with >>> ppiov->refcount == 1. >>> - The page_pool gives the page to the driver. The driver may >>> obtain/release references with page_pool_page_[get|put]_many(), but >>> the driver is likely not doing that unless it's doing its own page >>> recycling. >>> - The net stack obtains references via skb_frag_ref() -> >>> page_pool_page_get_many() >>> - The net stack drops references via skb_frag_unref() -> >>> napi_pp_put_page() -> page_pool_return_page() and friends. >>> >>> Thus, the issue where the unref path was skipping >>> page_pool_return_page() and friends should be resolved in this >>> iteration, let me know if you think otherwise, but I think this was an >>> issue limited to the last RFC. >> >> Then page_pool_iov_put_many() should and supposedly would never be >> called by non devmap code because all puts must circle back into >> ->release_page. Why adding it to into page_pool_page_put_many()? >> >> @@ -731,6 +731,29 @@ __page_pool_put_page(struct page_pool *pool, struct page *page, >> + if (page_is_page_pool_iov(page)) { >> ... >> + page_pool_page_put_many(page, 1); >> + return NULL; >> + } >> >> Well, I'm looking at this new branch from Patch 10, it can put >> the buffer, but what if we race at it's actually the final put? >> Looks like nobody is going to to bump up pages_state_release_cnt >> > > Good catch, I think indeed the release_cnt would be incorrect in this > case. I think the race is benign in the sense that the ppiov will be > freed correctly and available for allocation when the page_pool next > needs it; the issue is with the stats AFAICT. hold_cnt + release_cnt serves is used for refcounting. In this case it'll leak the pool when you try to destroy it. >> If you remove the branch, let it fall into ->release and rely >> on refcounting there, then the callback could also fix up >> release_cnt or ask pp to do it, like in the patch I linked above >> > > Sadly I don't think this is possible due to the reasons I mention in > the commit message of that patch. Prematurely releasing ppiov and not > having them be candidates for recycling shows me a 4-5x degradation in > performance. I don't think I follow. The concept is to only recycle a buffer (i.e. make it available for allocation) when its refs drop to zero, which is IMHO the only way it can work, and IIUC what this patchset is doing. That's also I suggest to do, but through a slightly different path. Let's say at some moment there are 2 refs (e.g. 1 for an skb and 1 for userspace/xarray). Say it first puts the skb: napi_pp_put_page() -> page_pool_return_page() -> mp_ops->release_page() -> need_to_free = put_buf() // not last ref, need_to_free==false, // don't recycle, don't increase release_cnt Then you put the last ref: page_pool_iov_put_many() -> page_pool_return_page() -> mp_ops->release_page() -> need_to_free = put_buf() // last ref, need_to_free==true, // recycle and release_cnt++ And that last put can even be recycled right into the pp / ptr_ring, in which case it doesn't need to touch release_cnt. Does it make sense? I don't see where 4-5x degradation would come from > What I could do here is detect that the refcount was dropped to 0 and > fix up the stats in that case. -- Pavel Begunkov