Received: by 2002:ab2:1149:0:b0:1f3:1f8c:d0c6 with SMTP id z9csp2415495lqz; Tue, 2 Apr 2024 17:43:38 -0700 (PDT) X-Forwarded-Encrypted: i=3; AJvYcCVQHq9gEfhwicUTCRK2BcMnL54AczElNH3rX7/uk8xDyRDqlc8DjCvZ0ehgLYHIEf+Ehf91o8cISSX6P93biBlmUU7orENJZ5GVa+o6UQ== X-Google-Smtp-Source: AGHT+IHeobmhySr9Oy73H7uEFmZhncNuirueFI7K5mCi8kFImWXjL8bqFSk+wlKQUdadgqtSbkp0 X-Received: by 2002:a50:955b:0:b0:56d:fdb3:bcc0 with SMTP id v27-20020a50955b000000b0056dfdb3bcc0mr750540eda.11.1712105018830; Tue, 02 Apr 2024 17:43:38 -0700 (PDT) ARC-Seal: i=2; a=rsa-sha256; t=1712105018; cv=pass; d=google.com; s=arc-20160816; b=TFVvcjinag5Qg2FilC2VoVyoi4m3/L+4cvL+tJn7R9ru2+iQUvav2GyWaBbtaC2Z9J o9wJCLiHC1G97sd6W5z0o0+qLN0TaEwye3/gxyebkFlnhVucCsnQ5iODz60YOkNDNQf5 JHY0UpzrurH/yqiPM+2NF/GcEyJ76fF5c9cX59bAZoU/wvarXsj/zkGOc2jfpWxTAZSS yaksUVuFhB7GqiWtaYwHfSXpn/P1VCzuDUxShukk1hh3xYraD8b3VIG9fMSLdxwFWZIi lMpRmm+74uitenVOWhsDX1o4LJ0aDtvDJ+g37cu/PRwL5x6K0LqYZvhIvxl0hiv03yA6 95Jg== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=cc:to:from:subject:message-id:references:mime-version :list-unsubscribe:list-subscribe:list-id:precedence:in-reply-to:date :dkim-signature; bh=x/8vQhiUVWB0TqxWDSZHHcHjMsWUWa56TaLL1CevYW4=; fh=mUIGeoXaXGS8sqCkILHn2grfF5vP9h7NXg8DkfCobX8=; b=ejXWc6r42Z/d2VIgC0F46VyXEghfTGfj0uqC1n42ubUJe5bOMX6R74csq5whw04m8B FN3lGzpkiLUFxoJtufsOOFYDuLdgL/cO3qV5T9NDfzk+2xnLQ7R/m2dVbjNjtNGNjUt5 PRXb2h5mfUEdYeVERnBcIDb8xR4/WHgaG31lUaU3GDSNqnEP3w8coTeAfJDNfw26ghbG aWraWSaRaLrqvYeOBhNQtoItAXiW3MIvZDOVIgRKczecHPGy2FUwA9lsSlAskEAE5hM+ zuxL4K78vUBs792irqu4QOvdRZyvepPNQnGRKpt3gukpFsmeJyi9tNLrVj8ZgRSpvo+2 Jz8Q==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@google.com header.s=20230601 header.b=HSgGeOF4; arc=pass (i=1 spf=pass spfdomain=flex--almasrymina.bounces.google.com dkim=pass dkdomain=google.com dmarc=pass fromdomain=google.com); spf=pass (google.com: domain of linux-kernel+bounces-128858-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.80.249 as permitted sender) smtp.mailfrom="linux-kernel+bounces-128858-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from am.mirrors.kernel.org (am.mirrors.kernel.org. [147.75.80.249]) by mx.google.com with ESMTPS id d28-20020a50cd5c000000b0056df9d6b14asi568611edj.442.2024.04.02.17.43.38 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 02 Apr 2024 17:43:38 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel+bounces-128858-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.80.249 as permitted sender) client-ip=147.75.80.249; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20230601 header.b=HSgGeOF4; arc=pass (i=1 spf=pass spfdomain=flex--almasrymina.bounces.google.com dkim=pass dkdomain=google.com dmarc=pass fromdomain=google.com); spf=pass (google.com: domain of linux-kernel+bounces-128858-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.80.249 as permitted sender) smtp.mailfrom="linux-kernel+bounces-128858-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by am.mirrors.kernel.org (Postfix) with ESMTPS id 266DE1F25877 for ; Wed, 3 Apr 2024 00:22:36 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 7939C18EB8; Wed, 3 Apr 2024 00:21:11 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=google.com header.i=@google.com header.b="HSgGeOF4" Received: from mail-yw1-f202.google.com (mail-yw1-f202.google.com [209.85.128.202]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 5E53D1361 for ; Wed, 3 Apr 2024 00:21:04 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=209.85.128.202 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1712103668; cv=none; b=hoo/paWKTpvIA0IK1IWONzZPwV/svECvk6eMm6kS71s+J1tFNgkaQ6xJQrn0A95NDMWZhpx1EzEXgvhKuEzTPooxQToSp9ZQlon94lHdPCqEWDtl/bCjMe/s4FJUFj6Qru+/QWcqUrH6ZD1jJSgmifalm7kMlPZywoxXeORu8wk= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1712103668; c=relaxed/simple; bh=Fc1L3ZAbVnnY4wUeWUoZvY8ELvZ/3DCLplQBdqNXtz0=; h=Date:In-Reply-To:Mime-Version:References:Message-ID:Subject:From: To:Cc:Content-Type; b=UosuNEI8iUU8gCqWv11XNdgGwiYJJKUEfmy02+VmqOGJ3fJ6Bg92HN+E05oe1UCUh6V1okR/osPrmOvFLJA+dg8Zb38UllS9YmzeP8MQttN5+O5x4xehBvqruw2lBewH8flGHaEFDu9XSd06svGJ2p48G5ycJo8NqTk79HKXqSU= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=reject dis=none) header.from=google.com; spf=pass smtp.mailfrom=flex--almasrymina.bounces.google.com; dkim=pass (2048-bit key) header.d=google.com header.i=@google.com header.b=HSgGeOF4; arc=none smtp.client-ip=209.85.128.202 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=reject dis=none) header.from=google.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=flex--almasrymina.bounces.google.com Received: by mail-yw1-f202.google.com with SMTP id 00721157ae682-6150e36ca0dso28488797b3.1 for ; Tue, 02 Apr 2024 17:21:04 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20230601; t=1712103663; x=1712708463; darn=vger.kernel.org; h=cc:to:from:subject:message-id:references:mime-version:in-reply-to :date:from:to:cc:subject:date:message-id:reply-to; bh=x/8vQhiUVWB0TqxWDSZHHcHjMsWUWa56TaLL1CevYW4=; b=HSgGeOF4J1H0IOiyTOuHpD+EG0vAWd92pRBYN0ty+2Xy4g64UMlg23vRfwGFzT10kn SQPC8uyrIZpK4sUV8fm5SE6TRXYkbTO9c8f4tKd7C+AViLtKZBU1uLqISMebAatXULNn hx+e4PsZFMRwYGGudQ0e/Ehfedg+ahZ5LiQ5KBbBy9AmyVq3Cm6OJsa+8HSH4V7W2/XP 3Hf/Adu1FOav8tPM3iHJz7HzK4NaA3v+DTE+X2oFSf5cDNS22ci0eaGcHSBh0mC+65Iw coUUM1G2c++OdmOrQl4U1xrvmnLYo8AHnBrCEWeK0OWIH0iqPzJ9MHKpxwCHTqogTL4u 4fVA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1712103663; x=1712708463; h=cc:to:from:subject:message-id:references:mime-version:in-reply-to :date:x-gm-message-state:from:to:cc:subject:date:message-id:reply-to; bh=x/8vQhiUVWB0TqxWDSZHHcHjMsWUWa56TaLL1CevYW4=; b=wJbM8N56bBqf0uELvTm71RP7dUO+T7G2c+fSOFOlmvfh8Jy8dI8KCO8k2cV7U45G2J K83xGc4Mac8aND3o72LrZGiIShRQf/nETr2EkTBVWU+bBso8rYVWNFSAAmlWfL8t5ZEt Z1ji69ER2/tiiBe3/Oktf7a/7CL20WmPmu+UyFJSIqeEdaBu69GBRoUUKbRCOvoV4x35 v/0MZrq51wGxKOW5ao1/eLWSmYo3rdYsVYkNRQtwxD9zyqXWUxpY+cqXzbJ27R3AKXXT 9IV2cyk/SoyrFRyR09Q02ypXnmEdhZIeWatJFV8cvPwg7r7HlpWPLvTHWChbe2oIetLf TLng== X-Forwarded-Encrypted: i=1; AJvYcCWm1AQ4HlEubJOvmmdMNCFbCI09YpCfOqrnL3GznuBOLwXSY9HRQeZ/tPtSLYpVzLIybVY8dt3su/9ubfRVEGBHDj84Y2omFory14Kx X-Gm-Message-State: AOJu0YwXlkjgBAmhVZNKjBF5fNV8LpRiyIMYoJIf8nZn3ExxGKGw168E 2RNaiC2CFVdXJl7RN5AnxT6uW1H3MGjlznX3+uhsrzv7bjsWcgEk5r2jyCmRDET/+cxyIKOYnRC umhTrnvwq+2y/iW5zRFfLtA== X-Received: from almasrymina.svl.corp.google.com ([2620:15c:2c4:200:1726:7620:90a1:78b9]) (user=almasrymina job=sendgmr) by 2002:a81:a1d0:0:b0:615:5b6f:b098 with SMTP id y199-20020a81a1d0000000b006155b6fb098mr88336ywg.7.1712103663526; Tue, 02 Apr 2024 17:21:03 -0700 (PDT) Date: Tue, 2 Apr 2024 17:20:39 -0700 In-Reply-To: <20240403002053.2376017-1-almasrymina@google.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: Mime-Version: 1.0 References: <20240403002053.2376017-1-almasrymina@google.com> X-Mailer: git-send-email 2.44.0.478.gd926399ef9-goog Message-ID: <20240403002053.2376017-3-almasrymina@google.com> Subject: [RFC PATCH net-next v8 02/14] net: page_pool: create hooks for custom page providers From: Mina Almasry To: netdev@vger.kernel.org, linux-kernel@vger.kernel.org, linux-doc@vger.kernel.org, linux-alpha@vger.kernel.org, linux-mips@vger.kernel.org, linux-parisc@vger.kernel.org, sparclinux@vger.kernel.org, linux-trace-kernel@vger.kernel.org, linux-arch@vger.kernel.org, bpf@vger.kernel.org, linux-kselftest@vger.kernel.org, linux-media@vger.kernel.org, dri-devel@lists.freedesktop.org Cc: Mina Almasry , "David S. Miller" , Eric Dumazet , Jakub Kicinski , Paolo Abeni , Jonathan Corbet , Richard Henderson , Ivan Kokshaysky , Matt Turner , Thomas Bogendoerfer , "James E.J. Bottomley" , Helge Deller , Andreas Larsson , Jesper Dangaard Brouer , Ilias Apalodimas , Steven Rostedt , Masami Hiramatsu , Mathieu Desnoyers , Arnd Bergmann , Alexei Starovoitov , Daniel Borkmann , Andrii Nakryiko , Martin KaFai Lau , Eduard Zingerman , Song Liu , Yonghong Song , John Fastabend , KP Singh , Stanislav Fomichev , Hao Luo , Jiri Olsa , Steffen Klassert , Herbert Xu , David Ahern , Willem de Bruijn , Shuah Khan , Sumit Semwal , "=?UTF-8?q?Christian=20K=C3=B6nig?=" , Amritha Nambiar , Maciej Fijalkowski , Alexander Mikhalitsyn , Kaiyuan Zhang , Christian Brauner , Simon Horman , David Howells , Florian Westphal , Yunsheng Lin , Kuniyuki Iwashima , Jens Axboe , Arseniy Krasnov , Aleksander Lobakin , Michael Lass , Jiri Pirko , Sebastian Andrzej Siewior , Lorenzo Bianconi , Richard Gobert , Sridhar Samudrala , Xuan Zhuo , Johannes Berg , Abel Wu , Breno Leitao , Pavel Begunkov , David Wei , Jason Gunthorpe , Shailend Chand , Harshitha Ramamurthy , Shakeel Butt , Jeroen de Borst , Praveen Kaligineedi Content-Type: text/plain; charset="UTF-8" From: Jakub Kicinski The page providers which try to reuse the same pages will need to hold onto the ref, even if page gets released from the pool - as in releasing the page from the pp just transfers the "ownership" reference from pp to the provider, and provider will wait for other references to be gone before feeding this page back into the pool. Signed-off-by: Jakub Kicinski Signed-off-by: Mina Almasry --- This is implemented by Jakub in his RFC: https://lore.kernel.org/netdev/f8270765-a27b-6ccf-33ea-cda097168d79@redhat.com/T/ I take no credit for the idea or implementation; I only added minor edits to make this workable with device memory TCP, and removed some hacky test code. This is a critical dependency of device memory TCP and thus I'm pulling it into this series to make it revewable and mergeable. RFC v3 -> v1 - Removed unusued mem_provider. (Yunsheng). - Replaced memory_provider & mp_priv with netdev_rx_queue (Jakub). --- include/net/page_pool/types.h | 12 ++++++++++ net/core/page_pool.c | 43 +++++++++++++++++++++++++++++++---- 2 files changed, 50 insertions(+), 5 deletions(-) diff --git a/include/net/page_pool/types.h b/include/net/page_pool/types.h index 5e43a08d3231..ffe5f31fb0da 100644 --- a/include/net/page_pool/types.h +++ b/include/net/page_pool/types.h @@ -52,6 +52,7 @@ struct pp_alloc_cache { * @dev: device, for DMA pre-mapping purposes * @netdev: netdev this pool will serve (leave as NULL if none or multiple) * @napi: NAPI which is the sole consumer of pages, otherwise NULL + * @queue: struct netdev_rx_queue this page_pool is being created for. * @dma_dir: DMA mapping direction * @max_len: max DMA sync memory size for PP_FLAG_DMA_SYNC_DEV * @offset: DMA sync address offset for PP_FLAG_DMA_SYNC_DEV @@ -64,6 +65,7 @@ struct page_pool_params { int nid; struct device *dev; struct napi_struct *napi; + struct netdev_rx_queue *queue; enum dma_data_direction dma_dir; unsigned int max_len; unsigned int offset; @@ -126,6 +128,13 @@ struct page_pool_stats { }; #endif +struct memory_provider_ops { + int (*init)(struct page_pool *pool); + void (*destroy)(struct page_pool *pool); + struct page *(*alloc_pages)(struct page_pool *pool, gfp_t gfp); + bool (*release_page)(struct page_pool *pool, struct page *page); +}; + struct page_pool { struct page_pool_params_fast p; @@ -176,6 +185,9 @@ struct page_pool { */ struct ptr_ring ring; + void *mp_priv; + const struct memory_provider_ops *mp_ops; + #ifdef CONFIG_PAGE_POOL_STATS /* recycle stats are per-cpu to avoid locking */ struct page_pool_recycle_stats __percpu *recycle_stats; diff --git a/net/core/page_pool.c b/net/core/page_pool.c index dd364d738c00..795b7ff1c01f 100644 --- a/net/core/page_pool.c +++ b/net/core/page_pool.c @@ -25,6 +25,8 @@ #include "page_pool_priv.h" +static DEFINE_STATIC_KEY_FALSE(page_pool_mem_providers); + #define DEFER_TIME (msecs_to_jiffies(1000)) #define DEFER_WARN_INTERVAL (60 * HZ) @@ -177,6 +179,7 @@ static int page_pool_init(struct page_pool *pool, int cpuid) { unsigned int ring_qsize = 1024; /* Default */ + int err; memcpy(&pool->p, ¶ms->fast, sizeof(pool->p)); memcpy(&pool->slow, ¶ms->slow, sizeof(pool->slow)); @@ -248,10 +251,25 @@ static int page_pool_init(struct page_pool *pool, /* Driver calling page_pool_create() also call page_pool_destroy() */ refcount_set(&pool->user_cnt, 1); + if (pool->mp_ops) { + err = pool->mp_ops->init(pool); + if (err) { + pr_warn("%s() mem-provider init failed %d\n", __func__, + err); + goto free_ptr_ring; + } + + static_branch_inc(&page_pool_mem_providers); + } + if (pool->p.flags & PP_FLAG_DMA_MAP) get_device(pool->p.dev); return 0; + +free_ptr_ring: + ptr_ring_cleanup(&pool->ring, NULL); + return err; } static void page_pool_uninit(struct page_pool *pool) @@ -546,7 +564,10 @@ struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp) return page; /* Slow-path: cache empty, do real allocation */ - page = __page_pool_alloc_pages_slow(pool, gfp); + if (static_branch_unlikely(&page_pool_mem_providers) && pool->mp_ops) + page = pool->mp_ops->alloc_pages(pool, gfp); + else + page = __page_pool_alloc_pages_slow(pool, gfp); return page; } EXPORT_SYMBOL(page_pool_alloc_pages); @@ -603,10 +624,13 @@ void __page_pool_release_page_dma(struct page_pool *pool, struct page *page) void page_pool_return_page(struct page_pool *pool, struct page *page) { int count; + bool put; - __page_pool_release_page_dma(pool, page); - - page_pool_clear_pp_info(page); + put = true; + if (static_branch_unlikely(&page_pool_mem_providers) && pool->mp_ops) + put = pool->mp_ops->release_page(pool, page); + else + __page_pool_release_page_dma(pool, page); /* This may be the last page returned, releasing the pool, so * it is not safe to reference pool afterwards. @@ -614,7 +638,10 @@ void page_pool_return_page(struct page_pool *pool, struct page *page) count = atomic_inc_return_relaxed(&pool->pages_state_release_cnt); trace_page_pool_state_release(pool, page, count); - put_page(page); + if (put) { + page_pool_clear_pp_info(page); + put_page(page); + } /* An optimization would be to call __free_pages(page, pool->p.order) * knowing page is not part of page-cache (thus avoiding a * __page_cache_release() call). @@ -889,6 +916,12 @@ static void __page_pool_destroy(struct page_pool *pool) page_pool_unlist(pool); page_pool_uninit(pool); + + if (pool->mp_ops) { + pool->mp_ops->destroy(pool); + static_branch_dec(&page_pool_mem_providers); + } + kfree(pool); } -- 2.44.0.478.gd926399ef9-goog