Received: by 2002:ab2:b82:0:b0:1f3:401:3cfb with SMTP id 2csp364409lqh; Thu, 28 Mar 2024 04:43:56 -0700 (PDT) X-Forwarded-Encrypted: i=3; AJvYcCXjn9QYmr7Sstg9sppOlYodBrdtP8Rz44bAsQ/Xi8mtQ8zMn4+d7pV9KOkKsZXtRMRuvia/0ksiHOQcBQ5H0czEkDT1js4Tv4PPZWCXpQ== X-Google-Smtp-Source: AGHT+IHJS2zeCZUpNA4rWjBFvaL9mtnh+Onpc7dYULUMa4z8ikkWy05KrXhI6xW/QZMp3lTAaKAs X-Received: by 2002:a05:6870:3118:b0:229:9aaa:6f27 with SMTP id v24-20020a056870311800b002299aaa6f27mr2773609oaa.8.1711626236174; Thu, 28 Mar 2024 04:43:56 -0700 (PDT) ARC-Seal: i=2; a=rsa-sha256; t=1711626236; cv=pass; d=google.com; s=arc-20160816; b=MJDffV2qmLpz0W5GYpos79B/DjcYmvStUoRXwdktw18+BmvhV+BfBi6o+qNIcBYrqo BXlVrSz+AYyX4XsCOtXTS0mCcPNfxNKE62oQ2LXaTY1ehB9VNXpLoHI13JLnsoPc//TZ 5rDS6kwB+A5uaXDvlbLNyrYqidZ/P/5FtkWsNjtUMM5QY3JZnPKSpvKk38PIzwvjYAp2 bQGLow+gOStn1yGAO6QsqEKeuiySVetb5oUlO4SJFKazRFcQkFTyj9GFKffJK4do2y9U 0DvYWR+n6QlrQ79m6TDMPxoB7dKTB2xPDfvfyyw+BPo4UDZwn8lCgWkgFtYxdNBGnGB3 /++g== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=in-reply-to:content-disposition:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:references:message-id:subject:cc :to:from:date:dkim-signature; bh=lV5rX0h6ccenEymRkHDK9J1Q58MDNFo0qZ9ZOjy0Sjo=; fh=Uz8sVn6GMck/pWYxDQhi5Yt0hIX50RXx4XcGxduRxi8=; b=NeKAR58jowMaquJHPA9w43aFW4JcD9WWTpqQrtaCj8TP2wZfWS8Lq81nTe/sB2u1V8 7BSZhsgZW4RgADXKEwBbXnUSxdtmSA7ZSmtyz7pQSu0gAz/kWiaZyCKytvPV15ckYH89 ouyEYHS8EuTvTK74p7VSa2t9f1F4w8Swoo3OdQAMZCAC2P97klDHVeL5DrFBtGbjbFUE NYpnVExIKxIed1nW2dNFr+zLY5C2lJg9qziQEf4H+zuduKvpEU/Br5IjmRQUUxKqcpEe /AFdoqIFutfUTw6CqxY93Nhb2rOWr4HJg/hsLWx4VWXBah3dUHsbc3BgvpbcM4f1Rz6u TtXQ==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@marvell.com header.s=pfpt0220 header.b="GVYZk/mI"; arc=pass (i=1 spf=pass spfdomain=marvell.com dkim=pass dkdomain=marvell.com dmarc=pass fromdomain=marvell.com); spf=pass (google.com: domain of linux-kernel+bounces-122843-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45e3:2400::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-122843-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=REJECT dis=NONE) header.from=marvell.com Return-Path: Received: from sv.mirrors.kernel.org (sv.mirrors.kernel.org. [2604:1380:45e3:2400::1]) by mx.google.com with ESMTPS id l128-20020a633e86000000b005dc958db2fcsi1299951pga.363.2024.03.28.04.43.55 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 28 Mar 2024 04:43:56 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel+bounces-122843-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45e3:2400::1 as permitted sender) client-ip=2604:1380:45e3:2400::1; Authentication-Results: mx.google.com; dkim=pass header.i=@marvell.com header.s=pfpt0220 header.b="GVYZk/mI"; arc=pass (i=1 spf=pass spfdomain=marvell.com dkim=pass dkdomain=marvell.com dmarc=pass fromdomain=marvell.com); spf=pass (google.com: domain of linux-kernel+bounces-122843-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45e3:2400::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-122843-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=REJECT dis=NONE) header.from=marvell.com Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by sv.mirrors.kernel.org (Postfix) with ESMTPS id B2D69295079 for ; Thu, 28 Mar 2024 11:43:55 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id A55F87EEF5; Thu, 28 Mar 2024 11:43:38 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=marvell.com header.i=@marvell.com header.b="GVYZk/mI" Received: from mx0b-0016f401.pphosted.com (mx0b-0016f401.pphosted.com [67.231.156.173]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id D68727C08E; Thu, 28 Mar 2024 11:43:33 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=67.231.156.173 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1711626216; cv=none; b=fWCnzHzgN/z3jGy2vT8IdgD93b+8rhoq99s9UyistLkFZUzpambKO7wXnfh9r0yDsl5L2PpOjGs1P9TTpzIeT2Zw/kUHFqeKrK1kuwSEG2JjnTCA7Vv23zzyracz5YKMo1E56BTR5U1jAnypqhIiRWzJDoxh6/SfDyyyhb9V8rU= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1711626216; c=relaxed/simple; bh=4uWG+/4LoiMtnnL9wJuvy86O86r6FRYv43t5tfPYu3E=; h=Date:From:To:CC:Subject:Message-ID:References:MIME-Version: Content-Type:Content-Disposition:In-Reply-To; b=qSY66rPNXXFO1NwfwA4EjORdCqHPvt43RHEAAq17Zlb0n8ktyWmd0kd2DxK4q7RM9zHjnDzAHBqr0O1w/POA+FIBSXQosbrabZQmb61ujb03OQOEofrmemxIiekpD+GfM3R4NZDoH317aUjBcdqJCtyDk0P++o90BklLiOYImI0= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=marvell.com; spf=pass smtp.mailfrom=marvell.com; dkim=pass (2048-bit key) header.d=marvell.com header.i=@marvell.com header.b=GVYZk/mI; arc=none smtp.client-ip=67.231.156.173 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=marvell.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=marvell.com Received: from pps.filterd (m0045851.ppops.net [127.0.0.1]) by mx0b-0016f401.pphosted.com (8.17.1.24/8.17.1.24) with ESMTP id 42SA17C0031387; Thu, 28 Mar 2024 04:42:53 -0700 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=marvell.com; h= date:from:to:cc:subject:message-id:references:mime-version :content-type:in-reply-to; s=pfpt0220; bh=lV5rX0h6ccenEymRkHDK9J 1Q58MDNFo0qZ9ZOjy0Sjo=; b=GVYZk/mIHDRSbKIfGx+PlC03LYWOJOJUKt3Le3 /tNBiJZ8JZQuQHJRQRJQ+XwRRF1Wd2WUQ32CVczc947jwp47RbE1/eb5t/qVcg4r PaR3PXZ40FwWcwKNdO9nxx0Xlg8t+gcsl/05U7J70FVsWNnF8EtEgU6IvH+0tVyt +hjJ2watlCCTSGGr25W0PZrUNyXYt90YRVDvcZ6z/uab+LaAXR77zZyHm+q1rUHV g+cAjPv9iLXqxbfJqDGGDX+C7YjjuOIwpuG5+dXBptlDxlNYbvb9//KGSgXyGnIE YBzHbmv0TwpSu+WzNm8FR1rmkF0HqBql73NCTZhss/ZcW8pQ== Received: from dc6wp-exch02.marvell.com ([4.21.29.225]) by mx0b-0016f401.pphosted.com (PPS) with ESMTPS id 3x4jd1pybt-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Thu, 28 Mar 2024 04:42:53 -0700 (PDT) Received: from DC6WP-EXCH02.marvell.com (10.76.176.209) by DC6WP-EXCH02.marvell.com (10.76.176.209) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.1544.4; Thu, 28 Mar 2024 04:42:52 -0700 Received: from maili.marvell.com (10.69.176.80) by DC6WP-EXCH02.marvell.com (10.76.176.209) with Microsoft SMTP Server id 15.2.1544.4 via Frontend Transport; Thu, 28 Mar 2024 04:42:52 -0700 Received: from maili.marvell.com (unknown [10.28.36.165]) by maili.marvell.com (Postfix) with SMTP id 3AA7C3F708A; Thu, 28 Mar 2024 04:42:45 -0700 (PDT) Date: Thu, 28 Mar 2024 17:12:45 +0530 From: Ratheesh Kannoth To: Julien Panis CC: "David S. Miller" , Eric Dumazet , Jakub Kicinski , Paolo Abeni , Russell King , Alexei Starovoitov , Daniel Borkmann , Jesper Dangaard Brouer , John Fastabend , Sumit Semwal , Christian =?iso-8859-1?Q?K=F6nig?= , Simon Horman , Andrew Lunn , , , , , , Subject: Re: [PATCH net-next v5 3/3] net: ethernet: ti: am65-cpsw: Add minimal XDP support Message-ID: <20240328114245.GA1560669@maili.marvell.com> References: <20240223-am65-cpsw-xdp-basic-v5-0-bc1739170bc6@baylibre.com> <20240223-am65-cpsw-xdp-basic-v5-3-bc1739170bc6@baylibre.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Disposition: inline In-Reply-To: <20240223-am65-cpsw-xdp-basic-v5-3-bc1739170bc6@baylibre.com> X-Proofpoint-GUID: lQpo8tLgN43PCuQH1csWc7MpqVJDCbfE X-Proofpoint-ORIG-GUID: lQpo8tLgN43PCuQH1csWc7MpqVJDCbfE X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.272,Aquarius:18.0.1011,Hydra:6.0.619,FMLib:17.11.176.26 definitions=2024-03-28_11,2024-03-27_01,2023-05-22_02 On 2024-03-28 at 14:56:42, Julien Panis (jpanis@baylibre.com) wrote: > This patch adds XDP (eXpress Data Path) support to TI AM65 CPSW > Ethernet driver. The following features are implemented: > - NETDEV_XDP_ACT_BASIC (XDP_PASS, XDP_TX, XDP_DROP, XDP_ABORTED) > - NETDEV_XDP_ACT_REDIRECT (XDP_REDIRECT) > - NETDEV_XDP_ACT_NDO_XMIT (ndo_xdp_xmit callback) > > The page pool memory model is used to get better performance. > Below are benchmark results obtained for the receiver with iperf3 default > parameters: > - Without page pool: 495 Mbits/sec > - With page pool: 505 Mbits/sec (actually 510 Mbits/sec, with a 5 Mbits/sec > loss due to extra processing in the hot path to handle XDP). > > Signed-off-by: Julien Panis > --- > drivers/net/ethernet/ti/am65-cpsw-nuss.c | 536 ++++++++++++++++++++++++++++--- > drivers/net/ethernet/ti/am65-cpsw-nuss.h | 13 + > 2 files changed, 499 insertions(+), 50 deletions(-) > > diff --git a/drivers/net/ethernet/ti/am65-cpsw-nuss.c b/drivers/net/ethernet/ti/am65-cpsw-nuss.c > index 9d2f4ac783e4..67239c35d346 100644 > --- a/drivers/net/ethernet/ti/am65-cpsw-nuss.c > +++ b/drivers/net/ethernet/ti/am65-cpsw-nuss.c > @@ -5,6 +5,7 @@ > * > */ > > +#include > #include > #include > #include > @@ -30,6 +31,7 @@ > #include > #include > #include > +#include > #include > > #include "cpsw_ale.h" > @@ -138,6 +140,17 @@ > > #define AM65_CPSW_DEFAULT_TX_CHNS 8 > > +/* CPPI streaming packet interface */ > +#define AM65_CPSW_CPPI_TX_FLOW_ID 0x3FFF > +#define AM65_CPSW_CPPI_TX_PKT_TYPE 0x7 > + > +/* XDP */ > +#define AM65_CPSW_XDP_CONSUMED 1 > +#define AM65_CPSW_XDP_PASS 0 > + > +/* Include headroom compatible with both skb and xdpf */ > +#define AM65_CPSW_HEADROOM max(NET_SKB_PAD, XDP_PACKET_HEADROOM) > + > static void am65_cpsw_port_set_sl_mac(struct am65_cpsw_port *slave, > const u8 *dev_addr) > { > @@ -369,6 +382,66 @@ static void am65_cpsw_init_host_port_emac(struct am65_cpsw_common *common); > static void am65_cpsw_init_port_switch_ale(struct am65_cpsw_port *port); > static void am65_cpsw_init_port_emac_ale(struct am65_cpsw_port *port); > > +static void am65_cpsw_destroy_xdp_rxqs(struct am65_cpsw_common *common) > +{ > + struct am65_cpsw_rx_chn *rx_chn = &common->rx_chns; > + struct xdp_rxq_info *rxq; > + int i; > + > + for (i = 0; i < common->port_num; i++) { > + rxq = &common->ports[i].xdp_rxq; > + > + if (xdp_rxq_info_is_reg(rxq)) > + xdp_rxq_info_unreg(rxq); > + } > + > + if (rx_chn->page_pool) { > + page_pool_destroy(rx_chn->page_pool); > + rx_chn->page_pool = NULL; > + } > +} > + > +static int am65_cpsw_create_xdp_rxqs(struct am65_cpsw_common *common) > +{ > + struct am65_cpsw_rx_chn *rx_chn = &common->rx_chns; > + struct page_pool_params pp_params = { > + .flags = PP_FLAG_DMA_MAP, > + .order = 0, > + .pool_size = AM65_CPSW_MAX_RX_DESC, > + .nid = dev_to_node(common->dev), > + .dev = common->dev, > + .dma_dir = DMA_BIDIRECTIONAL, > + .napi = &common->napi_rx, > + }; > + struct xdp_rxq_info *rxq; > + struct page_pool *pool; > + int i, ret; > + > + pool = page_pool_create(&pp_params); > + if (IS_ERR(pool)) > + return PTR_ERR(pool); > + > + rx_chn->page_pool = pool; > + > + for (i = 0; i < common->port_num; i++) { > + rxq = &common->ports[i].xdp_rxq; > + > + ret = xdp_rxq_info_reg(rxq, common->ports[i].ndev, i, 0); > + if (ret) > + goto err; > + > + ret = xdp_rxq_info_reg_mem_model(rxq, MEM_TYPE_PAGE_POOL, pool); > + if (ret) > + goto err; > + } > + > + return 0; > + > +err: > + am65_cpsw_destroy_xdp_rxqs(common); > + return ret; > +} > + > static void am65_cpsw_nuss_rx_cleanup(void *data, dma_addr_t desc_dma) > { > struct am65_cpsw_rx_chn *rx_chn = data; > @@ -440,9 +513,40 @@ static void am65_cpsw_nuss_tx_cleanup(void *data, dma_addr_t desc_dma) > dev_kfree_skb_any(skb); > } > > +static struct sk_buff *am65_cpsw_alloc_skb(struct am65_cpsw_rx_chn *rx_chn, > + struct net_device *ndev, > + unsigned int len, > + int desc_idx) > +{ > + struct sk_buff *skb; > + struct page *page; > + > + page = page_pool_dev_alloc_pages(rx_chn->page_pool); > + if (unlikely(!page)) > + return NULL; > + > + len += AM65_CPSW_HEADROOM; > + > + skb = build_skb(page_address(page), len); > + if (unlikely(!skb)) { > + page_pool_put_full_page(rx_chn->page_pool, page, ndev); Is it compiling ? third argument should be a bool. > + rx_chn->pages[desc_idx] = NULL; > + return NULL; > + } > + > + skb_reserve(skb, AM65_CPSW_HEADROOM + NET_IP_ALIGN); > + skb->dev = ndev; > + > + rx_chn->pages[desc_idx] = page; > + > + return skb; > +} > + > static int am65_cpsw_nuss_common_open(struct am65_cpsw_common *common) > { > struct am65_cpsw_host *host_p = am65_common_get_host(common); > + struct am65_cpsw_rx_chn *rx_chn = &common->rx_chns; > + struct am65_cpsw_tx_chn *tx_chn = common->tx_chns; > int port_idx, i, ret, tx; > struct sk_buff *skb; > u32 val, port_mask; > @@ -505,10 +609,14 @@ static int am65_cpsw_nuss_common_open(struct am65_cpsw_common *common) > > am65_cpsw_qos_tx_p0_rate_init(common); > > - for (i = 0; i < common->rx_chns.descs_num; i++) { > - skb = __netdev_alloc_skb_ip_align(NULL, > - AM65_CPSW_MAX_PACKET_SIZE, > - GFP_KERNEL); > + ret = am65_cpsw_create_xdp_rxqs(common); > + if (ret) { > + dev_err(common->dev, "Failed to create XDP rx queues\n"); > + return ret; > + } > + > + for (i = 0; i < rx_chn->descs_num; i++) { > + skb = am65_cpsw_alloc_skb(rx_chn, NULL, AM65_CPSW_MAX_PACKET_SIZE, i); > if (!skb) { > ret = -ENOMEM; > dev_err(common->dev, "cannot allocate skb\n"); > @@ -531,27 +639,27 @@ static int am65_cpsw_nuss_common_open(struct am65_cpsw_common *common) > } > } > > - ret = k3_udma_glue_enable_rx_chn(common->rx_chns.rx_chn); > + ret = k3_udma_glue_enable_rx_chn(rx_chn->rx_chn); > if (ret) { > dev_err(common->dev, "couldn't enable rx chn: %d\n", ret); > goto fail_rx; > } > > for (tx = 0; tx < common->tx_ch_num; tx++) { > - ret = k3_udma_glue_enable_tx_chn(common->tx_chns[tx].tx_chn); > + ret = k3_udma_glue_enable_tx_chn(tx_chn[tx].tx_chn); > if (ret) { > dev_err(common->dev, "couldn't enable tx chn %d: %d\n", > tx, ret); > tx--; > goto fail_tx; > } > - napi_enable(&common->tx_chns[tx].napi_tx); > + napi_enable(&tx_chn[tx].napi_tx); > } > > napi_enable(&common->napi_rx); > if (common->rx_irq_disabled) { > common->rx_irq_disabled = false; > - enable_irq(common->rx_chns.irq); > + enable_irq(rx_chn->irq); > } > > dev_dbg(common->dev, "cpsw_nuss started\n"); > @@ -559,22 +667,23 @@ static int am65_cpsw_nuss_common_open(struct am65_cpsw_common *common) > > fail_tx: > while (tx >= 0) { > - napi_disable(&common->tx_chns[tx].napi_tx); > - k3_udma_glue_disable_tx_chn(common->tx_chns[tx].tx_chn); > + napi_disable(&tx_chn[tx].napi_tx); > + k3_udma_glue_disable_tx_chn(tx_chn[tx].tx_chn); > tx--; > } > > - k3_udma_glue_disable_rx_chn(common->rx_chns.rx_chn); > + k3_udma_glue_disable_rx_chn(rx_chn->rx_chn); > > fail_rx: > - k3_udma_glue_reset_rx_chn(common->rx_chns.rx_chn, 0, > - &common->rx_chns, > + k3_udma_glue_reset_rx_chn(rx_chn->rx_chn, 0, rx_chn, > am65_cpsw_nuss_rx_cleanup, 0); > return ret; > } > > static int am65_cpsw_nuss_common_stop(struct am65_cpsw_common *common) > { > + struct am65_cpsw_rx_chn *rx_chn = &common->rx_chns; > + struct am65_cpsw_tx_chn *tx_chn = common->tx_chns; > int i; > > if (common->usage_count != 1) > @@ -590,26 +699,25 @@ static int am65_cpsw_nuss_common_stop(struct am65_cpsw_common *common) > reinit_completion(&common->tdown_complete); > > for (i = 0; i < common->tx_ch_num; i++) > - k3_udma_glue_tdown_tx_chn(common->tx_chns[i].tx_chn, false); > + k3_udma_glue_tdown_tx_chn(tx_chn[i].tx_chn, false); > > i = wait_for_completion_timeout(&common->tdown_complete, > msecs_to_jiffies(1000)); > if (!i) > dev_err(common->dev, "tx timeout\n"); > for (i = 0; i < common->tx_ch_num; i++) { > - napi_disable(&common->tx_chns[i].napi_tx); > - hrtimer_cancel(&common->tx_chns[i].tx_hrtimer); > + napi_disable(&tx_chn[i].napi_tx); > + hrtimer_cancel(&tx_chn[i].tx_hrtimer); > } > > for (i = 0; i < common->tx_ch_num; i++) { > - k3_udma_glue_reset_tx_chn(common->tx_chns[i].tx_chn, > - &common->tx_chns[i], > + k3_udma_glue_reset_tx_chn(tx_chn[i].tx_chn, &tx_chn[i], > am65_cpsw_nuss_tx_cleanup); > - k3_udma_glue_disable_tx_chn(common->tx_chns[i].tx_chn); > + k3_udma_glue_disable_tx_chn(tx_chn[i].tx_chn); > } > > reinit_completion(&common->tdown_complete); > - k3_udma_glue_tdown_rx_chn(common->rx_chns.rx_chn, true); > + k3_udma_glue_tdown_rx_chn(rx_chn->rx_chn, true); > > if (common->pdata.quirks & AM64_CPSW_QUIRK_DMA_RX_TDOWN_IRQ) { > i = wait_for_completion_timeout(&common->tdown_complete, msecs_to_jiffies(1000)); > @@ -621,17 +729,24 @@ static int am65_cpsw_nuss_common_stop(struct am65_cpsw_common *common) > hrtimer_cancel(&common->rx_hrtimer); > > for (i = 0; i < AM65_CPSW_MAX_RX_FLOWS; i++) > - k3_udma_glue_reset_rx_chn(common->rx_chns.rx_chn, i, > - &common->rx_chns, > + k3_udma_glue_reset_rx_chn(rx_chn->rx_chn, i, rx_chn, > am65_cpsw_nuss_rx_cleanup, !!i); > > - k3_udma_glue_disable_rx_chn(common->rx_chns.rx_chn); > + k3_udma_glue_disable_rx_chn(rx_chn->rx_chn); > > cpsw_ale_stop(common->ale); > > writel(0, common->cpsw_base + AM65_CPSW_REG_CTL); > writel(0, common->cpsw_base + AM65_CPSW_REG_STAT_PORT_EN); > > + for (i = 0; i < rx_chn->descs_num; i++) { > + if (rx_chn->pages[i]) { > + page_pool_put_full_page(rx_chn->page_pool, rx_chn->pages[i], false); > + rx_chn->pages[i] = NULL; > + } > + } > + am65_cpsw_destroy_xdp_rxqs(common); > + > dev_dbg(common->dev, "cpsw_nuss stopped\n"); > return 0; > } > @@ -749,6 +864,176 @@ static int am65_cpsw_nuss_ndo_slave_open(struct net_device *ndev) > return ret; > } > > +static int am65_cpsw_nuss_desc_idx(struct k3_cppi_desc_pool *desc_pool, void *desc, > + unsigned char dsize_log2) > +{ > + void *pool_addr = k3_cppi_desc_pool_cpuaddr(desc_pool); > + > + return (desc - pool_addr) >> dsize_log2; > +} > + > +static void am65_cpsw_nuss_set_buf_type(struct am65_cpsw_tx_chn *tx_chn, > + struct cppi5_host_desc_t *desc, > + enum am65_cpsw_tx_buf_type buf_type) > +{ > + int desc_idx; > + > + desc_idx = am65_cpsw_nuss_desc_idx(tx_chn->desc_pool, desc, tx_chn->dsize_log2); > + k3_cppi_desc_pool_desc_info_set(tx_chn->desc_pool, desc_idx, (void *)buf_type); > +} > + > +static enum am65_cpsw_tx_buf_type am65_cpsw_nuss_buf_type(struct am65_cpsw_tx_chn *tx_chn, > + dma_addr_t desc_dma) > +{ > + struct cppi5_host_desc_t *desc_tx; > + int desc_idx; > + > + desc_tx = k3_cppi_desc_pool_dma2virt(tx_chn->desc_pool, desc_dma); > + desc_idx = am65_cpsw_nuss_desc_idx(tx_chn->desc_pool, desc_tx, tx_chn->dsize_log2); > + > + return (enum am65_cpsw_tx_buf_type)k3_cppi_desc_pool_desc_info(tx_chn->desc_pool, > + desc_idx); > +} > + > +static int am65_cpsw_xdp_tx_frame(struct net_device *ndev, > + struct am65_cpsw_tx_chn *tx_chn, > + struct xdp_frame *xdpf, > + enum am65_cpsw_tx_buf_type buf_type) > +{ > + struct am65_cpsw_common *common = am65_ndev_to_common(ndev); > + struct am65_cpsw_port *port = am65_ndev_to_port(ndev); > + struct cppi5_host_desc_t *host_desc; > + struct netdev_queue *netif_txq; > + dma_addr_t dma_desc, dma_buf; > + u32 pkt_len = xdpf->len; > + void **swdata; > + int ret; > + > + host_desc = k3_cppi_desc_pool_alloc(tx_chn->desc_pool); > + if (unlikely(!host_desc)) { > + ndev->stats.tx_dropped++; > + return -ENOMEM; > + } > + > + am65_cpsw_nuss_set_buf_type(tx_chn, host_desc, buf_type); > + > + dma_buf = dma_map_single(tx_chn->dma_dev, xdpf->data, pkt_len, DMA_TO_DEVICE); > + if (unlikely(dma_mapping_error(tx_chn->dma_dev, dma_buf))) { > + ndev->stats.tx_dropped++; > + ret = -ENOMEM; > + goto pool_free; > + } > + > + cppi5_hdesc_init(host_desc, CPPI5_INFO0_HDESC_EPIB_PRESENT, AM65_CPSW_NAV_PS_DATA_SIZE); > + cppi5_hdesc_set_pkttype(host_desc, AM65_CPSW_CPPI_TX_PKT_TYPE); > + cppi5_hdesc_set_pktlen(host_desc, pkt_len); > + cppi5_desc_set_pktids(&host_desc->hdr, 0, AM65_CPSW_CPPI_TX_FLOW_ID); > + cppi5_desc_set_tags_ids(&host_desc->hdr, 0, port->port_id); > + > + k3_udma_glue_tx_dma_to_cppi5_addr(tx_chn->tx_chn, &dma_buf); > + cppi5_hdesc_attach_buf(host_desc, dma_buf, pkt_len, dma_buf, pkt_len); > + > + swdata = cppi5_hdesc_get_swdata(host_desc); > + *(swdata) = xdpf; > + > + /* Report BQL before sending the packet */ > + netif_txq = netdev_get_tx_queue(ndev, tx_chn->id); > + netdev_tx_sent_queue(netif_txq, pkt_len); > + > + dma_desc = k3_cppi_desc_pool_virt2dma(tx_chn->desc_pool, host_desc); > + if (AM65_CPSW_IS_CPSW2G(common)) { > + ret = k3_udma_glue_push_tx_chn(tx_chn->tx_chn, host_desc, dma_desc); > + } else { > + spin_lock_bh(&tx_chn->lock); > + ret = k3_udma_glue_push_tx_chn(tx_chn->tx_chn, host_desc, dma_desc); > + spin_unlock_bh(&tx_chn->lock); > + } > + if (ret) { > + /* Inform BQL */ > + netdev_tx_completed_queue(netif_txq, 1, pkt_len); > + ndev->stats.tx_errors++; > + goto dma_unmap; > + } > + > + return 0; > + > +dma_unmap: > + k3_udma_glue_tx_cppi5_to_dma_addr(tx_chn->tx_chn, &dma_buf); > + dma_unmap_single(tx_chn->dma_dev, dma_buf, pkt_len, DMA_TO_DEVICE); > +pool_free: > + k3_cppi_desc_pool_free(tx_chn->desc_pool, host_desc); > + return ret; > +} > + > +static int am65_cpsw_run_xdp(struct am65_cpsw_common *common, struct am65_cpsw_port *port, > + struct xdp_buff *xdp, int desc_idx, int cpu, int *len) > +{ > + struct am65_cpsw_rx_chn *rx_chn = &common->rx_chns; > + struct net_device *ndev = port->ndev; > + int ret = AM65_CPSW_XDP_CONSUMED; > + struct am65_cpsw_tx_chn *tx_chn; > + struct netdev_queue *netif_txq; > + struct xdp_frame *xdpf; > + struct bpf_prog *prog; > + struct page *page; > + u32 act; > + > + prog = READ_ONCE(port->xdp_prog); > + if (!prog) > + return AM65_CPSW_XDP_PASS; > + > + act = bpf_prog_run_xdp(prog, xdp); > + /* XDP prog might have changed packet data and boundaries */ > + *len = xdp->data_end - xdp->data; > + > + switch (act) { > + case XDP_PASS: > + ret = AM65_CPSW_XDP_PASS; > + goto out; > + case XDP_TX: > + tx_chn = &common->tx_chns[cpu % AM65_CPSW_MAX_TX_QUEUES]; > + netif_txq = netdev_get_tx_queue(ndev, tx_chn->id); > + > + xdpf = xdp_convert_buff_to_frame(xdp); > + if (unlikely(!xdpf)) > + break; > + > + __netif_tx_lock(netif_txq, cpu); > + ret = am65_cpsw_xdp_tx_frame(ndev, tx_chn, xdpf, > + AM65_CPSW_TX_BUF_TYPE_XDP_TX); > + __netif_tx_unlock(netif_txq); > + if (ret) > + break; > + > + ndev->stats.rx_bytes += *len; > + ndev->stats.rx_packets++; > + ret = AM65_CPSW_XDP_CONSUMED; > + goto out; > + case XDP_REDIRECT: > + if (unlikely(xdp_do_redirect(ndev, xdp, prog))) > + break; > + > + xdp_do_flush(); > + ndev->stats.rx_bytes += *len; > + ndev->stats.rx_packets++; > + goto out; > + default: > + bpf_warn_invalid_xdp_action(ndev, prog, act); > + fallthrough; > + case XDP_ABORTED: > + trace_xdp_exception(ndev, prog, act); > + fallthrough; > + case XDP_DROP: > + ndev->stats.rx_dropped++; > + } > + > + page = virt_to_head_page(xdp->data); > + page_pool_recycle_direct(rx_chn->page_pool, page); > + rx_chn->pages[desc_idx] = NULL; > +out: > + return ret; > +} > + > static void am65_cpsw_nuss_rx_ts(struct sk_buff *skb, u32 *psdata) > { > struct skb_shared_hwtstamps *ssh; > @@ -795,7 +1080,7 @@ static void am65_cpsw_nuss_rx_csum(struct sk_buff *skb, u32 csum_info) > } > > static int am65_cpsw_nuss_rx_packets(struct am65_cpsw_common *common, > - u32 flow_idx) > + u32 flow_idx, int cpu) > { > struct am65_cpsw_rx_chn *rx_chn = &common->rx_chns; > u32 buf_dma_len, pkt_len, port_id = 0, csum_info; > @@ -806,10 +1091,12 @@ static int am65_cpsw_nuss_rx_packets(struct am65_cpsw_common *common, > struct sk_buff *skb, *new_skb; > dma_addr_t desc_dma, buf_dma; > struct am65_cpsw_port *port; > + int headroom, desc_idx, ret; > struct net_device *ndev; > + struct xdp_buff xdp; > + struct page *page; > void **swdata; > u32 *psdata; > - int ret = 0; > > ret = k3_udma_glue_pop_rx_chn(rx_chn->rx_chn, flow_idx, &desc_dma); > if (ret) { > @@ -851,11 +1138,30 @@ static int am65_cpsw_nuss_rx_packets(struct am65_cpsw_common *common, > > k3_cppi_desc_pool_free(rx_chn->desc_pool, desc_rx); > > - new_skb = netdev_alloc_skb_ip_align(ndev, AM65_CPSW_MAX_PACKET_SIZE); > + desc_idx = am65_cpsw_nuss_desc_idx(rx_chn->desc_pool, desc_rx, rx_chn->dsize_log2); > + > + if (port->xdp_prog) { > + xdp_init_buff(&xdp, AM65_CPSW_MAX_PACKET_SIZE, &port->xdp_rxq); > + > + page = virt_to_page(skb->data); > + xdp_prepare_buff(&xdp, page_address(page), skb_headroom(skb), pkt_len, false); > + > + ret = am65_cpsw_run_xdp(common, port, &xdp, desc_idx, cpu, &pkt_len); > + if (ret != AM65_CPSW_XDP_PASS) > + return ret; > + > + /* Compute additional headroom to be reserved */ > + headroom = (xdp.data - xdp.data_hard_start) - skb_headroom(skb); > + skb_reserve(skb, headroom); > + } > + > + /* Pass skb to netstack if no XDP prog or returned XDP_PASS */ > + new_skb = am65_cpsw_alloc_skb(rx_chn, ndev, AM65_CPSW_MAX_PACKET_SIZE, desc_idx); > if (new_skb) { > ndev_priv = netdev_priv(ndev); > am65_cpsw_nuss_set_offload_fwd_mark(skb, ndev_priv->offload_fwd_mark); > skb_put(skb, pkt_len); > + skb_mark_for_recycle(skb); > skb->protocol = eth_type_trans(skb, ndev); > am65_cpsw_nuss_rx_csum(skb, csum_info); > napi_gro_receive(&common->napi_rx, skb); > @@ -901,6 +1207,7 @@ static int am65_cpsw_nuss_rx_poll(struct napi_struct *napi_rx, int budget) > { > struct am65_cpsw_common *common = am65_cpsw_napi_to_common(napi_rx); > int flow = AM65_CPSW_MAX_RX_FLOWS; > + int cpu = smp_processor_id(); > int cur_budget, ret; > int num_rx = 0; > > @@ -909,7 +1216,7 @@ static int am65_cpsw_nuss_rx_poll(struct napi_struct *napi_rx, int budget) > cur_budget = budget - num_rx; > > while (cur_budget--) { > - ret = am65_cpsw_nuss_rx_packets(common, flow); > + ret = am65_cpsw_nuss_rx_packets(common, flow, cpu); > if (ret) > break; > num_rx++; > @@ -938,8 +1245,8 @@ static int am65_cpsw_nuss_rx_poll(struct napi_struct *napi_rx, int budget) > } > > static struct sk_buff * > -am65_cpsw_nuss_tx_compl_packet(struct am65_cpsw_tx_chn *tx_chn, > - dma_addr_t desc_dma) > +am65_cpsw_nuss_tx_compl_packet_skb(struct am65_cpsw_tx_chn *tx_chn, > + dma_addr_t desc_dma) > { > struct am65_cpsw_ndev_priv *ndev_priv; > struct am65_cpsw_ndev_stats *stats; > @@ -968,6 +1275,39 @@ am65_cpsw_nuss_tx_compl_packet(struct am65_cpsw_tx_chn *tx_chn, > return skb; > } > > +static struct xdp_frame * > +am65_cpsw_nuss_tx_compl_packet_xdp(struct am65_cpsw_common *common, > + struct am65_cpsw_tx_chn *tx_chn, > + dma_addr_t desc_dma, > + struct net_device **ndev) > +{ > + struct am65_cpsw_ndev_priv *ndev_priv; > + struct am65_cpsw_ndev_stats *stats; > + struct cppi5_host_desc_t *desc_tx; > + struct am65_cpsw_port *port; > + struct xdp_frame *xdpf; > + u32 port_id = 0; > + void **swdata; > + > + desc_tx = k3_cppi_desc_pool_dma2virt(tx_chn->desc_pool, desc_dma); > + cppi5_desc_get_tags_ids(&desc_tx->hdr, NULL, &port_id); > + swdata = cppi5_hdesc_get_swdata(desc_tx); > + xdpf = *(swdata); > + am65_cpsw_nuss_xmit_free(tx_chn, desc_tx); > + > + port = am65_common_get_port(common, port_id); > + *ndev = port->ndev; > + > + ndev_priv = netdev_priv(*ndev); > + stats = this_cpu_ptr(ndev_priv->stats); > + u64_stats_update_begin(&stats->syncp); > + stats->tx_packets++; > + stats->tx_bytes += xdpf->len; > + u64_stats_update_end(&stats->syncp); > + > + return xdpf; > +} > + > static void am65_cpsw_nuss_tx_wake(struct am65_cpsw_tx_chn *tx_chn, struct net_device *ndev, > struct netdev_queue *netif_txq) > { > @@ -988,11 +1328,13 @@ static void am65_cpsw_nuss_tx_wake(struct am65_cpsw_tx_chn *tx_chn, struct net_d > static int am65_cpsw_nuss_tx_compl_packets(struct am65_cpsw_common *common, > int chn, unsigned int budget, bool *tdown) > { > + enum am65_cpsw_tx_buf_type buf_type; > struct device *dev = common->dev; > struct am65_cpsw_tx_chn *tx_chn; > struct netdev_queue *netif_txq; > unsigned int total_bytes = 0; > struct net_device *ndev; > + struct xdp_frame *xdpf; > struct sk_buff *skb; > dma_addr_t desc_dma; > int res, num_tx = 0; > @@ -1013,10 +1355,20 @@ static int am65_cpsw_nuss_tx_compl_packets(struct am65_cpsw_common *common, > break; > } > > - skb = am65_cpsw_nuss_tx_compl_packet(tx_chn, desc_dma); > - total_bytes = skb->len; > - ndev = skb->dev; > - napi_consume_skb(skb, budget); > + buf_type = am65_cpsw_nuss_buf_type(tx_chn, desc_dma); > + if (buf_type == AM65_CPSW_TX_BUF_TYPE_SKB) { > + skb = am65_cpsw_nuss_tx_compl_packet_skb(tx_chn, desc_dma); > + ndev = skb->dev; > + total_bytes = skb->len; > + napi_consume_skb(skb, budget); > + } else { > + xdpf = am65_cpsw_nuss_tx_compl_packet_xdp(common, tx_chn, desc_dma, &ndev); > + total_bytes = xdpf->len; > + if (buf_type == AM65_CPSW_TX_BUF_TYPE_XDP_TX) > + xdp_return_frame_rx_napi(xdpf); > + else > + xdp_return_frame(xdpf); > + } > num_tx++; > > netif_txq = netdev_get_tx_queue(ndev, chn); > @@ -1034,11 +1386,13 @@ static int am65_cpsw_nuss_tx_compl_packets(struct am65_cpsw_common *common, > static int am65_cpsw_nuss_tx_compl_packets_2g(struct am65_cpsw_common *common, > int chn, unsigned int budget, bool *tdown) > { > + enum am65_cpsw_tx_buf_type buf_type; > struct device *dev = common->dev; > struct am65_cpsw_tx_chn *tx_chn; > struct netdev_queue *netif_txq; > unsigned int total_bytes = 0; > struct net_device *ndev; > + struct xdp_frame *xdpf; > struct sk_buff *skb; > dma_addr_t desc_dma; > int res, num_tx = 0; > @@ -1057,11 +1411,20 @@ static int am65_cpsw_nuss_tx_compl_packets_2g(struct am65_cpsw_common *common, > break; > } > > - skb = am65_cpsw_nuss_tx_compl_packet(tx_chn, desc_dma); > - > - ndev = skb->dev; > - total_bytes += skb->len; > - napi_consume_skb(skb, budget); > + buf_type = am65_cpsw_nuss_buf_type(tx_chn, desc_dma); > + if (buf_type == AM65_CPSW_TX_BUF_TYPE_SKB) { > + skb = am65_cpsw_nuss_tx_compl_packet_skb(tx_chn, desc_dma); > + ndev = skb->dev; > + total_bytes += skb->len; > + napi_consume_skb(skb, budget); > + } else { > + xdpf = am65_cpsw_nuss_tx_compl_packet_xdp(common, tx_chn, desc_dma, &ndev); > + total_bytes += xdpf->len; > + if (buf_type == AM65_CPSW_TX_BUF_TYPE_XDP_TX) > + xdp_return_frame_rx_napi(xdpf); > + else > + xdp_return_frame(xdpf); > + } > num_tx++; > } > > @@ -1183,10 +1546,12 @@ static netdev_tx_t am65_cpsw_nuss_ndo_slave_xmit(struct sk_buff *skb, > goto busy_stop_q; > } > > + am65_cpsw_nuss_set_buf_type(tx_chn, first_desc, AM65_CPSW_TX_BUF_TYPE_SKB); > + > cppi5_hdesc_init(first_desc, CPPI5_INFO0_HDESC_EPIB_PRESENT, > AM65_CPSW_NAV_PS_DATA_SIZE); > - cppi5_desc_set_pktids(&first_desc->hdr, 0, 0x3FFF); > - cppi5_hdesc_set_pkttype(first_desc, 0x7); > + cppi5_desc_set_pktids(&first_desc->hdr, 0, AM65_CPSW_CPPI_TX_FLOW_ID); > + cppi5_hdesc_set_pkttype(first_desc, AM65_CPSW_CPPI_TX_PKT_TYPE); > cppi5_desc_set_tags_ids(&first_desc->hdr, 0, port->port_id); > > k3_udma_glue_tx_dma_to_cppi5_addr(tx_chn->tx_chn, &buf_dma); > @@ -1225,6 +1590,8 @@ static netdev_tx_t am65_cpsw_nuss_ndo_slave_xmit(struct sk_buff *skb, > goto busy_free_descs; > } > > + am65_cpsw_nuss_set_buf_type(tx_chn, next_desc, AM65_CPSW_TX_BUF_TYPE_SKB); > + > buf_dma = skb_frag_dma_map(tx_chn->dma_dev, frag, 0, frag_size, > DMA_TO_DEVICE); > if (unlikely(dma_mapping_error(tx_chn->dma_dev, buf_dma))) { > @@ -1488,6 +1855,58 @@ static void am65_cpsw_nuss_ndo_get_stats(struct net_device *dev, > stats->tx_dropped = dev->stats.tx_dropped; > } > > +static int am65_cpsw_xdp_prog_setup(struct net_device *ndev, struct bpf_prog *prog) > +{ > + struct am65_cpsw_port *port = am65_ndev_to_port(ndev); > + bool running = netif_running(ndev); > + struct bpf_prog *old_prog; > + > + if (running) > + am65_cpsw_nuss_ndo_slave_stop(ndev); > + > + old_prog = xchg(&port->xdp_prog, prog); > + if (old_prog) > + bpf_prog_put(old_prog); > + > + if (running) > + return am65_cpsw_nuss_ndo_slave_open(ndev); > + > + return 0; > +} > + > +static int am65_cpsw_ndo_bpf(struct net_device *ndev, struct netdev_bpf *bpf) > +{ > + switch (bpf->command) { > + case XDP_SETUP_PROG: > + return am65_cpsw_xdp_prog_setup(ndev, bpf->prog); > + default: > + return -EINVAL; > + } > +} > + > +static int am65_cpsw_ndo_xdp_xmit(struct net_device *ndev, int n, > + struct xdp_frame **frames, u32 flags) > +{ > + struct am65_cpsw_tx_chn *tx_chn; > + struct netdev_queue *netif_txq; > + int cpu = smp_processor_id(); > + int i, nxmit = 0; > + > + tx_chn = &am65_ndev_to_common(ndev)->tx_chns[cpu % AM65_CPSW_MAX_TX_QUEUES]; > + netif_txq = netdev_get_tx_queue(ndev, tx_chn->id); > + > + __netif_tx_lock(netif_txq, cpu); > + for (i = 0; i < n; i++) { > + if (am65_cpsw_xdp_tx_frame(ndev, tx_chn, frames[i], > + AM65_CPSW_TX_BUF_TYPE_XDP_NDO)) > + break; > + nxmit++; > + } > + __netif_tx_unlock(netif_txq); > + > + return nxmit; > +} > + > static const struct net_device_ops am65_cpsw_nuss_netdev_ops = { > .ndo_open = am65_cpsw_nuss_ndo_slave_open, > .ndo_stop = am65_cpsw_nuss_ndo_slave_stop, > @@ -1502,6 +1921,8 @@ static const struct net_device_ops am65_cpsw_nuss_netdev_ops = { > .ndo_eth_ioctl = am65_cpsw_nuss_ndo_slave_ioctl, > .ndo_setup_tc = am65_cpsw_qos_ndo_setup_tc, > .ndo_set_tx_maxrate = am65_cpsw_qos_ndo_tx_p0_set_maxrate, > + .ndo_bpf = am65_cpsw_ndo_bpf, > + .ndo_xdp_xmit = am65_cpsw_ndo_xdp_xmit, > }; > > static void am65_cpsw_disable_phy(struct phy *phy) > @@ -1772,7 +2193,7 @@ static int am65_cpsw_nuss_init_tx_chns(struct am65_cpsw_common *common) > .mode = K3_RINGACC_RING_MODE_RING, > .flags = 0 > }; > - u32 hdesc_size; > + u32 hdesc_size, hdesc_size_out; > int i, ret = 0; > > hdesc_size = cppi5_hdesc_calc_size(true, AM65_CPSW_NAV_PS_DATA_SIZE, > @@ -1816,6 +2237,10 @@ static int am65_cpsw_nuss_init_tx_chns(struct am65_cpsw_common *common) > goto err; > } > > + hdesc_size_out = k3_cppi_desc_pool_desc_size(tx_chn->desc_pool); > + tx_chn->dsize_log2 = __fls(hdesc_size_out); > + WARN_ON(hdesc_size_out != (1 << tx_chn->dsize_log2)); > + > tx_chn->irq = k3_udma_glue_tx_get_irq(tx_chn->tx_chn); > if (tx_chn->irq < 0) { > dev_err(dev, "Failed to get tx dma irq %d\n", > @@ -1862,8 +2287,8 @@ static void am65_cpsw_nuss_free_rx_chns(void *data) > static void am65_cpsw_nuss_remove_rx_chns(void *data) > { > struct am65_cpsw_common *common = data; > - struct am65_cpsw_rx_chn *rx_chn; > struct device *dev = common->dev; > + struct am65_cpsw_rx_chn *rx_chn; > > rx_chn = &common->rx_chns; > devm_remove_action(dev, am65_cpsw_nuss_free_rx_chns, common); > @@ -1873,11 +2298,7 @@ static void am65_cpsw_nuss_remove_rx_chns(void *data) > > netif_napi_del(&common->napi_rx); > > - if (!IS_ERR_OR_NULL(rx_chn->desc_pool)) > - k3_cppi_desc_pool_destroy(rx_chn->desc_pool); > - > - if (!IS_ERR_OR_NULL(rx_chn->rx_chn)) > - k3_udma_glue_release_rx_chn(rx_chn->rx_chn); > + am65_cpsw_nuss_free_rx_chns(common); > > common->rx_flow_id_base = -1; > } > @@ -1888,7 +2309,7 @@ static int am65_cpsw_nuss_init_rx_chns(struct am65_cpsw_common *common) > struct k3_udma_glue_rx_channel_cfg rx_cfg = { 0 }; > u32 max_desc_num = AM65_CPSW_MAX_RX_DESC; > struct device *dev = common->dev; > - u32 hdesc_size; > + u32 hdesc_size, hdesc_size_out; > u32 fdqring_id; > int i, ret = 0; > > @@ -1920,6 +2341,16 @@ static int am65_cpsw_nuss_init_rx_chns(struct am65_cpsw_common *common) > goto err; > } > > + hdesc_size_out = k3_cppi_desc_pool_desc_size(rx_chn->desc_pool); > + rx_chn->dsize_log2 = __fls(hdesc_size_out); > + WARN_ON(hdesc_size_out != (1 << rx_chn->dsize_log2)); > + > + rx_chn->page_pool = NULL; > + > + rx_chn->pages = devm_kcalloc(dev, rx_chn->descs_num, sizeof(*rx_chn->pages), GFP_KERNEL); > + if (!rx_chn->pages) > + return -ENOMEM; > + > common->rx_flow_id_base = > k3_udma_glue_rx_get_flow_id_base(rx_chn->rx_chn); > dev_info(dev, "set new flow-id-base %u\n", common->rx_flow_id_base); > @@ -2252,6 +2683,9 @@ am65_cpsw_nuss_init_port_ndev(struct am65_cpsw_common *common, u32 port_idx) > NETIF_F_HW_TC; > port->ndev->features = port->ndev->hw_features | > NETIF_F_HW_VLAN_CTAG_FILTER; > + port->ndev->xdp_features = NETDEV_XDP_ACT_BASIC | > + NETDEV_XDP_ACT_REDIRECT | > + NETDEV_XDP_ACT_NDO_XMIT; > port->ndev->vlan_features |= NETIF_F_SG; > port->ndev->netdev_ops = &am65_cpsw_nuss_netdev_ops; > port->ndev->ethtool_ops = &am65_cpsw_ethtool_ops_slave; > @@ -2315,6 +2749,8 @@ am65_cpsw_nuss_init_port_ndev(struct am65_cpsw_common *common, u32 port_idx) > if (ret) > dev_err(dev, "failed to add percpu stat free action %d\n", ret); > > + port->xdp_prog = NULL; > + > if (!common->dma_ndev) > common->dma_ndev = port->ndev; > > @@ -2940,9 +3376,9 @@ static int am65_cpsw_nuss_probe(struct platform_device *pdev) > struct device_node *node; > struct resource *res; > struct clk *clk; > + int ale_entries; > u64 id_temp; > int ret, i; > - int ale_entries; > > common = devm_kzalloc(dev, sizeof(struct am65_cpsw_common), GFP_KERNEL); > if (!common) > @@ -3154,10 +3590,10 @@ static int am65_cpsw_nuss_suspend(struct device *dev) > static int am65_cpsw_nuss_resume(struct device *dev) > { > struct am65_cpsw_common *common = dev_get_drvdata(dev); > + struct am65_cpsw_host *host_p = am65_common_get_host(common); > struct am65_cpsw_port *port; > struct net_device *ndev; > int i, ret; > - struct am65_cpsw_host *host_p = am65_common_get_host(common); > > ret = am65_cpsw_nuss_init_tx_chns(common); > if (ret) > diff --git a/drivers/net/ethernet/ti/am65-cpsw-nuss.h b/drivers/net/ethernet/ti/am65-cpsw-nuss.h > index 7da0492dc091..d8ce88dc9c89 100644 > --- a/drivers/net/ethernet/ti/am65-cpsw-nuss.h > +++ b/drivers/net/ethernet/ti/am65-cpsw-nuss.h > @@ -14,6 +14,7 @@ > #include > #include > #include > +#include > #include "am65-cpsw-qos.h" > > struct am65_cpts; > @@ -56,10 +57,18 @@ struct am65_cpsw_port { > bool rx_ts_enabled; > struct am65_cpsw_qos qos; > struct devlink_port devlink_port; > + struct bpf_prog *xdp_prog; > + struct xdp_rxq_info xdp_rxq; > /* Only for suspend resume context */ > u32 vid_context; > }; > > +enum am65_cpsw_tx_buf_type { > + AM65_CPSW_TX_BUF_TYPE_SKB, > + AM65_CPSW_TX_BUF_TYPE_XDP_TX, > + AM65_CPSW_TX_BUF_TYPE_XDP_NDO, > +}; > + > struct am65_cpsw_host { > struct am65_cpsw_common *common; > void __iomem *port_base; > @@ -80,6 +89,7 @@ struct am65_cpsw_tx_chn { > int irq; > u32 id; > u32 descs_num; > + unsigned char dsize_log2; > char tx_chn_name[128]; > u32 rate_mbps; > }; > @@ -89,7 +99,10 @@ struct am65_cpsw_rx_chn { > struct device *dma_dev; > struct k3_cppi_desc_pool *desc_pool; > struct k3_udma_glue_rx_channel *rx_chn; > + struct page_pool *page_pool; > + struct page **pages; > u32 descs_num; > + unsigned char dsize_log2; > int irq; > }; > > > -- > 2.37.3 >