Received: by 2002:a05:7412:b101:b0:e2:908c:2ebd with SMTP id az1csp2653272rdb; Wed, 15 Nov 2023 07:03:46 -0800 (PST) X-Google-Smtp-Source: AGHT+IE9FaFp8TBq0HBry6vq61VVvBH9Voz8W4X6KYNwdwcBtZsEYryh+wLNz1Td+JburkU2ceSU X-Received: by 2002:a05:6808:2385:b0:3ab:85a1:fb04 with SMTP id bp5-20020a056808238500b003ab85a1fb04mr19693639oib.9.1700060625831; Wed, 15 Nov 2023 07:03:45 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1700060625; cv=none; d=google.com; s=arc-20160816; b=OKf9ooVr6UfyOiPR2Pns2FOdY2wVNtMWeLPGbzUu4nvOEjOkniMbjSTA+5npq+/0Pw N854CRN+KtzHrOqcCYnqJPJbvPCX6askWNGAe+gdgHhHvE6lNnAdVfc0nWH2roEtDLvI vV3Ags9cfP5OtBqQJXYXYzsUJc9CcJNSVhtsHAxA2duqM1jD0G5Z+ACKez72ohfGbXnA NP+vcGN4qyVY3URyeFc+aPDlpJ3y8AJdvBZYNVBeXtGBHnPZCf4Awq2PYx3e5+X4MV8/ dFlDsEiPHaWPoAiaLyDnYuAwZBXbnPS/5bwQrswS7/mjEFK8M10l3ZYrZRk2ZiKa/VfR 7qcg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:ui-outboundreport:content-transfer-encoding :in-reply-to:from:references:cc:to:content-language:subject :user-agent:mime-version:date:message-id:dkim-signature; bh=wtBEFzO/JENu935pGP7HfUX5o3RrJrsEXlshX5+TLoo=; fh=eTPBf+jPJiu3/Nm5DCmdMcxS55hFe7gXdXip0v9ZAIM=; b=MpI2k2p7js4cqU0zqVoVd971UmDRx24nfUyGPTrwa3+6xMAaIecThzxM6gBSnhZOG6 FimV+wDlCycBdDruL4wm9e8y5pdDMA4yxoCQSDZ0mMHuoR0GkQzM+4im+eIoRxTPcMDI 1aX/xnrP9TXc7cW2umesUwX5J5xp8XKSf57u3ss3L4+9GTbQo0jSnmxTks/eVxDELa1D ETcZa5OQO0bwVS8HK3Z8zW7O8Lumvuk6Zyibk4J/0CsafiDwPMEO7s/pV5Q5wkQ872Fj TAigAC+1DIjcSHURirJh/lKNUDhN4+uMy9O/BA1dabn6W5S1nCMJn7Sxe+aU3L6V3WYH 8n1w== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@web.de header.s=s29768273 header.b=VLrvhC2G; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:7 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=web.de Return-Path: Received: from snail.vger.email (snail.vger.email. [2620:137:e000::3:7]) by mx.google.com with ESMTPS id a11-20020a9d470b000000b006c64d0dc3e6si3686458otf.340.2023.11.15.07.03.23 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 15 Nov 2023 07:03:45 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:7 as permitted sender) client-ip=2620:137:e000::3:7; Authentication-Results: mx.google.com; dkim=pass header.i=@web.de header.s=s29768273 header.b=VLrvhC2G; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:7 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=web.de Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by snail.vger.email (Postfix) with ESMTP id E4EB781D7A48; Wed, 15 Nov 2023 07:03:04 -0800 (PST) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.11 at snail.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1344364AbjKOPC6 (ORCPT + 99 others); Wed, 15 Nov 2023 10:02:58 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:58196 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1344361AbjKOPCv (ORCPT ); Wed, 15 Nov 2023 10:02:51 -0500 Received: from mout.web.de (mout.web.de [212.227.17.12]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 4E37AF9; Wed, 15 Nov 2023 07:02:43 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=web.de; s=s29768273; t=1700060530; x=1700665330; i=heiner.kallweit@web.de; bh=8/e4dCHBUrc3oSrxwesDZkjN6XEKV8a9NL1LegbJieo=; h=X-UI-Sender-Class:Date:Subject:To:Cc:References:From: In-Reply-To; b=VLrvhC2GqBrAYNQdyqpsxl9ITkfh9VxdFJw4cO3HytkV9sPZ4MtE9ZYi5qhfl1BJ W8RvXO5CUawW+nasyk+U0B2SnITiqpXfpX2ta4UD3/xKfvlMr4T3oHVdBcMTQ2Mg8 NNMxwBOqcgyU+ChRoE8uUItoZXMhPmqw5XH3gK0a2ub3DYqXQXGckbDpRTubhL44/ QeTXObqwZUvhfYMqbLq5caGfAtSutRu8C4hQeLmu1g5UpItGqSKa1SGcJI347z4w9 4Jq1CpXAbVcscQYmds6K+PypDyA9TCZeQaSdYXxh+pPshzsmvcXBkTmzJ/hou8tGd TJ7r3gUf5wlx2w680Q== X-UI-Sender-Class: 814a7b36-bfc1-4dae-8640-3722d8ec6cd6 Received: from [192.168.178.86] ([95.117.28.160]) by smtp.web.de (mrweb106 [213.165.67.124]) with ESMTPSA (Nemesis) id 1MCGWe-1rCfK41E82-009ktL; Wed, 15 Nov 2023 16:02:10 +0100 Message-ID: Date: Wed, 15 Nov 2023 16:02:07 +0100 MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH net-next v11 05/13] net:ethernet:realtek:rtase: Implement hardware configuration function Content-Language: en-US To: Justin Lai , kuba@kernel.org Cc: davem@davemloft.net, edumazet@google.com, pabeni@redhat.com, linux-kernel@vger.kernel.org, netdev@vger.kernel.org, andrew@lunn.ch, pkshih@realtek.com, larry.chiu@realtek.com References: <20231115133414.1221480-1-justinlai0215@realtek.com> <20231115133414.1221480-6-justinlai0215@realtek.com> From: Heiner Kallweit In-Reply-To: <20231115133414.1221480-6-justinlai0215@realtek.com> Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable X-Provags-ID: V03:K1:rETKS6Ti1naWGvUWP89w2w4QXtN4wDcx07nvX7/H34Ix15zLjmm nZ0voPnqAkWejHrQXMXFvxK3Vczrk8WJ08WkkPPrLLhQPrcZr5E712y0OGFzNCbEuiuNXoe /fOaOpmIdyamYfUpAEEVT8mgznc+RN5OCnYLADrEjwruxjQ1iABhxvPIwobX+lky77c4fEr oD0Vw82eowa3IQe5VY8XA== UI-OutboundReport: notjunk:1;M01:P0:vBBxnYHRv2s=;2ZNYq0XMnDP3dpXC7fAZjp5UrWa MkodrsgZML+LIPnqnmXHEmJHI+4/+dcsmB8EGoFmjQ5W5VFNZPZZwsuvV01O8+ZTBeZwvQY+v K4fqO4EK8a0DlotuV7UM1keKKomRS4lsMfRg2CsGkt1dAsNUxuVNA5eavGkRTVB52pOoUs7gY 8vQ9t7cf2gIVw7QydMG6ntaBxCHYDEoZkLVzFmYd55DaqEwdfrW/f1/rCWcHbrO1tCnfb/Z8j Yux/orwLGTRRx1B6/YtvE1Ojzq88lcYxpFntFW8k9A7HktBmUoMqey1R0VnjFc5Zt/Jfm69QN IceL5rd6+mqjPHKbFxP38F2m5P8LOtNCZ8430kKFPCZIcBmaHFbv6NhwDlh8PFxh+3bw1mW6n QuM7owxcPLZPN7vveBh3D+mAaXTq9O0qxwtejWUgIGQ5kc2uIjmaknCqo4ETj7+JbVKV1KlA2 zu5cWJ2I528lw/2mkdSU0G9TOQ8utwrLJz2iZF2h+7U1jtD9LUeM3ngP1SvzfFPJhIBnLjV2d eUm2zDaLAl9oQuIn2DFV851hFqCSqsmtKq1G+XbBCVHe3FHHjQd83eWxpDw+/srhC4o8TGDZk i6dAsVzjpFoyRLE01X8+UKLGDzhw6pfVCq3mQhNIj303wtjhj272oJi6E4+cAdz7gGvmJbNBS nkzQaZuDQb0k5u6cZn/cvUQAi/jWjd/bL94FibfLIEnLMduB2oJo46feqtZ2c0BOS1wYzXP/H nGrZKcFc4eSkNHdn6RXBPpey/dcuyYT55YVy7aFADWgHPuhCdLdWttG6Hab3zJ1OWOjUJTZIq ez X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,FREEMAIL_FROM, RCVD_IN_DNSWL_BLOCKED,RCVD_IN_MSPIKE_H3,RCVD_IN_MSPIKE_WL, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (snail.vger.email [0.0.0.0]); Wed, 15 Nov 2023 07:03:05 -0800 (PST) On 15.11.2023 14:34, Justin Lai wrote: > Implement rtase_hw_config to set default hardware settings, including > setting interrupt mitigation, tx/rx DMA burst, interframe gap time, > rx packet filter, near fifo threshold and fill descriptor ring and > tally counter address, and enable flow control. When filling the > rx descriptor ring, the first group of queues needs to be processed > separately because the positions of the first group of queues are not > regular with other subsequent groups. The other queues are all newly > added features, but we want to retain the original design. So they were > not put together. > > Signed-off-by: Justin Lai > --- > .../net/ethernet/realtek/rtase/rtase_main.c | 243 ++ > drivers/net/ethernet/realtek/rtase/tt.c | 2542 +++++++++++++++++ > drivers/net/ethernet/realtek/rtase/tt.h | 353 +++ > 3 files changed, 3138 insertions(+) > create mode 100644 drivers/net/ethernet/realtek/rtase/tt.c > create mode 100644 drivers/net/ethernet/realtek/rtase/tt.h > > diff --git a/drivers/net/ethernet/realtek/rtase/rtase_main.c b/drivers/n= et/ethernet/realtek/rtase/rtase_main.c > index 46d128a68844..4e04189050cc 100644 > --- a/drivers/net/ethernet/realtek/rtase/rtase_main.c > +++ b/drivers/net/ethernet/realtek/rtase/rtase_main.c > @@ -466,6 +466,25 @@ static int rtase_init_ring(const struct net_device = *dev) > return -ENOMEM; > } > > +static void rtase_interrupt_mitigation(const struct rtase_private *tp) > +{ > + u32 i; > + > + /* tx interrupt mitigation */ > + for (i =3D 0; i < tp->func_tx_queue_num; i++) > + rtase_w16(tp, RTASE_INT_MITI_TX + i * 2, tp->tx_int_mit); > + > + /* rx interrupt mitigation */ > + for (i =3D 0; i < tp->func_rx_queue_num; i++) > + rtase_w16(tp, RTASE_INT_MITI_RX + i * 2, tp->rx_int_mit); > +} > + > +static void rtase_tally_counter_addr_fill(const struct rtase_private *t= p) > +{ > + rtase_w32(tp, RTASE_DTCCR4, upper_32_bits(tp->tally_paddr)); > + rtase_w32(tp, RTASE_DTCCR0, lower_32_bits(tp->tally_paddr)); > +} > + > static void rtase_tally_counter_clear(const struct rtase_private *tp) > { > u32 cmd =3D lower_32_bits(tp->tally_paddr); > @@ -474,6 +493,123 @@ static void rtase_tally_counter_clear(const struct= rtase_private *tp) > rtase_w32(tp, RTASE_DTCCR0, cmd | COUNTER_RESET); > } > > +static void rtase_desc_addr_fill(const struct rtase_private *tp) > +{ > + const struct rtase_ring *ring; > + u16 i, cmd, val; > + int err; > + > + for (i =3D 0; i < tp->func_tx_queue_num; i++) { > + ring =3D &tp->tx_ring[i]; > + > + rtase_w32(tp, RTASE_TX_DESC_ADDR0, > + lower_32_bits(ring->phy_addr)); > + rtase_w32(tp, RTASE_TX_DESC_ADDR4, > + upper_32_bits(ring->phy_addr)); > + > + cmd =3D i | TX_DESC_CMD_WE | TX_DESC_CMD_CS; > + rtase_w16(tp, RTASE_TX_DESC_COMMAND, cmd); > + > + err =3D read_poll_timeout(rtase_r16, val, !(val & TX_DESC_CMD_CS), > + 10, 1000, false, tp, RTASE_TX_DESC_COMMAND); > + > + if (err =3D=3D -ETIMEDOUT) > + netdev_err(tp->dev, "error occurred in fill tx descriptor\n"); > + } > + > + for (i =3D 0; i < tp->func_rx_queue_num; i++) { > + ring =3D &tp->rx_ring[i]; > + > + if (i =3D=3D 0) { > + rtase_w32(tp, RTASE_Q0_RX_DESC_ADDR0, > + lower_32_bits(ring->phy_addr)); > + rtase_w32(tp, RTASE_Q0_RX_DESC_ADDR4, > + upper_32_bits(ring->phy_addr)); > + } else { > + rtase_w32(tp, (RTASE_Q1_RX_DESC_ADDR0 + ((i - 1) * 8)), > + lower_32_bits(ring->phy_addr)); > + rtase_w32(tp, (RTASE_Q1_RX_DESC_ADDR4 + ((i - 1) * 8)), > + upper_32_bits(ring->phy_addr)); > + } > + } > +} > + > +static void rtase_hw_set_features(const struct net_device *dev, > + netdev_features_t features) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + u16 rx_config, val; > + > + rx_config =3D rtase_r16(tp, RTASE_RX_CONFIG_0); > + if (features & NETIF_F_RXALL) > + rx_config |=3D (ACCEPT_ERR | ACCEPT_RUNT); > + else > + rx_config &=3D ~(ACCEPT_ERR | ACCEPT_RUNT); > + > + rtase_w16(tp, RTASE_RX_CONFIG_0, rx_config); > + > + val =3D rtase_r16(tp, RTASE_CPLUS_CMD); > + if (features & NETIF_F_RXCSUM) > + rtase_w16(tp, RTASE_CPLUS_CMD, val | RX_CHKSUM); > + else > + rtase_w16(tp, RTASE_CPLUS_CMD, val & ~RX_CHKSUM); > + > + rx_config =3D rtase_r16(tp, RTASE_RX_CONFIG_1); > + if (dev->features & NETIF_F_HW_VLAN_CTAG_RX) > + rx_config |=3D (INNER_VLAN_DETAG_EN | OUTER_VLAN_DETAG_EN); > + else > + rx_config &=3D ~(INNER_VLAN_DETAG_EN | OUTER_VLAN_DETAG_EN); > + > + rtase_w16(tp, RTASE_RX_CONFIG_1, rx_config); > +} > + > +static void rtase_set_mar(const struct rtase_private *tp) > +{ > + rtase_w32(tp, RTASE_MAR0, tp->mc_filter[0]); > + rtase_w32(tp, RTASE_MAR1, tp->mc_filter[1]); > +} > + > +static void rtase_hw_set_rx_packet_filter(struct net_device *dev) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + u32 mc_filter[2] =3D { 0xFFFFFFFF, 0xFFFFFFFF }; > + u16 rx_mode; > + > + rx_mode =3D rtase_r16(tp, RTASE_RX_CONFIG_0) & ~ACCEPT_MASK; > + rx_mode |=3D ACCEPT_BROADCAST | ACCEPT_MYPHYS; > + > + if (dev->flags & IFF_PROMISC) { > + rx_mode |=3D ACCEPT_MULTICAST | ACCEPT_ALLPHYS; > + } else if ((netdev_mc_count(dev) > MULTICAST_FILTER_LIMIT) || > + (dev->flags & IFF_ALLMULTI)) { > + /* too many to filter perfectly -- accept all multicasts */ This comment seems to exist since RTL8139. The filtering seems to be hash-based, so there's no perfect filtering anyway. What's the benefit of the filter limit? > + rx_mode |=3D ACCEPT_MULTICAST; > + } else { > + struct netdev_hw_addr *hw_addr; > + > + mc_filter[0] =3D 0; > + mc_filter[1] =3D 0; > + > + netdev_for_each_mc_addr(hw_addr, dev) { > + u32 bit_nr =3D eth_hw_addr_crc(hw_addr); > + u32 idx =3D u32_get_bits(bit_nr, BIT(31)); > + u32 bit =3D u32_get_bits(bit_nr, MULTICAST_FILTER_MASK); > + > + mc_filter[idx] |=3D BIT(bit); > + rx_mode |=3D ACCEPT_MULTICAST; > + } > + } > + > + if (dev->features & NETIF_F_RXALL) > + rx_mode |=3D ACCEPT_ERR | ACCEPT_RUNT; > + > + tp->mc_filter[0] =3D swab32(mc_filter[1]); > + tp->mc_filter[1] =3D swab32(mc_filter[0]); > + > + rtase_set_mar(tp); Why this function at all? You can set the registers directly here. Related question: Why store mc_filter[] in tp? It's used nowhere else AFAICS. > + rtase_w16(tp, RTASE_RX_CONFIG_0, rx_mode); > +} > + > static void rtase_irq_dis_and_clear(const struct rtase_private *tp) > { > const struct rtase_int_vector *ivec =3D &tp->int_vector[0]; > @@ -545,6 +681,113 @@ static void rtase_hw_reset(const struct net_device= *dev) > rtase_nic_reset(dev); > } > > +static void rtase_set_rx_queue(const struct rtase_private *tp) > +{ > + u16 reg_data; > + > + reg_data =3D rtase_r16(tp, RTASE_FCR); > + switch (tp->func_rx_queue_num) { > + case 1: > + u16p_replace_bits(®_data, 0x1, FCR_RXQ_MASK); > + break; > + case 2: > + u16p_replace_bits(®_data, 0x2, FCR_RXQ_MASK); > + break; > + case 4: > + u16p_replace_bits(®_data, 0x3, FCR_RXQ_MASK); > + break; > + } > + rtase_w16(tp, RTASE_FCR, reg_data); > +} > + > +static void rtase_set_tx_queue(const struct rtase_private *tp) > +{ > + u16 reg_data; > + > + reg_data =3D rtase_r16(tp, RTASE_TX_CONFIG_1); > + switch (tp->tx_queue_ctrl) { > + case 1: > + u16p_replace_bits(®_data, 0x0, TC_MODE_MASK); > + break; > + case 2: > + u16p_replace_bits(®_data, 0x1, TC_MODE_MASK); > + break; > + case 3: > + case 4: > + u16p_replace_bits(®_data, 0x2, TC_MODE_MASK); > + break; > + default: > + u16p_replace_bits(®_data, 0x3, TC_MODE_MASK); > + break; > + } > + rtase_w16(tp, RTASE_TX_CONFIG_1, reg_data); > +} > + > +static void rtase_hw_config(struct net_device *dev) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + u32 reg_data32; > + u16 reg_data16; > + > + rtase_hw_reset(dev); > + > + /* Set Rx DMA burst */ > + reg_data16 =3D rtase_r16(tp, RTASE_RX_CONFIG_0); > + reg_data16 &=3D ~(RX_SINGLE_TAG | RX_SINGLE_FETCH); > + u16p_replace_bits(®_data16, RX_DMA_BURST_256, RX_MX_DMA_MASK); > + rtase_w16(tp, RTASE_RX_CONFIG_0, reg_data16); > + > + /* New Rx Descritpor */ > + reg_data16 =3D rtase_r16(tp, RTASE_RX_CONFIG_1); > + reg_data16 |=3D RX_NEW_DESC_FORMAT_EN | PCIE_NEW_FLOW; > + u16p_replace_bits(®_data16, 0xF, RX_MAX_FETCH_DESC_MASK); > + rtase_w16(tp, RTASE_RX_CONFIG_1, reg_data16); > + > + rtase_set_rx_queue(tp); > + > + /* interrupt mitigation */ > + rtase_interrupt_mitigation(tp); > + > + /* set tx DMA burst size and interframe gap time */ > + reg_data32 =3D rtase_r32(tp, RTASE_TX_CONFIG_0); > + u32p_replace_bits(®_data32, TX_DMA_BURST_UNLIMITED, TX_DMA_MASK); > + u32p_replace_bits(®_data32, INTERFRAMEGAP, TX_INTER_FRAME_GAP_MASK)= ; > + rtase_w32(tp, RTASE_TX_CONFIG_0, reg_data32); > + > + /* new tx Descriptor */ > + reg_data16 =3D rtase_r16(tp, RTASE_TFUN_CTRL); > + rtase_w16(tp, RTASE_TFUN_CTRL, reg_data16 | TX_NEW_DESC_FORMAT_EN); > + > + /* tx Fetch Desc Number */ > + rtase_w8(tp, RTASE_TDFNR, 0x10); > + > + /* tag num select */ > + reg_data16 =3D rtase_r16(tp, RTASE_MTPS); > + u16p_replace_bits(®_data16, 0x4, TAG_NUM_SEL_MASK); > + rtase_w16(tp, RTASE_MTPS, reg_data16); > + > + rtase_set_tx_queue(tp); > + > + /* TOK condition */ > + rtase_w16(tp, RTASE_TOKSEL, 0x5555); > + > + rtase_tally_counter_addr_fill(tp); > + rtase_desc_addr_fill(tp); > + rtase_hw_set_features(dev, dev->features); > + > + /* enable flow control */ > + reg_data16 =3D rtase_r16(tp, RTASE_CPLUS_CMD); > + reg_data16 |=3D (FORCE_TXFLOW_EN | FORCE_RXFLOW_EN); > + rtase_w16(tp, RTASE_CPLUS_CMD, reg_data16); > + /* set Near FIFO Threshold - rx missed issue. */ > + rtase_w16(tp, RTASE_RFIFONFULL, 0x190); > + > + rtase_w16(tp, RTASE_RMS, tp->rx_buf_sz); > + > + /* set Rx packet filter */ > + rtase_hw_set_rx_packet_filter(dev); > +} > + > static void rtase_nic_enable(const struct net_device *dev) > { > const struct rtase_private *tp =3D netdev_priv(dev); > diff --git a/drivers/net/ethernet/realtek/rtase/tt.c b/drivers/net/ether= net/realtek/rtase/tt.c > new file mode 100644 > index 000000000000..5ea4d51fcc47 > --- /dev/null > +++ b/drivers/net/ethernet/realtek/rtase/tt.c > @@ -0,0 +1,2542 @@ > +// SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause > +/* > + * rtase is the Linux device driver released for Realtek Automotive Sw= itch > + * controllers with PCI-Express interface. > + * > + * Copyright(c) 2023 Realtek Semiconductor Corp. > + * > + * Below is a simplified block diagram of the chip and its relevant in= terfaces. > + * > + * ************************* > + * * * > + * * CPU network device * > + * * * > + * * +-------------+ * > + * * | PCIE Host | * > + * ***********++************ > + * || > + * PCIE > + * || > + * ********************++********************** > + * * | PCIE Endpoint | * > + * * +---------------+ * > + * * | GMAC | * > + * * +--++--+ Realtek * > + * * || RTL90xx Series * > + * * || * > + * * +-------------++----------------+ * > + * * | | MAC | | * > + * * | +-----+ | * > + * * | | * > + * * | Ethernet Switch Core | * > + * * | | * > + * * | +-----+ +-----+ | * > + * * | | MAC |...........| MAC | | * > + * * +---+-----+-----------+-----+---+ * > + * * | PHY |...........| PHY | * > + * * +--++-+ +--++-+ * > + * *************||****************||*********** > + * > + * The block of the Realtek RTL90xx series is our entire chip architec= ture, > + * the GMAC is connected to the switch core, and there is no PHY in be= tween. > + * In addition, this driver is mainly used to control GMAC, but does n= ot > + * control the switch core, so it is not the same as DSA. > + */ > + > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > + > +#include "rtase.h" > + > +#define RTK_OPTS1_DEBUG_VALUE 0x0BADBEEF > +#define RTK_MAGIC_NUMBER 0x0BADBADBADBADBAD > + > +static const struct pci_device_id rtase_pci_tbl[] =3D { > + {PCI_VDEVICE(REALTEK, 0x906A)}, > + {} > +}; > + > +MODULE_DEVICE_TABLE(pci, rtase_pci_tbl); > + > +MODULE_AUTHOR("Realtek ARD Software Team"); > +MODULE_DESCRIPTION("Network Driver for the PCIe interface of Realtek Au= tomotive Ethernet Switch"); > +MODULE_LICENSE("Dual BSD/GPL"); > + > +struct rtase_counters { > + __le64 tx_packets; > + __le64 rx_packets; > + __le64 tx_errors; > + __le32 rx_errors; > + __le16 rx_missed; > + __le16 align_errors; > + __le32 tx_one_collision; > + __le32 tx_multi_collision; > + __le64 rx_unicast; > + __le64 rx_broadcast; > + __le32 rx_multicast; > + __le16 tx_aborted; > + __le16 tx_underun; > +} __packed; > + You defined this struct in patch 1 already in another source code file. Why redefine it? > +static void rtase_w8(const struct rtase_private *tp, u16 reg, u8 val8) > +{ > + writeb(val8, tp->mmio_addr + reg); > +} > + > +static void rtase_w16(const struct rtase_private *tp, u16 reg, u16 val1= 6) > +{ > + writew(val16, tp->mmio_addr + reg); > +} > + > +static void rtase_w32(const struct rtase_private *tp, u16 reg, u32 val3= 2) > +{ > + writel(val32, tp->mmio_addr + reg); > +} > + > +static u8 rtase_r8(const struct rtase_private *tp, u16 reg) > +{ > + return readb(tp->mmio_addr + reg); > +} > + > +static u16 rtase_r16(const struct rtase_private *tp, u16 reg) > +{ > + return readw(tp->mmio_addr + reg); > +} > + > +static u32 rtase_r32(const struct rtase_private *tp, u16 reg) > +{ > + return readl(tp->mmio_addr + reg); > +} > + > +static void rtase_set_rxbufsize(struct rtase_private *tp) > +{ > + tp->rx_buf_sz =3D RX_BUF_SIZE; > +} > + > +static int rtase_alloc_desc(struct rtase_private *tp) > +{ > + struct pci_dev *pdev =3D tp->pdev; > + u32 i; > + > + /* rx and tx descriptors needs 256 bytes alignment. > + * dma_alloc_coherent provides more. > + */ > + for (i =3D 0; i < tp->func_tx_queue_num; i++) { > + tp->tx_ring[i].desc =3D dma_alloc_coherent(&pdev->dev, > + RTASE_TX_RING_DESC_SIZE, > + &tp->tx_ring[i].phy_addr, > + GFP_KERNEL); > + if (!tp->tx_ring[i].desc) > + return -ENOMEM; > + } > + > + for (i =3D 0; i < tp->func_rx_queue_num; i++) { > + tp->rx_ring[i].desc =3D > + dma_alloc_coherent(&pdev->dev, RTASE_RX_RING_DESC_SIZE, > + &tp->rx_ring[i].phy_addr, > + GFP_KERNEL); > + if (!tp->rx_ring[i].desc) > + return -ENOMEM; > + } > + > + return 0; > +} > + > +static void rtase_free_desc(struct rtase_private *tp) > +{ > + struct pci_dev *pdev =3D tp->pdev; > + u32 i; > + > + for (i =3D 0; i < tp->func_tx_queue_num; i++) { > + if (!tp->tx_ring[i].desc) > + continue; > + > + dma_free_coherent(&pdev->dev, RTASE_TX_RING_DESC_SIZE, > + tp->tx_ring[i].desc, > + tp->tx_ring[i].phy_addr); > + tp->tx_ring[i].desc =3D NULL; > + } > + > + for (i =3D 0; i < tp->func_rx_queue_num; i++) { > + if (!tp->rx_ring[i].desc) > + continue; > + > + dma_free_coherent(&pdev->dev, RTASE_RX_RING_DESC_SIZE, > + tp->rx_ring[i].desc, > + tp->rx_ring[i].phy_addr); > + tp->rx_ring[i].desc =3D NULL; > + } > +} > + > +static void rtase_unmap_tx_skb(struct pci_dev *pdev, u32 len, > + struct tx_desc *desc) > +{ > + dma_unmap_single(&pdev->dev, le64_to_cpu(desc->addr), len, > + DMA_TO_DEVICE); > + desc->opts1 =3D cpu_to_le32(RTK_OPTS1_DEBUG_VALUE); > + desc->opts2 =3D 0x00; > + desc->addr =3D cpu_to_le64(RTK_MAGIC_NUMBER); > +} > + > +static void rtase_tx_clear_range(struct rtase_ring *ring, u32 start, u3= 2 n) > +{ > + const struct rtase_private *tp =3D ring->ivec->tp; > + struct net_device *dev =3D tp->dev; > + struct tx_desc *desc_base =3D ring->desc; > + u32 i; > + > + for (i =3D 0; i < n; i++) { > + u32 entry =3D (start + i) % NUM_DESC; > + struct tx_desc *desc =3D desc_base + entry; > + u32 len =3D ring->mis.len[entry]; > + struct sk_buff *skb; > + > + if (len =3D=3D 0) > + continue; > + > + rtase_unmap_tx_skb(tp->pdev, len, desc); > + ring->mis.len[entry] =3D 0; > + skb =3D ring->skbuff[entry]; > + if (!skb) > + continue; > + > + dev->stats.tx_dropped++; > + dev_kfree_skb_any(skb); > + ring->skbuff[entry] =3D NULL; > + } > +} > + > +static void rtase_tx_clear(struct rtase_private *tp) > +{ > + struct rtase_ring *ring; > + u16 i; > + > + for (i =3D 0; i < tp->func_tx_queue_num; i++) { > + ring =3D &tp->tx_ring[i]; > + rtase_tx_clear_range(ring, ring->dirty_idx, NUM_DESC); > + ring->cur_idx =3D 0; > + ring->dirty_idx =3D 0; > + } > +} > + > +static void rtase_mark_to_asic(union rx_desc *desc, u32 rx_buf_sz) > +{ > + u32 eor =3D le32_to_cpu(desc->desc_cmd.opts1) & RING_END; > + > + desc->desc_status.opts2 =3D 0; > + /* force memory writes to complete before releasing descriptor */ > + dma_wmb(); > + WRITE_ONCE(desc->desc_cmd.opts1, > + cpu_to_le32(DESC_OWN | eor | rx_buf_sz)); > +} > + > +static bool rtase_tx_avail(struct rtase_ring *ring) > +{ > + u32 avail_num =3D READ_ONCE(ring->dirty_idx) + NUM_DESC - > + READ_ONCE(ring->cur_idx); > + > + return avail_num > MAX_SKB_FRAGS; > +} > + > +static int tx_handler(struct rtase_ring *ring, int budget) > +{ > + const struct rtase_private *tp =3D ring->ivec->tp; > + struct net_device *dev =3D tp->dev; > + int workdone =3D 0; > + u32 dirty_tx; > + u32 tx_left; > + > + dirty_tx =3D ring->dirty_idx; > + tx_left =3D READ_ONCE(ring->cur_idx) - dirty_tx; > + > + while (tx_left > 0) { > + u32 entry =3D dirty_tx % NUM_DESC; > + struct tx_desc *desc =3D ring->desc + > + sizeof(struct tx_desc) * entry; > + u32 len =3D ring->mis.len[entry]; > + u32 status; > + > + status =3D le32_to_cpu(desc->opts1); > + > + if (status & DESC_OWN) > + break; > + > + rtase_unmap_tx_skb(tp->pdev, len, desc); > + ring->mis.len[entry] =3D 0; > + if (ring->skbuff[entry]) { > + dev_consume_skb_any(ring->skbuff[entry]); > + ring->skbuff[entry] =3D NULL; > + } > + > + dev->stats.tx_bytes +=3D len; > + dev->stats.tx_packets++; > + dirty_tx++; > + tx_left--; > + workdone++; > + > + if (workdone =3D=3D budget) > + break; > + } > + > + if (ring->dirty_idx !=3D dirty_tx) { > + WRITE_ONCE(ring->dirty_idx, dirty_tx); > + > + if (__netif_subqueue_stopped(dev, ring->index) && > + rtase_tx_avail(ring)) > + netif_start_subqueue(dev, ring->index); > + Really no memory barriers needed in this function? > + if (ring->cur_idx !=3D dirty_tx) > + rtase_w8(tp, RTASE_TPPOLL, BIT(ring->index)); > + } > + > + return workdone; > +} > + > +static void rtase_tx_desc_init(struct rtase_private *tp, u16 idx) > +{ > + struct rtase_ring *ring =3D &tp->tx_ring[idx]; > + struct tx_desc *desc; > + u32 i; > + > + memset(ring->desc, 0x0, RTASE_TX_RING_DESC_SIZE); > + memset(ring->skbuff, 0x0, sizeof(ring->skbuff)); > + ring->cur_idx =3D 0; > + ring->dirty_idx =3D 0; > + ring->index =3D idx; > + > + for (i =3D 0; i < NUM_DESC; i++) { > + ring->mis.len[i] =3D 0; > + if ((NUM_DESC - 1) =3D=3D i) { > + desc =3D ring->desc + sizeof(struct tx_desc) * i; > + desc->opts1 =3D cpu_to_le32(RING_END); > + } > + } > + > + ring->ring_handler =3D tx_handler; > + if (idx < 4) { > + ring->ivec =3D &tp->int_vector[idx]; > + list_add_tail(&ring->ring_entry, > + &tp->int_vector[idx].ring_list); > + } else { > + ring->ivec =3D &tp->int_vector[0]; > + list_add_tail(&ring->ring_entry, &tp->int_vector[0].ring_list); > + } > +} > + > +static void rtase_map_to_asic(union rx_desc *desc, dma_addr_t mapping, > + u32 rx_buf_sz) > +{ > + desc->desc_cmd.addr =3D cpu_to_le64(mapping); > + /* make sure the physical address has been updated */ > + wmb(); > + rtase_mark_to_asic(desc, rx_buf_sz); > +} > + > +static void rtase_make_unusable_by_asic(union rx_desc *desc) > +{ > + desc->desc_cmd.addr =3D cpu_to_le64(RTK_MAGIC_NUMBER); > + desc->desc_cmd.opts1 &=3D ~cpu_to_le32(DESC_OWN | RSVD_MASK); > +} > + > +static int rtase_alloc_rx_skb(const struct rtase_ring *ring, > + struct sk_buff **p_sk_buff, union rx_desc *desc, > + dma_addr_t *rx_phy_addr, u8 in_intr) > +{ > + struct rtase_int_vector *ivec =3D ring->ivec; > + const struct rtase_private *tp =3D ivec->tp; > + struct sk_buff *skb =3D NULL; > + struct page *page; > + dma_addr_t mapping; > + void *buf_addr; > + int ret =3D 0; > + > + page =3D page_pool_dev_alloc_pages(tp->page_pool); > + if (!page) { > + netdev_err(tp->dev, "failed to alloc page\n"); > + goto err_out; > + } > + > + buf_addr =3D page_address(page); > + mapping =3D page_pool_get_dma_addr(page); > + > + skb =3D build_skb(buf_addr, PAGE_SIZE); > + if (!skb) { > + page_pool_put_full_page(tp->page_pool, page, true); > + netdev_err(tp->dev, "failed to build skb\n"); > + goto err_out; > + } > + > + *p_sk_buff =3D skb; > + *rx_phy_addr =3D mapping; > + rtase_map_to_asic(desc, mapping, tp->rx_buf_sz); > + > + return ret; > + > +err_out: > + if (skb) > + dev_kfree_skb(skb); > + > + ret =3D -ENOMEM; > + rtase_make_unusable_by_asic(desc); > + > + return ret; > +} > + > +static u32 rtase_rx_ring_fill(struct rtase_ring *ring, u32 ring_start, > + u32 ring_end, u8 in_intr) > +{ > + union rx_desc *desc_base =3D ring->desc; > + u32 cur; > + > + for (cur =3D ring_start; ring_end - cur > 0; cur++) { > + u32 i =3D cur % NUM_DESC; > + union rx_desc *desc =3D desc_base + i; > + int ret; > + > + if (ring->skbuff[i]) > + continue; > + > + ret =3D rtase_alloc_rx_skb(ring, &ring->skbuff[i], desc, > + &ring->mis.data_phy_addr[i], > + in_intr); > + if (ret) > + break; > + } > + > + return cur - ring_start; > +} > + > +static void rtase_mark_as_last_descriptor(union rx_desc *desc) > +{ > + desc->desc_cmd.opts1 |=3D cpu_to_le32(RING_END); > +} > + > +static void rtase_rx_ring_clear(struct rtase_ring *ring) > +{ > + union rx_desc *desc; > + u32 i; > + > + for (i =3D 0; i < NUM_DESC; i++) { > + desc =3D ring->desc + sizeof(union rx_desc) * i; > + > + if (!ring->skbuff[i]) > + continue; > + > + dev_kfree_skb(ring->skbuff[i]); > + > + ring->skbuff[i] =3D NULL; > + > + rtase_make_unusable_by_asic(desc); > + } > +} > + > +static int rtase_fragmented_frame(u32 status) > +{ > + return (status & (RX_FIRST_FRAG | RX_LAST_FRAG)) !=3D > + (RX_FIRST_FRAG | RX_LAST_FRAG); > +} > + > +static void rtase_rx_csum(const struct rtase_private *tp, struct sk_buf= f *skb, > + const union rx_desc *desc) > +{ > + u32 opts2 =3D le32_to_cpu(desc->desc_status.opts2); > + > + /* rx csum offload */ > + if (((opts2 & RX_V4F) && !(opts2 & RX_IPF)) || (opts2 & RX_V6F)) { > + if (((opts2 & RX_TCPT) && !(opts2 & RX_TCPF)) || > + ((opts2 & RX_UDPT) && !(opts2 & RX_UDPF))) { > + skb->ip_summed =3D CHECKSUM_UNNECESSARY; > + } else { > + skb->ip_summed =3D CHECKSUM_NONE; > + } > + } else { > + skb->ip_summed =3D CHECKSUM_NONE; > + } > +} > + > +static void rtase_rx_vlan_skb(union rx_desc *desc, struct sk_buff *skb) > +{ > + u32 opts2 =3D le32_to_cpu(desc->desc_status.opts2); > + > + if (!(opts2 & RX_VLAN_TAG)) > + return; > + > + __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), swab16(opts2 & VLAN_TA= G_MASK)); > +} > + > +static void rtase_rx_skb(const struct rtase_ring *ring, struct sk_buff = *skb) > +{ > + struct rtase_int_vector *ivec =3D ring->ivec; > + > + napi_gro_receive(&ivec->napi, skb); > +} > + > +static int rx_handler(struct rtase_ring *ring, int budget) > +{ > + const struct rtase_private *tp =3D ring->ivec->tp; > + u32 pkt_size, cur_rx, delta, entry, status; > + struct net_device *dev =3D tp->dev; > + union rx_desc *desc_base =3D ring->desc; > + struct sk_buff *skb; > + union rx_desc *desc; > + int workdone =3D 0; > + > + if (!ring->desc) > + return workdone; > + > + cur_rx =3D ring->cur_idx; > + entry =3D cur_rx % NUM_DESC; > + desc =3D &desc_base[entry]; > + > + do { > + /* make sure discriptor has been updated */ > + rmb(); > + status =3D le32_to_cpu(desc->desc_status.opts1); > + > + if (status & DESC_OWN) > + break; > + > + if (unlikely(status & RX_RES)) { > + if (net_ratelimit()) > + netdev_warn(dev, "Rx ERROR. status =3D %08x\n", > + status); > + > + dev->stats.rx_errors++; > + > + if (status & (RX_RWT | RX_RUNT)) > + dev->stats.rx_length_errors++; > + > + if (status & RX_CRC) > + dev->stats.rx_crc_errors++; > + > + if (dev->features & NETIF_F_RXALL) > + goto process_pkt; > + > + rtase_mark_to_asic(desc, tp->rx_buf_sz); > + goto skip_process_pkt; > + } > + > +process_pkt: > + pkt_size =3D status & RX_PKT_SIZE_MASK; > + if (likely(!(dev->features & NETIF_F_RXFCS))) > + pkt_size -=3D ETH_FCS_LEN; > + > + /* the driver does not support incoming fragmented > + * frames. they are seen as a symptom of over-mtu > + * sized frames > + */ > + if (unlikely(rtase_fragmented_frame(status))) { > + dev->stats.rx_dropped++; > + dev->stats.rx_length_errors++; > + rtase_mark_to_asic(desc, tp->rx_buf_sz); > + continue; > + } > + > + skb =3D ring->skbuff[entry]; > + dma_sync_single_for_cpu(&tp->pdev->dev, > + ring->mis.data_phy_addr[entry], > + tp->rx_buf_sz, DMA_FROM_DEVICE); > + > + ring->skbuff[entry] =3D NULL; > + > + if (dev->features & NETIF_F_RXCSUM) > + rtase_rx_csum(tp, skb, desc); > + > + skb->dev =3D dev; > + skb_put(skb, pkt_size); > + skb_mark_for_recycle(skb); > + skb->protocol =3D eth_type_trans(skb, dev); > + > + if (skb->pkt_type =3D=3D PACKET_MULTICAST) > + dev->stats.multicast++; > + > + rtase_rx_vlan_skb(desc, skb); > + rtase_rx_skb(ring, skb); > + > + dev->stats.rx_bytes +=3D pkt_size; > + dev->stats.rx_packets++; > + > +skip_process_pkt: > + workdone++; > + cur_rx++; > + entry =3D cur_rx % NUM_DESC; > + desc =3D ring->desc + sizeof(union rx_desc) * entry; > + prefetch(desc); > + } while (workdone !=3D budget); > + > + ring->cur_idx =3D cur_rx; > + delta =3D rtase_rx_ring_fill(ring, ring->dirty_idx, ring->cur_idx, 1); > + > + if (!delta && workdone) > + netdev_info(dev, "no Rx buffer allocated\n"); > + > + ring->dirty_idx +=3D delta; > + > + if ((ring->dirty_idx + NUM_DESC) =3D=3D ring->cur_idx) > + netdev_emerg(dev, "Rx buffers exhausted\n"); > + > + return workdone; > +} > + > +static void rtase_rx_desc_init(struct rtase_private *tp, u16 idx) > +{ > + struct rtase_ring *ring =3D &tp->rx_ring[idx]; > + u16 i; > + > + memset(ring->desc, 0x0, RTASE_RX_RING_DESC_SIZE); > + memset(ring->skbuff, 0x0, sizeof(ring->skbuff)); > + ring->cur_idx =3D 0; > + ring->dirty_idx =3D 0; > + ring->index =3D idx; > + > + for (i =3D 0; i < NUM_DESC; i++) > + ring->mis.data_phy_addr[i] =3D 0; > + > + ring->ring_handler =3D rx_handler; > + ring->ivec =3D &tp->int_vector[idx]; > + list_add_tail(&ring->ring_entry, &tp->int_vector[idx].ring_list); > +} > + > +static void rtase_rx_clear(struct rtase_private *tp) > +{ > + u32 i; > + > + for (i =3D 0; i < tp->func_rx_queue_num; i++) > + rtase_rx_ring_clear(&tp->rx_ring[i]); > + > + page_pool_destroy(tp->page_pool); > + tp->page_pool =3D NULL; > +} > + > +static int rtase_init_ring(const struct net_device *dev) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + struct page_pool *page_pool; > + struct page_pool_params pp_params =3D { > + .flags =3D PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV, > + .order =3D 0, > + .pool_size =3D NUM_DESC * tp->func_rx_queue_num, > + .nid =3D dev_to_node(&tp->pdev->dev), > + .dev =3D &tp->pdev->dev, > + .dma_dir =3D DMA_FROM_DEVICE, > + .max_len =3D PAGE_SIZE, > + .offset =3D 0, > + }; > + u32 num; > + u16 i; > + > + page_pool =3D page_pool_create(&pp_params); > + if (IS_ERR(page_pool)) { > + netdev_err(tp->dev, "failed to create page pool\n"); > + return -ENOMEM; > + } > + > + tp->page_pool =3D page_pool; > + > + for (i =3D 0; i < tp->func_tx_queue_num; i++) > + rtase_tx_desc_init(tp, i); > + > + for (i =3D 0; i < tp->func_rx_queue_num; i++) { > + rtase_rx_desc_init(tp, i); > + num =3D rtase_rx_ring_fill(&tp->rx_ring[i], 0, NUM_DESC, 0); > + if (num !=3D NUM_DESC) > + goto err_out; > + > + rtase_mark_as_last_descriptor(tp->rx_ring[i].desc + > + sizeof(union rx_desc) * > + (NUM_DESC - 1)); > + } > + > + return 0; > + > +err_out: > + rtase_rx_clear(tp); > + return -ENOMEM; > +} > + > +static void rtase_interrupt_mitigation(const struct rtase_private *tp) > +{ > + u32 i; > + > + /* tx interrupt mitigation */ > + for (i =3D 0; i < tp->func_tx_queue_num; i++) > + rtase_w16(tp, RTASE_INT_MITI_TX + i * 2, tp->tx_int_mit); > + > + /* rx interrupt mitigation */ > + for (i =3D 0; i < tp->func_rx_queue_num; i++) > + rtase_w16(tp, RTASE_INT_MITI_RX + i * 2, tp->rx_int_mit); > +} > + > +static void rtase_tally_counter_addr_fill(const struct rtase_private *t= p) > +{ > + rtase_w32(tp, RTASE_DTCCR4, upper_32_bits(tp->tally_paddr)); > + rtase_w32(tp, RTASE_DTCCR0, lower_32_bits(tp->tally_paddr)); > +} > + > +static void rtase_tally_counter_clear(const struct rtase_private *tp) > +{ > + u32 cmd =3D lower_32_bits(tp->tally_paddr); > + > + rtase_w32(tp, RTASE_DTCCR4, upper_32_bits(tp->tally_paddr)); > + rtase_w32(tp, RTASE_DTCCR0, cmd | COUNTER_RESET); > +} > + > +static void rtase_desc_addr_fill(const struct rtase_private *tp) > +{ > + const struct rtase_ring *ring; > + u16 i, cmd, val; > + int err; > + > + for (i =3D 0; i < tp->func_tx_queue_num; i++) { > + ring =3D &tp->tx_ring[i]; > + > + rtase_w32(tp, RTASE_TX_DESC_ADDR0, > + lower_32_bits(ring->phy_addr)); > + rtase_w32(tp, RTASE_TX_DESC_ADDR4, > + upper_32_bits(ring->phy_addr)); > + > + cmd =3D i | TX_DESC_CMD_WE | TX_DESC_CMD_CS; > + rtase_w16(tp, RTASE_TX_DESC_COMMAND, cmd); > + > + err =3D read_poll_timeout(rtase_r16, val, !(val & TX_DESC_CMD_CS), > + 10, 1000, false, tp, RTASE_TX_DESC_COMMAND); > + > + if (err =3D=3D -ETIMEDOUT) > + netdev_err(tp->dev, "error occurred in fill tx descriptor\n"); > + } > + > + for (i =3D 0; i < tp->func_rx_queue_num; i++) { > + ring =3D &tp->rx_ring[i]; > + > + if (i =3D=3D 0) { > + rtase_w32(tp, RTASE_Q0_RX_DESC_ADDR0, > + lower_32_bits(ring->phy_addr)); > + rtase_w32(tp, RTASE_Q0_RX_DESC_ADDR4, > + upper_32_bits(ring->phy_addr)); > + } else { > + rtase_w32(tp, (RTASE_Q1_RX_DESC_ADDR0 + ((i - 1) * 8)), > + lower_32_bits(ring->phy_addr)); > + rtase_w32(tp, (RTASE_Q1_RX_DESC_ADDR4 + ((i - 1) * 8)), > + upper_32_bits(ring->phy_addr)); > + } > + } > +} > + > +static void rtase_hw_set_features(const struct net_device *dev, > + netdev_features_t features) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + u16 rx_config, val; > + > + rx_config =3D rtase_r16(tp, RTASE_RX_CONFIG_0); > + if (features & NETIF_F_RXALL) > + rx_config |=3D (ACCEPT_ERR | ACCEPT_RUNT); > + else > + rx_config &=3D ~(ACCEPT_ERR | ACCEPT_RUNT); > + > + rtase_w16(tp, RTASE_RX_CONFIG_0, rx_config); > + > + val =3D rtase_r16(tp, RTASE_CPLUS_CMD); > + if (features & NETIF_F_RXCSUM) > + rtase_w16(tp, RTASE_CPLUS_CMD, val | RX_CHKSUM); > + else > + rtase_w16(tp, RTASE_CPLUS_CMD, val & ~RX_CHKSUM); > + > + rx_config =3D rtase_r16(tp, RTASE_RX_CONFIG_1); > + if (dev->features & NETIF_F_HW_VLAN_CTAG_RX) > + rx_config |=3D (INNER_VLAN_DETAG_EN | OUTER_VLAN_DETAG_EN); > + else > + rx_config &=3D ~(INNER_VLAN_DETAG_EN | OUTER_VLAN_DETAG_EN); > + > + rtase_w16(tp, RTASE_RX_CONFIG_1, rx_config); > +} > + > +static void rtase_set_mar(const struct rtase_private *tp) > +{ > + rtase_w32(tp, RTASE_MAR0, tp->mc_filter[0]); > + rtase_w32(tp, RTASE_MAR1, tp->mc_filter[1]); > +} > + > +static void rtase_hw_set_rx_packet_filter(struct net_device *dev) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + u32 mc_filter[2] =3D { 0xFFFFFFFF, 0xFFFFFFFF }; > + u16 rx_mode; > + > + rx_mode =3D rtase_r16(tp, RTASE_RX_CONFIG_0) & ~ACCEPT_MASK; > + rx_mode |=3D ACCEPT_BROADCAST | ACCEPT_MYPHYS; > + > + if (dev->flags & IFF_PROMISC) { > + rx_mode |=3D ACCEPT_MULTICAST | ACCEPT_ALLPHYS; > + } else if ((netdev_mc_count(dev) > MULTICAST_FILTER_LIMIT) || > + (dev->flags & IFF_ALLMULTI)) { > + /* too many to filter perfectly -- accept all multicasts */ > + rx_mode |=3D ACCEPT_MULTICAST; > + } else { > + struct netdev_hw_addr *hw_addr; > + > + mc_filter[0] =3D 0; > + mc_filter[1] =3D 0; > + > + netdev_for_each_mc_addr(hw_addr, dev) { > + u32 bit_nr =3D eth_hw_addr_crc(hw_addr); > + u32 idx =3D u32_get_bits(bit_nr, BIT(31)); > + u32 bit =3D u32_get_bits(bit_nr, MULTICAST_FILTER_MASK); > + > + mc_filter[idx] |=3D BIT(bit); > + rx_mode |=3D ACCEPT_MULTICAST; > + } > + } > + > + if (dev->features & NETIF_F_RXALL) > + rx_mode |=3D ACCEPT_ERR | ACCEPT_RUNT; > + > + tp->mc_filter[0] =3D swab32(mc_filter[1]); > + tp->mc_filter[1] =3D swab32(mc_filter[0]); > + > + rtase_set_mar(tp); > + rtase_w16(tp, RTASE_RX_CONFIG_0, rx_mode); > +} > + This function is define in rt_main already. You seem to duplicate a lot of code. Why? > +static void rtase_irq_dis_and_clear(const struct rtase_private *tp) > +{ > + const struct rtase_int_vector *ivec =3D &tp->int_vector[0]; > + u32 val1; > + u16 val2; > + u8 i; > + > + rtase_w32(tp, ivec->imr_addr, 0); > + val1 =3D rtase_r32(tp, ivec->isr_addr); > + rtase_w32(tp, ivec->isr_addr, val1); > + > + for (i =3D 1; i < tp->int_nums; i++) { > + ivec =3D &tp->int_vector[i]; > + rtase_w16(tp, ivec->imr_addr, 0); > + val2 =3D rtase_r16(tp, ivec->isr_addr); > + rtase_w16(tp, ivec->isr_addr, val2); > + } > +} > + > +static void rtase_poll_timeout(const struct rtase_private *tp, u32 cond= , > + u32 sleep_us, u64 timeout_us, u16 reg) > +{ > + int err; > + u8 val; > + > + err =3D read_poll_timeout(rtase_r8, val, val & cond, sleep_us, timeout= _us, > + false, tp, reg); > + > + if (err =3D=3D -ETIMEDOUT) > + netdev_err(tp->dev, "poll reg 0x00%x timeout\n", reg); > +} > + > +static void rtase_nic_reset(const struct net_device *dev) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + u16 rx_config; > + u8 val; > + > + rx_config =3D rtase_r16(tp, RTASE_RX_CONFIG_0); > + rtase_w16(tp, RTASE_RX_CONFIG_0, rx_config & ~ACCEPT_MASK); > + > + val =3D rtase_r8(tp, RTASE_MISC); > + rtase_w8(tp, RTASE_MISC, val | RX_DV_GATE_EN); > + > + val =3D rtase_r8(tp, RTASE_CHIP_CMD); > + rtase_w8(tp, RTASE_CHIP_CMD, val | STOP_REQ); > + mdelay(2); > + > + rtase_poll_timeout(tp, STOP_REQ_DONE, 100, 150000, RTASE_CHIP_CMD); > + > + rtase_poll_timeout(tp, TX_FIFO_EMPTY, 100, 100000, RTASE_FIFOR); > + > + rtase_poll_timeout(tp, RX_FIFO_EMPTY, 100, 100000, RTASE_FIFOR); > + > + val =3D rtase_r8(tp, RTASE_CHIP_CMD); > + rtase_w8(tp, RTASE_CHIP_CMD, val & ~(TE | RE)); > + val =3D rtase_r8(tp, RTASE_CHIP_CMD); > + rtase_w8(tp, RTASE_CHIP_CMD, val & ~STOP_REQ); > + > + rtase_w16(tp, RTASE_RX_CONFIG_0, rx_config); > +} > + > +static void rtase_hw_reset(const struct net_device *dev) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + > + rtase_irq_dis_and_clear(tp); > + > + rtase_nic_reset(dev); > +} > + > +static void rtase_set_rx_queue(const struct rtase_private *tp) > +{ > + u16 reg_data; > + > + reg_data =3D rtase_r16(tp, RTASE_FCR); > + switch (tp->func_rx_queue_num) { > + case 1: > + u16p_replace_bits(®_data, 0x1, FCR_RXQ_MASK); > + break; > + case 2: > + u16p_replace_bits(®_data, 0x2, FCR_RXQ_MASK); > + break; > + case 4: > + u16p_replace_bits(®_data, 0x3, FCR_RXQ_MASK); > + break; > + } > + rtase_w16(tp, RTASE_FCR, reg_data); > +} > + > +static void rtase_set_tx_queue(const struct rtase_private *tp) > +{ > + u16 reg_data; > + > + reg_data =3D rtase_r16(tp, RTASE_TX_CONFIG_1); > + switch (tp->tx_queue_ctrl) { > + case 1: > + u16p_replace_bits(®_data, 0x0, TC_MODE_MASK); > + break; > + case 2: > + u16p_replace_bits(®_data, 0x1, TC_MODE_MASK); > + break; > + case 3: > + case 4: > + u16p_replace_bits(®_data, 0x2, TC_MODE_MASK); > + break; > + default: > + u16p_replace_bits(®_data, 0x3, TC_MODE_MASK); > + break; > + } > + rtase_w16(tp, RTASE_TX_CONFIG_1, reg_data); > +} > + > +static void rtase_hw_config(struct net_device *dev) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + u32 reg_data32; > + u16 reg_data16; > + > + rtase_hw_reset(dev); > + > + /* Set Rx DMA burst */ > + reg_data16 =3D rtase_r16(tp, RTASE_RX_CONFIG_0); > + reg_data16 &=3D ~(RX_SINGLE_TAG | RX_SINGLE_FETCH); > + u16p_replace_bits(®_data16, RX_DMA_BURST_256, RX_MX_DMA_MASK); > + rtase_w16(tp, RTASE_RX_CONFIG_0, reg_data16); > + > + /* New Rx Descritpor */ > + reg_data16 =3D rtase_r16(tp, RTASE_RX_CONFIG_1); > + reg_data16 |=3D RX_NEW_DESC_FORMAT_EN | PCIE_NEW_FLOW; > + u16p_replace_bits(®_data16, 0xF, RX_MAX_FETCH_DESC_MASK); > + rtase_w16(tp, RTASE_RX_CONFIG_1, reg_data16); > + > + rtase_set_rx_queue(tp); > + > + /* interrupt mitigation */ > + rtase_interrupt_mitigation(tp); > + > + /* set tx DMA burst size and interframe gap time */ > + reg_data32 =3D rtase_r32(tp, RTASE_TX_CONFIG_0); > + u32p_replace_bits(®_data32, TX_DMA_BURST_UNLIMITED, TX_DMA_MASK); > + u32p_replace_bits(®_data32, INTERFRAMEGAP, TX_INTER_FRAME_GAP_MASK)= ; > + rtase_w32(tp, RTASE_TX_CONFIG_0, reg_data32); > + > + /* new tx Descriptor */ > + reg_data16 =3D rtase_r16(tp, RTASE_TFUN_CTRL); > + rtase_w16(tp, RTASE_TFUN_CTRL, reg_data16 | TX_NEW_DESC_FORMAT_EN); > + > + /* tx Fetch Desc Number */ > + rtase_w8(tp, RTASE_TDFNR, 0x10); > + > + /* tag num select */ > + reg_data16 =3D rtase_r16(tp, RTASE_MTPS); > + u16p_replace_bits(®_data16, 0x4, TAG_NUM_SEL_MASK); > + rtase_w16(tp, RTASE_MTPS, reg_data16); > + > + rtase_set_tx_queue(tp); > + > + /* TOK condition */ > + rtase_w16(tp, RTASE_TOKSEL, 0x5555); > + > + rtase_tally_counter_addr_fill(tp); > + rtase_desc_addr_fill(tp); > + rtase_hw_set_features(dev, dev->features); > + > + /* enable flow control */ > + reg_data16 =3D rtase_r16(tp, RTASE_CPLUS_CMD); > + reg_data16 |=3D (FORCE_TXFLOW_EN | FORCE_RXFLOW_EN); > + rtase_w16(tp, RTASE_CPLUS_CMD, reg_data16); > + /* set Near FIFO Threshold - rx missed issue. */ > + rtase_w16(tp, RTASE_RFIFONFULL, 0x190); > + > + rtase_w16(tp, RTASE_RMS, tp->rx_buf_sz); > + > + /* set Rx packet filter */ > + rtase_hw_set_rx_packet_filter(dev); > +} > + > +static void rtase_nic_enable(const struct net_device *dev) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + u16 rcr =3D rtase_r16(tp, RTASE_RX_CONFIG_1); > + u8 val; > + > + /* PCIe PLA reload */ > + rtase_w16(tp, RTASE_RX_CONFIG_1, rcr & ~PCIE_RELOAD_En); > + rtase_w16(tp, RTASE_RX_CONFIG_1, rcr | PCIE_RELOAD_En); > + > + /* set PCIe TE & RE */ > + val =3D rtase_r8(tp, RTASE_CHIP_CMD); > + rtase_w8(tp, RTASE_CHIP_CMD, val | TE | RE); > + > + /* clear rxdv_gated_en */ > + val =3D rtase_r8(tp, RTASE_MISC); > + rtase_w8(tp, RTASE_MISC, val & ~RX_DV_GATE_EN); > +} > + > +static void rtase_enable_hw_interrupt(const struct rtase_private *tp) > +{ > + const struct rtase_int_vector *ivec =3D &tp->int_vector[0]; > + u32 i; > + > + rtase_w32(tp, ivec->imr_addr, ivec->imr); > + > + for (i =3D 1; i < tp->int_nums; i++) { > + ivec =3D &tp->int_vector[i]; > + rtase_w16(tp, ivec->imr_addr, ivec->imr); > + } > +} > + > +static void rtase_hw_start(const struct net_device *dev) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + > + rtase_nic_enable(dev); > + rtase_enable_hw_interrupt(tp); > +} > + > +/* the interrupt handler does RXQ0 and TXQ0, TXQ4~7 interrutp status > + */ > +static irqreturn_t rtase_interrupt(int irq, void *dev_instance) > +{ > + const struct rtase_private *tp; > + struct rtase_int_vector *ivec; > + u32 status; > + > + ivec =3D dev_instance; > + tp =3D ivec->tp; > + status =3D rtase_r32(tp, ivec->isr_addr); > + > + rtase_w32(tp, ivec->imr_addr, 0x0); > + rtase_w32(tp, ivec->isr_addr, status & ~FOVW); > + > + if (napi_schedule_prep(&ivec->napi)) > + __napi_schedule(&ivec->napi); > + > + return IRQ_HANDLED; > +} > + > +/* the interrupt handler does RXQ1&TXQ1 or RXQ2&TXQ2 or RXQ3&TXQ3 inte= rrupt > + * status according to interrupt vector > + */ > +static irqreturn_t rtase_q_interrupt(int irq, void *dev_instance) > +{ > + const struct rtase_private *tp; > + struct rtase_int_vector *ivec; > + u16 status; > + > + ivec =3D dev_instance; > + tp =3D ivec->tp; > + status =3D rtase_r16(tp, ivec->isr_addr); > + > + rtase_w16(tp, ivec->imr_addr, 0x0); > + rtase_w16(tp, ivec->isr_addr, status); > + > + if (napi_schedule_prep(&ivec->napi)) > + __napi_schedule(&ivec->napi); > + > + return IRQ_HANDLED; > +} > + > +static int rtase_poll(struct napi_struct *napi, int budget) > +{ > + const struct rtase_int_vector *ivec; > + const struct rtase_private *tp; > + struct rtase_ring *ring; > + int total_workdone =3D 0; > + > + ivec =3D container_of(napi, struct rtase_int_vector, napi); > + tp =3D ivec->tp; > + > + list_for_each_entry(ring, &ivec->ring_list, ring_entry) { > + total_workdone +=3D ring->ring_handler(ring, budget); > + } > + > + if (total_workdone >=3D budget) > + return budget; > + > + if (napi_complete_done(napi, total_workdone)) { > + if (!ivec->index) > + rtase_w32(tp, ivec->imr_addr, ivec->imr); > + else > + rtase_w16(tp, ivec->imr_addr, ivec->imr); > + } > + > + return total_workdone; > +} > + > +static int rtase_open(struct net_device *dev) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + struct rtase_int_vector *ivec =3D &tp->int_vector[0]; > + const struct pci_dev *pdev =3D tp->pdev; > + int ret; > + u16 i; > + > + rtase_set_rxbufsize(tp); > + > + ret =3D rtase_alloc_desc(tp); > + if (ret) > + goto err_free_all_allocated_mem; > + > + ret =3D rtase_init_ring(dev); > + if (ret) > + goto err_free_all_allocated_mem; > + > + INIT_DELAYED_WORK(&tp->task, NULL); > + Why not just init once in probe()? > + rtase_hw_config(dev); > + > + if (tp->sw_flag & SWF_MSIX_ENABLED) { > + ret =3D request_irq(ivec->irq, rtase_interrupt, 0, > + dev->name, ivec); > + > + /* request other interrupts to handle multiqueue */ > + for (i =3D 1; i < tp->int_nums; i++) { > + if (ret) > + continue; > + > + ivec =3D &tp->int_vector[i]; > + if (ivec->status !=3D 1) > + continue; > + > + snprintf(ivec->name, sizeof(ivec->name), "%s_int%i", tp->dev->name, = i); > + ret =3D request_irq(ivec->irq, rtase_q_interrupt, 0, > + ivec->name, ivec); > + } > + } else if (tp->sw_flag & SWF_MSI_ENABLED) { > + ret =3D request_irq(pdev->irq, rtase_interrupt, 0, dev->name, > + ivec); > + } else { > + ret =3D request_irq(pdev->irq, rtase_interrupt, IRQF_SHARED, > + dev->name, ivec); > + } > + > + if (ret !=3D 0) { > + netdev_err(dev, "can't request MSIX interrupt. Error: %d\n", ret); > + goto err_free_all_allocated_mem; > + } > + > + rtase_hw_start(dev); > + > + netif_carrier_on(dev); > + netif_wake_queue(dev); > + > + goto out; > + > +err_free_all_allocated_mem: > + rtase_free_desc(tp); > + > +out: > + return ret; > +} > + > +static void rtase_down(struct net_device *dev) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + u32 i; > + > + netif_stop_queue(dev); > + > + /* give a racing hard_start_xmit a few cycles to complete */ > + synchronize_rcu(); > + > + netif_carrier_off(dev); > + > + rtase_hw_reset(dev); > + > + for (i =3D 0; i < tp->int_nums; i++) > + synchronize_irq(tp->int_vector[i].irq); > + > + rtase_tx_clear(tp); > + > + rtase_rx_clear(tp); > +} > + > +static int rtase_close(struct net_device *dev) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + const struct pci_dev *pdev =3D tp->pdev; > + u32 i; > + > + rtase_down(dev); > + > + if (tp->sw_flag & SWF_MSIX_ENABLED) { > + for (i =3D 0; i < tp->int_nums; i++) > + free_irq(tp->int_vector[i].irq, &tp->int_vector[i]); > + > + } else { > + free_irq(pdev->irq, &tp->int_vector[0]); > + } > + > + rtase_free_desc(tp); > + > + return 0; > +} > + > +static u32 rtase_tx_vlan_tag(const struct rtase_private *tp, > + const struct sk_buff *skb) > +{ > + return (skb_vlan_tag_present(skb)) ? > + (TX_VLAN_TAG | swab16(skb_vlan_tag_get(skb))) : 0x00; > +} > + > +static u32 rtase_tx_csum(struct sk_buff *skb, const struct net_device *= dev) > +{ > + u8 ip_protocol; > + u32 csum_cmd; > + > + switch (vlan_get_protocol(skb)) { > + case htons(ETH_P_IP): > + csum_cmd =3D TX_IPCS_C; > + ip_protocol =3D ip_hdr(skb)->protocol; > + break; > + > + case htons(ETH_P_IPV6): > + csum_cmd =3D TX_IPV6F_C; > + ip_protocol =3D ipv6_hdr(skb)->nexthdr; > + break; > + > + default: > + ip_protocol =3D IPPROTO_RAW; > + break; > + } > + > + if (ip_protocol =3D=3D IPPROTO_TCP) > + csum_cmd |=3D TX_TCPCS_C; > + else if (ip_protocol =3D=3D IPPROTO_UDP) > + csum_cmd |=3D TX_UDPCS_C; > + else > + WARN_ON_ONCE(1); > + > + csum_cmd |=3D u32_encode_bits(skb_transport_offset(skb), TCPHO_MASK); > + > + return csum_cmd; > +} > + > +static int rtase_xmit_frags(struct rtase_ring *ring, struct sk_buff *sk= b, > + u32 opts1, u32 opts2) > +{ > + const struct skb_shared_info *info =3D skb_shinfo(skb); > + const struct rtase_private *tp =3D ring->ivec->tp; > + const u8 nr_frags =3D info->nr_frags; > + struct tx_desc *txd =3D NULL; > + u32 cur_frag, entry; > + u64 pkt_len_cnt =3D 0; > + > + entry =3D ring->cur_idx; > + for (cur_frag =3D 0; cur_frag < nr_frags; cur_frag++) { > + const skb_frag_t *frag =3D &info->frags[cur_frag]; > + dma_addr_t mapping; > + u32 status, len; > + void *addr; > + > + entry =3D (entry + 1) % NUM_DESC; > + > + txd =3D ring->desc + sizeof(struct tx_desc) * entry; > + len =3D skb_frag_size(frag); > + addr =3D skb_frag_address(frag); > + mapping =3D dma_map_single(&tp->pdev->dev, addr, len, > + DMA_TO_DEVICE); > + > + if (unlikely(dma_mapping_error(&tp->pdev->dev, mapping))) { > + if (unlikely(net_ratelimit())) > + netdev_err(tp->dev, > + "Failed to map TX fragments DMA!\n"); > + > + goto err_out; > + } > + > + if (((entry + 1) % NUM_DESC) =3D=3D 0) > + status =3D (opts1 | len | RING_END); > + else > + status =3D opts1 | len; > + > + if (cur_frag =3D=3D (nr_frags - 1)) { > + ring->skbuff[entry] =3D skb; > + status |=3D TX_LAST_FRAG; > + } > + > + ring->mis.len[entry] =3D len; > + txd->addr =3D cpu_to_le64(mapping); > + txd->opts2 =3D cpu_to_le32(opts2); > + > + /* make sure the operating fields have been updated */ > + wmb(); > + txd->opts1 =3D cpu_to_le32(status); > + pkt_len_cnt +=3D len; > + } > + > + return cur_frag; > + > +err_out: > + rtase_tx_clear_range(ring, ring->cur_idx + 1, cur_frag); > + return -EIO; > +} > + > +static netdev_tx_t rtase_start_xmit(struct sk_buff *skb, > + struct net_device *dev) > +{ > + struct skb_shared_info *shinfo =3D skb_shinfo(skb); > + struct rtase_private *tp =3D netdev_priv(dev); > + u32 q_idx, entry, len, opts1, opts2; > + u32 mss =3D shinfo->gso_size; > + struct rtase_ring *ring; > + struct tx_desc *txd; > + dma_addr_t mapping; > + bool stop_queue; > + int frags; > + > + /* multiqueues */ > + q_idx =3D skb_get_queue_mapping(skb); > + ring =3D &tp->tx_ring[q_idx]; > + > + if (unlikely(!rtase_tx_avail(ring))) { > + if (net_ratelimit()) > + netdev_err(dev, "BUG! Tx Ring full when queue awake!\n"); > + goto err_stop; > + } > + > + entry =3D ring->cur_idx % NUM_DESC; > + txd =3D ring->desc + sizeof(struct tx_desc) * entry; > + > + opts1 =3D DESC_OWN; > + opts2 =3D rtase_tx_vlan_tag(tp, skb); > + > + /* tcp segmentation offload (or tcp large send) */ > + if (mss) { > + if (shinfo->gso_type & SKB_GSO_TCPV4) { > + opts1 |=3D GIANT_SEND_V4; > + } else if (shinfo->gso_type & SKB_GSO_TCPV6) { > + if (skb_cow_head(skb, 0)) > + goto err_dma_0; > + > + tcp_v6_gso_csum_prep(skb); > + opts1 |=3D GIANT_SEND_V6; > + } else { > + WARN_ON_ONCE(1); > + } > + > + opts1 |=3D u32_encode_bits(skb_transport_offset(skb), TCPHO_MASK); > + opts2 |=3D u32_encode_bits(mss, MSS_MASK); > + } else if (skb->ip_summed =3D=3D CHECKSUM_PARTIAL) { > + opts2 |=3D rtase_tx_csum(skb, dev); > + } > + > + frags =3D rtase_xmit_frags(ring, skb, opts1, opts2); > + if (unlikely(frags < 0)) > + goto err_dma_0; > + > + if (frags) { > + len =3D skb_headlen(skb); > + opts1 |=3D TX_FIRST_FRAG; > + } else { > + len =3D skb->len; > + ring->skbuff[entry] =3D skb; > + opts1 |=3D TX_FIRST_FRAG | TX_LAST_FRAG; > + } > + > + if (((entry + 1) % NUM_DESC) =3D=3D 0) > + opts1 |=3D (len | RING_END); > + else > + opts1 |=3D len; > + > + mapping =3D dma_map_single(&tp->pdev->dev, skb->data, len, > + DMA_TO_DEVICE); > + > + if (unlikely(dma_mapping_error(&tp->pdev->dev, mapping))) { > + if (unlikely(net_ratelimit())) > + netdev_err(dev, "Failed to map TX DMA!\n"); > + > + goto err_dma_1; > + } > + > + ring->mis.len[entry] =3D len; > + txd->addr =3D cpu_to_le64(mapping); > + txd->opts2 =3D cpu_to_le32(opts2); > + txd->opts1 =3D cpu_to_le32(opts1 & ~DESC_OWN); > + > + /* make sure the operating fields have been updated */ > + wmb(); > + > + txd->opts1 =3D cpu_to_le32(opts1); > + > + skb_tx_timestamp(skb); > + > + /* tx needs to see descriptor changes before updated cur_idx */ > + smp_wmb(); > + > + WRITE_ONCE(ring->cur_idx, ring->cur_idx + frags + 1); > + > + stop_queue =3D !rtase_tx_avail(ring); > + if (unlikely(stop_queue)) > + netif_stop_subqueue(dev, q_idx); > + > + /* set polling bit */ > + rtase_w8(tp, RTASE_TPPOLL, BIT(ring->index)); > + > + if (unlikely(stop_queue)) { > + /* make sure cur_idx and dirty_idx have been updated */ > + smp_rmb(); > + if (rtase_tx_avail(ring)) > + netif_start_subqueue(dev, q_idx); > + } > + > + return NETDEV_TX_OK; > + > +err_dma_1: > + ring->skbuff[entry] =3D NULL; > + rtase_tx_clear_range(ring, ring->cur_idx + 1, frags); > + > +err_dma_0: > + dev->stats.tx_dropped++; > + dev_kfree_skb_any(skb); > + return NETDEV_TX_OK; > + > +err_stop: > + netif_stop_queue(dev); > + dev->stats.tx_dropped++; > + return NETDEV_TX_BUSY; > +} > + > +static void rtase_set_rx_mode(struct net_device *dev) > +{ > + rtase_hw_set_rx_packet_filter(dev); > +} > + > +static void rtase_enable_eem_write(const struct rtase_private *tp) > +{ > + u8 val; > + > + val =3D rtase_r8(tp, RTASE_EEM); > + rtase_w8(tp, RTASE_EEM, val | EEM_UNLOCK); > +} > + > +static void rtase_disable_eem_write(const struct rtase_private *tp) > +{ > + u8 val; > + > + val =3D rtase_r8(tp, RTASE_EEM); > + rtase_w8(tp, RTASE_EEM, val & ~EEM_UNLOCK); > +} > + > +static void rtase_rar_set(const struct rtase_private *tp, const u8 *add= r) > +{ > + u32 rar_low, rar_high; > + > + rar_low =3D (u32)addr[0] | ((u32)addr[1] << 8) | > + ((u32)addr[2] << 16) | ((u32)addr[3] << 24); > + > + rar_high =3D (u32)addr[4] | ((u32)addr[5] << 8); > + > + rtase_enable_eem_write(tp); > + rtase_w32(tp, RTASE_MAC0, rar_low); > + rtase_w32(tp, RTASE_MAC4, rar_high); > + rtase_disable_eem_write(tp); > + rtase_w16(tp, RTASE_LBK_CTRL, LBK_ATLD | LBK_CLR); > +} > + > +static int rtase_set_mac_address(struct net_device *dev, void *p) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + int ret; > + > + ret =3D eth_mac_addr(dev, p); > + if (ret) > + return ret; > + > + rtase_rar_set(tp, dev->dev_addr); > + > + return 0; > +} > + > +static int rtase_change_mtu(struct net_device *dev, int new_mtu) > +{ > + dev->mtu =3D new_mtu; > + > + netdev_update_features(dev); > + > + return 0; > +} > + > +static void rtase_wait_for_quiescence(const struct net_device *dev) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + struct rtase_int_vector *ivec; > + u32 i; > + > + for (i =3D 0; i < tp->int_nums; i++) { > + ivec =3D &tp->int_vector[i]; > + synchronize_irq(ivec->irq); > + /* wait for any pending NAPI task to complete */ > + napi_disable(&ivec->napi); > + } > + > + rtase_irq_dis_and_clear(tp); > + > + for (i =3D 0; i < tp->int_nums; i++) { > + ivec =3D &tp->int_vector[i]; > + napi_enable(&ivec->napi); > + } > +} > + > +static void rtase_sw_reset(struct net_device *dev) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + int ret; > + > + netif_stop_queue(dev); > + netif_carrier_off(dev); > + rtase_hw_reset(dev); > + > + /* let's wait a bit while any (async) irq lands on */ > + rtase_wait_for_quiescence(dev); > + rtase_tx_clear(tp); > + rtase_rx_clear(tp); > + > + ret =3D rtase_init_ring(dev); > + if (ret) { > + netdev_err(dev, "unable to init ring\n"); > + rtase_free_desc(tp); > + return; > + } > + > + rtase_hw_config(dev); > + /* always link, so start to transmit & receive */ > + rtase_hw_start(dev); > + > + netif_carrier_on(dev); > + netif_wake_queue(dev); > +} > + > +static void rtase_dump_tally_counter(const struct rtase_private *tp) > +{ > + dma_addr_t paddr =3D tp->tally_paddr; > + u32 cmd =3D lower_32_bits(paddr); > + u32 val; > + int err; > + > + rtase_w32(tp, RTASE_DTCCR4, upper_32_bits(paddr)); > + rtase_w32(tp, RTASE_DTCCR0, cmd); > + rtase_w32(tp, RTASE_DTCCR0, cmd | COUNTER_DUMP); > + > + err =3D read_poll_timeout(rtase_r32, val, !(val & COUNTER_DUMP), 10, 2= 50, > + false, tp, RTASE_DTCCR0); > + > + if (err =3D=3D -ETIMEDOUT) > + netdev_err(tp->dev, "error occurred in dump tally counter\n"); > +} > + > +static void rtase_dump_state(const struct net_device *dev) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + const struct rtase_counters *counters; > + int max_reg_size =3D RTASE_PCI_REGS_SIZE; > + const struct rtase_ring *ring; > + u32 dword_rd; > + int n =3D 0; > + > + ring =3D &tp->tx_ring[0]; > + netdev_err(dev, "Tx descriptor info:\n"); > + netdev_err(dev, "Tx curIdx =3D 0x%x\n", ring->cur_idx); > + netdev_err(dev, "Tx dirtyIdx =3D 0x%x\n", ring->dirty_idx); > + netdev_err(dev, "Tx phyAddr =3D 0x%llx\n", ring->phy_addr); > + > + ring =3D &tp->rx_ring[0]; > + netdev_err(dev, "Rx descriptor info:\n"); > + netdev_err(dev, "Rx curIdx =3D 0x%x\n", ring->cur_idx); > + netdev_err(dev, "Rx dirtyIdx =3D 0x%x\n", ring->dirty_idx); > + netdev_err(dev, "Rx phyAddr =3D 0x%llx\n", ring->phy_addr); > + > + netdev_err(dev, "Device Registers:\n"); > + netdev_err(dev, "Chip Command =3D 0x%02x\n", rtase_r8(tp, RTASE_CHIP_C= MD)); > + netdev_err(dev, "IMR =3D %08x\n", rtase_r32(tp, RTASE_IMR0)); > + netdev_err(dev, "ISR =3D %08x\n", rtase_r32(tp, RTASE_ISR0)); > + netdev_err(dev, "Boot Ctrl Reg(0xE004) =3D %04x\n", > + rtase_r16(tp, RTASE_BOOT_CTL)); > + netdev_err(dev, "EPHY ISR(0xE014) =3D %04x\n", > + rtase_r16(tp, RTASE_EPHY_ISR)); > + netdev_err(dev, "EPHY IMR(0xE016) =3D %04x\n", > + rtase_r16(tp, RTASE_EPHY_IMR)); > + netdev_err(dev, "CLKSW SET REG(0xE018) =3D %04x\n", > + rtase_r16(tp, RTASE_CLKSW_SET)); > + > + netdev_err(dev, "Dump PCI Registers:\n"); > + > + while (n < max_reg_size) { > + if ((n % RTASE_DWORD_MOD) =3D=3D 0) > + netdev_err(tp->dev, "0x%03x:\n", n); > + > + pci_read_config_dword(tp->pdev, n, &dword_rd); > + netdev_err(tp->dev, "%08x\n", dword_rd); > + n +=3D 4; > + } > + > + netdev_err(dev, "Dump tally counter:\n"); > + counters =3D tp->tally_vaddr; > + rtase_dump_tally_counter(tp); > + > + netdev_err(dev, "tx_packets %lld\n", > + le64_to_cpu(counters->tx_packets)); > + netdev_err(dev, "rx_packets %lld\n", > + le64_to_cpu(counters->rx_packets)); > + netdev_err(dev, "tx_errors %lld\n", > + le64_to_cpu(counters->tx_errors)); > + netdev_err(dev, "rx_missed %lld\n", > + le64_to_cpu(counters->rx_missed)); > + netdev_err(dev, "align_errors %lld\n", > + le64_to_cpu(counters->align_errors)); > + netdev_err(dev, "tx_one_collision %lld\n", > + le64_to_cpu(counters->tx_one_collision)); > + netdev_err(dev, "tx_multi_collision %lld\n", > + le64_to_cpu(counters->tx_multi_collision)); > + netdev_err(dev, "rx_unicast %lld\n", > + le64_to_cpu(counters->rx_unicast)); > + netdev_err(dev, "rx_broadcast %lld\n", > + le64_to_cpu(counters->rx_broadcast)); > + netdev_err(dev, "rx_multicast %lld\n", > + le64_to_cpu(counters->rx_multicast)); > + netdev_err(dev, "tx_aborted %lld\n", > + le64_to_cpu(counters->tx_aborted)); > + netdev_err(dev, "tx_underun %lld\n", > + le64_to_cpu(counters->tx_underun)); > +} > + > +static void rtase_tx_timeout(struct net_device *dev, unsigned int txque= ue) > +{ > + rtase_dump_state(dev); > + rtase_sw_reset(dev); > +} > + > +static void rtase_get_stats64(struct net_device *dev, > + struct rtnl_link_stats64 *stats) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + const struct rtase_counters *counters =3D tp->tally_vaddr; > + > + if (!counters) > + return; > + > + netdev_stats_to_stats64(stats, &dev->stats); > + dev_fetch_sw_netstats(stats, dev->tstats); > + > + /* fetch additional counter values missing in stats collected by drive= r > + * from tally counter > + */ > + rtase_dump_tally_counter(tp); > + > + stats->tx_errors =3D le64_to_cpu(counters->tx_errors); > + stats->collisions =3D le32_to_cpu(counters->tx_multi_collision); > + stats->tx_aborted_errors =3D le16_to_cpu(counters->tx_aborted); > + stats->rx_missed_errors =3D le16_to_cpu(counters->rx_missed); > +} > + > +static void rtase_enable_vlan_filter(const struct rtase_private *tp, bo= ol enabled) > +{ > + u16 tmp; > + > + if (enabled =3D=3D 1) { > + tmp =3D rtase_r16(tp, RTASE_FCR); > + if (!(tmp & FCR_VLAN_FTR_EN)) > + rtase_w16(tp, RTASE_FCR, tmp | FCR_VLAN_FTR_EN); > + > + tmp =3D rtase_r16(tp, RTASE_PCPR); > + if (!(tmp & PCPR_VLAN_FTR_EN)) > + rtase_w16(tp, RTASE_PCPR, tmp | PCPR_VLAN_FTR_EN); > + } else { > + tmp =3D rtase_r16(tp, RTASE_FCR); > + if (tmp & FCR_VLAN_FTR_EN) > + rtase_w16(tp, RTASE_FCR, tmp & ~FCR_VLAN_FTR_EN); > + > + tmp =3D rtase_r16(tp, RTASE_PCPR); > + if (!(tmp & PCPR_VLAN_FTR_EN)) > + rtase_w16(tp, RTASE_PCPR, tmp & ~PCPR_VLAN_FTR_EN); > + } > +} > + > +static int rtase_vlan_rx_add_vid(struct net_device *dev, __be16 protoco= l, > + u16 vid) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + u16 tmp_mem, i; > + > + if (be16_to_cpu(protocol) !=3D ETH_P_8021Q) > + return -EINVAL; > + > + for (i =3D 0; i < RTASE_VLAN_FILTER_ENTRY_NUM; i++) { > + u16 addr, mask; > + > + if (!(tp->vlan_filter_ctrl & BIT(i))) { > + tp->vlan_filter_ctrl |=3D BIT(i); > + tp->vlan_filter_vid[i] =3D vid; > + rtase_w32(tp, RTASE_VLAN_ENTRY_0 + i * 4, > + vid | VLAN_ENTRY_CAREBIT); > + /* each 16-bit register contains two VLAN entries */ > + addr =3D RTASE_VLAN_ENTRY_MEM_0 + (i & ~0x1); > + mask =3D 0x1 << ((i & 0x1) * 8); > + tmp_mem =3D rtase_r16(tp, addr); > + tmp_mem |=3D mask; > + rtase_w16(tp, addr, tmp_mem); > + break; > + } > + } > + > + if (i =3D=3D RTASE_VLAN_FILTER_ENTRY_NUM) > + return -ENOENT; > + > + rtase_enable_vlan_filter(tp, true); > + > + return 0; > +} > + > +static int rtase_vlan_rx_kill_vid(struct net_device *dev, __be16 protoc= ol, > + u16 vid) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + u16 tmp_mem, i; > + > + if (be16_to_cpu(protocol) !=3D ETH_P_8021Q) > + return -EINVAL; > + > + for (i =3D 0; i < RTASE_VLAN_FILTER_ENTRY_NUM; i++) { > + u16 addr, mask; > + > + if (tp->vlan_filter_vid[i] =3D=3D vid) { > + tp->vlan_filter_ctrl &=3D ~BIT(i); > + tp->vlan_filter_vid[i] =3D 0; > + rtase_w32(tp, RTASE_VLAN_ENTRY_0 + i * 4, 0); > + > + /* each 16-bit register contains two VLAN entries */ > + addr =3D RTASE_VLAN_ENTRY_MEM_0 + (i & ~0x1); > + mask =3D ~(0x1 << ((i & 0x1) * 8)); > + tmp_mem =3D rtase_r16(tp, addr); > + tmp_mem &=3D mask; > + rtase_w16(tp, addr, tmp_mem); > + break; > + } > + } > + > + /* check vlan filter enabled */ > + if (!tp->vlan_filter_ctrl) > + rtase_enable_vlan_filter(tp, false); > + > + return 0; > +} > + > +#ifdef CONFIG_NET_POLL_CONTROLLER > +/* Polling 'interrupt' - used by things like netconsole to send skbs > + * without having to re-enable interrupts. It's not called while > + * the interrupt routine is executing. > + */ > +static void rtase_netpoll(struct net_device *dev) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + const struct pci_dev *pdev =3D tp->pdev; > + > + disable_irq(pdev->irq); > + rtase_interrupt(pdev->irq, dev); > + enable_irq(pdev->irq); > +} > +#endif > + > +static void rtase_set_hw_cbs(const struct rtase_private *tp, u32 queue) > +{ > + u32 idle =3D tp->tx_qos[queue].idleslope * RTASE_1T_CLOCK; > + u32 val, i; > + > + val =3D u32_encode_bits(idle / RTASE_1T_POWER, RTASE_IDLESLOPE_INT_MAS= K); > + idle %=3D RTASE_1T_POWER; > + > + for (i =3D 1; i <=3D RTASE_IDLESLOPE_INT_SHIFT; i++) { > + idle *=3D 2; > + if ((idle / RTASE_1T_POWER) =3D=3D 1) > + val |=3D BIT(RTASE_IDLESLOPE_INT_SHIFT - i); > + > + idle %=3D RTASE_1T_POWER; > + } > + rtase_w32(tp, RTASE_TXQCRDT_0 + queue * 4, val); > +} > + > +static void rtase_setup_tc_cbs(struct rtase_private *tp, > + const struct tc_cbs_qopt_offload *qopt) > +{ > + u32 queue =3D qopt->queue; > + > + tp->tx_qos[queue].hicredit =3D qopt->hicredit; > + tp->tx_qos[queue].locredit =3D qopt->locredit; > + tp->tx_qos[queue].idleslope =3D qopt->idleslope; > + tp->tx_qos[queue].sendslope =3D qopt->sendslope; > + > + /* set hardware cbs */ > + rtase_set_hw_cbs(tp, queue); > +} > + > +static int rtase_setup_tc(struct net_device *dev, enum tc_setup_type ty= pe, > + void *type_data) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + > + switch (type) { > + case TC_SETUP_QDISC_CBS: > + rtase_setup_tc_cbs(tp, type_data); > + break; > + default: > + return -EOPNOTSUPP; > + } > + > + return 0; > +} > + > +static netdev_features_t rtase_fix_features(struct net_device *dev, > + netdev_features_t features) > +{ > + netdev_features_t features_fix =3D features; > + > + if (dev->mtu > MSS_MAX) > + features_fix &=3D ~NETIF_F_ALL_TSO; > + > + if (dev->mtu > ETH_DATA_LEN) > + features_fix &=3D ~NETIF_F_ALL_TSO; > + > + return features_fix; > +} > + > +static int rtase_set_features(struct net_device *dev, > + netdev_features_t features) > +{ > + netdev_features_t features_set =3D features; > + > + features_set &=3D NETIF_F_RXALL | NETIF_F_RXCSUM | > + NETIF_F_HW_VLAN_CTAG_RX; > + > + if (features_set ^ dev->features) > + rtase_hw_set_features(dev, features_set); > + > + return 0; > +} > + > +static const struct net_device_ops rtase_netdev_ops =3D { > + .ndo_open =3D rtase_open, > + .ndo_stop =3D rtase_close, > + .ndo_start_xmit =3D rtase_start_xmit, > + .ndo_set_rx_mode =3D rtase_set_rx_mode, > + .ndo_set_mac_address =3D rtase_set_mac_address, > + .ndo_change_mtu =3D rtase_change_mtu, > + .ndo_tx_timeout =3D rtase_tx_timeout, > + .ndo_get_stats64 =3D rtase_get_stats64, > + .ndo_vlan_rx_add_vid =3D rtase_vlan_rx_add_vid, > + .ndo_vlan_rx_kill_vid =3D rtase_vlan_rx_kill_vid, > +#ifdef CONFIG_NET_POLL_CONTROLLER > + .ndo_poll_controller =3D rtase_netpoll, > +#endif > + .ndo_setup_tc =3D rtase_setup_tc, > + .ndo_fix_features =3D rtase_fix_features, > + .ndo_set_features =3D rtase_set_features, > +}; > + > +static void rtase_get_mac_address(struct net_device *dev) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + u8 mac_addr[ETH_ALEN] __aligned(2) =3D {}; > + u32 i; > + > + for (i =3D 0; i < ETH_ALEN; i++) > + mac_addr[i] =3D rtase_r8(tp, RTASE_MAC0 + i); > + > + if (!is_valid_ether_addr(mac_addr)) { > + eth_random_addr(mac_addr); > + dev->addr_assign_type =3D NET_ADDR_RANDOM; > + netdev_warn(dev, "Random ether addr %pM\n", mac_addr); > + } > + > + eth_hw_addr_set(dev, mac_addr); > + rtase_rar_set(tp, mac_addr); > + > + /* keep the original MAC address */ > + ether_addr_copy(tp->org_mac_addr, dev->dev_addr); > + ether_addr_copy(dev->perm_addr, dev->dev_addr); > +} > + > +static void rtase_get_drvinfo(struct net_device *dev, > + struct ethtool_drvinfo *drvinfo) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + > + strscpy(drvinfo->driver, KBUILD_MODNAME, 32); > + strscpy(drvinfo->bus_info, pci_name(tp->pdev), 32); > +} > + > +static int rtase_get_settings(struct net_device *dev, > + struct ethtool_link_ksettings *cmd) > +{ > + u32 supported =3D SUPPORTED_MII | SUPPORTED_Pause | SUPPORTED_Asym_Pau= se; > + > + ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported, > + supported); > + cmd->base.speed =3D SPEED_5000; > + cmd->base.duplex =3D DUPLEX_FULL; > + cmd->base.port =3D PORT_MII; > + cmd->base.autoneg =3D AUTONEG_DISABLE; > + > + return 0; > +} > + > +static void rtase_get_pauseparam(struct net_device *dev, > + struct ethtool_pauseparam *pause) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + u16 value =3D rtase_r16(tp, RTASE_CPLUS_CMD); > + > + pause->autoneg =3D AUTONEG_DISABLE; > + > + if ((value & (FORCE_TXFLOW_EN | FORCE_RXFLOW_EN)) =3D=3D > + (FORCE_TXFLOW_EN | FORCE_RXFLOW_EN)) { > + pause->rx_pause =3D 1; > + pause->tx_pause =3D 1; > + } else if ((value & FORCE_TXFLOW_EN)) { > + pause->tx_pause =3D 1; > + } else if ((value & FORCE_RXFLOW_EN)) { > + pause->rx_pause =3D 1; > + } > +} > + > +static int rtase_set_pauseparam(struct net_device *dev, > + struct ethtool_pauseparam *pause) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + u16 value =3D rtase_r16(tp, RTASE_CPLUS_CMD); > + > + if (pause->autoneg) > + return -EOPNOTSUPP; > + > + value &=3D ~(FORCE_TXFLOW_EN | FORCE_RXFLOW_EN); > + > + if (pause->tx_pause) > + value |=3D FORCE_TXFLOW_EN; > + > + if (pause->rx_pause) > + value |=3D FORCE_RXFLOW_EN; > + > + rtase_w16(tp, RTASE_CPLUS_CMD, value); > + return 0; > +} > + > +static const char rtase_gstrings[][ETH_GSTRING_LEN] =3D { > + "tx_packets", > + "rx_packets", > + "tx_errors", > + "rx_errors", > + "rx_missed", > + "align_errors", > + "tx_single_collisions", > + "tx_multi_collisions", > + "unicast", > + "broadcast", > + "multicast", > + "tx_aborted", > + "tx_underrun", > +}; > + > +static void rtase_get_strings(struct net_device *dev, u32 stringset, u8= *data) > +{ > + switch (stringset) { > + case ETH_SS_STATS: > + memcpy(data, *rtase_gstrings, sizeof(rtase_gstrings)); > + break; > + } > +} > + > +static int rtase_get_sset_count(struct net_device *dev, int sset) > +{ > + int ret =3D -EOPNOTSUPP; > + > + switch (sset) { > + case ETH_SS_STATS: > + ret =3D ARRAY_SIZE(rtase_gstrings); > + break; > + } > + > + return ret; > +} > + > +static void rtase_get_ethtool_stats(struct net_device *dev, > + struct ethtool_stats *stats, u64 *data) > +{ > + struct rtase_private *tp =3D netdev_priv(dev); > + const struct rtase_counters *counters; > + > + ASSERT_RTNL(); > + > + counters =3D tp->tally_vaddr; > + if (!counters) > + return; > + > + rtase_dump_tally_counter(tp); > + > + data[0] =3D le64_to_cpu(counters->tx_packets); > + data[1] =3D le64_to_cpu(counters->rx_packets); > + data[2] =3D le64_to_cpu(counters->tx_errors); > + data[3] =3D le32_to_cpu(counters->rx_errors); > + data[4] =3D le16_to_cpu(counters->rx_missed); > + data[5] =3D le16_to_cpu(counters->align_errors); > + data[6] =3D le32_to_cpu(counters->tx_one_collision); > + data[7] =3D le32_to_cpu(counters->tx_multi_collision); > + data[8] =3D le64_to_cpu(counters->rx_unicast); > + data[9] =3D le64_to_cpu(counters->rx_broadcast); > + data[10] =3D le32_to_cpu(counters->rx_multicast); > + data[11] =3D le16_to_cpu(counters->tx_aborted); > + data[12] =3D le16_to_cpu(counters->tx_underun); > +} > + > +static const struct ethtool_ops rtase_ethtool_ops =3D { > + .get_drvinfo =3D rtase_get_drvinfo, > + .get_link =3D ethtool_op_get_link, > + .get_link_ksettings =3D rtase_get_settings, > + .get_pauseparam =3D rtase_get_pauseparam, > + .set_pauseparam =3D rtase_set_pauseparam, > + .get_strings =3D rtase_get_strings, > + .get_sset_count =3D rtase_get_sset_count, > + .get_ethtool_stats =3D rtase_get_ethtool_stats, > + .get_ts_info =3D ethtool_op_get_ts_info, > +}; > + > +static void rtase_init_netdev_ops(struct net_device *dev) > +{ > + dev->netdev_ops =3D &rtase_netdev_ops; > + dev->ethtool_ops =3D &rtase_ethtool_ops; > +} > + > +static void rtase_reset_interrupt(struct pci_dev *pdev, > + const struct rtase_private *tp) > +{ > + if (tp->sw_flag & SWF_MSIX_ENABLED) > + pci_disable_msix(pdev); > + else > + pci_disable_msi(pdev); > +} > + > +static int rtase_alloc_msix(struct pci_dev *pdev, struct rtase_private = *tp) > +{ > + int ret; > + u16 i; > + > + memset(tp->msix_entry, 0x0, RTASE_NUM_MSIX * sizeof(struct msix_entry)= ); > + > + for (i =3D 0; i < RTASE_NUM_MSIX; i++) > + tp->msix_entry[i].entry =3D i; > + > + ret =3D pci_enable_msix_range(pdev, tp->msix_entry, tp->int_nums, > + tp->int_nums); > + > + if (ret =3D=3D tp->int_nums) { > + for (i =3D 0; i < tp->int_nums; i++) { > + tp->int_vector[i].irq =3D pci_irq_vector(pdev, i); > + tp->int_vector[i].status =3D 1; > + } > + } > + > + return ret; > +} > + > +static int rtase_alloc_interrupt(struct pci_dev *pdev, > + struct rtase_private *tp) > +{ > + int ret; > + > + ret =3D rtase_alloc_msix(pdev, tp); > + if (ret !=3D tp->int_nums) { > + ret =3D pci_enable_msi(pdev); > + if (ret) > + dev_err(&pdev->dev, > + "unable to alloc interrupt.(MSI)\n"); > + else > + tp->sw_flag |=3D SWF_MSI_ENABLED; > + } else { > + tp->sw_flag |=3D SWF_MSIX_ENABLED; > + } > + > + return ret; > +} > + > +static void rtase_init_hardware(const struct rtase_private *tp) > +{ > + u16 i; > + > + for (i =3D 0; i < RTASE_VLAN_FILTER_ENTRY_NUM; i++) > + rtase_w32(tp, RTASE_VLAN_ENTRY_0 + i * 4, 0); > +} > + > +static void rtase_init_int_vector(struct rtase_private *tp) > +{ > + u16 i; > + > + /* interrupt vector 0 */ > + tp->int_vector[0].tp =3D tp; > + tp->int_vector[0].index =3D 0; > + tp->int_vector[0].imr_addr =3D RTASE_IMR0; > + tp->int_vector[0].isr_addr =3D RTASE_ISR0; > + tp->int_vector[0].imr =3D ROK | RDU | TOK | TOK4 | TOK5 | TOK6 | TOK7; > + tp->int_vector[0].poll =3D rtase_poll; > + > + memset(tp->int_vector[0].name, 0x0, sizeof(tp->int_vector[0].name)); > + INIT_LIST_HEAD(&tp->int_vector[0].ring_list); > + > + netif_napi_add(tp->dev, &tp->int_vector[0].napi, > + tp->int_vector[0].poll); > + napi_enable(&tp->int_vector[0].napi); > + > + /* interrupt vector 1 ~ 3 */ > + for (i =3D 1; i < tp->int_nums; i++) { > + tp->int_vector[i].tp =3D tp; > + tp->int_vector[i].index =3D i; > + tp->int_vector[i].imr_addr =3D RTASE_IMR1 + (i - 1) * 4; > + tp->int_vector[i].isr_addr =3D RTASE_ISR1 + (i - 1) * 4; > + tp->int_vector[i].imr =3D Q_ROK | Q_RDU | Q_TOK; > + tp->int_vector[i].poll =3D rtase_poll; > + > + memset(tp->int_vector[i].name, 0x0, sizeof(tp->int_vector[0].name)); > + INIT_LIST_HEAD(&tp->int_vector[i].ring_list); > + > + netif_napi_add(tp->dev, &tp->int_vector[i].napi, > + tp->int_vector[i].poll); > + napi_enable(&tp->int_vector[i].napi); > + } > +} > + > +static u16 rtase_calc_time_mitigation(u32 time_us) > +{ > + u16 int_miti; > + u8 msb, time_count, time_unit; > + > + time_us =3D min_t(int, time_us, MITI_MAX_TIME); > + > + msb =3D fls(time_us); > + if (msb >=3D MITI_COUNT_BIT_NUM) { > + time_unit =3D msb - MITI_COUNT_BIT_NUM; > + time_count =3D time_us >> (msb - MITI_COUNT_BIT_NUM); > + } else { > + time_unit =3D 0; > + time_count =3D time_us; > + } > + > + int_miti =3D u16_encode_bits(time_count, MITI_TIME_COUNT_MASK) | > + u16_encode_bits(time_unit, MITI_TIME_UNIT_MASK); > + > + return int_miti; > +} > + > +static u16 rtase_calc_packet_num_mitigation(u16 pkt_num) > +{ > + u16 int_miti; > + u8 msb, pkt_num_count, pkt_num_unit; > + > + pkt_num =3D min_t(int, pkt_num, MITI_MAX_PKT_NUM); > + > + if (pkt_num > 60) { > + pkt_num_unit =3D MITI_MAX_PKT_NUM_IDX; > + pkt_num_count =3D pkt_num / MITI_MAX_PKT_NUM_UNIT; > + } else { > + msb =3D fls(pkt_num); > + if (msb >=3D MITI_COUNT_BIT_NUM) { > + pkt_num_unit =3D msb - MITI_COUNT_BIT_NUM; > + pkt_num_count =3D pkt_num >> (msb - MITI_COUNT_BIT_NUM); > + } else { > + pkt_num_unit =3D 0; > + pkt_num_count =3D pkt_num; > + } > + } > + > + int_miti =3D u16_encode_bits(pkt_num_count, MITI_PKT_NUM_COUNT_MASK) | > + u16_encode_bits(pkt_num_unit, MITI_PKT_NUM_UNIT_MASK); > + > + return int_miti; > +} > + > +static void rtase_init_software_variable(struct pci_dev *pdev, > + struct rtase_private *tp) > +{ > + u16 int_miti; > + > + tp->tx_queue_ctrl =3D RTASE_TXQ_CTRL; > + tp->func_tx_queue_num =3D RTASE_FUNC_TXQ_NUM; > + tp->func_rx_queue_num =3D RTASE_FUNC_RXQ_NUM; > + tp->int_nums =3D RTASE_INTERRUPT_NUM; > + > + int_miti =3D rtase_calc_time_mitigation(MITI_DEFAULT_TIME) | > + rtase_calc_packet_num_mitigation(MITI_DEFAULT_PKT_NUM); > + tp->tx_int_mit =3D int_miti; > + tp->rx_int_mit =3D int_miti; > + > + tp->sw_flag =3D 0; > + > + rtase_init_int_vector(tp); > + > + /* MTU range: 60 - hw-specific max */ > + tp->dev->min_mtu =3D ETH_ZLEN; > + tp->dev->max_mtu =3D MAX_JUMBO_SIZE; > +} > + > +static bool rtase_check_mac_version_valid(struct rtase_private *tp) > +{ > + u32 hw_ver =3D rtase_r32(tp, RTASE_TX_CONFIG_0) & HW_VER_MASK; > + bool known_ver =3D false; > + > + switch (hw_ver) { > + case 0x00800000: > + case 0x04000000: > + case 0x04800000: > + known_ver =3D true; > + break; > + } > + > + return known_ver; > +} > + > +static int rtase_init_board(struct pci_dev *pdev, struct net_device **d= ev_out, > + void __iomem **ioaddr_out) > +{ > + struct net_device *dev; > + void __iomem *ioaddr; > + int ret =3D -ENOMEM; > + > + /* dev zeroed in alloc_etherdev */ > + dev =3D alloc_etherdev_mq(sizeof(struct rtase_private), > + RTASE_FUNC_TXQ_NUM); > + if (!dev) > + goto err_out; > + > + SET_NETDEV_DEV(dev, &pdev->dev); > + > + ret =3D pci_enable_device(pdev); > + if (ret < 0) > + goto err_out_free_dev; > + > + /* make sure PCI base addr 1 is MMIO */ > + if (!(pci_resource_flags(pdev, 2) & IORESOURCE_MEM)) { > + ret =3D -ENODEV; > + goto err_out_disable; > + } > + > + /* check for weird/broken PCI region reporting */ > + if (pci_resource_len(pdev, 2) < RTASE_REGS_SIZE) { > + ret =3D -ENODEV; > + goto err_out_disable; > + } > + > + ret =3D pci_request_regions(pdev, KBUILD_MODNAME); > + if (ret < 0) > + goto err_out_disable; > + > + if (!dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64))) > + dev->features |=3D NETIF_F_HIGHDMA; > + else if (dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32))) > + goto err_out_free_res; > + else > + dev_info(&pdev->dev, "DMA_BIT_MASK: 32\n"); > + > + pci_set_master(pdev); > + > + /* ioremap MMIO region */ > + ioaddr =3D ioremap(pci_resource_start(pdev, 2), > + pci_resource_len(pdev, 2)); > + if (!ioaddr) { > + ret =3D -EIO; > + goto err_out_free_res; > + } > + > + *ioaddr_out =3D ioaddr; > + *dev_out =3D dev; > + goto out; > + > +err_out_free_res: > + pci_release_regions(pdev); > + > +err_out_disable: > + pci_disable_device(pdev); > + > +err_out_free_dev: > + free_netdev(dev); > + > +err_out: > + *ioaddr_out =3D NULL; > + *dev_out =3D NULL; > + > +out: > + return ret; > +} > + > +static void rtase_release_board(struct pci_dev *pdev, struct net_device= *dev, > + void __iomem *ioaddr) > +{ > + const struct rtase_private *tp =3D netdev_priv(dev); > + > + rtase_rar_set(tp, tp->org_mac_addr); > + iounmap(ioaddr); > + > + if ((tp->sw_flag & SWF_MSIX_ENABLED)) > + pci_disable_msix(pdev); > + else > + pci_disable_msi(pdev); > + > + pci_release_regions(pdev); > + pci_disable_device(pdev); > + free_netdev(dev); > +} > + > +static int rtase_init_one(struct pci_dev *pdev, > + const struct pci_device_id *ent) > +{ > + struct net_device *dev =3D NULL; > + void __iomem *ioaddr =3D NULL; > + struct rtase_private *tp; > + int ret; > + > + if (!pdev->is_physfn && pdev->is_virtfn) { > + dev_err(&pdev->dev, "This module does not support a virtual function.= "); > + return -EINVAL; > + } > + > + dev_dbg(&pdev->dev, "Automotive Switch Ethernet driver loaded\n"); > + > + ret =3D rtase_init_board(pdev, &dev, &ioaddr); > + if (ret !=3D 0) > + return ret; > + > + tp =3D netdev_priv(dev); > + tp->mmio_addr =3D ioaddr; > + tp->dev =3D dev; > + tp->pdev =3D pdev; > + > + /* identify chip attached to board */ > + if (!rtase_check_mac_version_valid(tp)) { > + return dev_err_probe(&pdev->dev, -ENODEV, > + "unknown chip version, contact rtase maintainers (see MAINTAIN= ERS file)\n"); > + } > + > + dev->tstats =3D netdev_alloc_pcpu_stats(struct pcpu_sw_netstats); > + if (!dev->tstats) > + goto err_out_1; > + > + rtase_init_software_variable(pdev, tp); > + rtase_init_hardware(tp); > + > + ret =3D rtase_alloc_interrupt(pdev, tp); > + if (ret < 0) { > + dev_err(&pdev->dev, "unable to alloc MSIX/MSI\n"); > + goto err_out_1; > + } > + > + rtase_init_netdev_ops(dev); > + > + dev->features |=3D NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX; > + > + dev->features |=3D NETIF_F_IP_CSUM; > + dev->features |=3D NETIF_F_RXCSUM | NETIF_F_SG | NETIF_F_TSO; > + dev->features |=3D NETIF_F_IPV6_CSUM | NETIF_F_TSO6; > + dev->hw_features =3D NETIF_F_SG | NETIF_F_IP_CSUM | NETIF_F_TSO | > + NETIF_F_RXCSUM | NETIF_F_HW_VLAN_CTAG_TX | > + NETIF_F_HW_VLAN_CTAG_RX; > + dev->hw_features |=3D NETIF_F_RXALL; > + dev->hw_features |=3D NETIF_F_RXFCS; > + dev->hw_features |=3D NETIF_F_IPV6_CSUM | NETIF_F_TSO6; > + dev->vlan_features =3D NETIF_F_SG | NETIF_F_IP_CSUM | NETIF_F_TSO | > + NETIF_F_HIGHDMA; > + dev->priv_flags |=3D IFF_LIVE_ADDR_CHANGE; > + netif_set_tso_max_size(dev, LSO_64K); > + netif_set_tso_max_segs(dev, NIC_MAX_PHYS_BUF_COUNT_LSO2); > + > + rtase_get_mac_address(dev); > + > + tp->tally_vaddr =3D dma_alloc_coherent(&pdev->dev, > + sizeof(*tp->tally_vaddr), > + &tp->tally_paddr, > + GFP_KERNEL); > + if (!tp->tally_vaddr) { > + ret =3D -ENOMEM; > + goto err_out; > + } > + > + rtase_tally_counter_clear(tp); > + > + pci_set_drvdata(pdev, dev); > + > + ret =3D register_netdev(dev); > + if (ret !=3D 0) > + goto err_out; > + > + netdev_dbg(dev, "%pM, IRQ %d\n", dev->dev_addr, dev->irq); > + > + netif_carrier_off(dev); > + > + goto out; > + > +err_out: > + if (tp->tally_vaddr) { > + dma_free_coherent(&pdev->dev, > + sizeof(*tp->tally_vaddr), > + tp->tally_vaddr, > + tp->tally_paddr); > + > + tp->tally_vaddr =3D NULL; > + } > + > +err_out_1: > + rtase_release_board(pdev, dev, ioaddr); > + > +out: > + return ret; > +} > + > +static void rtase_remove_one(struct pci_dev *pdev) > +{ > + struct net_device *dev =3D pci_get_drvdata(pdev); > + struct rtase_private *tp =3D netdev_priv(dev); > + struct rtase_int_vector *ivec; > + u32 i; > + > + for (i =3D 0; i < tp->int_nums; i++) { > + ivec =3D &tp->int_vector[i]; > + netif_napi_del(&ivec->napi); > + } > + > + unregister_netdev(dev); > + rtase_reset_interrupt(pdev, tp); > + if (tp->tally_vaddr) { > + dma_free_coherent(&pdev->dev, > + sizeof(*tp->tally_vaddr), > + tp->tally_vaddr, > + tp->tally_paddr); > + tp->tally_vaddr =3D NULL; > + } > + > + rtase_release_board(pdev, dev, tp->mmio_addr); > + pci_set_drvdata(pdev, NULL); > +} > + > +static void rtase_shutdown(struct pci_dev *pdev) > +{ > + struct net_device *dev =3D pci_get_drvdata(pdev); > + const struct rtase_private *tp =3D netdev_priv(dev); > + > + if (netif_running(dev)) > + rtase_close(dev); > + > + rtase_reset_interrupt(pdev, tp); > +} > + > +#ifdef CONFIG_PM > +static int rtase_suspend(struct pci_dev *pdev, pm_message_t state) > +{ > + struct net_device *dev =3D pci_get_drvdata(pdev); > + > + if (!netif_running(dev)) > + goto out; > + > + netif_stop_queue(dev); > + netif_carrier_off(dev); > + netif_device_detach(dev); > + rtase_hw_reset(dev); > + > +out: > + pci_save_state(pdev); > + > + return 0; > +} > + > +static int rtase_resume(struct pci_dev *pdev) > +{ > + struct net_device *dev =3D pci_get_drvdata(pdev); > + struct rtase_private *tp =3D netdev_priv(dev); > + int ret; > + > + pci_set_power_state(pdev, PCI_D0); > + pci_restore_state(pdev); > + pci_enable_wake(pdev, PCI_D0, 0); > + > + /* restore last modified mac address */ > + rtase_rar_set(tp, dev->dev_addr); > + > + if (!netif_running(dev)) > + goto out; > + > + rtase_wait_for_quiescence(dev); > + > + rtase_tx_clear(tp); > + rtase_rx_clear(tp); > + > + ret =3D rtase_init_ring(dev); > + if (ret) { > + netdev_err(dev, "unable to init ring\n"); > + rtase_free_desc(tp); > + return -ENOMEM; > + } > + > + rtase_hw_config(dev); > + /* always link, so start to transmit & receive */ > + rtase_hw_start(dev); > + > + netif_carrier_on(dev); > + netif_wake_queue(dev); > + netif_device_attach(dev); > + > +out: > + return 0; > +} > +#endif /* CONFIG_PM */ > + > +static struct pci_driver rtase_pci_driver =3D { > + .name =3D KBUILD_MODNAME, > + .id_table =3D rtase_pci_tbl, > + .probe =3D rtase_init_one, > + .remove =3D rtase_remove_one, > + .shutdown =3D rtase_shutdown, > +#ifdef CONFIG_PM > + .suspend =3D rtase_suspend, > + .resume =3D rtase_resume, > +#endif > +}; > + > +module_pci_driver(rtase_pci_driver); > diff --git a/drivers/net/ethernet/realtek/rtase/tt.h b/drivers/net/ether= net/realtek/rtase/tt.h > new file mode 100644 > index 000000000000..9239c518c504 > --- /dev/null > +++ b/drivers/net/ethernet/realtek/rtase/tt.h > @@ -0,0 +1,353 @@ > +/* SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause */ > +/* > + * rtase is the Linux device driver released for Realtek Automotive Sw= itch > + * controllers with PCI-Express interface. > + * > + * Copyright(c) 2023 Realtek Semiconductor Corp. > + */ > + > +#ifndef _RTASE_H_ > +#define _RTASE_H_ > + > +/* the low 32 bit address of receive buffer must be 8-byte alignment. *= / > +#define RTK_RX_ALIGN 8 > + > +#define HW_VER_MASK 0x7C800000 > + > +#define RX_DMA_BURST_256 4 > +#define TX_DMA_BURST_UNLIMITED 7 > +#define RX_BUF_SIZE (PAGE_SIZE - \ > + SKB_DATA_ALIGN(sizeof(struct skb_shared_info))) > +#define MAX_JUMBO_SIZE (RX_BUF_SIZE - VLAN_ETH_HLEN - ETH_FCS_L= EN) > + > +/* 3 means InterFrameGap =3D the shortest one */ > +#define INTERFRAMEGAP 0x03 > + > +#define RTASE_REGS_SIZE 256 > +#define RTASE_PCI_REGS_SIZE 0x100 > + > +#define MULTICAST_FILTER_MASK GENMASK(30, 26) > +#define MULTICAST_FILTER_LIMIT 32 > + > +#define RTASE_VLAN_FILTER_ENTRY_NUM 32 > +#define RTASE_NUM_TX_QUEUE 8 > +#define RTASE_NUM_RX_QUEUE 4 > + > +#define RTASE_TXQ_CTRL 1 > +#define RTASE_FUNC_TXQ_NUM 1 > +#define RTASE_FUNC_RXQ_NUM 1 > +#define RTASE_INTERRUPT_NUM 1 > + > +#define MITI_TIME_COUNT_MASK GENMASK(3, 0) > +#define MITI_TIME_UNIT_MASK GENMASK(7, 4) > +#define MITI_DEFAULT_TIME 128 > +#define MITI_MAX_TIME 491520 > +#define MITI_PKT_NUM_COUNT_MASK GENMASK(11, 8) > +#define MITI_PKT_NUM_UNIT_MASK GENMASK(13, 12) > +#define MITI_DEFAULT_PKT_NUM 64 > +#define MITI_MAX_PKT_NUM_IDX 3 > +#define MITI_MAX_PKT_NUM_UNIT 16 > +#define MITI_MAX_PKT_NUM 240 > +#define MITI_COUNT_BIT_NUM 4 > + > +#define RTASE_NUM_MSIX 4 > + > +#define RTASE_DWORD_MOD 16 > + > +/**********************************************************************= *******/ > +enum rtase_registers { > + RTASE_MAC0 =3D 0x0000, > + RTASE_MAC4 =3D 0x0004, > + RTASE_MAR0 =3D 0x0008, > + RTASE_MAR1 =3D 0x000C, > + RTASE_DTCCR0 =3D 0x0010, > + RTASE_DTCCR4 =3D 0x0014, > +#define COUNTER_RESET BIT(0) > +#define COUNTER_DUMP BIT(3) > + > + RTASE_FCR =3D 0x0018, > +#define FCR_RXQ_MASK GENMASK(5, 4) > +#define FCR_VLAN_FTR_EN BIT(1) > + > + RTASE_LBK_CTRL =3D 0x001A, > +#define LBK_ATLD BIT(1) > +#define LBK_CLR BIT(0) > + > + RTASE_TX_DESC_ADDR0 =3D 0x0020, > + RTASE_TX_DESC_ADDR4 =3D 0x0024, > + RTASE_TX_DESC_COMMAND =3D 0x0028, > +#define TX_DESC_CMD_CS BIT(15) > +#define TX_DESC_CMD_WE BIT(14) > + > + RTASE_BOOT_CTL =3D 0x6004, > + RTASE_CLKSW_SET =3D 0x6018, > + > + RTASE_CHIP_CMD =3D 0x0037, > +#define STOP_REQ BIT(7) > +#define STOP_REQ_DONE BIT(6) > +#define RE BIT(3) > +#define TE BIT(2) > + > + RTASE_IMR0 =3D 0x0038, > + RTASE_ISR0 =3D 0x003C, > +#define TOK7 BIT(30) > +#define TOK6 BIT(28) > +#define TOK5 BIT(26) > +#define TOK4 BIT(24) > +#define FOVW BIT(6) > +#define RDU BIT(4) > +#define TOK BIT(2) > +#define ROK BIT(0) > + > + RTASE_IMR1 =3D 0x0800, > + RTASE_ISR1 =3D 0x0802, > +#define Q_TOK BIT(4) > +#define Q_RDU BIT(1) > +#define Q_ROK BIT(0) > + > + RTASE_EPHY_ISR =3D 0x6014, > + RTASE_EPHY_IMR =3D 0x6016, > + > + RTASE_TX_CONFIG_0 =3D 0x0040, > +#define TX_INTER_FRAME_GAP_MASK GENMASK(25, 24) > + /* DMA burst value (0-7) is shift this many bits */ > +#define TX_DMA_MASK GENMASK(10, 8) > + > + RTASE_RX_CONFIG_0 =3D 0x0044, > +#define RX_SINGLE_FETCH BIT(14) > +#define RX_SINGLE_TAG BIT(13) > +#define RX_MX_DMA_MASK GENMASK(10, 8) > +#define ACPT_FLOW BIT(7) > +#define ACCEPT_ERR BIT(5) > +#define ACCEPT_RUNT BIT(4) > +#define ACCEPT_BROADCAST BIT(3) > +#define ACCEPT_MULTICAST BIT(2) > +#define ACCEPT_MYPHYS BIT(1) > +#define ACCEPT_ALLPHYS BIT(0) > +#define ACCEPT_MASK (ACPT_FLOW | ACCEPT_ERR | ACCEPT_RUNT | \ > + ACCEPT_BROADCAST | ACCEPT_MULTICAST | \ > + ACCEPT_MYPHYS | ACCEPT_ALLPHYS) > + > + RTASE_RX_CONFIG_1 =3D 0x0046, > +#define RX_MAX_FETCH_DESC_MASK GENMASK(15, 11) > +#define RX_NEW_DESC_FORMAT_EN BIT(8) > +#define OUTER_VLAN_DETAG_EN BIT(7) > +#define INNER_VLAN_DETAG_EN BIT(6) > +#define PCIE_NEW_FLOW BIT(2) > +#define PCIE_RELOAD_En BIT(0) > + > + RTASE_EEM =3D 0x0050, > +#define EEM_UNLOCK 0xC0 > + > + RTASE_TDFNR =3D 0x0057, > + RTASE_TPPOLL =3D 0x0090, > + RTASE_PDR =3D 0x00B0, > + RTASE_FIFOR =3D 0x00D3, > +#define TX_FIFO_EMPTY BIT(5) > +#define RX_FIFO_EMPTY BIT(4) > + > + RTASE_PCPR =3D 0x00D8, > +#define PCPR_VLAN_FTR_EN BIT(6) > + > + RTASE_RMS =3D 0x00DA, > + RTASE_CPLUS_CMD =3D 0x00E0, > +#define FORCE_RXFLOW_EN BIT(11) > +#define FORCE_TXFLOW_EN BIT(10) > +#define RX_CHKSUM BIT(5) > + > + RTASE_Q0_RX_DESC_ADDR0 =3D 0x00E4, > + RTASE_Q0_RX_DESC_ADDR4 =3D 0x00E8, > + RTASE_Q1_RX_DESC_ADDR0 =3D 0x4000, > + RTASE_Q1_RX_DESC_ADDR4 =3D 0x4004, > + RTASE_MTPS =3D 0x00EC, > +#define TAG_NUM_SEL_MASK GENMASK(10, 8) > + > + RTASE_MISC =3D 0x00F2, > +#define RX_DV_GATE_EN BIT(3) > + > + RTASE_TFUN_CTRL =3D 0x0400, > +#define TX_NEW_DESC_FORMAT_EN BIT(0) > + > + RTASE_TX_CONFIG_1 =3D 0x203E, > +#define TC_MODE_MASK GENMASK(11, 10) > + > + RTASE_TOKSEL =3D 0x2046, > + RTASE_TXQCRDT_0 =3D 0x2500, > + RTASE_RFIFONFULL =3D 0x4406, > + RTASE_INT_MITI_TX =3D 0x0A00, > + RTASE_INT_MITI_RX =3D 0x0A80, > + > + RTASE_VLAN_ENTRY_MEM_0 =3D 0x7234, > + RTASE_VLAN_ENTRY_0 =3D 0xAC80, > +}; > + > +enum desc_status_bit { > + DESC_OWN =3D BIT(31), /* Descriptor is owned by NIC */ > + RING_END =3D BIT(30), /* End of descriptor ring */ > +}; > + > +enum sw_flag_content { > + SWF_MSI_ENABLED =3D BIT(1), > + SWF_MSIX_ENABLED =3D BIT(2), > +}; > + > +#define RSVD_MASK 0x3FFFC000 > + > +struct tx_desc { > + __le32 opts1; > + __le32 opts2; > + __le64 addr; > + __le32 opts3; > + __le32 reserved1; > + __le32 reserved2; > + __le32 reserved3; > +} __packed; > + > +/*------ offset 0 of tx descriptor ------*/ > +#define TX_FIRST_FRAG BIT(29) /* Tx First segment of a packet */ > +#define TX_LAST_FRAG BIT(28) /* Tx Final segment of a packet */ > +#define GIANT_SEND_V4 BIT(26) /* TCP Giant Send Offload V4 (GSOv4) */ > +#define GIANT_SEND_V6 BIT(25) /* TCP Giant Send Offload V6 (GSOv6) */ > +#define TX_VLAN_TAG BIT(17) /* Add VLAN tag */ > + > +/*------ offset 4 of tx descriptor ------*/ > +#define TX_UDPCS_C BIT(31) /* Calculate UDP/IP checksum */ > +#define TX_TCPCS_C BIT(30) /* Calculate TCP/IP checksum */ > +#define TX_IPCS_C BIT(29) /* Calculate IP checksum */ > +#define TX_IPV6F_C BIT(28) /* Indicate it is an IPv6 packet */ > + > +union rx_desc { > + struct { > + __le64 header_buf_addr; > + __le32 reserved1; > + __le32 opts_header_len; > + __le64 addr; > + __le32 reserved2; > + __le32 opts1; > + } __packed desc_cmd; > + > + struct { > + __le32 reserved1; > + __le32 reserved2; > + __le32 rss; > + __le32 opts4; > + __le32 reserved3; > + __le32 opts3; > + __le32 opts2; > + __le32 opts1; > + } __packed desc_status; > +} __packed; > + > +/*------ offset 28 of rx descriptor ------*/ > +#define RX_FIRST_FRAG BIT(25) /* Rx First segment of a packet */ > +#define RX_LAST_FRAG BIT(24) /* Rx Final segment of a packet */ > +#define RX_RES BIT(20) > +#define RX_RUNT BIT(19) > +#define RX_RWT BIT(18) > +#define RX_CRC BIT(16) > +#define RX_V6F BIT(31) > +#define RX_V4F BIT(30) > +#define RX_UDPT BIT(29) > +#define RX_TCPT BIT(28) > +#define RX_IPF BIT(26) /* IP checksum failed */ > +#define RX_UDPF BIT(25) /* UDP/IP checksum failed */ > +#define RX_TCPF BIT(24) /* TCP/IP checksum failed */ > +#define RX_LBK_FIFO_FULL BIT(17) /* Loopback FIFO Full */ > +#define RX_VLAN_TAG BIT(16) /* VLAN tag available */ > + > +#define NUM_DESC 1024 > +#define RTASE_TX_RING_DESC_SIZE (NUM_DESC * sizeof(struct tx_desc)) > +#define RTASE_RX_RING_DESC_SIZE (NUM_DESC * sizeof(union rx_desc)) > +#define VLAN_ENTRY_CAREBIT 0xF0000000 > +#define VLAN_TAG_MASK GENMASK(15, 0) > +#define RX_PKT_SIZE_MASK GENMASK(13, 0) > + > +/* txqos hardware definitions */ > +#define RTASE_1T_CLOCK 64 > +#define RTASE_1T_POWER 10000000 > +#define RTASE_IDLESLOPE_INT_SHIFT 25 > +#define RTASE_IDLESLOPE_INT_MASK GENMASK(31, 25) > + > +#define IVEC_NAME_SIZE (IFNAMSIZ + 10) > + > +struct rtase_int_vector { > + struct rtase_private *tp; > + unsigned int irq; > + u8 status; > + char name[IVEC_NAME_SIZE]; > + u16 index; > + u16 imr_addr; > + u16 isr_addr; > + u32 imr; > + struct list_head ring_list; > + struct napi_struct napi; > + int (*poll)(struct napi_struct *napi, int budget); > +}; > + > +struct rtase_ring { > + struct rtase_int_vector *ivec; > + void *desc; > + dma_addr_t phy_addr; > + u32 cur_idx; > + u32 dirty_idx; > + u16 index; > + > + struct sk_buff *skbuff[NUM_DESC]; > + union { > + u32 len[NUM_DESC]; > + dma_addr_t data_phy_addr[NUM_DESC]; > + } mis; > + > + struct list_head ring_entry; > + int (*ring_handler)(struct rtase_ring *ring, int budget); > +}; > + > +struct rtase_txqos { > + int hicredit; > + int locredit; > + int idleslope; > + int sendslope; > +}; > + > +struct rtase_private { > + void __iomem *mmio_addr; > + u32 sw_flag; > + u32 mc_filter[2]; > + > + struct pci_dev *pdev; > + struct net_device *dev; > + u32 rx_buf_sz; > + > + struct page_pool *page_pool; > + struct rtase_ring tx_ring[RTASE_NUM_TX_QUEUE]; > + struct rtase_txqos tx_qos[RTASE_NUM_TX_QUEUE]; > + struct rtase_ring rx_ring[RTASE_NUM_RX_QUEUE]; > + struct rtase_counters *tally_vaddr; > + dma_addr_t tally_paddr; > + > + u32 vlan_filter_ctrl; > + u16 vlan_filter_vid[RTASE_VLAN_FILTER_ENTRY_NUM]; > + > + struct delayed_work task; I don't see this member anywhere used. What is it good for? And related: Why the delayed version? > + u8 org_mac_addr[ETH_ALEN]; How's the org_mac_addr different from perm_addr? > + struct msix_entry msix_entry[RTASE_NUM_MSIX]; > + struct rtase_int_vector int_vector[RTASE_NUM_MSIX]; > + > + u16 tx_queue_ctrl; > + u16 func_tx_queue_num; > + u16 func_rx_queue_num; > + u16 int_nums; > + u16 tx_int_mit; > + u16 rx_int_mit; > +}; > + > +#define LSO_64K 64000 > + > +#define NIC_MAX_PHYS_BUF_COUNT_LSO2 (16 * 4) > + > +#define TCPHO_MASK GENMASK(24, 18) > + > +#define MSS_MAX 0x07FF /* MSS value */ > +#define MSS_MASK GENMASK(28, 18) > + > +#endif /* _RTASE_H_ */