Received: by 2002:a05:7412:2a8c:b0:e2:908c:2ebd with SMTP id u12csp1822990rdh; Tue, 26 Sep 2023 04:54:57 -0700 (PDT) X-Google-Smtp-Source: AGHT+IH+vhOBH8KdzTB8mMx9BiCS8Azv+QvEghW6YQ7wpMeVMxy0So+yJjo8v1w+YNAszCtzmd3Z X-Received: by 2002:a05:6a20:734b:b0:15f:16f5:858e with SMTP id v11-20020a056a20734b00b0015f16f5858emr5014345pzc.58.1695729297497; Tue, 26 Sep 2023 04:54:57 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1695729297; cv=none; d=google.com; s=arc-20160816; b=UXyXQebDLGSiFysrchZeggKs9Yhgba2AgjtPOZQpOSDQuB6ZBJL53akrMbJzzUJMTV kEulXUHgg1yNsSDK6dYbIhg7fxSs6MduHlRA/TsBb8PfhXl4qTlvRiJJcdpGeFe2fa5Z ObCo+k0wjqXBmmn1GNeoPC44KTwcWtdH0pdJ3kU16bvnSSgvJnFDyKIrRjoNmA8DDSGm aV2hiQXCLKG+lB4kL81gE2KhoBGUFQ3ZbELJiEc7+iZjCBbdNzVvjolhjN/FaPHY58sx uOY7B4QoScbcfUiRHNAb2e5IwfyvL2PpuI461dU8X+IKLNhCRl0q9jIH6nZyGbtkJv86 A/YQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from; bh=taxq76ZEZB1YnYXRXwCD65x4BADFfrfiejM69jTGYyE=; fh=J4St9tMjBInT3r1VVHSF1O7f2qS3KxSaZBHCVggmRHk=; b=xZ0aC3I00CVH8xhb/GoSeB1hUHSLnbXO5bQb+CUuZHZLARSFJUr/R8t8J22NN/b34R Eg5gBj1Am8v2n2rQEAx7r52tzQDXRmim/ae1NCezvCSBU0Gy7GVPws3fcU+ed5gYd3G6 +pTjtSVZUbewWZfLgY4wsuPqWGer19d6C3FOhIaWUOfcWDidEnZaLEwmvVoKOn5MOjWg 8aBr8ofp57TqaFr1ILQPtfqa42O+ZIEGBTrdBdqjsJa6lyKeRb+cX5wyczz6EVlKOwDw Hx3OEHe6SN1bqIqdVy0PhKRpRoIRUJgi3kMyfP/v7BM4TF8bnM5kiGf8hvrpIYedAwZ8 dv6w== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.36 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from pete.vger.email (pete.vger.email. [23.128.96.36]) by mx.google.com with ESMTPS id o5-20020a056a0015c500b006910070695fsi13196833pfu.31.2023.09.26.04.54.56 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 26 Sep 2023 04:54:57 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.36 as permitted sender) client-ip=23.128.96.36; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.36 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by pete.vger.email (Postfix) with ESMTP id 60FD28022C2C; Tue, 26 Sep 2023 04:51:11 -0700 (PDT) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.10 at pete.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S234368AbjIZLvC (ORCPT + 99 others); Tue, 26 Sep 2023 07:51:02 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:58192 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234086AbjIZLua (ORCPT ); Tue, 26 Sep 2023 07:50:30 -0400 Received: from rtits2.realtek.com.tw (rtits2.realtek.com [211.75.126.72]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 7007AE6; Tue, 26 Sep 2023 04:50:23 -0700 (PDT) X-SpamFilter-By: ArmorX SpamTrap 5.78 with qID 38QBo3GI1882027, This message is accepted by code: ctloc85258 Received: from mail.realtek.com (rtexh36505.realtek.com.tw[172.21.6.25]) by rtits2.realtek.com.tw (8.15.2/2.92/5.92) with ESMTPS id 38QBo3GI1882027 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Tue, 26 Sep 2023 19:50:03 +0800 Received: from RTEXMBS04.realtek.com.tw (172.21.6.97) by RTEXH36505.realtek.com.tw (172.21.6.25) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2375.32; Tue, 26 Sep 2023 19:49:56 +0800 Received: from RTDOMAIN (172.21.210.160) by RTEXMBS04.realtek.com.tw (172.21.6.97) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2375.7; Tue, 26 Sep 2023 19:49:55 +0800 From: Justin Lai To: CC: , , , , , , , , Justin Lai Subject: [PATCH net-next v8 05/13] net:ethernet:realtek:rtase: Implement hardware configuration function Date: Tue, 26 Sep 2023 19:49:35 +0800 Message-ID: <20230926114943.16340-6-justinlai0215@realtek.com> X-Mailer: git-send-email 2.34.1 In-Reply-To: <20230926114943.16340-1-justinlai0215@realtek.com> References: <20230926114943.16340-1-justinlai0215@realtek.com> MIME-Version: 1.0 Content-Transfer-Encoding: 7BIT Content-Type: text/plain; charset=US-ASCII X-Originating-IP: [172.21.210.160] X-ClientProxiedBy: RTEXH36505.realtek.com.tw (172.21.6.25) To RTEXMBS04.realtek.com.tw (172.21.6.97) X-KSE-ServerInfo: RTEXMBS04.realtek.com.tw, 9 X-KSE-AntiSpam-Interceptor-Info: fallback X-KSE-Antivirus-Interceptor-Info: fallback X-KSE-AntiSpam-Interceptor-Info: fallback X-KSE-ServerInfo: RTEXH36505.realtek.com.tw, 9 X-KSE-AntiSpam-Interceptor-Info: fallback X-KSE-Antivirus-Interceptor-Info: fallback X-KSE-AntiSpam-Interceptor-Info: fallback X-Spam-Status: No, score=-0.8 required=5.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on pete.vger.email Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (pete.vger.email [0.0.0.0]); Tue, 26 Sep 2023 04:51:11 -0700 (PDT) Implement rtase_hw_config to set default hardware settings, including setting interrupt mitigation, tx/rx DMA burst, interframe gap time, rx packet filter, near fifo threshold and fill descriptor ring and tally counter address, and enable flow control. When filling the rx descriptor ring, the first group of queues needs to be processed separately because the positions of the first group of queues are not regular with other subsequent groups. The other queues are all newly added features, but we want to retain the original design. So they were not put together. Signed-off-by: Justin Lai --- .../net/ethernet/realtek/rtase/rtase_main.c | 245 ++++++++++++++++++ 1 file changed, 245 insertions(+) diff --git a/drivers/net/ethernet/realtek/rtase/rtase_main.c b/drivers/net/ethernet/realtek/rtase/rtase_main.c index f8c3c317c737..9b93cb578834 100644 --- a/drivers/net/ethernet/realtek/rtase/rtase_main.c +++ b/drivers/net/ethernet/realtek/rtase/rtase_main.c @@ -456,6 +456,25 @@ static int rtase_init_ring(const struct net_device *dev) return -ENOMEM; } +static void rtase_interrupt_mitigation(const struct rtase_private *tp) +{ + u32 i; + + /* tx interrupt mitigation */ + for (i = 0; i < tp->func_tx_queue_num; i++) + rtase_w16(tp, RTASE_INT_MITI_TX + i * 2, tp->tx_int_mit); + + /* rx interrupt mitigation */ + for (i = 0; i < tp->func_rx_queue_num; i++) + rtase_w16(tp, RTASE_INT_MITI_RX + i * 2, tp->rx_int_mit); +} + +static void rtase_tally_counter_addr_fill(const struct rtase_private *tp) +{ + rtase_w32(tp, RTASE_DTCCR4, upper_32_bits(le64_to_cpu(tp->tally_paddr))); + rtase_w32(tp, RTASE_DTCCR0, lower_32_bits(le64_to_cpu(tp->tally_paddr))); +} + static void rtase_tally_counter_clear(const struct rtase_private *tp) { u32 cmd = lower_32_bits(le64_to_cpu(tp->tally_paddr)); @@ -464,6 +483,125 @@ static void rtase_tally_counter_clear(const struct rtase_private *tp) rtase_w32(tp, RTASE_DTCCR0, cmd | COUNTER_RESET); } +static void rtase_desc_addr_fill(const struct rtase_private *tp) +{ + const struct rtase_ring *ring; + u16 i, cmd, val; + int err; + + for (i = 0; i < tp->func_tx_queue_num; i++) { + ring = &tp->tx_ring[i]; + + rtase_w32(tp, RTASE_TX_DESC_ADDR0, + lower_32_bits(le64_to_cpu(ring->phy_addr))); + rtase_w32(tp, RTASE_TX_DESC_ADDR4, + upper_32_bits(le64_to_cpu(ring->phy_addr))); + + cmd = i | TX_DESC_CMD_WE | TX_DESC_CMD_CS; + rtase_w16(tp, RTASE_TX_DESC_COMMAND, cmd); + + err = read_poll_timeout(rtase_r16, val, !(val & TX_DESC_CMD_CS), + 10, 1000, false, tp, RTASE_TX_DESC_COMMAND); + + if (err == -ETIMEDOUT) + netdev_err(tp->dev, "error occurred in fill tx descriptor\n"); + } + + for (i = 0; i < tp->func_rx_queue_num; i++) { + ring = &tp->rx_ring[i]; + + if (i == 0) { + rtase_w32(tp, RTASE_Q0_RX_DESC_ADDR0, + lower_32_bits(le64_to_cpu(ring->phy_addr))); + rtase_w32(tp, RTASE_Q0_RX_DESC_ADDR4, + upper_32_bits(le64_to_cpu(ring->phy_addr))); + } else { + rtase_w32(tp, (RTASE_Q1_RX_DESC_ADDR0 + ((i - 1) * 8)), + lower_32_bits(le64_to_cpu(ring->phy_addr))); + rtase_w32(tp, (RTASE_Q1_RX_DESC_ADDR4 + ((i - 1) * 8)), + upper_32_bits(le64_to_cpu(ring->phy_addr))); + } + } +} + +static int rtase_hw_set_features(const struct net_device *dev, + netdev_features_t features) +{ + const struct rtase_private *tp = netdev_priv(dev); + u16 rx_config, val; + + rx_config = rtase_r16(tp, RTASE_RX_CONFIG_0); + if (features & NETIF_F_RXALL) + rx_config |= (ACCEPT_ERR | ACCEPT_RUNT); + else + rx_config &= ~(ACCEPT_ERR | ACCEPT_RUNT); + + rtase_w16(tp, RTASE_RX_CONFIG_0, rx_config); + + val = rtase_r16(tp, RTASE_CPLUS_CMD); + if (features & NETIF_F_RXCSUM) + rtase_w16(tp, RTASE_CPLUS_CMD, val | RX_CHKSUM); + else + rtase_w16(tp, RTASE_CPLUS_CMD, val & ~RX_CHKSUM); + + rx_config = rtase_r16(tp, RTASE_RX_CONFIG_1); + if (dev->features & NETIF_F_HW_VLAN_CTAG_RX) + rx_config |= (INNER_VLAN_DETAG_EN | OUTER_VLAN_DETAG_EN); + else + rx_config &= ~(INNER_VLAN_DETAG_EN | OUTER_VLAN_DETAG_EN); + + rtase_w16(tp, RTASE_RX_CONFIG_1, rx_config); + + return 0; +} + +static void rtase_set_mar(const struct rtase_private *tp) +{ + rtase_w32(tp, RTASE_MAR0, tp->mc_filter[0]); + rtase_w32(tp, RTASE_MAR1, tp->mc_filter[1]); +} + +void rtase_hw_set_rx_packet_filter(struct net_device *dev) +{ + struct rtase_private *tp = netdev_priv(dev); + u32 mc_filter[2] = { 0xFFFFFFFF, 0xFFFFFFFF }; + u16 rx_mode; + + rx_mode = rtase_r16(tp, RTASE_RX_CONFIG_0) & ~ACCEPT_MASK; + rx_mode |= ACCEPT_BROADCAST | ACCEPT_MYPHYS; + + if (dev->flags & IFF_PROMISC) { + rx_mode |= ACCEPT_MULTICAST | ACCEPT_ALLPHYS; + } else if ((netdev_mc_count(dev) > MULTICAST_FILTER_LIMIT) || + (dev->flags & IFF_ALLMULTI)) { + /* too many to filter perfectly -- accept all multicasts */ + rx_mode |= ACCEPT_MULTICAST; + } else { + struct netdev_hw_addr *hw_addr; + + mc_filter[0] = 0; + mc_filter[1] = 0; + + netdev_for_each_mc_addr(hw_addr, dev) { + u32 bit_nr = eth_hw_addr_crc(hw_addr); + u32 idx = u32_get_bits(bit_nr, BIT(31)); + u32 bit = u32_get_bits(bit_nr, MULTICAST_FILTER_MASK); + + mc_filter[idx] |= BIT(bit); + rx_mode |= ACCEPT_MULTICAST; + } + } + + if (dev->features & NETIF_F_RXALL) + rx_mode |= ACCEPT_ERR | ACCEPT_RUNT; + + tp->mc_filter[0] = swab32(mc_filter[1]); + tp->mc_filter[1] = swab32(mc_filter[0]); + + rtase_set_mar(tp); + rtase_w16(tp, RTASE_RX_CONFIG_0, rx_mode); +} + static void rtase_irq_dis_and_clear(const struct rtase_private *tp) { const struct rtase_int_vector *ivec = &tp->int_vector[0]; @@ -535,6 +673,113 @@ void rtase_hw_reset(const struct net_device *dev) rtase_nic_reset(dev); } +static void rtase_set_rx_queue(const struct rtase_private *tp) +{ + u16 reg_data; + + reg_data = rtase_r16(tp, RTASE_FCR); + switch (tp->func_rx_queue_num) { + case 1: + u16p_replace_bits(®_data, 0x1, FCR_RXQ_MASK); + break; + case 2: + u16p_replace_bits(®_data, 0x2, FCR_RXQ_MASK); + break; + case 4: + u16p_replace_bits(®_data, 0x3, FCR_RXQ_MASK); + break; + } + rtase_w16(tp, RTASE_FCR, reg_data); +} + +static void rtase_set_tx_queue(const struct rtase_private *tp) +{ + u16 reg_data; + + reg_data = rtase_r16(tp, RTASE_TX_CONFIG_1); + switch (tp->tx_queue_ctrl) { + case 1: + u16p_replace_bits(®_data, 0x0, TC_MODE_MASK); + break; + case 2: + u16p_replace_bits(®_data, 0x1, TC_MODE_MASK); + break; + case 3: + case 4: + u16p_replace_bits(®_data, 0x2, TC_MODE_MASK); + break; + default: + u16p_replace_bits(®_data, 0x3, TC_MODE_MASK); + break; + } + rtase_w16(tp, RTASE_TX_CONFIG_1, reg_data); +} + +static void rtase_hw_config(struct net_device *dev) +{ + const struct rtase_private *tp = netdev_priv(dev); + u32 reg_data32; + u16 reg_data16; + + rtase_hw_reset(dev); + + /* Set Rx DMA burst */ + reg_data16 = rtase_r16(tp, RTASE_RX_CONFIG_0); + reg_data16 &= ~(RX_SINGLE_TAG | RX_SINGLE_FETCH); + u16p_replace_bits(®_data16, RX_DMA_BURST_256, RX_MX_DMA_MASK); + rtase_w16(tp, RTASE_RX_CONFIG_0, reg_data16); + + /* New Rx Descritpor */ + reg_data16 = rtase_r16(tp, RTASE_RX_CONFIG_1); + reg_data16 |= RX_NEW_DESC_FORMAT_EN | PCIE_NEW_FLOW; + u16p_replace_bits(®_data16, 0xF, RX_MAX_FETCH_DESC_MASK); + rtase_w16(tp, RTASE_RX_CONFIG_1, reg_data16); + + rtase_set_rx_queue(tp); + + /* interrupt mitigation */ + rtase_interrupt_mitigation(tp); + + /* set tx DMA burst size and interframe gap time */ + reg_data32 = rtase_r32(tp, RTASE_TX_CONFIG_0); + u32p_replace_bits(®_data32, TX_DMA_BURST_UNLIMITED, TX_DMA_MASK); + u32p_replace_bits(®_data32, INTERFRAMEGAP, TX_INTER_FRAME_GAP_MASK); + rtase_w32(tp, RTASE_TX_CONFIG_0, reg_data32); + + /* new tx Descriptor */ + reg_data16 = rtase_r16(tp, RTASE_TFUN_CTRL); + rtase_w16(tp, RTASE_TFUN_CTRL, reg_data16 | TX_NEW_DESC_FORMAT_EN); + + /* tx Fetch Desc Number */ + rtase_w8(tp, RTASE_TDFNR, 0x10); + + /* tag num select */ + reg_data16 = rtase_r16(tp, RTASE_MTPS); + u16p_replace_bits(®_data16, 0x4, TAG_NUM_SEL_MASK); + rtase_w16(tp, RTASE_MTPS, reg_data16); + + rtase_set_tx_queue(tp); + + /* TOK condition */ + rtase_w16(tp, RTASE_TOKSEL, 0x5555); + + rtase_tally_counter_addr_fill(tp); + rtase_desc_addr_fill(tp); + rtase_hw_set_features(dev, dev->features); + + /* enable flow control */ + reg_data16 = rtase_r16(tp, RTASE_CPLUS_CMD); + reg_data16 |= (FORCE_TXFLOW_EN | FORCE_RXFLOW_EN); + rtase_w16(tp, RTASE_CPLUS_CMD, reg_data16); + /* set Near FIFO Threshold - rx missed issue. */ + rtase_w16(tp, RTASE_RFIFONFULL, 0x190); + + rtase_w16(tp, RTASE_RMS, tp->rx_buf_sz); + + /* set Rx packet filter */ + rtase_hw_set_rx_packet_filter(dev); +} + static void rtase_nic_enable(const struct net_device *dev) { const struct rtase_private *tp = netdev_priv(dev); -- 2.34.1