Received: by 2002:a25:8b12:0:0:0:0:0 with SMTP id i18csp147623ybl; Tue, 13 Aug 2019 17:45:15 -0700 (PDT) X-Google-Smtp-Source: APXvYqy8OeUKI3milK/WzU1M8M7EOSd6/QMRpyH9gBlv38UBdLC+ZetT1G+HfqXOwZPZ9p+/1lRM X-Received: by 2002:a17:902:1024:: with SMTP id b33mr31181292pla.325.1565743515057; Tue, 13 Aug 2019 17:45:15 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1565743515; cv=none; d=google.com; s=arc-20160816; b=mM1/8HRfTUV50NoWmMJsnHpvc1OjTy9ohnl5vCYLZMRynN2XCtwUoJ5tOIckAx0MR1 wxS1NDxg3Y9S8DrJHvlAWxHhGR9xRrSYa13Upse3eS7yf3DdNMlyme1IgmatQOn5c7m1 K8WliDlupagBH4ofXUVg7AVIMpJlsQZO8p6Yo4gd/7dpm3y6gGYaA+OESJqgC80Ed+HU xuZZGma7KBr0NKj8taRSHfByBs70ayh1rSiWrXe3uyy1ypw4YDK9OTrf9ahTNUbBGqoQ P21OqDQfWY8kcjBHYDWQYJsLk5VUSjFdokwJe9yBhrPHHxGcczWtrslAb8XK90ijuggl plHQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=xTgTghZIlx7I03WklruelXMJsYuL++iO37GrwFIicHk=; b=Op+eItuyTd6OXORhFzOptHQkqN9fbt1ylMFWE8GKttlvIN1oCIZlx6Ag/M17lCgfS+ xOOlzI/y4KULUhJ0qkHxp5Zqycj9mnBB4kDZvlTcLxyc1MN6yVI5mo2/tH72MlduaVn6 PrGL0j6fKsYjLG8MGt+zrPOpCEs3VTEUp1oRMbh+T8IM6WWWxzaka+fH12YlxZOTPCKo 3aHIqE5U1QIqTIUtWX2HPMrP3U9zX165APLRiEkgLb/MGQQkgAmYkQVVZgBiCzaXt5Vi 3y+ygcdQMMJk/EM5mqJUlOQY51hHdbFEA2pVe1HfSxHHzA87IQ/HAFfnxgZD0/a+sqrU 6g7w== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=SRJqoOow; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id j19si2852686pfd.53.2019.08.13.17.44.58; Tue, 13 Aug 2019 17:45:15 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=SRJqoOow; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726688AbfHNAoE (ORCPT + 99 others); Tue, 13 Aug 2019 20:44:04 -0400 Received: from mail-oi1-f194.google.com ([209.85.167.194]:44269 "EHLO mail-oi1-f194.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726515AbfHNAoE (ORCPT ); Tue, 13 Aug 2019 20:44:04 -0400 Received: by mail-oi1-f194.google.com with SMTP id k22so1375809oiw.11 for ; Tue, 13 Aug 2019 17:44:01 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=xTgTghZIlx7I03WklruelXMJsYuL++iO37GrwFIicHk=; b=SRJqoOowrRgCds5aSQx3XU6KOHzcsjNxoJQiJv0fsYy9/FZKa9TtOs5veqLNNA2652 aYIWl8vtRh5+eVNtxly2UIuSHnkZHIcRgfi4pKyTN2HPbt8F+xJ5vdMj8KBrs0q7k0NX IDOodsq/sKE2kSNGZ+x4Wo7ZVLAnMuci9XUOU8YsURZO4SKRyeP5YRtkIsafnz810SOo 1trmOENVtDElTOoH6VwoUZN0wlwRANljAyzcEa+h/q9qbwlFiTVMDmRIYdwCZtBDs6gE lM8vDlz1bs1SoBmv+Nn3r2a+DETMwp+GmTc/gnWxR/PiitYJh/tWgB6iHLAHNleWIH97 rCwQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=xTgTghZIlx7I03WklruelXMJsYuL++iO37GrwFIicHk=; b=dkFHSwY61QbFTiMKOX24ZWGVi/vv4Cp97skqos22WHaww/Z9j2j8KiAJNtzitrBI0p Tv2WzvMZU3FA8/Ep8EYHq4EDKe+YAeL1TmANf7mKmES265WHLqbdiW0QYLek+QVFdPuf voYGI0k4XQq9AiY6KyDr8TmGU06uOw4A6DNGkEqsIbmoV52p3VP0dT+v8QHnbtmOFfq1 hE/IGOX5Qu2ekArOlFPkgFdR4Wa2ddwJXGZ0aVmuKCH69bNzt6rH0+Fgt91n+ezxraWW DRO2xB7t05+9TsMSYeMH7zstP0IdpnFJzufyCTBeaxGoA4vZJRnpIAz9AjNG+TD9La40 qxrQ== X-Gm-Message-State: APjAAAXULZDX7g52ELNwzuRyeh+a0ajf6VVp7/sY2d9wNwZ2W4Fzzloj rAG97w2tGDxzwQM4U61AfNggi1E6BD9rb3UjONPhUQ== X-Received: by 2002:aca:6104:: with SMTP id v4mr3463968oib.172.1565743441057; Tue, 13 Aug 2019 17:44:01 -0700 (PDT) MIME-Version: 1.0 References: <20190807112432.26521-1-sibis@codeaurora.org> <20190807112432.26521-3-sibis@codeaurora.org> <9ea6ab4c-6357-653b-f91c-9b649205f0ab@codeaurora.org> In-Reply-To: <9ea6ab4c-6357-653b-f91c-9b649205f0ab@codeaurora.org> From: Saravana Kannan Date: Tue, 13 Aug 2019 17:43:24 -0700 Message-ID: Subject: Re: [PATCH 2/2] interconnect: qcom: Add OSM L3 interconnect provider support To: Sibi Sankar Cc: Rob Herring , Georgi Djakov , Bjorn Andersson , agross@kernel.org, LKML , "open list:OPEN FIRMWARE AND FLATTENED DEVICE TREE BINDINGS" , linux-arm-msm@vger.kernel.org, Mark Rutland , Evan Green , David Dai Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Aug 8, 2019 at 10:37 AM Sibi Sankar wrote: > > Hey Saravana, > > Thanks for the review! > > On 8/8/19 2:51 AM, Saravana Kannan wrote: > > On Wed, Aug 7, 2019 at 4:24 AM Sibi Sankar wrote: > >> > >> On some Qualcomm SoCs, Operating State Manager (OSM) controls the > >> resources of scaling L3 caches. Add a driver to handle bandwidth > >> requests to OSM L3 from CPU/GPU. > >> > >> Signed-off-by: Sibi Sankar > >> --- > >> drivers/interconnect/qcom/Kconfig | 7 + > >> drivers/interconnect/qcom/Makefile | 2 + > >> drivers/interconnect/qcom/osm-l3.c | 292 +++++++++++++++++++++++++++++ > >> 3 files changed, 301 insertions(+) > >> create mode 100644 drivers/interconnect/qcom/osm-l3.c > >> > >> diff --git a/drivers/interconnect/qcom/Kconfig b/drivers/interconnect/qcom/Kconfig > >> index d5e70ebc24108..f6c2a11a1a2c9 100644 > >> --- a/drivers/interconnect/qcom/Kconfig > >> +++ b/drivers/interconnect/qcom/Kconfig > >> @@ -5,6 +5,13 @@ config INTERCONNECT_QCOM > >> help > >> Support for Qualcomm's Network-on-Chip interconnect hardware. > >> > >> +config INTERCONNECT_QCOM_OSM_L3 > >> + tristate "Qualcomm OSM L3 interconnect driver" > >> + depends on INTERCONNECT_QCOM || COMPILE_TEST > >> + help > >> + Say y here to support the Operating State Manager (OSM) interconnect > >> + driver which controls the scaling of L3 caches on Qualcomm SoCs. > >> + > >> config INTERCONNECT_QCOM_SDM845 > >> tristate "Qualcomm SDM845 interconnect driver" > >> depends on INTERCONNECT_QCOM > >> diff --git a/drivers/interconnect/qcom/Makefile b/drivers/interconnect/qcom/Makefile > >> index 1c1cea690f922..9078af5fed109 100644 > >> --- a/drivers/interconnect/qcom/Makefile > >> +++ b/drivers/interconnect/qcom/Makefile > >> @@ -1,5 +1,7 @@ > >> # SPDX-License-Identifier: GPL-2.0 > >> > >> +icc-osm-l3-objs := osm-l3.o > >> qnoc-sdm845-objs := sdm845.o > >> > >> +obj-$(CONFIG_INTERCONNECT_QCOM_OSM_L3) += icc-osm-l3.o > >> obj-$(CONFIG_INTERCONNECT_QCOM_SDM845) += qnoc-sdm845.o > >> diff --git a/drivers/interconnect/qcom/osm-l3.c b/drivers/interconnect/qcom/osm-l3.c > >> new file mode 100644 > >> index 0000000000000..1e7dfce6f4f9b > >> --- /dev/null > >> +++ b/drivers/interconnect/qcom/osm-l3.c > >> @@ -0,0 +1,292 @@ > >> +// SPDX-License-Identifier: GPL-2.0 > >> +/* > >> + * Copyright (c) 2019, The Linux Foundation. All rights reserved. > >> + * > >> + */ > >> + > >> +#include > >> +#include > >> +#include > >> +#include > >> +#include > >> +#include > >> +#include > >> +#include > >> +#include > >> +#include > >> +#include > >> + > >> +#define LUT_MAX_ENTRIES 40U > >> +#define LUT_SRC GENMASK(31, 30) > >> +#define LUT_L_VAL GENMASK(7, 0) > >> +#define LUT_ROW_SIZE 32 > >> +#define CLK_HW_DIV 2 > >> + > >> +/* Register offsets */ > >> +#define REG_ENABLE 0x0 > >> +#define REG_FREQ_LUT 0x110 > >> +#define REG_PERF_STATE 0x920 > >> + > >> +#define OSM_L3_MAX_LINKS 1 > >> + > >> +#define to_qcom_provider(_provider) \ > >> + container_of(_provider, struct qcom_icc_provider, provider) > >> + > >> +enum { > >> + SDM845_MASTER_OSM_L3_APPS = SLAVE_TCU + 1, > >> + SDM845_MASTER_OSM_L3_GPU, > >> + SDM845_SLAVE_OSM_L3, > >> +}; > >> + > >> +struct qcom_icc_provider { > >> + void __iomem *base; > >> + unsigned int max_state; > >> + unsigned long lut_tables[LUT_MAX_ENTRIES]; > >> + struct icc_provider provider; > >> +}; > >> + > >> +/** > >> + * struct qcom_icc_node - Qualcomm specific interconnect nodes > >> + * @name: the node name used in debugfs > >> + * @links: an array of nodes where we can go next while traversing > >> + * @id: a unique node identifier > >> + * @num_links: the total number of @links > >> + * @buswidth: width of the interconnect between a node and the bus > >> + */ > >> +struct qcom_icc_node { > >> + const char *name; > >> + u16 links[OSM_L3_MAX_LINKS]; > >> + u16 id; > >> + u16 num_links; > >> + u16 buswidth; > >> +}; > >> + > >> +struct qcom_icc_desc { > >> + struct qcom_icc_node **nodes; > >> + size_t num_nodes; > >> +}; > >> + > >> +#define DEFINE_QNODE(_name, _id, _buswidth, ...) \ > >> + static struct qcom_icc_node _name = { \ > >> + .name = #_name, \ > >> + .id = _id, \ > >> + .buswidth = _buswidth, \ > >> + .num_links = ARRAY_SIZE(((int[]){ __VA_ARGS__ })), \ > >> + .links = { __VA_ARGS__ }, \ > >> + } > >> + > >> +DEFINE_QNODE(osm_apps_l3, SDM845_MASTER_OSM_L3_APPS, 16, SDM845_SLAVE_OSM_L3); > >> +DEFINE_QNODE(osm_gpu_l3, SDM845_MASTER_OSM_L3_GPU, 16, SDM845_SLAVE_OSM_L3); > >> +DEFINE_QNODE(osm_l3, SDM845_SLAVE_OSM_L3, 16); > >> + > >> +static struct qcom_icc_node *sdm845_osm_l3_nodes[] = { > >> + [MASTER_OSM_L3_APPS] = &osm_apps_l3, > >> + [MASTER_OSM_L3_GPU] = &osm_gpu_l3, > >> + [SLAVE_OSM_L3] = &osm_l3, > >> +}; > >> + > >> +static struct qcom_icc_desc sdm845_osm_l3 = { > >> + .nodes = sdm845_osm_l3_nodes, > >> + .num_nodes = ARRAY_SIZE(sdm845_osm_l3_nodes), > >> +}; > >> + > >> +static int qcom_icc_aggregate(struct icc_node *node, u32 avg_bw, > >> + u32 peak_bw, u32 *agg_avg, u32 *agg_peak) > >> +{ > >> + *agg_avg += avg_bw; > >> + *agg_peak = max_t(u32, *agg_peak, peak_bw); > >> + > >> + return 0; > >> +} > >> + > >> +static int qcom_icc_set(struct icc_node *src, struct icc_node *dst) > >> +{ > >> + struct icc_provider *provider; > >> + struct qcom_icc_provider *qp; > >> + struct qcom_icc_node *qn; > >> + struct icc_node *n; > >> + unsigned int index; > >> + u32 agg_peak = 0; > >> + u32 agg_avg = 0; > >> + u64 rate; > >> + > >> + qn = src->data; > >> + provider = src->provider; > >> + qp = to_qcom_provider(provider); > >> + > >> + list_for_each_entry(n, &provider->nodes, node_list) > >> + qcom_icc_aggregate(n, n->avg_bw, n->peak_bw, > >> + &agg_avg, &agg_peak); > >> + > >> + rate = max(agg_avg, agg_peak); > >> + rate = icc_units_to_bps(rate); > >> + do_div(rate, qn->buswidth); > >> + > >> + for (index = 0; index < qp->max_state; index++) { > >> + if (qp->lut_tables[index] >= rate) > >> + break; > >> + } > >> + > >> + writel_relaxed(index, qp->base + REG_PERF_STATE); > >> + > >> + return 0; > >> +} > >> + > >> +static int qcom_osm_l3_probe(struct platform_device *pdev) > >> +{ > >> + u32 info, src, lval, i, prev_freq = 0, freq; > >> + static unsigned long hw_rate, xo_rate; > >> + const struct qcom_icc_desc *desc; > >> + struct icc_onecell_data *data; > >> + struct icc_provider *provider; > >> + struct qcom_icc_node **qnodes; > >> + struct qcom_icc_provider *qp; > >> + struct icc_node *node; > >> + size_t num_nodes; > >> + struct clk *clk; > >> + int ret; > >> + > >> + clk = clk_get(&pdev->dev, "xo"); > >> + if (IS_ERR(clk)) > >> + return PTR_ERR(clk); > >> + > >> + xo_rate = clk_get_rate(clk); > >> + clk_put(clk); > >> + > >> + clk = clk_get(&pdev->dev, "alternate"); > >> + if (IS_ERR(clk)) > >> + return PTR_ERR(clk); > >> + > >> + hw_rate = clk_get_rate(clk) / CLK_HW_DIV; > >> + clk_put(clk); > >> + > >> + qp = devm_kzalloc(&pdev->dev, sizeof(*qp), GFP_KERNEL); > >> + if (!qp) > >> + return -ENOMEM; > >> + > >> + qp->base = devm_platform_ioremap_resource(pdev, 0); > >> + if (IS_ERR(qp->base)) > >> + return PTR_ERR(qp->base); > >> + > >> + /* HW should be in enabled state to proceed */ > >> + if (!(readl_relaxed(qp->base + REG_ENABLE) & 0x1)) { > >> + dev_err(&pdev->dev, "error hardware not enabled\n"); > >> + return -ENODEV; > >> + } > >> + > >> + for (i = 0; i < LUT_MAX_ENTRIES; i++) { > >> + info = readl_relaxed(qp->base + REG_FREQ_LUT + > >> + i * LUT_ROW_SIZE); > >> + src = FIELD_GET(LUT_SRC, info); > >> + lval = FIELD_GET(LUT_L_VAL, info); > >> + if (src) > >> + freq = xo_rate * lval; > >> + else > >> + freq = hw_rate; > >> + > >> + /* > >> + * Two of the same frequencies with the same core counts means > >> + * end of table > >> + */ > >> + if (i > 0 && prev_freq == freq) > >> + break; > >> + > >> + qp->lut_tables[i] = freq; > >> + prev_freq = freq; > >> + } > >> + qp->max_state = i; > >> + > >> + desc = of_device_get_match_data(&pdev->dev); > >> + if (!desc) > >> + return -EINVAL; > >> + > >> + qnodes = desc->nodes; > >> + num_nodes = desc->num_nodes; > >> + > >> + data = devm_kcalloc(&pdev->dev, num_nodes, sizeof(*node), GFP_KERNEL); > >> + if (!data) > >> + return -ENOMEM; > >> + > >> + provider = &qp->provider; > >> + provider->dev = &pdev->dev; > >> + provider->set = qcom_icc_set; > >> + provider->aggregate = qcom_icc_aggregate; > >> + provider->xlate = of_icc_xlate_onecell; > >> + INIT_LIST_HEAD(&provider->nodes); > >> + provider->data = data; > >> + > >> + ret = icc_provider_add(provider); > >> + if (ret) { > >> + dev_err(&pdev->dev, "error adding interconnect provider\n"); > >> + return ret; > >> + } > >> + > >> + for (i = 0; i < num_nodes; i++) { > >> + size_t j; > >> + > >> + node = icc_node_create(qnodes[i]->id); > >> + if (IS_ERR(node)) { > >> + ret = PTR_ERR(node); > >> + goto err; > >> + } > >> + > >> + node->name = qnodes[i]->name; > >> + node->data = qnodes[i]; > >> + icc_node_add(node, provider); > >> + > >> + dev_dbg(&pdev->dev, "registered node %p %s %d\n", node, > >> + qnodes[i]->name, node->id); > >> + > >> + /* populate links */ > >> + for (j = 0; j < qnodes[i]->num_links; j++) > >> + icc_link_create(node, qnodes[i]->links[j]); > >> + > >> + data->nodes[i] = node; > >> + } > >> + data->num_nodes = num_nodes; > >> + > >> + platform_set_drvdata(pdev, qp); > >> + > >> + return ret; > >> +err: > >> + list_for_each_entry(node, &provider->nodes, node_list) { > >> + icc_node_del(node); > >> + icc_node_destroy(node->id); > >> + } > >> + > >> + icc_provider_del(provider); > >> + return ret; > >> +} > >> + > >> +static int qcom_osm_l3_remove(struct platform_device *pdev) > >> +{ > >> + struct qcom_icc_provider *qp = platform_get_drvdata(pdev); > >> + struct icc_provider *provider = &qp->provider; > >> + struct icc_node *n; > >> + > >> + list_for_each_entry(n, &provider->nodes, node_list) { > >> + icc_node_del(n); > >> + icc_node_destroy(n->id); > >> + } > >> + > >> + return icc_provider_del(provider); > >> +} > >> + > >> +static const struct of_device_id osm_l3_of_match[] = { > >> + { .compatible = "qcom,sdm845-osm-l3", .data = &sdm845_osm_l3 }, > >> + { }, > >> +}; > >> +MODULE_DEVICE_TABLE(of, osm_l3_of_match); > >> + > >> +static struct platform_driver osm_l3_driver = { > >> + .probe = qcom_osm_l3_probe, > >> + .remove = qcom_osm_l3_remove, > >> + .driver = { > >> + .name = "osm-l3", > >> + .of_match_table = osm_l3_of_match, > >> + }, > >> +}; > >> +module_platform_driver(osm_l3_driver); > >> + > >> +MODULE_DESCRIPTION("Qualcomm OSM L3 interconnect driver"); > >> +MODULE_LICENSE("GPL v2"); > > > > Did a quick scan of the code and it's not clear how you connect the L3 > > interconnect provider to the rest of the interconnect. I don't see any > > DT or code references to the rest of the interconnects in the system. > > If GPU is making a bandwidth request all the way to L3, how do you > > make sure the interconnects between GPU and L3 also scale up properly. > > For SDM845 OSM L3 provider the icc nodes endpoints are isolated from > rsc icc node endpoints i.e GPU would need to vote on this path in > addition to voting for DDR. On future SoCs if the need to scale > interconnect between GPU rsc nodes along with the OSM l3 nodes arises, > it can be trivially extended by linking the osm icc nodes with global > icc node ids of the gpu rsc nodes. Sorry, I'm not sure we are talking about the same thing. To keep the discussion focused, let's completely ignore GPU's DDR needs for now. Talking about GPU <-> L3 traffic, unless the GPU is directly connected to the L3 with no additional interconnects in between (unlikely), this L3 provider need to connect the L3 icc node with the rest of the interconnect nodes. Otherwise, even if the L3 itself is provisioned to handle, say 10 GB/s, the interconnects connecting the GPU to L3 might be left at 1 GB/s. So, can we please connect this node to the rest of the interconnect nodes in SDM845? Thanks, Saravana