Received: by 2002:ac0:e34a:0:0:0:0:0 with SMTP id g10csp406837imn; Thu, 28 Jul 2022 04:33:33 -0700 (PDT) X-Google-Smtp-Source: AGRyM1slESXye5uGUs6M6jgVLgiNgzj5e7di0ZhNAEHL4QA61CBB10Fq9p4HExEBJwWu2le9kyEh X-Received: by 2002:a17:90b:17c6:b0:1f2:e246:6712 with SMTP id me6-20020a17090b17c600b001f2e2466712mr9736894pjb.63.1659008013039; Thu, 28 Jul 2022 04:33:33 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1659008013; cv=none; d=google.com; s=arc-20160816; b=tm856Yv9lAjTmen23vwYb59cWp/r9VK84mIenDAHgDgg8G2/6l2PZqvlUkB2AeTov3 0KPrkeDeMR/wJUstZ4cLrrWmWwWEHNcqIhuxyfqN2tkY5TllKC83tskLJujKNfBT7sJB Z56WMB3lfAOTYtzPE/sAmK9eZcJK43rzGDH0/H9J9CAnwwkhMvnolhss6t3iZTvtYXdV bFTC0l2tSfgybwc5SuA+aF8IEBdaNJVfHxzz6Nd5cDFXg9SDjY2R48SiU58XQckZSRR6 512ttvQwDCvrN9hEPpMeTy/UaqEYCILWMp76OBt22rZQkvNOPgp/TdfVkIryMstmtPqT k9Sg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:to:references:message-id :content-transfer-encoding:cc:date:in-reply-to:from:subject :mime-version:dkim-signature; bh=n3PmPErZ3qz515kkpqX3FC9Qmz7SVOo0xd78m90Zoi4=; b=VoH+aX9omSX9dE7EKyd1fpfKGmWhQGJutLOy2RRc/Kzz6hQwlqebk52sDOANVFPgVQ Q8n/9+dkvEyo5qLBsrcOk7j/ZM/drBBxzfXe37umYVJFa7oZtJBOI1Dx8q/mpGInFiza FNG/8P2C3DYmg6f/bLNSxXf3TTsj39pQ5qGl549Jl5gk3ZS7d+bLQiVPiotcsPAQ3+f8 0oYLjoRoeDPxR+Crcv/kcj2mfvKH6Pxvtnw9wH2dzXuoGbEFGoEJun359aDpRg7qYdk/ uPIYYufK9j9sGciuWdOHJCfJ3c+xuDtAEn+UZl8Uuna1jCHyVAOh5wl9QpPHdDyL43BO Vc9w== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20210112 header.b=C2csZZ9X; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id o13-20020a056a0015cd00b00525b44e124bsi415670pfu.259.2022.07.28.04.33.18; Thu, 28 Jul 2022 04:33:33 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20210112 header.b=C2csZZ9X; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236656AbiG1LD5 (ORCPT + 99 others); Thu, 28 Jul 2022 07:03:57 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33448 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234566AbiG1LDx (ORCPT ); Thu, 28 Jul 2022 07:03:53 -0400 Received: from mail-wm1-x334.google.com (mail-wm1-x334.google.com [IPv6:2a00:1450:4864:20::334]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 830E4664D7; Thu, 28 Jul 2022 04:03:50 -0700 (PDT) Received: by mail-wm1-x334.google.com with SMTP id b6so756988wmq.5; Thu, 28 Jul 2022 04:03:50 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=mime-version:subject:from:in-reply-to:date:cc :content-transfer-encoding:message-id:references:to; bh=n3PmPErZ3qz515kkpqX3FC9Qmz7SVOo0xd78m90Zoi4=; b=C2csZZ9XG2SLPWemOFjJyQVMVsPCI1UxaA+hNyVd/MWMBhh7Y22upv99ftjTN3ltLD +BKmCPnG2WRZUqDgLimOUFPzCrFpK21VGZreM6Jasf+peiCG0rJGpyOBbpnWzvjALLqt hgX8641+ThwuNx9POQaYt7YkggIjFkPDXY+wrrXuQWZc60KSHUmcGu+/M6TFjU4P+zsf RVoB225sDTw0PXuwQyLMXl+R+9tbLfk4oNVHr/CWM35xHvCZbySqSCIjs2KCYzKCCjf0 h49uu0Ch81GrKtaAVsvXpLSDsxRKxWT0LD+AIobnlXCssULMOsycbbUfMzvQnbuHjwUp UDQQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:mime-version:subject:from:in-reply-to:date:cc :content-transfer-encoding:message-id:references:to; bh=n3PmPErZ3qz515kkpqX3FC9Qmz7SVOo0xd78m90Zoi4=; b=4HIdbjGY2VGZX1AjdBRnqbV3Bw3Gbv0/U9D4mUHMvznzxIxl1cB8LiwvM3uwQiBrvJ 5alfBTVEQdQuq/wakiaQO7yemOzyijZ8sMNVP5kbNu0D1zsOlaurepg/p8TP+3CwAekb i5MZEWSdUeL/qlxWAKcqTzV06mVtxT4cJGDK5aH4RUYkWjDLMR+U5f8K7XtFJSckxrsx 1ATFstSfiuyH8TjawFrIJm4ixUdEje+DMOFfT7L0I9TGaYvkrx2aYoMm23pvKYNMm6w7 qH+To5jzO+MikbaN57PA/X4pDlVULc1xnRYAEVR8R/2VtKr9ZXwoQ5J5AS00QoMzVoph pKvA== X-Gm-Message-State: AJIora8TtDzfLuqBISRqgSOvTsafLEu5jM68kYHegwl1o1XDL/ZRHhOJ UsGJDHvtJ01jvVTbmoO9orekGoaWRkaKhgru X-Received: by 2002:a05:600c:3d93:b0:3a3:3a93:fb16 with SMTP id bi19-20020a05600c3d9300b003a33a93fb16mr6393409wmb.190.1659006228676; Thu, 28 Jul 2022 04:03:48 -0700 (PDT) Received: from smtpclient.apple ([87.200.95.144]) by smtp.gmail.com with ESMTPSA id v62-20020a1cac41000000b003a31f1edfa7sm5276765wme.41.2022.07.28.04.03.45 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Thu, 28 Jul 2022 04:03:48 -0700 (PDT) Content-Type: text/plain; charset=us-ascii Mime-Version: 1.0 (Mac OS X Mail 16.0 \(3696.100.31\)) Subject: Re: [PATCH v2 1/4] perf/amlogic: Add support for Amlogic meson G12 SoC DDR PMU driver From: Christian Hewitt In-Reply-To: <20220726230329.2844101-1-jiucheng.xu@amlogic.com> Date: Thu, 28 Jul 2022 15:03:43 +0400 Cc: LKML , linux-arm-kernel@lists.infradead.org, "open list:ARM/Amlogic Meson..." , devicetree , Rob Herring , Krzysztof Kozlowski , Will Deacon , Mark Rutland , Neil Armstrong , Kevin Hilman , Jerome Brunet , Martin Blumenstingl , Chris Healy Content-Transfer-Encoding: quoted-printable Message-Id: References: <20220726230329.2844101-1-jiucheng.xu@amlogic.com> To: Jiucheng Xu X-Mailer: Apple Mail (2.3696.100.31) X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,FREEMAIL_FROM, RCVD_IN_DNSWL_NONE,SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org > On 27 Jul 2022, at 3:03 am, Jiucheng Xu = wrote: >=20 > This patch adds support Amlogic meson G12 series SoC > DDR bandwidth PMU driver framework and interfaces. >=20 > The PMU not only can monitor the total DDR bandwidth, > but also the bandwidth which is from individual IP module. >=20 > Example usage: >=20 > $ perf stat -a -e aml_ddr_bw/total_rw_bytes/ -I 1000 sleep 10 >=20 > - or - >=20 > $ perf stat -a -e \ > aml_ddr_bw/total_rw_bytes/,\ > aml_ddr_bw/chan_1_rw_bytes,arm=3D1/ -I 1000 \ > sleep 10 >=20 > g12 SoC support 4 channels to monitor DDR bandwidth > simultaneously. Each channel can monitor up to 4 IP modules > simultaneously. >=20 > For Instance, If you want to get the sum of DDR bandwidth > from CPU, GPU, USB3.0 and VDEC. You can use the following > command parameters to display. >=20 > $ perf stat -a -e \ > aml_ddr_bw/chan_2_rw_bytes,arm=3D1,gpu=3D1,usb3_0=3D1,nna=3D1/ -I = 1000 \ > sleep 10 >=20 > Other events are supported, and advertised via perf list. >=20 > Signed-off-by: Jiucheng Xu > Reported-by: kernel test robot > --- > Changes v1 -> v2: > - Remove inline to let GCC make the decisions > - Remove spinlock > - Remove ddr_cnt_accumulate() > - Remove the message which only indicate a bug > - Remove all dev_warn() message > - Use hweight64() helper instead of whole loop > - Remove setting of hwc > - Use for_each_set_bit() helper for bit loop > - Use sysfs_emit() in sysfs show > - Remove checking for bugs > - Replace irq_set_affinity_hint() to irq_set_affinity() > - Remove #ifdef CONFIG_OF > - Use devm_platform_ioremap_resource() instead of > platform_get_resource()&ioremap() > - Use platform_get_irq() instead of platform_get_resource()&ioremap() > - Replace IRQF_SHARED to IRQF_NOBALANCING > - Remove meaningless log like "init ok" > - Use compatible instead of creating new property to distinguish > different platform. > - Use the is_visible callback to avoid exposing unsupported fmt_attr > - Use module_platform_driver_probe() instead of module_init/exit > --- > MAINTAINERS | 7 + > drivers/perf/Kconfig | 2 + > drivers/perf/Makefile | 1 + > drivers/perf/amlogic/Kconfig | 10 + > drivers/perf/amlogic/Makefile | 5 + > drivers/perf/amlogic/aml_ddr_pmu_core.c | 551 ++++++++++++++++++++++++ > drivers/perf/amlogic/aml_ddr_pmu_g12.c | 388 +++++++++++++++++ > include/soc/amlogic/aml_ddr_pmu.h | 76 ++++ > 8 files changed, 1040 insertions(+) > create mode 100644 drivers/perf/amlogic/Kconfig > create mode 100644 drivers/perf/amlogic/Makefile > create mode 100644 drivers/perf/amlogic/aml_ddr_pmu_core.c > create mode 100644 drivers/perf/amlogic/aml_ddr_pmu_g12.c > create mode 100644 include/soc/amlogic/aml_ddr_pmu.h >=20 > diff --git a/MAINTAINERS b/MAINTAINERS > index f679152bdbad..cb6ee59a4f44 100644 > --- a/MAINTAINERS > +++ b/MAINTAINERS > @@ -1050,6 +1050,13 @@ S: Maintained > F: Documentation/hid/amd-sfh* > F: drivers/hid/amd-sfh-hid/ >=20 > +AMLOGIC DDR PMU DRIVER > +M: Jiucheng Xu > +S: Supported > +W: http://www.amlogic.com > +F: drivers/perf/amlogic/ > +F: include/soc/amlogic/ > + > AMPHION VPU CODEC V4L2 DRIVER > M: Ming Qian > M: Shijie Qin > diff --git a/drivers/perf/Kconfig b/drivers/perf/Kconfig > index 1e2d69453771..68200c798128 100644 > --- a/drivers/perf/Kconfig > +++ b/drivers/perf/Kconfig > @@ -192,4 +192,6 @@ config MARVELL_CN10K_DDR_PMU > Enable perf support for Marvell DDR Performance monitoring > event on CN10K platform. >=20 > +source "drivers/perf/amlogic/Kconfig" > + > endmenu > diff --git a/drivers/perf/Makefile b/drivers/perf/Makefile > index 57a279c61df5..ef82b27e36d0 100644 > --- a/drivers/perf/Makefile > +++ b/drivers/perf/Makefile > @@ -1,4 +1,5 @@ > # SPDX-License-Identifier: GPL-2.0 > +obj-$(CONFIG_AML_DDR_PMU) +=3D amlogic/ Upstream kernel Amlogic drivers normally use MESON naming, so the config item would be e.g. CONFIG_MESON_DDR_PMU. > obj-$(CONFIG_ARM_CCI_PMU) +=3D arm-cci.o > obj-$(CONFIG_ARM_CCN) +=3D arm-ccn.o > obj-$(CONFIG_ARM_CMN) +=3D arm-cmn.o > diff --git a/drivers/perf/amlogic/Kconfig = b/drivers/perf/amlogic/Kconfig > new file mode 100644 > index 000000000000..0e421c15c8c7 > --- /dev/null > +++ b/drivers/perf/amlogic/Kconfig > @@ -0,0 +1,10 @@ > +# SPDX-License-Identifier: GPL-2.0-only > +config AML_DDR_PMU > + tristate "Amlogic DDR Bandwidth Performance Monitor" > + depends on ARCH_MESON || COMPILE_TEST > + help > + Provides support for the DDR performance monitor > + in Amlogic SoCs, which can give information about > + memory throughput and other related events. It > + supports multiple channels to monitor the memory > + bandwidth simultaneously. > diff --git a/drivers/perf/amlogic/Makefile = b/drivers/perf/amlogic/Makefile > new file mode 100644 > index 000000000000..874b885aa5cc > --- /dev/null > +++ b/drivers/perf/amlogic/Makefile > @@ -0,0 +1,5 @@ > +# SPDX-License-Identifier: GPL-2.0-only > + > +obj-$(CONFIG_AML_DDR_PMU) +=3D aml_ddr_pmu.o > + > +aml_ddr_pmu-y :=3D aml_ddr_pmu_core.o aml_ddr_pmu_g12.o I would similarly expect to see meson_ filenames to reflect the MESON driver name, e.g. meson_ddr_pmu > diff --git a/drivers/perf/amlogic/aml_ddr_pmu_core.c = b/drivers/perf/amlogic/aml_ddr_pmu_core.c > new file mode 100644 > index 000000000000..4e2f7f0d9af7 > --- /dev/null > +++ b/drivers/perf/amlogic/aml_ddr_pmu_core.c > @@ -0,0 +1,551 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * Copyright (c) 2022 Amlogic, Inc. All rights reserved. > + */ > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > + > +#include > + > +#define DDR_PERF_DEV_NAME "aml_ddr_bw" > +#define MAX_AXI_PORTS_OF_CHANNEL 4 /* A DMC channel can = monitor max 4 axi ports */ > + > +#define to_ddr_pmu(p) container_of(p, struct ddr_pmu, = pmu) > + > +#define hw_info_to_pmu(p) container_of(p, struct ddr_pmu, info) > + > +static void dmc_pmu_enable(struct ddr_pmu *pmu) > +{ > + if (!pmu->pmu_enabled) > + pmu->info.ops->enable(&pmu->info); > + > + pmu->pmu_enabled =3D true; > +} > + > +static void dmc_pmu_disable(struct ddr_pmu *pmu) > +{ > + if (pmu->pmu_enabled) > + pmu->info.ops->disable(&pmu->info); > + > + pmu->pmu_enabled =3D false; > +} > + > +static void aml_ddr_config_axi_id(struct ddr_pmu *pmu, int axi_id, = int chann) And I would similarly expect to see meson_ structures to reflect the MESON driver name, e.g. meson_ddr_config_axi_id > +{ > + pmu->info.ops->config_axi_id(&pmu->info, axi_id, chann); > +} > + > +static void aml_ddr_set_filter(struct perf_event *event, u8 axi_id) > +{ > + struct ddr_pmu *pmu =3D to_ddr_pmu(event->pmu); > + int chann; > + > + if (event->attr.config > ALL_CHAN_COUNTER_ID && > + event->attr.config < COUNTER_MAX_ID) { > + chann =3D event->attr.config - CHAN1_COUNTER_ID; > + > + aml_ddr_config_axi_id(pmu, axi_id, chann); > + } > +} > + > +static void ddr_cnt_addition(struct dmc_counter *sum, > + struct dmc_counter *add1, > + struct dmc_counter *add2, > + int chann_nr) > +{ > + int i; > + u64 cnt1, cnt2; > + > + sum->all_cnt =3D add1->all_cnt + add2->all_cnt; > + sum->all_req =3D add1->all_req + add2->all_req; > + for (i =3D 0; i < chann_nr; i++) { > + cnt1 =3D add1->channel_cnt[i]; > + cnt2 =3D add2->channel_cnt[i]; > + > + sum->channel_cnt[i] =3D cnt1 + cnt2; > + } > +} > + > +static void aml_ddr_perf_event_update(struct perf_event *event) > +{ > + struct ddr_pmu *pmu =3D to_ddr_pmu(event->pmu); > + u64 new_raw_count =3D 0; > + struct dmc_counter dc =3D {0}, sum_dc =3D {0}; > + int idx; > + > + /* get the remain counters in register. */ > + pmu->info.ops->get_counters(&pmu->info, &dc); > + > + ddr_cnt_addition(&sum_dc, &pmu->counters, &dc, = pmu->info.chann_nr); > + > + switch (event->attr.config) { > + case ALL_CHAN_COUNTER_ID: > + new_raw_count =3D sum_dc.all_cnt; > + break; > + case CHAN1_COUNTER_ID: > + case CHAN2_COUNTER_ID: > + case CHAN3_COUNTER_ID: > + case CHAN4_COUNTER_ID: > + case CHAN5_COUNTER_ID: > + case CHAN6_COUNTER_ID: > + case CHAN7_COUNTER_ID: > + case CHAN8_COUNTER_ID: > + idx =3D event->attr.config - CHAN1_COUNTER_ID; > + new_raw_count =3D sum_dc.channel_cnt[idx]; > + break; > + } > + > + local64_set(&event->count, new_raw_count); > +} > + > +static int aml_ddr_perf_event_init(struct perf_event *event) > +{ > + struct ddr_pmu *pmu =3D to_ddr_pmu(event->pmu); > + u64 config1 =3D event->attr.config1; > + > + if (event->attr.type !=3D event->pmu->type) > + return -ENOENT; > + > + if (is_sampling_event(event) || event->attach_state & = PERF_ATTACH_TASK) > + return -EOPNOTSUPP; > + > + if (event->cpu < 0) > + return -EOPNOTSUPP; > + > + /* check if the number of parameters is too much */ > + if (event->attr.config !=3D ALL_CHAN_COUNTER_ID && > + hweight64(config1) > MAX_AXI_PORTS_OF_CHANNEL) > + return -EOPNOTSUPP; > + > + event->cpu =3D pmu->cpu; > + > + return 0; > +} > + > +static void aml_ddr_perf_event_start(struct perf_event *event, int = flags) > +{ > + struct ddr_pmu *pmu =3D to_ddr_pmu(event->pmu); > + > + memset(&pmu->counters, 0, sizeof(pmu->counters)); > + dmc_pmu_enable(pmu); > +} > + > +static int aml_ddr_perf_event_add(struct perf_event *event, int = flags) > +{ > + u64 config1 =3D event->attr.config1; > + int i; > + > + for_each_set_bit(i, (const unsigned long *)&config1, = sizeof(config1)) > + aml_ddr_set_filter(event, i); > + > + if (flags & PERF_EF_START) > + aml_ddr_perf_event_start(event, flags); > + > + return 0; > +} > + > +static void aml_ddr_perf_event_stop(struct perf_event *event, int = flags) > +{ > + struct ddr_pmu *pmu =3D to_ddr_pmu(event->pmu); > + > + if (flags & PERF_EF_UPDATE) > + aml_ddr_perf_event_update(event); > + > + dmc_pmu_disable(pmu); > +} > + > +static void aml_ddr_perf_event_del(struct perf_event *event, int = flags) > +{ > + aml_ddr_perf_event_stop(event, PERF_EF_UPDATE); > +} > + > +static ssize_t aml_ddr_perf_cpumask_show(struct device *dev, > + struct device_attribute *attr, > + char *buf) > +{ > + struct ddr_pmu *pmu =3D dev_get_drvdata(dev); > + > + return cpumap_print_to_pagebuf(true, buf, cpumask_of(pmu->cpu)); > +} > + > +static struct device_attribute aml_ddr_perf_cpumask_attr =3D > +__ATTR(cpumask, 0444, aml_ddr_perf_cpumask_show, NULL); > + > +static struct attribute *aml_ddr_perf_cpumask_attrs[] =3D { > + &aml_ddr_perf_cpumask_attr.attr, > + NULL, > +}; > + > +static const struct attribute_group ddr_perf_cpumask_attr_group =3D { > + .attrs =3D aml_ddr_perf_cpumask_attrs, > +}; > + > +static ssize_t > +pmu_event_show(struct device *dev, struct device_attribute *attr, > + char *page) > +{ > + struct perf_pmu_events_attr *pmu_attr; > + > + pmu_attr =3D container_of(attr, struct perf_pmu_events_attr, = attr); > + return sysfs_emit(page, "event=3D0x%02llx\n", pmu_attr->id); > +} > + > +static ssize_t > +event_show_unit(struct device *dev, struct device_attribute *attr, > + char *page) > +{ > + struct perf_pmu_events_attr *pmu_attr; > + > + pmu_attr =3D container_of(attr, struct perf_pmu_events_attr, = attr); > + return sysfs_emit(page, "MB\n"); > +} > + > +static ssize_t > +event_show_scale(struct device *dev, struct device_attribute *attr, > + char *page) > +{ > + struct perf_pmu_events_attr *pmu_attr; > + > + pmu_attr =3D container_of(attr, struct perf_pmu_events_attr, = attr); > + > + /* one count =3D 16byte =3D 1.52587890625e-05 MB */ > + return sysfs_emit(page, "1.52587890625e-05\n"); > +} > + > +#define AML_DDR_PMU_EVENT_ATTR(_name, _id) = \ > +{ = \ > + .attr =3D __ATTR(_name, 0444, pmu_event_show, NULL), = \ > + .id =3D _id, = \ > +} > + > +#define AML_DDR_PMU_EVENT_UNIT_ATTR(_name) = \ > + __ATTR(_name.unit, 0444, event_show_unit, NULL) > + > +#define AML_DDR_PMU_EVENT_SCALE_ATTR(_name) = \ > + __ATTR(_name.scale, 0444, event_show_scale, NULL) > + > +static struct device_attribute event_unit_attrs[] =3D { > + AML_DDR_PMU_EVENT_UNIT_ATTR(total_rw_bytes), > + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_1_rw_bytes), > + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_2_rw_bytes), > + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_3_rw_bytes), > + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_4_rw_bytes), > + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_5_rw_bytes), > + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_6_rw_bytes), > + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_7_rw_bytes), > + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_8_rw_bytes), > +}; > + > +static struct device_attribute event_scale_attrs[] =3D { > + AML_DDR_PMU_EVENT_SCALE_ATTR(total_rw_bytes), > + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_1_rw_bytes), > + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_2_rw_bytes), > + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_3_rw_bytes), > + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_4_rw_bytes), > + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_5_rw_bytes), > + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_6_rw_bytes), > + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_7_rw_bytes), > + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_8_rw_bytes), > +}; > + > +static struct perf_pmu_events_attr event_attrs[] =3D { > + AML_DDR_PMU_EVENT_ATTR(total_rw_bytes, ALL_CHAN_COUNTER_ID), > + AML_DDR_PMU_EVENT_ATTR(chan_1_rw_bytes, CHAN1_COUNTER_ID), > + AML_DDR_PMU_EVENT_ATTR(chan_2_rw_bytes, CHAN2_COUNTER_ID), > + AML_DDR_PMU_EVENT_ATTR(chan_3_rw_bytes, CHAN3_COUNTER_ID), > + AML_DDR_PMU_EVENT_ATTR(chan_4_rw_bytes, CHAN4_COUNTER_ID), > + AML_DDR_PMU_EVENT_ATTR(chan_5_rw_bytes, CHAN5_COUNTER_ID), > + AML_DDR_PMU_EVENT_ATTR(chan_6_rw_bytes, CHAN6_COUNTER_ID), > + AML_DDR_PMU_EVENT_ATTR(chan_7_rw_bytes, CHAN7_COUNTER_ID), > + AML_DDR_PMU_EVENT_ATTR(chan_8_rw_bytes, CHAN8_COUNTER_ID), > +}; > + > +/* three attrs are combined an event */ > +static struct attribute *ddr_perf_events_attrs[COUNTER_MAX_ID * 3]; > + > +static struct attribute_group ddr_perf_events_attr_group =3D { > + .name =3D "events", > + .attrs =3D ddr_perf_events_attrs, > +}; > + > +/* an attr represents axi id, the max id is less than 256 */ > +static struct attribute *ddr_perf_format_attrs[256]; > + > +static umode_t aml_ddr_perf_format_attr_visible(struct kobject *kobj, > + struct attribute *attr, > + int n) > +{ > + return attr->mode; > +} > + > +static struct attribute_group ddr_perf_format_attr_group =3D { > + .name =3D "format", > + .attrs =3D ddr_perf_format_attrs, > + .is_visible =3D aml_ddr_perf_format_attr_visible, > +}; > + > +static ssize_t aml_ddr_perf_identifier_show(struct device *dev, > + struct device_attribute = *attr, > + char *page) > +{ > + struct ddr_pmu *pmu =3D dev_get_drvdata(dev); > + > + return sysfs_emit(page, "%s\n", pmu->name); > +} > + > +static struct device_attribute aml_ddr_perf_identifier_attr =3D > +__ATTR(identifier, 0444, aml_ddr_perf_identifier_show, NULL); > + > +static struct attribute *aml_ddr_perf_identifier_attrs[] =3D { > + &aml_ddr_perf_identifier_attr.attr, > + NULL, > +}; > + > +static const struct attribute_group ddr_perf_identifier_attr_group =3D = { > + .attrs =3D aml_ddr_perf_identifier_attrs, > +}; > + > +static const struct attribute_group *attr_groups[] =3D { > + &ddr_perf_events_attr_group, > + &ddr_perf_format_attr_group, > + &ddr_perf_cpumask_attr_group, > + &ddr_perf_identifier_attr_group, > + NULL, > +}; > + > +static irqreturn_t dmc_irq_handler(int irq, void *dev_id) > +{ > + struct dmc_hw_info *info =3D dev_id; > + struct ddr_pmu *pmu; > + struct dmc_counter counters, *sum_cnter; > + int i; > + > + pmu =3D hw_info_to_pmu(info); > + > + if (info->ops->irq_handler(info, &counters) !=3D 0) > + goto out; > + > + sum_cnter =3D &pmu->counters; > + sum_cnter->all_cnt +=3D counters.all_cnt; > + sum_cnter->all_req +=3D counters.all_req; > + > + for (i =3D 0; i < pmu->info.chann_nr; i++) > + sum_cnter->channel_cnt[i] +=3D counters.channel_cnt[i]; > + > + if (pmu->pmu_enabled) > + /* > + * the timer interrupt only supprt > + * one shot mode, we have to re-enable > + * it in ISR to support continue mode. > + */ > + info->ops->enable(info); > + > + dev_dbg(pmu->dev, "counts: %llu %llu %llu, %llu, %llu, %llu\t\t" > + "sum: %llu %llu %llu, %llu, %llu, %llu\n", > + counters.all_req, > + counters.all_cnt, > + counters.channel_cnt[0], > + counters.channel_cnt[1], > + counters.channel_cnt[2], > + counters.channel_cnt[3], > + > + pmu->counters.all_req, > + pmu->counters.all_cnt, > + pmu->counters.channel_cnt[0], > + pmu->counters.channel_cnt[1], > + pmu->counters.channel_cnt[2], > + pmu->counters.channel_cnt[3]); > +out: > + return IRQ_HANDLED; > +} > + > +static int ddr_perf_offline_cpu(unsigned int cpu, struct hlist_node = *node) > +{ > + struct ddr_pmu *pmu =3D hlist_entry_safe(node, struct ddr_pmu, = node); > + int target; > + > + if (cpu !=3D pmu->cpu) > + return 0; > + > + target =3D cpumask_any_but(cpu_online_mask, cpu); > + if (target >=3D nr_cpu_ids) > + return 0; > + > + perf_pmu_migrate_context(&pmu->pmu, cpu, target); > + pmu->cpu =3D target; > + > + WARN_ON(irq_set_affinity(pmu->info.irq_num, = cpumask_of(pmu->cpu))); > + > + return 0; > +} > + > +static void fill_event_attr(struct ddr_pmu *pmu) > +{ > + int i, j, k; > + struct attribute **dst =3D ddr_perf_events_attrs; > + > + j =3D 0; > + k =3D 0; > + > + /* fill ALL_CHAN_COUNTER_ID event */ > + dst[j++] =3D &event_attrs[k].attr.attr; > + dst[j++] =3D &event_unit_attrs[k].attr; > + dst[j++] =3D &event_scale_attrs[k].attr; > + > + k++; > + > + /* fill each channel event */ > + for (i =3D 0; i < pmu->info.chann_nr; i++, k++) { > + dst[j++] =3D &event_attrs[k].attr.attr; > + dst[j++] =3D &event_unit_attrs[k].attr; > + dst[j++] =3D &event_scale_attrs[k].attr; > + } > + > + dst[j] =3D NULL; /* mark end */ > +} > + > +static void fmt_attr_fill(struct attribute **fmt_attr) > +{ > + int i; > + > + for (i =3D 0; fmt_attr[i]; i++) > + ddr_perf_format_attrs[i] =3D fmt_attr[i]; > + > + ddr_perf_format_attrs[i] =3D NULL; > +} > + > +static int ddr_pmu_parse_dt(struct platform_device *pdev, struct = dmc_hw_info *info) > +{ > + /*struct pinctrl *p;*/ > + void __iomem *base; > + int i, ret =3D -EINVAL; > + > + for (i =3D 0; i < info->dmc_nr; i++) { > + /* resource 0 for ddr register base */ > + base =3D devm_platform_ioremap_resource(pdev, i); > + if (IS_ERR(base)) { > + dev_err(&pdev->dev, "couldn't ioremap ddr reg = %d\n", i); > + return PTR_ERR(base); > + } > + info->ddr_reg[i] =3D base; > + } > + > + /* resource i for pll register base */ > + base =3D devm_platform_ioremap_resource(pdev, i); > + if (IS_ERR(base)) { > + dev_err(&pdev->dev, "couldn't ioremap for pll reg\n"); > + return PTR_ERR(base); > + } > + info->pll_reg =3D base; > + > + ret =3D platform_get_irq(pdev, 0); > + if (ret < 0) { > + dev_err(&pdev->dev, "couldn't get irq\n"); > + return ret; > + } > + info->irq_num =3D ret; > + > + ret =3D devm_request_irq(&pdev->dev, info->irq_num, = dmc_irq_handler, > + IRQF_NOBALANCING, dev_name(&pdev->dev), > + (void *)info); > + if (ret < 0) > + dev_err(&pdev->dev, "ddr request irq failed\n"); > + > + return ret; > +} > + > +int aml_ddr_pmu_create(struct platform_device *pdev, struct ddr_pmu = *pmu) > +{ > + int ret; > + char *name; > + struct pmu tmp_pmu =3D { > + .module =3D THIS_MODULE, > + .capabilities =3D PERF_PMU_CAP_NO_EXCLUDE, > + .task_ctx_nr =3D perf_invalid_context, > + .attr_groups =3D attr_groups, > + .event_init =3D aml_ddr_perf_event_init, > + .add =3D aml_ddr_perf_event_add, > + .del =3D aml_ddr_perf_event_del, > + .start =3D aml_ddr_perf_event_start, > + .stop =3D aml_ddr_perf_event_stop, > + .read =3D aml_ddr_perf_event_update, > + }; > + > + pmu->pmu =3D tmp_pmu; > + > + ret =3D ddr_pmu_parse_dt(pdev, &pmu->info); > + if (ret < 0) > + return ret; > + > + fmt_attr_fill(pmu->info.fmt_attr); > + > + pmu->cpu =3D raw_smp_processor_id(); > + > + name =3D devm_kasprintf(&pdev->dev, GFP_KERNEL, = DDR_PERF_DEV_NAME); > + if (!name) { > + dev_err(&pdev->dev, "couldn't allocat name\n"); > + return -ENOMEM; > + } > + > + ret =3D cpuhp_setup_state_multi(CPUHP_AP_ONLINE_DYN, name, NULL, > + ddr_perf_offline_cpu); > + if (ret < 0) { > + dev_err(&pdev->dev, "cpuhp_setup_state_multi failed\n"); > + return ret; > + } > + pmu->cpuhp_state =3D ret; > + > + /* Register the pmu instance for cpu hotplug */ > + ret =3D cpuhp_state_add_instance_nocalls(pmu->cpuhp_state, = &pmu->node); > + if (ret) { > + dev_err(&pdev->dev, "Error %d registering hotplug\n", = ret); > + goto cpuhp_instance_err; > + } > + > + fill_event_attr(pmu); > + > + ret =3D perf_pmu_register(&pmu->pmu, name, -1); > + if (ret) { > + dev_err(&pdev->dev, "perf pmu register failed\n"); > + goto pmu_register_err; > + } > + > + pmu->name =3D name; > + pmu->dev =3D &pdev->dev; > + pmu->pmu_enabled =3D false; > + > + platform_set_drvdata(pdev, pmu); > + > + return 0; > + > +pmu_register_err: > + cpuhp_state_remove_instance_nocalls(pmu->cpuhp_state, = &pmu->node); > +cpuhp_instance_err: > + cpuhp_remove_state(pmu->cpuhp_state); > + > + return ret; > +} > +EXPORT_SYMBOL(aml_ddr_pmu_create); > + > +int aml_ddr_pmu_remove(struct platform_device *pdev) > +{ > + struct ddr_pmu *pmu =3D platform_get_drvdata(pdev); > + > + perf_pmu_unregister(&pmu->pmu); > + cpuhp_state_remove_instance_nocalls(pmu->cpuhp_state, = &pmu->node); > + cpuhp_remove_state(pmu->cpuhp_state); > + > + return 0; > +} > +EXPORT_SYMBOL(aml_ddr_pmu_remove); > diff --git a/drivers/perf/amlogic/aml_ddr_pmu_g12.c = b/drivers/perf/amlogic/aml_ddr_pmu_g12.c > new file mode 100644 > index 000000000000..e98acdc72797 > --- /dev/null > +++ b/drivers/perf/amlogic/aml_ddr_pmu_g12.c > @@ -0,0 +1,388 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * Copyright (c) 2022 Amlogic, Inc. All rights reserved. > + */ > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > + > +#include > + > +#define PORT_MAJOR 32 > +#define DEFAULT_XTAL_FREQ 24000000UL > + > +#define DMC_QOS_IRQ BIT(30) > + > +/* DMC bandwidth monitor register address offset */ > +#define DMC_MON_G12_CTRL0 (0x20 << 2) > +#define DMC_MON_G12_CTRL1 (0x21 << 2) > +#define DMC_MON_G12_CTRL2 (0x22 << 2) > +#define DMC_MON_G12_CTRL3 (0x23 << 2) > +#define DMC_MON_G12_CTRL4 (0x24 << 2) > +#define DMC_MON_G12_CTRL5 (0x25 << 2) > +#define DMC_MON_G12_CTRL6 (0x26 << 2) > +#define DMC_MON_G12_CTRL7 (0x27 << 2) > +#define DMC_MON_G12_CTRL8 (0x28 << 2) > + > +#define DMC_MON_G12_ALL_REQ_CNT (0x29 << 2) > +#define DMC_MON_G12_ALL_GRANT_CNT (0x2a << 2) > +#define DMC_MON_G12_ONE_GRANT_CNT (0x2b << 2) > +#define DMC_MON_G12_SEC_GRANT_CNT (0x2c << 2) > +#define DMC_MON_G12_THD_GRANT_CNT (0x2d << 2) > +#define DMC_MON_G12_FOR_GRANT_CNT (0x2e << 2) > +#define DMC_MON_G12_TIMER (0x2f << 2) > + > +/* Each bit represent a axi line */ > +PMU_FORMAT_ATTR(event, "config:0-7"); > +PMU_FORMAT_ATTR(arm, "config1:0"); > +PMU_FORMAT_ATTR(gpu, "config1:1"); > +PMU_FORMAT_ATTR(pcie, "config1:2"); > +PMU_FORMAT_ATTR(hdcp, "config1:3"); > +PMU_FORMAT_ATTR(hevc_front, "config1:4"); > +PMU_FORMAT_ATTR(usb3_0, "config1:6"); > +PMU_FORMAT_ATTR(device, "config1:7"); > +PMU_FORMAT_ATTR(hevc_back, "config1:8"); > +PMU_FORMAT_ATTR(h265enc, "config1:9"); > +PMU_FORMAT_ATTR(vpu_read1, "config1:16"); > +PMU_FORMAT_ATTR(vpu_read2, "config1:17"); > +PMU_FORMAT_ATTR(vpu_read3, "config1:18"); > +PMU_FORMAT_ATTR(vpu_write1, "config1:19"); > +PMU_FORMAT_ATTR(vpu_write2, "config1:20"); > +PMU_FORMAT_ATTR(vdec, "config1:21"); > +PMU_FORMAT_ATTR(hcodec, "config1:22"); > +PMU_FORMAT_ATTR(ge2d, "config1:23"); > + > +PMU_FORMAT_ATTR(spicc1, "config1:32"); > +PMU_FORMAT_ATTR(usb0, "config1:33"); > +PMU_FORMAT_ATTR(dma, "config1:34"); > +PMU_FORMAT_ATTR(arb0, "config1:35"); > +PMU_FORMAT_ATTR(sd_emmc_b, "config1:36"); > +PMU_FORMAT_ATTR(usb1, "config1:37"); > +PMU_FORMAT_ATTR(audio, "config1:38"); > +PMU_FORMAT_ATTR(aififo, "config1:39"); > +PMU_FORMAT_ATTR(parser, "config1:41"); > +PMU_FORMAT_ATTR(ao_cpu, "config1:42"); > +PMU_FORMAT_ATTR(sd_emmc_c, "config1:43"); > +PMU_FORMAT_ATTR(spicc2, "config1:44"); > +PMU_FORMAT_ATTR(ethernet, "config1:45"); > +PMU_FORMAT_ATTR(sana, "config1:46"); > + > +/* for sm1 and g12b */ > +PMU_FORMAT_ATTR(nna, "config1:10"); > + > +/* for g12b only */ > +PMU_FORMAT_ATTR(gdc, "config1:11"); > +PMU_FORMAT_ATTR(mipi_isp, "config1:12"); > +PMU_FORMAT_ATTR(arm1, "config1:13"); > +PMU_FORMAT_ATTR(sd_emmc_a, "config1:40"); > + > +static struct attribute *g12_pmu_format_attrs[] =3D { > + &format_attr_event.attr, > + &format_attr_arm.attr, > + &format_attr_gpu.attr, > + &format_attr_nna.attr, > + &format_attr_gdc.attr, > + &format_attr_arm1.attr, > + &format_attr_mipi_isp.attr, > + &format_attr_sd_emmc_a.attr, > + &format_attr_pcie.attr, > + &format_attr_hdcp.attr, > + &format_attr_hevc_front.attr, > + &format_attr_usb3_0.attr, > + &format_attr_device.attr, > + &format_attr_hevc_back.attr, > + &format_attr_h265enc.attr, > + &format_attr_vpu_read1.attr, > + &format_attr_vpu_read2.attr, > + &format_attr_vpu_read3.attr, > + &format_attr_vpu_write1.attr, > + &format_attr_vpu_write2.attr, > + &format_attr_vdec.attr, > + &format_attr_hcodec.attr, > + &format_attr_ge2d.attr, > + &format_attr_spicc1.attr, > + &format_attr_usb0.attr, > + &format_attr_dma.attr, > + &format_attr_arb0.attr, > + &format_attr_sd_emmc_b.attr, > + &format_attr_usb1.attr, > + &format_attr_audio.attr, > + &format_attr_aififo.attr, > + &format_attr_parser.attr, > + &format_attr_ao_cpu.attr, > + &format_attr_sd_emmc_c.attr, > + &format_attr_spicc2.attr, > + &format_attr_ethernet.attr, > + &format_attr_sana.attr, > + NULL, > +}; > + > +/* calculate ddr clock */ > +static unsigned long dmc_g12_get_freq_quick(struct dmc_hw_info *info) > +{ > + unsigned int val; > + unsigned int n, m, od1; > + unsigned int od_div =3D 0xfff; > + unsigned long freq =3D 0; > + > + val =3D readl(info->pll_reg); > + val =3D val & 0xfffff; > + switch ((val >> 16) & 7) { > + case 0: > + od_div =3D 2; > + break; > + > + case 1: > + od_div =3D 3; > + break; > + > + case 2: > + od_div =3D 4; > + break; > + > + case 3: > + od_div =3D 6; > + break; > + > + case 4: > + od_div =3D 8; > + break; > + > + default: > + break; > + } > + > + m =3D val & 0x1ff; > + n =3D ((val >> 10) & 0x1f); > + od1 =3D (((val >> 19) & 0x1)) =3D=3D 1 ? 2 : 1; > + freq =3D DEFAULT_XTAL_FREQ / 1000; /* avoid overflow */ > + if (n) > + freq =3D ((((freq * m) / n) >> od1) / od_div) * 1000; > + > + return freq; > +} > + > +#ifdef DEBUG > +static void g12_dump_reg(struct dmc_hw_info *db) > +{ > + int s =3D 0, i; > + unsigned int r; > + > + for (i =3D 0; i < 9; i++) { > + r =3D readl(db->ddr_reg[0] + (DMC_MON_G12_CTRL0 + (i << = 2))); > + pr_notice("DMC_MON_CTRL%d: %08x\n", i, r); > + } > + r =3D readl(db->ddr_reg[0] + DMC_MON_G12_ALL_REQ_CNT); > + pr_notice("DMC_MON_ALL_REQ_CNT: %08x\n", r); > + r =3D readl(db->ddr_reg[0] + DMC_MON_G12_ALL_GRANT_CNT); > + pr_notice("DMC_MON_ALL_GRANT_CNT:%08x\n", r); > + r =3D readl(db->ddr_reg[0] + DMC_MON_G12_ONE_GRANT_CNT); > + pr_notice("DMC_MON_ONE_GRANT_CNT:%08x\n", r); > + r =3D readl(db->ddr_reg[0] + DMC_MON_G12_SEC_GRANT_CNT); > + pr_notice("DMC_MON_SEC_GRANT_CNT:%08x\n", r); > + r =3D readl(db->ddr_reg[0] + DMC_MON_G12_THD_GRANT_CNT); > + pr_notice("DMC_MON_THD_GRANT_CNT:%08x\n", r); > + r =3D readl(db->ddr_reg[0] + DMC_MON_G12_FOR_GRANT_CNT); > + pr_notice("DMC_MON_FOR_GRANT_CNT:%08x\n", r); > + r =3D readl(db->ddr_reg[0] + DMC_MON_G12_TIMER); > + pr_notice("DMC_MON_TIMER: %08x\n", r); > +} > +#endif > + > +static void dmc_g12_counter_enable(struct dmc_hw_info *info) > +{ > + unsigned int val; > + unsigned long clock_count =3D dmc_g12_get_freq_quick(info) / 10; = /* 100ms */ > + > + writel(clock_count, info->ddr_reg[0] + DMC_MON_G12_TIMER); > + > + val =3D readl(info->ddr_reg[0] + DMC_MON_G12_CTRL0); > + > + /* enable all channel */ > + val =3D BIT(31) | /* enable bit */ > + BIT(20) | /* use timer */ > + 0x0f; /* 4 channels */ > + > + writel(val, info->ddr_reg[0] + DMC_MON_G12_CTRL0); > + > +#ifdef DEBUG > + g12_dump_reg(info); > +#endif > +} > + > +static void dmc_g12_config_fiter(struct dmc_hw_info *info, > + int port, int channel) > +{ > + u32 val; > + u32 rp[MAX_CHANNEL_NUM] =3D {DMC_MON_G12_CTRL1, = DMC_MON_G12_CTRL3, > + DMC_MON_G12_CTRL5, = DMC_MON_G12_CTRL7}; > + u32 rs[MAX_CHANNEL_NUM] =3D {DMC_MON_G12_CTRL2, = DMC_MON_G12_CTRL4, > + DMC_MON_G12_CTRL6, = DMC_MON_G12_CTRL8}; > + int subport =3D -1; > + > + /* clear all port mask */ > + if (port < 0) { > + writel(0, info->ddr_reg[0] + rp[channel]); > + writel(0, info->ddr_reg[0] + rs[channel]); > + return; > + } > + > + if (port >=3D PORT_MAJOR) > + subport =3D port - PORT_MAJOR; > + > + if (subport < 0) { > + val =3D readl(info->ddr_reg[0] + rp[channel]); > + val |=3D (1 << port); > + writel(val, info->ddr_reg[0] + rp[channel]); > + val =3D 0xffff; > + writel(val, info->ddr_reg[0] + rs[channel]); > + } else { > + val =3D BIT(23); /* select device */ > + writel(val, info->ddr_reg[0] + rp[channel]); > + val =3D readl(info->ddr_reg[0] + rs[channel]); > + val |=3D (1 << subport); > + writel(val, info->ddr_reg[0] + rs[channel]); > + } > +} > + > +static void dmc_g12_config_axi_id(struct dmc_hw_info *info, int = axi_id, int channel) > +{ > + if (channel > info->chann_nr) > + return; > + > + dmc_g12_config_fiter(info, axi_id, channel); > +} > + > +static void dmc_g12_counter_disable(struct dmc_hw_info *info) > +{ > + int i; > + > + /* clear timer */ > + writel(0, info->ddr_reg[0] + DMC_MON_G12_CTRL0); > + writel(0, info->ddr_reg[0] + DMC_MON_G12_TIMER); > + > + writel(0, info->ddr_reg[0] + DMC_MON_G12_ALL_REQ_CNT); > + writel(0, info->ddr_reg[0] + DMC_MON_G12_ALL_GRANT_CNT); > + writel(0, info->ddr_reg[0] + DMC_MON_G12_ONE_GRANT_CNT); > + writel(0, info->ddr_reg[0] + DMC_MON_G12_SEC_GRANT_CNT); > + writel(0, info->ddr_reg[0] + DMC_MON_G12_THD_GRANT_CNT); > + writel(0, info->ddr_reg[0] + DMC_MON_G12_FOR_GRANT_CNT); > + > + /* clear port channel mapping */ > + for (i =3D 0; i < info->chann_nr; i++) > + dmc_g12_config_fiter(info, -1, i); > +} > + > +static void dmc_g12_get_counters(struct dmc_hw_info *info, > + struct dmc_counter *counter) > +{ > + int i; > + unsigned int reg; > + > + counter->all_cnt =3D readl(info->ddr_reg[0] + = DMC_MON_G12_ALL_GRANT_CNT); > + counter->all_req =3D readl(info->ddr_reg[0] + = DMC_MON_G12_ALL_REQ_CNT); > + > + for (i =3D 0; i < info->chann_nr; i++) { > + reg =3D DMC_MON_G12_ONE_GRANT_CNT + (i << 2); > + counter->channel_cnt[i] =3D readl(info->ddr_reg[0] + = reg); > + } > +} > + > +static int dmc_g12_irq_handler(struct dmc_hw_info *info, > + struct dmc_counter *counter) > +{ > + unsigned int val; > + int ret =3D -EINVAL; > + > + val =3D readl(info->ddr_reg[0] + DMC_MON_G12_CTRL0); > + if (val & DMC_QOS_IRQ) { > + dmc_g12_get_counters(info, counter); > + /* clear irq flags */ > + writel(val, info->ddr_reg[0] + DMC_MON_G12_CTRL0); > + ret =3D 0; > + } > + return ret; > +} > + > +static struct dmc_pmu_hw_ops g12_ops =3D { > + .enable =3D dmc_g12_counter_enable, > + .disable =3D dmc_g12_counter_disable, > + .irq_handler =3D dmc_g12_irq_handler, > + .get_counters =3D dmc_g12_get_counters, > + .config_axi_id =3D dmc_g12_config_axi_id, > +}; > + > +static int __init g12_ddr_pmu_probe(struct platform_device *pdev) > +{ > + struct ddr_pmu *pmu; > + > + if (of_device_is_compatible(pdev->dev.of_node, > + "amlogic,g12a-ddr-pmu")) { > + format_attr_nna.attr.mode =3D 0; > + format_attr_gdc.attr.mode =3D 0; > + format_attr_arm1.attr.mode =3D 0; > + format_attr_mipi_isp.attr.mode =3D 0; > + } else if (of_device_is_compatible(pdev->dev.of_node, > + "amlogic,sm1-ddr-pmu")) { > + format_attr_gdc.attr.mode =3D 0; > + format_attr_arm1.attr.mode =3D 0; > + format_attr_mipi_isp.attr.mode =3D 0; > + } > + > + pmu =3D devm_kzalloc(&pdev->dev, sizeof(struct ddr_pmu), = GFP_KERNEL); > + if (!pmu) > + return -ENOMEM; > + > + /* > + * G12 series Soc have single dmc controller and > + * 4x ddr bandwidth monitor channels > + */ > + pmu->info.dmc_nr =3D 1; > + pmu->info.chann_nr =3D 4; > + pmu->info.ops =3D &g12_ops; > + pmu->info.fmt_attr =3D g12_pmu_format_attrs; > + > + return aml_ddr_pmu_create(pdev, pmu); > +} > + > +static int __exit g12_ddr_pmu_remove(struct platform_device *pdev) > +{ > + aml_ddr_pmu_remove(pdev); > + > + return 0; > +} > + > +static const struct of_device_id aml_ddr_pmu_dt_match[] =3D { > + { > + .compatible =3D "amlogic,g12-ddr-pmu", > + }, > + { > + .compatible =3D "amlogic,g12a-ddr-pmu", > + }, > + { > + .compatible =3D "amlogic,g12b-ddr-pmu", > + }, > + { > + .compatible =3D "amlogic,sm1-ddr-pmu", > + }, > + {} > +}; > + > +static struct platform_driver g12_ddr_pmu_driver =3D { > + .driver =3D { > + .name =3D "amlogic,ddr-pmu", > + .of_match_table =3D aml_ddr_pmu_dt_match, > + }, > + .remove =3D __exit_p(g12_ddr_pmu_remove), > +}; > + > +module_platform_driver_probe(g12_ddr_pmu_driver, g12_ddr_pmu_probe); > +MODULE_AUTHOR("Jiucheng Xu"); > +MODULE_LICENSE("GPL"); > +MODULE_DESCRIPTION("Amlogic G12 series SoC DDR PMU"); > diff --git a/include/soc/amlogic/aml_ddr_pmu.h = b/include/soc/amlogic/aml_ddr_pmu.h > new file mode 100644 > index 000000000000..283f5773ecd1 > --- /dev/null > +++ b/include/soc/amlogic/aml_ddr_pmu.h > @@ -0,0 +1,76 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +/* > + * Copyright (c) 2022 Amlogic, Inc. All rights reserved. > + */ > + > +#ifndef __AML_DDR_PMU_H__ > +#define __AML_DDR_PMU_H__ > + > +#define MAX_CHANNEL_NUM 8 > + > +enum { > + ALL_CHAN_COUNTER_ID, > + CHAN1_COUNTER_ID, > + CHAN2_COUNTER_ID, > + CHAN3_COUNTER_ID, > + CHAN4_COUNTER_ID, > + CHAN5_COUNTER_ID, > + CHAN6_COUNTER_ID, > + CHAN7_COUNTER_ID, > + CHAN8_COUNTER_ID, > + COUNTER_MAX_ID, > +}; > + > +struct dmc_hw_info; > + > +struct dmc_counter { > + u64 all_cnt; /* The count of all requests come in/out ddr = controller */ > + union { > + u64 all_req; > + struct { > + u64 all_idle_cnt; > + u64 all_16bit_cnt; > + }; > + }; > + u64 channel_cnt[MAX_CHANNEL_NUM]; /* To save a DMC = bandwidth-monitor channel counter */ > +}; > + > +struct dmc_pmu_hw_ops { > + void (*enable)(struct dmc_hw_info *info); > + void (*disable)(struct dmc_hw_info *info); > + /* Bind an axi line to a bandwidth-monitor channel */ > + void (*config_axi_id)(struct dmc_hw_info *info, int axi_id, int = chann); > + int (*irq_handler)(struct dmc_hw_info *info, > + struct dmc_counter *counter); > + void (*get_counters)(struct dmc_hw_info *info, > + struct dmc_counter *counter); > +}; > + > +struct dmc_hw_info { > + struct dmc_pmu_hw_ops *ops; > + void __iomem *ddr_reg[4]; > + unsigned long timer_value; /* Timer value in TIMER register = */ > + void __iomem *pll_reg; > + int irq_num; /* irq vector number */ > + int dmc_nr; /* The number of dmc controller = */ > + int chann_nr; /* The number of dmc bandwidth = monitor channels */ > + int id; /* The number of supported = channels */ > + struct attribute **fmt_attr; > +}; > + > +struct ddr_pmu { > + struct pmu pmu; > + struct dmc_hw_info info; > + struct dmc_counter counters; /* save counters from hw */ > + bool pmu_enabled; > + struct device *dev; > + char *name; > + struct hlist_node node; > + enum cpuhp_state cpuhp_state; > + int cpu; /* for cpu hotplug */ > +}; > + > +int aml_ddr_pmu_create(struct platform_device *pdev, struct ddr_pmu = *pmu); > +int aml_ddr_pmu_remove(struct platform_device *pdev); > + > +#endif /* __AML_DDR_PMU_H__ */