2022-08-14 18:34:57

by Besar Wicaksono

[permalink] [raw]
Subject: [PATCH v4 2/2] perf: arm_cspmu: Add support for NVIDIA SCF and MCF attribute

Add support for NVIDIA System Cache Fabric (SCF) and Memory Control
Fabric (MCF) PMU attributes for CoreSight PMU implementation in
NVIDIA devices.

Signed-off-by: Besar Wicaksono <[email protected]>
---
Documentation/admin-guide/perf/index.rst | 1 +
Documentation/admin-guide/perf/nvidia-pmu.rst | 120 ++++++
drivers/perf/arm_cspmu/Makefile | 3 +-
drivers/perf/arm_cspmu/arm_cspmu.c | 7 +
drivers/perf/arm_cspmu/nvidia_cspmu.c | 367 ++++++++++++++++++
drivers/perf/arm_cspmu/nvidia_cspmu.h | 17 +
6 files changed, 514 insertions(+), 1 deletion(-)
create mode 100644 Documentation/admin-guide/perf/nvidia-pmu.rst
create mode 100644 drivers/perf/arm_cspmu/nvidia_cspmu.c
create mode 100644 drivers/perf/arm_cspmu/nvidia_cspmu.h

diff --git a/Documentation/admin-guide/perf/index.rst b/Documentation/admin-guide/perf/index.rst
index 69b23f087c05..cf05fed1f67f 100644
--- a/Documentation/admin-guide/perf/index.rst
+++ b/Documentation/admin-guide/perf/index.rst
@@ -17,3 +17,4 @@ Performance monitor support
xgene-pmu
arm_dsu_pmu
thunderx2-pmu
+ nvidia-pmu
diff --git a/Documentation/admin-guide/perf/nvidia-pmu.rst b/Documentation/admin-guide/perf/nvidia-pmu.rst
new file mode 100644
index 000000000000..c41b93965824
--- /dev/null
+++ b/Documentation/admin-guide/perf/nvidia-pmu.rst
@@ -0,0 +1,120 @@
+=========================================================
+NVIDIA Tegra SoC Uncore Performance Monitoring Unit (PMU)
+=========================================================
+
+The NVIDIA Tegra SoC includes various system PMUs to measure key performance
+metrics like memory bandwidth, latency, and utilization:
+
+* Scalable Coherency Fabric (SCF)
+* Memory Controller Fabric (MCF) GPU physical interface
+* MCF GPU virtual interface
+* MCF NVLINK interface
+* MCF PCIE interface
+
+PMU Driver
+----------
+
+The PMUs in this document are based on ARM CoreSight PMU Architecture as
+described in document: ARM IHI 0091. Since this is a standard architecture, the
+PMUs are managed by a common driver "arm-cs-arch-pmu". This driver describes
+the available events and configuration of each PMU in sysfs. Please see the
+sections below to get the sysfs path of each PMU. Like other uncore PMU driver,
+the driver provides "cpumask" sysfs attribute to show the CPU id used to handle
+the PMU event. There is also "associated_cpus" sysfs attribute, which contains a
+list of CPUs associated with the PMU instance.
+
+SCF PMU
+-------
+
+The SCF PMU monitors system level cache events, CPU traffic, and
+strongly-ordered PCIE traffic to local/remote memory.
+
+The events and configuration options of this PMU device are described in sysfs,
+see /sys/bus/event_sources/devices/nvidia_scf_pmu_<socket-id>.
+
+Example usage::
+
+ perf stat -a -e nvidia_scf_pmu_0/config=0x0/
+
+This will count the events in socket 0.
+
+MCF GPU Physical PMU
+--------------------
+
+The MCF GPU physical PMU monitors ATS translated traffic from GPU to
+local/remote memory via Nvlink C2C.
+
+The events and configuration options of this PMU device are described in sysfs,
+see /sys/bus/event_sources/devices/nvidia_mcf_gpu_pmu_<socket-id>.
+
+Multiple GPUs can be connected to the SoC. The user can use "gpu" bitmap
+parameter to select the GPU(s) to monitor, i.e. "gpu=0xF" corresponds to GPU 0
+to 3. /sys/bus/event_sources/devices/nvidia_mcf_gpu_pmu_<socket-id>/format/gpu
+shows the valid bits that can be set in the "gpu" parameter.
+
+Example usage::
+
+ perf stat -a -e nvidia_mcf_gpu_pmu_0/config=0x0,gpu=0x3/
+
+This will count the events on GPU 0 and 1 that are connected to SoC in socket 0.
+
+MCF GPU Virtual PMU
+-------------------
+
+The MCF GPU virtual PMU monitors SMMU inline translated traffic (as opposed to
+ATS) from GPU to local/remote memory via Nvlink C2C.
+
+The events and configuration options of this PMU device are described in sysfs,
+see /sys/bus/event_sources/devices/nvidia_mcf_gpuvir_pmu_<socket-id>.
+
+Multiple GPUs can be connected to the SoC. The user can use "gpu" bitmap
+parameter to select the GPU(s) to monitor, i.e. "gpu=0xF" corresponds to GPU 0
+to 3. /sys/bus/event_sources/devices/nvidia_mcf_gpuvir_pmu_<socket-id>/format/gpu
+shows the valid bits that can be set in the "gpu" parameter.
+
+Example usage::
+
+ perf stat -a -e nvidia_mcf_gpuvir_pmu_0/config=0x0,gpu=0x3/
+
+This will count the events on GPU 0 and 1 that are connected to SoC in socket 0.
+
+MCF NVLINK PMU
+--------------
+
+The MCF NVLINK PMU monitors I/O coherent traffic from external socket to local
+memory.
+
+The events and configuration options of this PMU device are described in sysfs,
+see /sys/bus/event_sources/devices/nvidia_mcf_nvlink_pmu_<socket-id>.
+
+Each SoC socket can be connected to one or more sockets via NVLINK. The user can
+use "rem_socket" bitmap parameter to select the remote socket(s) to monitor,
+i.e. "rem_socket=0xE" corresponds to socket 1 to 3.
+/sys/bus/event_sources/devices/nvidia_mcf_nvlink_pmu_<socket-id>/format/rem_socket
+shows the valid bits that can be set in the "rem_socket" parameter.
+
+Example usage::
+
+ perf stat -a -e nvidia_mcf_nvlink_pmu_0/config=0x0,rem_socket=0x6/
+
+This will count the events from remote socket 1 and 2 to socket 0.
+
+MCF PCIE PMU
+------------
+
+The MCF PCIE PMU monitors traffic from PCIE root ports to local/remote memory.
+
+The events and configuration options of this PMU device are described in sysfs,
+see /sys/bus/event_sources/devices/nvidia_mcf_pcie_pmu_<socket-id>.
+
+Each SoC socket can support multiple root ports. The user can use
+"root_port" bitmap parameter to select the port(s) to monitor, i.e.
+"root_port=0xF" corresponds to root port 0 to 3.
+/sys/bus/event_sources/devices/nvidia_mcf_pcie_pmu_<socket-id>/format/root_port
+shows the valid bits that can be set in the "root_port" parameter.
+
+Example usage::
+
+ perf stat -a -e nvidia_mcf_pcie_pmu_0/config=0x0,root_port=0x3/
+
+This will count the events from root port 0 and 1 of socket 0.
diff --git a/drivers/perf/arm_cspmu/Makefile b/drivers/perf/arm_cspmu/Makefile
index cdc3455f74d8..1b586064bd77 100644
--- a/drivers/perf/arm_cspmu/Makefile
+++ b/drivers/perf/arm_cspmu/Makefile
@@ -3,4 +3,5 @@
# SPDX-License-Identifier: GPL-2.0

obj-$(CONFIG_ARM_CORESIGHT_PMU_ARCH_SYSTEM_PMU) += \
- arm_cspmu.o
+ arm_cspmu.o \
+ nvidia_cspmu.o
diff --git a/drivers/perf/arm_cspmu/arm_cspmu.c b/drivers/perf/arm_cspmu/arm_cspmu.c
index 410876f86eb0..7a0beb515e53 100644
--- a/drivers/perf/arm_cspmu/arm_cspmu.c
+++ b/drivers/perf/arm_cspmu/arm_cspmu.c
@@ -31,6 +31,7 @@
#include <acpi/processor.h>

#include "arm_cspmu.h"
+#include "nvidia_cspmu.h"

#define PMUNAME "arm_cspmu"
#define DRVNAME "arm-cs-arch-pmu"
@@ -118,6 +119,9 @@ static_assert(
ops->callback = arm_cspmu_ ## callback; \
} while (0)

+/* JEDEC-assigned JEP106 identification code */
+#define ARM_CSPMU_IMPL_ID_NVIDIA 0x36B
+
static unsigned long arm_cspmu_cpuhp_state;

/*
@@ -369,6 +373,9 @@ struct impl_match {
};

static const struct impl_match impl_match[] = {
+ { .pmiidr = ARM_CSPMU_IMPL_ID_NVIDIA,
+ .mask = ARM_CSPMU_PMIIDR_IMPLEMENTER,
+ .impl_init_ops = nv_cspmu_init_ops },
{}
};

diff --git a/drivers/perf/arm_cspmu/nvidia_cspmu.c b/drivers/perf/arm_cspmu/nvidia_cspmu.c
new file mode 100644
index 000000000000..261f20680bc1
--- /dev/null
+++ b/drivers/perf/arm_cspmu/nvidia_cspmu.c
@@ -0,0 +1,367 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (c) 2022, NVIDIA CORPORATION & AFFILIATES.
+ *
+ */
+
+/* Support for NVIDIA specific attributes. */
+
+#include "nvidia_cspmu.h"
+
+#define NV_MCF_PCIE_PORT_COUNT 10ULL
+#define NV_MCF_PCIE_FILTER_ID_MASK GENMASK_ULL(NV_MCF_PCIE_PORT_COUNT - 1, 0)
+
+#define NV_MCF_GPU_PORT_COUNT 2ULL
+#define NV_MCF_GPU_FILTER_ID_MASK GENMASK_ULL(NV_MCF_GPU_PORT_COUNT - 1, 0)
+
+#define NV_MCF_NVL_PORT_COUNT 4ULL
+#define NV_MCF_NVL_FILTER_ID_MASK GENMASK_ULL(NV_MCF_NVL_PORT_COUNT - 1, 0)
+
+#define NV_SCF_MCF_PRODID_MASK GENMASK(31, 0)
+
+#define NV_FORMAT_NAME_GENERIC 0
+
+#define to_nv_cspmu_ctx(cspmu) ((struct nv_cspmu_ctx *)(cspmu->impl.ctx))
+
+#define NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _num, _suff, _config) \
+ ARM_CSPMU_EVENT_ATTR(_pref##_num##_suff, _config)
+
+#define NV_CSPMU_EVENT_ATTR_4(_pref, _suff, _config) \
+ NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _0_, _suff, _config), \
+ NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _1_, _suff, _config + 1), \
+ NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _2_, _suff, _config + 2), \
+ NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _3_, _suff, _config + 3)
+
+struct nv_cspmu_ctx {
+ const char *name;
+ u32 filter_mask;
+ struct attribute **event_attr;
+ struct attribute **format_attr;
+};
+
+static struct attribute *scf_pmu_event_attrs[] = {
+ ARM_CSPMU_EVENT_ATTR(bus_cycles, 0x1d),
+
+ ARM_CSPMU_EVENT_ATTR(scf_cache_allocate, 0xF0),
+ ARM_CSPMU_EVENT_ATTR(scf_cache_refill, 0xF1),
+ ARM_CSPMU_EVENT_ATTR(scf_cache, 0xF2),
+ ARM_CSPMU_EVENT_ATTR(scf_cache_wb, 0xF3),
+
+ NV_CSPMU_EVENT_ATTR_4(socket, rd_data, 0x101),
+ NV_CSPMU_EVENT_ATTR_4(socket, dl_rsp, 0x105),
+ NV_CSPMU_EVENT_ATTR_4(socket, wb_data, 0x109),
+ NV_CSPMU_EVENT_ATTR_4(socket, ev_rsp, 0x10d),
+ NV_CSPMU_EVENT_ATTR_4(socket, prb_data, 0x111),
+
+ NV_CSPMU_EVENT_ATTR_4(socket, rd_outstanding, 0x115),
+ NV_CSPMU_EVENT_ATTR_4(socket, dl_outstanding, 0x119),
+ NV_CSPMU_EVENT_ATTR_4(socket, wb_outstanding, 0x11d),
+ NV_CSPMU_EVENT_ATTR_4(socket, wr_outstanding, 0x121),
+ NV_CSPMU_EVENT_ATTR_4(socket, ev_outstanding, 0x125),
+ NV_CSPMU_EVENT_ATTR_4(socket, prb_outstanding, 0x129),
+
+ NV_CSPMU_EVENT_ATTR_4(socket, rd_access, 0x12d),
+ NV_CSPMU_EVENT_ATTR_4(socket, dl_access, 0x131),
+ NV_CSPMU_EVENT_ATTR_4(socket, wb_access, 0x135),
+ NV_CSPMU_EVENT_ATTR_4(socket, wr_access, 0x139),
+ NV_CSPMU_EVENT_ATTR_4(socket, ev_access, 0x13d),
+ NV_CSPMU_EVENT_ATTR_4(socket, prb_access, 0x141),
+
+ NV_CSPMU_EVENT_ATTR_4(ocu, gmem_rd_data, 0x145),
+ NV_CSPMU_EVENT_ATTR_4(ocu, gmem_rd_access, 0x149),
+ NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wb_access, 0x14d),
+ NV_CSPMU_EVENT_ATTR_4(ocu, gmem_rd_outstanding, 0x151),
+ NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wr_outstanding, 0x155),
+
+ NV_CSPMU_EVENT_ATTR_4(ocu, rem_rd_data, 0x159),
+ NV_CSPMU_EVENT_ATTR_4(ocu, rem_rd_access, 0x15d),
+ NV_CSPMU_EVENT_ATTR_4(ocu, rem_wb_access, 0x161),
+ NV_CSPMU_EVENT_ATTR_4(ocu, rem_rd_outstanding, 0x165),
+ NV_CSPMU_EVENT_ATTR_4(ocu, rem_wr_outstanding, 0x169),
+
+ ARM_CSPMU_EVENT_ATTR(gmem_rd_data, 0x16d),
+ ARM_CSPMU_EVENT_ATTR(gmem_rd_access, 0x16e),
+ ARM_CSPMU_EVENT_ATTR(gmem_rd_outstanding, 0x16f),
+ ARM_CSPMU_EVENT_ATTR(gmem_dl_rsp, 0x170),
+ ARM_CSPMU_EVENT_ATTR(gmem_dl_access, 0x171),
+ ARM_CSPMU_EVENT_ATTR(gmem_dl_outstanding, 0x172),
+ ARM_CSPMU_EVENT_ATTR(gmem_wb_data, 0x173),
+ ARM_CSPMU_EVENT_ATTR(gmem_wb_access, 0x174),
+ ARM_CSPMU_EVENT_ATTR(gmem_wb_outstanding, 0x175),
+ ARM_CSPMU_EVENT_ATTR(gmem_ev_rsp, 0x176),
+ ARM_CSPMU_EVENT_ATTR(gmem_ev_access, 0x177),
+ ARM_CSPMU_EVENT_ATTR(gmem_ev_outstanding, 0x178),
+ ARM_CSPMU_EVENT_ATTR(gmem_wr_data, 0x179),
+ ARM_CSPMU_EVENT_ATTR(gmem_wr_outstanding, 0x17a),
+ ARM_CSPMU_EVENT_ATTR(gmem_wr_access, 0x17b),
+
+ NV_CSPMU_EVENT_ATTR_4(socket, wr_data, 0x17c),
+
+ NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wr_data, 0x180),
+ NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wb_data, 0x184),
+ NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wr_access, 0x188),
+ NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wb_outstanding, 0x18c),
+
+ NV_CSPMU_EVENT_ATTR_4(ocu, rem_wr_data, 0x190),
+ NV_CSPMU_EVENT_ATTR_4(ocu, rem_wb_data, 0x194),
+ NV_CSPMU_EVENT_ATTR_4(ocu, rem_wr_access, 0x198),
+ NV_CSPMU_EVENT_ATTR_4(ocu, rem_wb_outstanding, 0x19c),
+
+ ARM_CSPMU_EVENT_ATTR(gmem_wr_total_bytes, 0x1a0),
+ ARM_CSPMU_EVENT_ATTR(remote_socket_wr_total_bytes, 0x1a1),
+ ARM_CSPMU_EVENT_ATTR(remote_socket_rd_data, 0x1a2),
+ ARM_CSPMU_EVENT_ATTR(remote_socket_rd_outstanding, 0x1a3),
+ ARM_CSPMU_EVENT_ATTR(remote_socket_rd_access, 0x1a4),
+
+ ARM_CSPMU_EVENT_ATTR(cmem_rd_data, 0x1a5),
+ ARM_CSPMU_EVENT_ATTR(cmem_rd_access, 0x1a6),
+ ARM_CSPMU_EVENT_ATTR(cmem_rd_outstanding, 0x1a7),
+ ARM_CSPMU_EVENT_ATTR(cmem_dl_rsp, 0x1a8),
+ ARM_CSPMU_EVENT_ATTR(cmem_dl_access, 0x1a9),
+ ARM_CSPMU_EVENT_ATTR(cmem_dl_outstanding, 0x1aa),
+ ARM_CSPMU_EVENT_ATTR(cmem_wb_data, 0x1ab),
+ ARM_CSPMU_EVENT_ATTR(cmem_wb_access, 0x1ac),
+ ARM_CSPMU_EVENT_ATTR(cmem_wb_outstanding, 0x1ad),
+ ARM_CSPMU_EVENT_ATTR(cmem_ev_rsp, 0x1ae),
+ ARM_CSPMU_EVENT_ATTR(cmem_ev_access, 0x1af),
+ ARM_CSPMU_EVENT_ATTR(cmem_ev_outstanding, 0x1b0),
+ ARM_CSPMU_EVENT_ATTR(cmem_wr_data, 0x1b1),
+ ARM_CSPMU_EVENT_ATTR(cmem_wr_outstanding, 0x1b2),
+
+ NV_CSPMU_EVENT_ATTR_4(ocu, cmem_rd_data, 0x1b3),
+ NV_CSPMU_EVENT_ATTR_4(ocu, cmem_rd_access, 0x1b7),
+ NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wb_access, 0x1bb),
+ NV_CSPMU_EVENT_ATTR_4(ocu, cmem_rd_outstanding, 0x1bf),
+ NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wr_outstanding, 0x1c3),
+
+ ARM_CSPMU_EVENT_ATTR(ocu_prb_access, 0x1c7),
+ ARM_CSPMU_EVENT_ATTR(ocu_prb_data, 0x1c8),
+ ARM_CSPMU_EVENT_ATTR(ocu_prb_outstanding, 0x1c9),
+
+ ARM_CSPMU_EVENT_ATTR(cmem_wr_access, 0x1ca),
+
+ NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wr_access, 0x1cb),
+ NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wb_data, 0x1cf),
+ NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wr_data, 0x1d3),
+ NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wb_outstanding, 0x1d7),
+
+ ARM_CSPMU_EVENT_ATTR(cmem_wr_total_bytes, 0x1db),
+
+ ARM_CSPMU_EVENT_ATTR(cycles, ARM_CSPMU_EVT_CYCLES_DEFAULT),
+ NULL,
+};
+
+static struct attribute *mcf_pmu_event_attrs[] = {
+ ARM_CSPMU_EVENT_ATTR(rd_bytes_loc, 0x0),
+ ARM_CSPMU_EVENT_ATTR(rd_bytes_rem, 0x1),
+ ARM_CSPMU_EVENT_ATTR(wr_bytes_loc, 0x2),
+ ARM_CSPMU_EVENT_ATTR(wr_bytes_rem, 0x3),
+ ARM_CSPMU_EVENT_ATTR(total_bytes_loc, 0x4),
+ ARM_CSPMU_EVENT_ATTR(total_bytes_rem, 0x5),
+ ARM_CSPMU_EVENT_ATTR(rd_req_loc, 0x6),
+ ARM_CSPMU_EVENT_ATTR(rd_req_rem, 0x7),
+ ARM_CSPMU_EVENT_ATTR(wr_req_loc, 0x8),
+ ARM_CSPMU_EVENT_ATTR(wr_req_rem, 0x9),
+ ARM_CSPMU_EVENT_ATTR(total_req_loc, 0xa),
+ ARM_CSPMU_EVENT_ATTR(total_req_rem, 0xb),
+ ARM_CSPMU_EVENT_ATTR(rd_cum_outs_loc, 0xc),
+ ARM_CSPMU_EVENT_ATTR(rd_cum_outs_rem, 0xd),
+ ARM_CSPMU_EVENT_ATTR(cycles, ARM_CSPMU_EVT_CYCLES_DEFAULT),
+ NULL,
+};
+
+static struct attribute *generic_pmu_event_attrs[] = {
+ ARM_CSPMU_EVENT_ATTR(cycles, ARM_CSPMU_EVT_CYCLES_DEFAULT),
+ NULL,
+};
+
+static struct attribute *scf_pmu_format_attrs[] = {
+ ARM_CSPMU_FORMAT_EVENT_ATTR,
+ NULL,
+};
+
+static struct attribute *mcf_pcie_pmu_format_attrs[] = {
+ ARM_CSPMU_FORMAT_EVENT_ATTR,
+ ARM_CSPMU_FORMAT_ATTR(root_port, "config1:0-9"),
+ NULL,
+};
+
+static struct attribute *mcf_gpu_pmu_format_attrs[] = {
+ ARM_CSPMU_FORMAT_EVENT_ATTR,
+ ARM_CSPMU_FORMAT_ATTR(gpu, "config1:0-1"),
+ NULL,
+};
+
+static struct attribute *mcf_nvlink_pmu_format_attrs[] = {
+ ARM_CSPMU_FORMAT_EVENT_ATTR,
+ ARM_CSPMU_FORMAT_ATTR(rem_socket, "config1:0-3"),
+ NULL,
+};
+
+static struct attribute *generic_pmu_format_attrs[] = {
+ ARM_CSPMU_FORMAT_EVENT_ATTR,
+ ARM_CSPMU_FORMAT_FILTER_ATTR,
+ NULL,
+};
+
+static struct attribute **
+nv_cspmu_get_event_attrs(const struct arm_cspmu *cspmu)
+{
+ const struct nv_cspmu_ctx *ctx = to_nv_cspmu_ctx(cspmu);
+
+ return ctx->event_attr;
+}
+
+static struct attribute **
+nv_cspmu_get_format_attrs(const struct arm_cspmu *cspmu)
+{
+ const struct nv_cspmu_ctx *ctx = to_nv_cspmu_ctx(cspmu);
+
+ return ctx->format_attr;
+}
+
+static const char *
+nv_cspmu_get_name(const struct arm_cspmu *cspmu)
+{
+ const struct nv_cspmu_ctx *ctx = to_nv_cspmu_ctx(cspmu);
+
+ return ctx->name;
+}
+
+static u32 nv_cspmu_event_filter(const struct perf_event *event)
+{
+ const struct nv_cspmu_ctx *ctx =
+ to_nv_cspmu_ctx(to_arm_cspmu(event->pmu));
+
+ return event->attr.config1 & ctx->filter_mask;
+}
+
+enum nv_cspmu_name_fmt {
+ NAME_FMT_GENERIC,
+ NAME_FMT_PROC
+};
+
+struct nv_cspmu_match {
+ u32 prodid;
+ u32 prodid_mask;
+ u64 filter_mask;
+ const char *name_pattern;
+ enum nv_cspmu_name_fmt name_fmt;
+ struct attribute **event_attr;
+ struct attribute **format_attr;
+};
+
+static const struct nv_cspmu_match nv_cspmu_match[] = {
+ { .prodid = 0x103,
+ .prodid_mask = NV_SCF_MCF_PRODID_MASK,
+ .filter_mask = NV_MCF_PCIE_FILTER_ID_MASK,
+ .name_pattern = "nvidia_mcf_pcie_pmu_%u",
+ .name_fmt = NAME_FMT_PROC,
+ .event_attr = mcf_pmu_event_attrs,
+ .format_attr = mcf_pcie_pmu_format_attrs },
+ { .prodid = 0x104,
+ .prodid_mask = NV_SCF_MCF_PRODID_MASK,
+ .filter_mask = NV_MCF_GPU_FILTER_ID_MASK,
+ .name_pattern = "nvidia_mcf_gpuvir_pmu_%u",
+ .name_fmt = NAME_FMT_PROC,
+ .event_attr = mcf_pmu_event_attrs,
+ .format_attr = mcf_gpu_pmu_format_attrs },
+ { .prodid = 0x105,
+ .prodid_mask = NV_SCF_MCF_PRODID_MASK,
+ .filter_mask = NV_MCF_GPU_FILTER_ID_MASK,
+ .name_pattern = "nvidia_mcf_gpu_pmu_%u",
+ .name_fmt = NAME_FMT_PROC,
+ .event_attr = mcf_pmu_event_attrs,
+ .format_attr = mcf_gpu_pmu_format_attrs },
+ { .prodid = 0x106,
+ .prodid_mask = NV_SCF_MCF_PRODID_MASK,
+ .filter_mask = NV_MCF_NVL_FILTER_ID_MASK,
+ .name_pattern = "nvidia_mcf_nvlink_pmu_%u",
+ .name_fmt = NAME_FMT_PROC,
+ .event_attr = mcf_pmu_event_attrs,
+ .format_attr = mcf_nvlink_pmu_format_attrs },
+ { .prodid = 0x2CF,
+ .prodid_mask = NV_SCF_MCF_PRODID_MASK,
+ .filter_mask = 0x0,
+ .name_pattern = "nvidia_scf_pmu_%u",
+ .name_fmt = NAME_FMT_PROC,
+ .event_attr = scf_pmu_event_attrs,
+ .format_attr = scf_pmu_format_attrs },
+ { .prodid = 0,
+ .prodid_mask = 0,
+ .filter_mask = ARM_CSPMU_FILTER_MASK,
+ .name_pattern = "nvidia_uncore_pmu_%u",
+ .name_fmt = NAME_FMT_GENERIC,
+ .event_attr = generic_pmu_event_attrs,
+ .format_attr = generic_pmu_format_attrs },
+};
+
+static char *nv_cspmu_format_name(const struct arm_cspmu *cspmu,
+ const struct nv_cspmu_match *match)
+{
+ char *name;
+ struct device *dev = cspmu->dev;
+
+ static atomic_t pmu_generic_idx = {0};
+
+ switch (match->name_fmt) {
+ case NAME_FMT_PROC:
+ name = devm_kasprintf(dev, GFP_KERNEL, match->name_pattern,
+ cspmu->apmt_node->proc_affinity);
+ break;
+ case NAME_FMT_GENERIC:
+ name = devm_kasprintf(dev, GFP_KERNEL, match->name_pattern,
+ atomic_fetch_inc(&pmu_generic_idx));
+ break;
+ default:
+ name = NULL;
+ break;
+ }
+
+ return name;
+}
+
+int nv_cspmu_init_ops(struct arm_cspmu *cspmu)
+{
+ u32 prodid;
+ struct nv_cspmu_ctx *ctx;
+ struct device *dev = cspmu->dev;
+ struct arm_cspmu_impl_ops *impl_ops = &cspmu->impl.ops;
+ const struct nv_cspmu_match *match = nv_cspmu_match;
+
+ ctx = devm_kzalloc(dev, sizeof(struct nv_cspmu_ctx), GFP_KERNEL);
+ if (!ctx)
+ return -ENOMEM;
+
+ prodid = FIELD_GET(ARM_CSPMU_PMIIDR_PRODUCTID, cspmu->impl.pmiidr);
+
+ /* Find matching PMU. */
+ for (; match->prodid; match++) {
+ const u32 prodid_mask = match->prodid_mask;
+
+ if ((match->prodid & prodid_mask) == (prodid & prodid_mask))
+ break;
+ }
+
+ ctx->name = nv_cspmu_format_name(cspmu, match);
+ ctx->filter_mask = match->filter_mask;
+ ctx->event_attr = match->event_attr;
+ ctx->format_attr = match->format_attr;
+
+ cspmu->impl.ctx = ctx;
+
+ /* NVIDIA specific callbacks. */
+ impl_ops->event_filter = nv_cspmu_event_filter;
+ impl_ops->get_event_attrs = nv_cspmu_get_event_attrs;
+ impl_ops->get_format_attrs = nv_cspmu_get_format_attrs;
+ impl_ops->get_name = nv_cspmu_get_name;
+
+ /* Set others to NULL to use default callback. */
+ impl_ops->event_type = NULL;
+ impl_ops->event_attr_is_visible = NULL;
+ impl_ops->get_identifier = NULL;
+ impl_ops->is_cycle_counter_event = NULL;
+
+ return 0;
+}
+EXPORT_SYMBOL_GPL(nv_cspmu_init_ops);
diff --git a/drivers/perf/arm_cspmu/nvidia_cspmu.h b/drivers/perf/arm_cspmu/nvidia_cspmu.h
new file mode 100644
index 000000000000..eefba85644f6
--- /dev/null
+++ b/drivers/perf/arm_cspmu/nvidia_cspmu.h
@@ -0,0 +1,17 @@
+/* SPDX-License-Identifier: GPL-2.0
+ *
+ * Copyright (c) 2022, NVIDIA CORPORATION & AFFILIATES.
+ *
+ */
+
+/* Support for NVIDIA specific attributes. */
+
+#ifndef __NVIDIA_CSPMU_H__
+#define __NVIDIA_CSPMU_H__
+
+#include "arm_cspmu.h"
+
+/* Allocate NVIDIA descriptor. */
+int nv_cspmu_init_ops(struct arm_cspmu *cspmu);
+
+#endif /* __NVIDIA_CSPMU_H__ */
--
2.17.1


2022-09-27 12:39:30

by Suzuki K Poulose

[permalink] [raw]
Subject: Re: [PATCH v4 2/2] perf: arm_cspmu: Add support for NVIDIA SCF and MCF attribute

On 14/08/2022 19:23, Besar Wicaksono wrote:
> Add support for NVIDIA System Cache Fabric (SCF) and Memory Control
> Fabric (MCF) PMU attributes for CoreSight PMU implementation in
> NVIDIA devices.
>
> Signed-off-by: Besar Wicaksono <[email protected]>
> ---
> Documentation/admin-guide/perf/index.rst | 1 +
> Documentation/admin-guide/perf/nvidia-pmu.rst | 120 ++++++
> drivers/perf/arm_cspmu/Makefile | 3 +-
> drivers/perf/arm_cspmu/arm_cspmu.c | 7 +
> drivers/perf/arm_cspmu/nvidia_cspmu.c | 367 ++++++++++++++++++
> drivers/perf/arm_cspmu/nvidia_cspmu.h | 17 +
> 6 files changed, 514 insertions(+), 1 deletion(-)
> create mode 100644 Documentation/admin-guide/perf/nvidia-pmu.rst
> create mode 100644 drivers/perf/arm_cspmu/nvidia_cspmu.c
> create mode 100644 drivers/perf/arm_cspmu/nvidia_cspmu.h
>
> diff --git a/Documentation/admin-guide/perf/index.rst b/Documentation/admin-guide/perf/index.rst
> index 69b23f087c05..cf05fed1f67f 100644
> --- a/Documentation/admin-guide/perf/index.rst
> +++ b/Documentation/admin-guide/perf/index.rst
> @@ -17,3 +17,4 @@ Performance monitor support
> xgene-pmu
> arm_dsu_pmu
> thunderx2-pmu
> + nvidia-pmu
> diff --git a/Documentation/admin-guide/perf/nvidia-pmu.rst b/Documentation/admin-guide/perf/nvidia-pmu.rst
> new file mode 100644
> index 000000000000..c41b93965824
> --- /dev/null
> +++ b/Documentation/admin-guide/perf/nvidia-pmu.rst
> @@ -0,0 +1,120 @@
> +=========================================================
> +NVIDIA Tegra SoC Uncore Performance Monitoring Unit (PMU)
> +=========================================================
> +
> +The NVIDIA Tegra SoC includes various system PMUs to measure key performance
> +metrics like memory bandwidth, latency, and utilization:
> +
> +* Scalable Coherency Fabric (SCF)
> +* Memory Controller Fabric (MCF) GPU physical interface
> +* MCF GPU virtual interface
> +* MCF NVLINK interface
> +* MCF PCIE interface
> +
> +PMU Driver
> +----------
> +
> +The PMUs in this document are based on ARM CoreSight PMU Architecture as
> +described in document: ARM IHI 0091. Since this is a standard architecture, the
> +PMUs are managed by a common driver "arm-cs-arch-pmu". This driver describes
> +the available events and configuration of each PMU in sysfs. Please see the
> +sections below to get the sysfs path of each PMU. Like other uncore PMU driver,
> +the driver provides "cpumask" sysfs attribute to show the CPU id used to handle
> +the PMU event. There is also "associated_cpus" sysfs attribute, which contains a
> +list of CPUs associated with the PMU instance.
> +
> +SCF PMU
> +-------
> +
> +The SCF PMU monitors system level cache events, CPU traffic, and
> +strongly-ordered PCIE traffic to local/remote memory.
> +
> +The events and configuration options of this PMU device are described in sysfs,
> +see /sys/bus/event_sources/devices/nvidia_scf_pmu_<socket-id>.
> +
> +Example usage::
> +
> + perf stat -a -e nvidia_scf_pmu_0/config=0x0/
> +
> +This will count the events in socket 0.
> +
> +MCF GPU Physical PMU
> +--------------------
> +
> +The MCF GPU physical PMU monitors ATS translated traffic from GPU to
> +local/remote memory via Nvlink C2C.
> +
> +The events and configuration options of this PMU device are described in sysfs,
> +see /sys/bus/event_sources/devices/nvidia_mcf_gpu_pmu_<socket-id>.
> +
> +Multiple GPUs can be connected to the SoC. The user can use "gpu" bitmap
> +parameter to select the GPU(s) to monitor, i.e. "gpu=0xF" corresponds to GPU 0
> +to 3. /sys/bus/event_sources/devices/nvidia_mcf_gpu_pmu_<socket-id>/format/gpu
> +shows the valid bits that can be set in the "gpu" parameter.
> +
> +Example usage::
> +
> + perf stat -a -e nvidia_mcf_gpu_pmu_0/config=0x0,gpu=0x3/
> +
> +This will count the events on GPU 0 and 1 that are connected to SoC in socket 0.
> +
> +MCF GPU Virtual PMU
> +-------------------
> +
> +The MCF GPU virtual PMU monitors SMMU inline translated traffic (as opposed to
> +ATS) from GPU to local/remote memory via Nvlink C2C.
> +
> +The events and configuration options of this PMU device are described in sysfs,
> +see /sys/bus/event_sources/devices/nvidia_mcf_gpuvir_pmu_<socket-id>.
> +
> +Multiple GPUs can be connected to the SoC. The user can use "gpu" bitmap
> +parameter to select the GPU(s) to monitor, i.e. "gpu=0xF" corresponds to GPU 0
> +to 3. /sys/bus/event_sources/devices/nvidia_mcf_gpuvir_pmu_<socket-id>/format/gpu
> +shows the valid bits that can be set in the "gpu" parameter.
> +
> +Example usage::
> +
> + perf stat -a -e nvidia_mcf_gpuvir_pmu_0/config=0x0,gpu=0x3/
> +
> +This will count the events on GPU 0 and 1 that are connected to SoC in socket 0.
> +
> +MCF NVLINK PMU
> +--------------
> +
> +The MCF NVLINK PMU monitors I/O coherent traffic from external socket to local
> +memory.
> +
> +The events and configuration options of this PMU device are described in sysfs,
> +see /sys/bus/event_sources/devices/nvidia_mcf_nvlink_pmu_<socket-id>.
> +
> +Each SoC socket can be connected to one or more sockets via NVLINK. The user can
> +use "rem_socket" bitmap parameter to select the remote socket(s) to monitor,
> +i.e. "rem_socket=0xE" corresponds to socket 1 to 3.
> +/sys/bus/event_sources/devices/nvidia_mcf_nvlink_pmu_<socket-id>/format/rem_socket
> +shows the valid bits that can be set in the "rem_socket" parameter.
> +
> +Example usage::
> +
> + perf stat -a -e nvidia_mcf_nvlink_pmu_0/config=0x0,rem_socket=0x6/
> +
> +This will count the events from remote socket 1 and 2 to socket 0.
> +
> +MCF PCIE PMU
> +------------
> +
> +The MCF PCIE PMU monitors traffic from PCIE root ports to local/remote memory.
> +
> +The events and configuration options of this PMU device are described in sysfs,
> +see /sys/bus/event_sources/devices/nvidia_mcf_pcie_pmu_<socket-id>.
> +
> +Each SoC socket can support multiple root ports. The user can use
> +"root_port" bitmap parameter to select the port(s) to monitor, i.e.
> +"root_port=0xF" corresponds to root port 0 to 3.
> +/sys/bus/event_sources/devices/nvidia_mcf_pcie_pmu_<socket-id>/format/root_port
> +shows the valid bits that can be set in the "root_port" parameter.
> +
> +Example usage::
> +
> + perf stat -a -e nvidia_mcf_pcie_pmu_0/config=0x0,root_port=0x3/
> +
> +This will count the events from root port 0 and 1 of socket 0.
> diff --git a/drivers/perf/arm_cspmu/Makefile b/drivers/perf/arm_cspmu/Makefile
> index cdc3455f74d8..1b586064bd77 100644
> --- a/drivers/perf/arm_cspmu/Makefile
> +++ b/drivers/perf/arm_cspmu/Makefile
> @@ -3,4 +3,5 @@
> # SPDX-License-Identifier: GPL-2.0
>
> obj-$(CONFIG_ARM_CORESIGHT_PMU_ARCH_SYSTEM_PMU) += \
> - arm_cspmu.o
> + arm_cspmu.o \
> + nvidia_cspmu.o
> diff --git a/drivers/perf/arm_cspmu/arm_cspmu.c b/drivers/perf/arm_cspmu/arm_cspmu.c
> index 410876f86eb0..7a0beb515e53 100644
> --- a/drivers/perf/arm_cspmu/arm_cspmu.c
> +++ b/drivers/perf/arm_cspmu/arm_cspmu.c
> @@ -31,6 +31,7 @@
> #include <acpi/processor.h>
>
> #include "arm_cspmu.h"
> +#include "nvidia_cspmu.h"
>
> #define PMUNAME "arm_cspmu"
> #define DRVNAME "arm-cs-arch-pmu"
> @@ -118,6 +119,9 @@ static_assert(
> ops->callback = arm_cspmu_ ## callback; \
> } while (0)
>
> +/* JEDEC-assigned JEP106 identification code */
> +#define ARM_CSPMU_IMPL_ID_NVIDIA 0x36B
> +
> static unsigned long arm_cspmu_cpuhp_state;
>
> /*
> @@ -369,6 +373,9 @@ struct impl_match {
> };
>
> static const struct impl_match impl_match[] = {
> + { .pmiidr = ARM_CSPMU_IMPL_ID_NVIDIA,
> + .mask = ARM_CSPMU_PMIIDR_IMPLEMENTER,
> + .impl_init_ops = nv_cspmu_init_ops },

Super minor nit: Coding style. Could we use :

{
.field = value,
...
},

> {}
> };
>
> diff --git a/drivers/perf/arm_cspmu/nvidia_cspmu.c b/drivers/perf/arm_cspmu/nvidia_cspmu.c
> new file mode 100644
> index 000000000000..261f20680bc1
> --- /dev/null
> +++ b/drivers/perf/arm_cspmu/nvidia_cspmu.c
> @@ -0,0 +1,367 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (c) 2022, NVIDIA CORPORATION & AFFILIATES.
> + *
> + */
> +
> +/* Support for NVIDIA specific attributes. */
> +
> +#include "nvidia_cspmu.h"
> +
> +#define NV_MCF_PCIE_PORT_COUNT 10ULL
> +#define NV_MCF_PCIE_FILTER_ID_MASK GENMASK_ULL(NV_MCF_PCIE_PORT_COUNT - 1, 0)
> +
> +#define NV_MCF_GPU_PORT_COUNT 2ULL
> +#define NV_MCF_GPU_FILTER_ID_MASK GENMASK_ULL(NV_MCF_GPU_PORT_COUNT - 1, 0)
> +
> +#define NV_MCF_NVL_PORT_COUNT 4ULL
> +#define NV_MCF_NVL_FILTER_ID_MASK GENMASK_ULL(NV_MCF_NVL_PORT_COUNT - 1, 0)
> +
> +#define NV_SCF_MCF_PRODID_MASK GENMASK(31, 0)
> +
> +#define NV_FORMAT_NAME_GENERIC 0
> +
> +#define to_nv_cspmu_ctx(cspmu) ((struct nv_cspmu_ctx *)(cspmu->impl.ctx))
> +
> +#define NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _num, _suff, _config) \
> + ARM_CSPMU_EVENT_ATTR(_pref##_num##_suff, _config)
> +
> +#define NV_CSPMU_EVENT_ATTR_4(_pref, _suff, _config) \
> + NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _0_, _suff, _config), \
> + NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _1_, _suff, _config + 1), \
> + NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _2_, _suff, _config + 2), \
> + NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _3_, _suff, _config + 3)
> +
> +struct nv_cspmu_ctx {
> + const char *name;
> + u32 filter_mask;
> + struct attribute **event_attr;
> + struct attribute **format_attr;
> +};
> +
> +static struct attribute *scf_pmu_event_attrs[] = {
> + ARM_CSPMU_EVENT_ATTR(bus_cycles, 0x1d),
> +
> + ARM_CSPMU_EVENT_ATTR(scf_cache_allocate, 0xF0),
> + ARM_CSPMU_EVENT_ATTR(scf_cache_refill, 0xF1),
> + ARM_CSPMU_EVENT_ATTR(scf_cache, 0xF2),
> + ARM_CSPMU_EVENT_ATTR(scf_cache_wb, 0xF3),
> +
> + NV_CSPMU_EVENT_ATTR_4(socket, rd_data, 0x101),
> + NV_CSPMU_EVENT_ATTR_4(socket, dl_rsp, 0x105),
> + NV_CSPMU_EVENT_ATTR_4(socket, wb_data, 0x109),
> + NV_CSPMU_EVENT_ATTR_4(socket, ev_rsp, 0x10d),
> + NV_CSPMU_EVENT_ATTR_4(socket, prb_data, 0x111),
> +
> + NV_CSPMU_EVENT_ATTR_4(socket, rd_outstanding, 0x115),
> + NV_CSPMU_EVENT_ATTR_4(socket, dl_outstanding, 0x119),
> + NV_CSPMU_EVENT_ATTR_4(socket, wb_outstanding, 0x11d),
> + NV_CSPMU_EVENT_ATTR_4(socket, wr_outstanding, 0x121),
> + NV_CSPMU_EVENT_ATTR_4(socket, ev_outstanding, 0x125),
> + NV_CSPMU_EVENT_ATTR_4(socket, prb_outstanding, 0x129),
> +
> + NV_CSPMU_EVENT_ATTR_4(socket, rd_access, 0x12d),
> + NV_CSPMU_EVENT_ATTR_4(socket, dl_access, 0x131),
> + NV_CSPMU_EVENT_ATTR_4(socket, wb_access, 0x135),
> + NV_CSPMU_EVENT_ATTR_4(socket, wr_access, 0x139),
> + NV_CSPMU_EVENT_ATTR_4(socket, ev_access, 0x13d),
> + NV_CSPMU_EVENT_ATTR_4(socket, prb_access, 0x141),
> +
> + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_rd_data, 0x145),
> + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_rd_access, 0x149),
> + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wb_access, 0x14d),
> + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_rd_outstanding, 0x151),
> + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wr_outstanding, 0x155),
> +
> + NV_CSPMU_EVENT_ATTR_4(ocu, rem_rd_data, 0x159),
> + NV_CSPMU_EVENT_ATTR_4(ocu, rem_rd_access, 0x15d),
> + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wb_access, 0x161),
> + NV_CSPMU_EVENT_ATTR_4(ocu, rem_rd_outstanding, 0x165),
> + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wr_outstanding, 0x169),
> +
> + ARM_CSPMU_EVENT_ATTR(gmem_rd_data, 0x16d),
> + ARM_CSPMU_EVENT_ATTR(gmem_rd_access, 0x16e),
> + ARM_CSPMU_EVENT_ATTR(gmem_rd_outstanding, 0x16f),
> + ARM_CSPMU_EVENT_ATTR(gmem_dl_rsp, 0x170),
> + ARM_CSPMU_EVENT_ATTR(gmem_dl_access, 0x171),
> + ARM_CSPMU_EVENT_ATTR(gmem_dl_outstanding, 0x172),
> + ARM_CSPMU_EVENT_ATTR(gmem_wb_data, 0x173),
> + ARM_CSPMU_EVENT_ATTR(gmem_wb_access, 0x174),
> + ARM_CSPMU_EVENT_ATTR(gmem_wb_outstanding, 0x175),
> + ARM_CSPMU_EVENT_ATTR(gmem_ev_rsp, 0x176),
> + ARM_CSPMU_EVENT_ATTR(gmem_ev_access, 0x177),
> + ARM_CSPMU_EVENT_ATTR(gmem_ev_outstanding, 0x178),
> + ARM_CSPMU_EVENT_ATTR(gmem_wr_data, 0x179),
> + ARM_CSPMU_EVENT_ATTR(gmem_wr_outstanding, 0x17a),
> + ARM_CSPMU_EVENT_ATTR(gmem_wr_access, 0x17b),
> +
> + NV_CSPMU_EVENT_ATTR_4(socket, wr_data, 0x17c),
> +
> + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wr_data, 0x180),
> + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wb_data, 0x184),
> + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wr_access, 0x188),
> + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wb_outstanding, 0x18c),
> +
> + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wr_data, 0x190),
> + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wb_data, 0x194),
> + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wr_access, 0x198),
> + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wb_outstanding, 0x19c),
> +
> + ARM_CSPMU_EVENT_ATTR(gmem_wr_total_bytes, 0x1a0),
> + ARM_CSPMU_EVENT_ATTR(remote_socket_wr_total_bytes, 0x1a1),
> + ARM_CSPMU_EVENT_ATTR(remote_socket_rd_data, 0x1a2),
> + ARM_CSPMU_EVENT_ATTR(remote_socket_rd_outstanding, 0x1a3),
> + ARM_CSPMU_EVENT_ATTR(remote_socket_rd_access, 0x1a4),
> +
> + ARM_CSPMU_EVENT_ATTR(cmem_rd_data, 0x1a5),
> + ARM_CSPMU_EVENT_ATTR(cmem_rd_access, 0x1a6),
> + ARM_CSPMU_EVENT_ATTR(cmem_rd_outstanding, 0x1a7),
> + ARM_CSPMU_EVENT_ATTR(cmem_dl_rsp, 0x1a8),
> + ARM_CSPMU_EVENT_ATTR(cmem_dl_access, 0x1a9),
> + ARM_CSPMU_EVENT_ATTR(cmem_dl_outstanding, 0x1aa),
> + ARM_CSPMU_EVENT_ATTR(cmem_wb_data, 0x1ab),
> + ARM_CSPMU_EVENT_ATTR(cmem_wb_access, 0x1ac),
> + ARM_CSPMU_EVENT_ATTR(cmem_wb_outstanding, 0x1ad),
> + ARM_CSPMU_EVENT_ATTR(cmem_ev_rsp, 0x1ae),
> + ARM_CSPMU_EVENT_ATTR(cmem_ev_access, 0x1af),
> + ARM_CSPMU_EVENT_ATTR(cmem_ev_outstanding, 0x1b0),
> + ARM_CSPMU_EVENT_ATTR(cmem_wr_data, 0x1b1),
> + ARM_CSPMU_EVENT_ATTR(cmem_wr_outstanding, 0x1b2),
> +
> + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_rd_data, 0x1b3),
> + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_rd_access, 0x1b7),
> + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wb_access, 0x1bb),
> + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_rd_outstanding, 0x1bf),
> + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wr_outstanding, 0x1c3),
> +
> + ARM_CSPMU_EVENT_ATTR(ocu_prb_access, 0x1c7),
> + ARM_CSPMU_EVENT_ATTR(ocu_prb_data, 0x1c8),
> + ARM_CSPMU_EVENT_ATTR(ocu_prb_outstanding, 0x1c9),
> +
> + ARM_CSPMU_EVENT_ATTR(cmem_wr_access, 0x1ca),
> +
> + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wr_access, 0x1cb),
> + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wb_data, 0x1cf),
> + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wr_data, 0x1d3),
> + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wb_outstanding, 0x1d7),
> +
> + ARM_CSPMU_EVENT_ATTR(cmem_wr_total_bytes, 0x1db),
> +
> + ARM_CSPMU_EVENT_ATTR(cycles, ARM_CSPMU_EVT_CYCLES_DEFAULT),
> + NULL,
> +};
> +
> +static struct attribute *mcf_pmu_event_attrs[] = {
> + ARM_CSPMU_EVENT_ATTR(rd_bytes_loc, 0x0),
> + ARM_CSPMU_EVENT_ATTR(rd_bytes_rem, 0x1),
> + ARM_CSPMU_EVENT_ATTR(wr_bytes_loc, 0x2),
> + ARM_CSPMU_EVENT_ATTR(wr_bytes_rem, 0x3),
> + ARM_CSPMU_EVENT_ATTR(total_bytes_loc, 0x4),
> + ARM_CSPMU_EVENT_ATTR(total_bytes_rem, 0x5),
> + ARM_CSPMU_EVENT_ATTR(rd_req_loc, 0x6),
> + ARM_CSPMU_EVENT_ATTR(rd_req_rem, 0x7),
> + ARM_CSPMU_EVENT_ATTR(wr_req_loc, 0x8),
> + ARM_CSPMU_EVENT_ATTR(wr_req_rem, 0x9),
> + ARM_CSPMU_EVENT_ATTR(total_req_loc, 0xa),
> + ARM_CSPMU_EVENT_ATTR(total_req_rem, 0xb),
> + ARM_CSPMU_EVENT_ATTR(rd_cum_outs_loc, 0xc),
> + ARM_CSPMU_EVENT_ATTR(rd_cum_outs_rem, 0xd),
> + ARM_CSPMU_EVENT_ATTR(cycles, ARM_CSPMU_EVT_CYCLES_DEFAULT),
> + NULL,
> +};
> +
> +static struct attribute *generic_pmu_event_attrs[] = {
> + ARM_CSPMU_EVENT_ATTR(cycles, ARM_CSPMU_EVT_CYCLES_DEFAULT),
> + NULL,
> +};
> +
> +static struct attribute *scf_pmu_format_attrs[] = {
> + ARM_CSPMU_FORMAT_EVENT_ATTR,
> + NULL,
> +};
> +
> +static struct attribute *mcf_pcie_pmu_format_attrs[] = {
> + ARM_CSPMU_FORMAT_EVENT_ATTR,
> + ARM_CSPMU_FORMAT_ATTR(root_port, "config1:0-9"),
> + NULL,
> +};
> +
> +static struct attribute *mcf_gpu_pmu_format_attrs[] = {
> + ARM_CSPMU_FORMAT_EVENT_ATTR,
> + ARM_CSPMU_FORMAT_ATTR(gpu, "config1:0-1"),
> + NULL,
> +};
> +
> +static struct attribute *mcf_nvlink_pmu_format_attrs[] = {
> + ARM_CSPMU_FORMAT_EVENT_ATTR,
> + ARM_CSPMU_FORMAT_ATTR(rem_socket, "config1:0-3"),
> + NULL,
> +};
> +
> +static struct attribute *generic_pmu_format_attrs[] = {
> + ARM_CSPMU_FORMAT_EVENT_ATTR,
> + ARM_CSPMU_FORMAT_FILTER_ATTR,
> + NULL,
> +};
> +
> +static struct attribute **
> +nv_cspmu_get_event_attrs(const struct arm_cspmu *cspmu)
> +{
> + const struct nv_cspmu_ctx *ctx = to_nv_cspmu_ctx(cspmu);
> +
> + return ctx->event_attr;
> +}
> +
> +static struct attribute **
> +nv_cspmu_get_format_attrs(const struct arm_cspmu *cspmu)
> +{
> + const struct nv_cspmu_ctx *ctx = to_nv_cspmu_ctx(cspmu);
> +
> + return ctx->format_attr;
> +}
> +
> +static const char *
> +nv_cspmu_get_name(const struct arm_cspmu *cspmu)
> +{
> + const struct nv_cspmu_ctx *ctx = to_nv_cspmu_ctx(cspmu);
> +
> + return ctx->name;
> +}
> +
> +static u32 nv_cspmu_event_filter(const struct perf_event *event)
> +{
> + const struct nv_cspmu_ctx *ctx =
> + to_nv_cspmu_ctx(to_arm_cspmu(event->pmu));
> +
> + return event->attr.config1 & ctx->filter_mask;
> +}
> +
> +enum nv_cspmu_name_fmt {
> + NAME_FMT_GENERIC,
> + NAME_FMT_PROC
> +};
> +
> +struct nv_cspmu_match {
> + u32 prodid;
> + u32 prodid_mask;
> + u64 filter_mask;
> + const char *name_pattern;
> + enum nv_cspmu_name_fmt name_fmt;
> + struct attribute **event_attr;
> + struct attribute **format_attr;
> +};
> +
> +static const struct nv_cspmu_match nv_cspmu_match[] = {

Similar coding style nit below.


Otherwise,

Acked-by: Suzuki K Poulose <[email protected]>

2022-09-28 02:18:29

by Besar Wicaksono

[permalink] [raw]
Subject: RE: [PATCH v4 2/2] perf: arm_cspmu: Add support for NVIDIA SCF and MCF attribute



> -----Original Message-----
> From: Suzuki K Poulose <[email protected]>
> Sent: Tuesday, September 27, 2022 6:43 AM
> To: Besar Wicaksono <[email protected]>; [email protected];
> [email protected]; [email protected]; [email protected]
> Cc: [email protected]; [email protected];
> [email protected]; [email protected];
> [email protected]; [email protected]; Thierry Reding
> <[email protected]>; Jonathan Hunter <[email protected]>; Vikram
> Sethi <[email protected]>; [email protected];
> [email protected]; [email protected]
> Subject: Re: [PATCH v4 2/2] perf: arm_cspmu: Add support for NVIDIA SCF
> and MCF attribute
>
> External email: Use caution opening links or attachments
>
>
> On 14/08/2022 19:23, Besar Wicaksono wrote:
> > Add support for NVIDIA System Cache Fabric (SCF) and Memory Control
> > Fabric (MCF) PMU attributes for CoreSight PMU implementation in
> > NVIDIA devices.
> >
> > Signed-off-by: Besar Wicaksono <[email protected]>
> > ---
> > Documentation/admin-guide/perf/index.rst | 1 +
> > Documentation/admin-guide/perf/nvidia-pmu.rst | 120 ++++++
> > drivers/perf/arm_cspmu/Makefile | 3 +-
> > drivers/perf/arm_cspmu/arm_cspmu.c | 7 +
> > drivers/perf/arm_cspmu/nvidia_cspmu.c | 367
> ++++++++++++++++++
> > drivers/perf/arm_cspmu/nvidia_cspmu.h | 17 +
> > 6 files changed, 514 insertions(+), 1 deletion(-)
> > create mode 100644 Documentation/admin-guide/perf/nvidia-pmu.rst
> > create mode 100644 drivers/perf/arm_cspmu/nvidia_cspmu.c
> > create mode 100644 drivers/perf/arm_cspmu/nvidia_cspmu.h
> >
> > diff --git a/Documentation/admin-guide/perf/index.rst
> b/Documentation/admin-guide/perf/index.rst
> > index 69b23f087c05..cf05fed1f67f 100644
> > --- a/Documentation/admin-guide/perf/index.rst
> > +++ b/Documentation/admin-guide/perf/index.rst
> > @@ -17,3 +17,4 @@ Performance monitor support
> > xgene-pmu
> > arm_dsu_pmu
> > thunderx2-pmu
> > + nvidia-pmu
> > diff --git a/Documentation/admin-guide/perf/nvidia-pmu.rst
> b/Documentation/admin-guide/perf/nvidia-pmu.rst
> > new file mode 100644
> > index 000000000000..c41b93965824
> > --- /dev/null
> > +++ b/Documentation/admin-guide/perf/nvidia-pmu.rst
> > @@ -0,0 +1,120 @@
> >
> +=========================================================
> > +NVIDIA Tegra SoC Uncore Performance Monitoring Unit (PMU)
> >
> +=========================================================
> > +
> > +The NVIDIA Tegra SoC includes various system PMUs to measure key
> performance
> > +metrics like memory bandwidth, latency, and utilization:
> > +
> > +* Scalable Coherency Fabric (SCF)
> > +* Memory Controller Fabric (MCF) GPU physical interface
> > +* MCF GPU virtual interface
> > +* MCF NVLINK interface
> > +* MCF PCIE interface
> > +
> > +PMU Driver
> > +----------
> > +
> > +The PMUs in this document are based on ARM CoreSight PMU
> Architecture as
> > +described in document: ARM IHI 0091. Since this is a standard
> architecture, the
> > +PMUs are managed by a common driver "arm-cs-arch-pmu". This driver
> describes
> > +the available events and configuration of each PMU in sysfs. Please see
> the
> > +sections below to get the sysfs path of each PMU. Like other uncore PMU
> driver,
> > +the driver provides "cpumask" sysfs attribute to show the CPU id used to
> handle
> > +the PMU event. There is also "associated_cpus" sysfs attribute, which
> contains a
> > +list of CPUs associated with the PMU instance.
> > +
> > +SCF PMU
> > +-------
> > +
> > +The SCF PMU monitors system level cache events, CPU traffic, and
> > +strongly-ordered PCIE traffic to local/remote memory.
> > +
> > +The events and configuration options of this PMU device are described in
> sysfs,
> > +see /sys/bus/event_sources/devices/nvidia_scf_pmu_<socket-id>.
> > +
> > +Example usage::
> > +
> > + perf stat -a -e nvidia_scf_pmu_0/config=0x0/
> > +
> > +This will count the events in socket 0.
> > +
> > +MCF GPU Physical PMU
> > +--------------------
> > +
> > +The MCF GPU physical PMU monitors ATS translated traffic from GPU to
> > +local/remote memory via Nvlink C2C.
> > +
> > +The events and configuration options of this PMU device are described in
> sysfs,
> > +see /sys/bus/event_sources/devices/nvidia_mcf_gpu_pmu_<socket-id>.
> > +
> > +Multiple GPUs can be connected to the SoC. The user can use "gpu"
> bitmap
> > +parameter to select the GPU(s) to monitor, i.e. "gpu=0xF" corresponds to
> GPU 0
> > +to 3. /sys/bus/event_sources/devices/nvidia_mcf_gpu_pmu_<socket-
> id>/format/gpu
> > +shows the valid bits that can be set in the "gpu" parameter.
> > +
> > +Example usage::
> > +
> > + perf stat -a -e nvidia_mcf_gpu_pmu_0/config=0x0,gpu=0x3/
> > +
> > +This will count the events on GPU 0 and 1 that are connected to SoC in
> socket 0.
> > +
> > +MCF GPU Virtual PMU
> > +-------------------
> > +
> > +The MCF GPU virtual PMU monitors SMMU inline translated traffic (as
> opposed to
> > +ATS) from GPU to local/remote memory via Nvlink C2C.
> > +
> > +The events and configuration options of this PMU device are described in
> sysfs,
> > +see /sys/bus/event_sources/devices/nvidia_mcf_gpuvir_pmu_<socket-
> id>.
> > +
> > +Multiple GPUs can be connected to the SoC. The user can use "gpu"
> bitmap
> > +parameter to select the GPU(s) to monitor, i.e. "gpu=0xF" corresponds to
> GPU 0
> > +to 3. /sys/bus/event_sources/devices/nvidia_mcf_gpuvir_pmu_<socket-
> id>/format/gpu
> > +shows the valid bits that can be set in the "gpu" parameter.
> > +
> > +Example usage::
> > +
> > + perf stat -a -e nvidia_mcf_gpuvir_pmu_0/config=0x0,gpu=0x3/
> > +
> > +This will count the events on GPU 0 and 1 that are connected to SoC in
> socket 0.
> > +
> > +MCF NVLINK PMU
> > +--------------
> > +
> > +The MCF NVLINK PMU monitors I/O coherent traffic from external socket
> to local
> > +memory.
> > +
> > +The events and configuration options of this PMU device are described in
> sysfs,
> > +see /sys/bus/event_sources/devices/nvidia_mcf_nvlink_pmu_<socket-
> id>.
> > +
> > +Each SoC socket can be connected to one or more sockets via NVLINK.
> The user can
> > +use "rem_socket" bitmap parameter to select the remote socket(s) to
> monitor,
> > +i.e. "rem_socket=0xE" corresponds to socket 1 to 3.
> > +/sys/bus/event_sources/devices/nvidia_mcf_nvlink_pmu_<socket-
> id>/format/rem_socket
> > +shows the valid bits that can be set in the "rem_socket" parameter.
> > +
> > +Example usage::
> > +
> > + perf stat -a -e nvidia_mcf_nvlink_pmu_0/config=0x0,rem_socket=0x6/
> > +
> > +This will count the events from remote socket 1 and 2 to socket 0.
> > +
> > +MCF PCIE PMU
> > +------------
> > +
> > +The MCF PCIE PMU monitors traffic from PCIE root ports to local/remote
> memory.
> > +
> > +The events and configuration options of this PMU device are described in
> sysfs,
> > +see /sys/bus/event_sources/devices/nvidia_mcf_pcie_pmu_<socket-
> id>.
> > +
> > +Each SoC socket can support multiple root ports. The user can use
> > +"root_port" bitmap parameter to select the port(s) to monitor, i.e.
> > +"root_port=0xF" corresponds to root port 0 to 3.
> > +/sys/bus/event_sources/devices/nvidia_mcf_pcie_pmu_<socket-
> id>/format/root_port
> > +shows the valid bits that can be set in the "root_port" parameter.
> > +
> > +Example usage::
> > +
> > + perf stat -a -e nvidia_mcf_pcie_pmu_0/config=0x0,root_port=0x3/
> > +
> > +This will count the events from root port 0 and 1 of socket 0.
> > diff --git a/drivers/perf/arm_cspmu/Makefile
> b/drivers/perf/arm_cspmu/Makefile
> > index cdc3455f74d8..1b586064bd77 100644
> > --- a/drivers/perf/arm_cspmu/Makefile
> > +++ b/drivers/perf/arm_cspmu/Makefile
> > @@ -3,4 +3,5 @@
> > # SPDX-License-Identifier: GPL-2.0
> >
> > obj-$(CONFIG_ARM_CORESIGHT_PMU_ARCH_SYSTEM_PMU) += \
> > - arm_cspmu.o
> > + arm_cspmu.o \
> > + nvidia_cspmu.o
> > diff --git a/drivers/perf/arm_cspmu/arm_cspmu.c
> b/drivers/perf/arm_cspmu/arm_cspmu.c
> > index 410876f86eb0..7a0beb515e53 100644
> > --- a/drivers/perf/arm_cspmu/arm_cspmu.c
> > +++ b/drivers/perf/arm_cspmu/arm_cspmu.c
> > @@ -31,6 +31,7 @@
> > #include <acpi/processor.h>
> >
> > #include "arm_cspmu.h"
> > +#include "nvidia_cspmu.h"
> >
> > #define PMUNAME "arm_cspmu"
> > #define DRVNAME "arm-cs-arch-pmu"
> > @@ -118,6 +119,9 @@ static_assert(
> > ops->callback = arm_cspmu_ ## callback; \
> > } while (0)
> >
> > +/* JEDEC-assigned JEP106 identification code */
> > +#define ARM_CSPMU_IMPL_ID_NVIDIA 0x36B
> > +
> > static unsigned long arm_cspmu_cpuhp_state;
> >
> > /*
> > @@ -369,6 +373,9 @@ struct impl_match {
> > };
> >
> > static const struct impl_match impl_match[] = {
> > + { .pmiidr = ARM_CSPMU_IMPL_ID_NVIDIA,
> > + .mask = ARM_CSPMU_PMIIDR_IMPLEMENTER,
> > + .impl_init_ops = nv_cspmu_init_ops },
>
> Super minor nit: Coding style. Could we use :
>
> {
> .field = value,
> ...
> },
>
> > {}
> > };
> >
> > diff --git a/drivers/perf/arm_cspmu/nvidia_cspmu.c
> b/drivers/perf/arm_cspmu/nvidia_cspmu.c
> > new file mode 100644
> > index 000000000000..261f20680bc1
> > --- /dev/null
> > +++ b/drivers/perf/arm_cspmu/nvidia_cspmu.c
> > @@ -0,0 +1,367 @@
> > +// SPDX-License-Identifier: GPL-2.0
> > +/*
> > + * Copyright (c) 2022, NVIDIA CORPORATION & AFFILIATES.
> > + *
> > + */
> > +
> > +/* Support for NVIDIA specific attributes. */
> > +
> > +#include "nvidia_cspmu.h"
> > +
> > +#define NV_MCF_PCIE_PORT_COUNT 10ULL
> > +#define NV_MCF_PCIE_FILTER_ID_MASK
> GENMASK_ULL(NV_MCF_PCIE_PORT_COUNT - 1, 0)
> > +
> > +#define NV_MCF_GPU_PORT_COUNT 2ULL
> > +#define NV_MCF_GPU_FILTER_ID_MASK
> GENMASK_ULL(NV_MCF_GPU_PORT_COUNT - 1, 0)
> > +
> > +#define NV_MCF_NVL_PORT_COUNT 4ULL
> > +#define NV_MCF_NVL_FILTER_ID_MASK
> GENMASK_ULL(NV_MCF_NVL_PORT_COUNT - 1, 0)
> > +
> > +#define NV_SCF_MCF_PRODID_MASK GENMASK(31, 0)
> > +
> > +#define NV_FORMAT_NAME_GENERIC 0
> > +
> > +#define to_nv_cspmu_ctx(cspmu) ((struct nv_cspmu_ctx *)(cspmu-
> >impl.ctx))
> > +
> > +#define NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _num, _suff, _config)
> \
> > + ARM_CSPMU_EVENT_ATTR(_pref##_num##_suff, _config)
> > +
> > +#define NV_CSPMU_EVENT_ATTR_4(_pref, _suff, _config) \
> > + NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _0_, _suff, _config), \
> > + NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _1_, _suff, _config + 1), \
> > + NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _2_, _suff, _config + 2), \
> > + NV_CSPMU_EVENT_ATTR_4_INNER(_pref, _3_, _suff, _config + 3)
> > +
> > +struct nv_cspmu_ctx {
> > + const char *name;
> > + u32 filter_mask;
> > + struct attribute **event_attr;
> > + struct attribute **format_attr;
> > +};
> > +
> > +static struct attribute *scf_pmu_event_attrs[] = {
> > + ARM_CSPMU_EVENT_ATTR(bus_cycles, 0x1d),
> > +
> > + ARM_CSPMU_EVENT_ATTR(scf_cache_allocate, 0xF0),
> > + ARM_CSPMU_EVENT_ATTR(scf_cache_refill, 0xF1),
> > + ARM_CSPMU_EVENT_ATTR(scf_cache, 0xF2),
> > + ARM_CSPMU_EVENT_ATTR(scf_cache_wb, 0xF3),
> > +
> > + NV_CSPMU_EVENT_ATTR_4(socket, rd_data, 0x101),
> > + NV_CSPMU_EVENT_ATTR_4(socket, dl_rsp, 0x105),
> > + NV_CSPMU_EVENT_ATTR_4(socket, wb_data, 0x109),
> > + NV_CSPMU_EVENT_ATTR_4(socket, ev_rsp, 0x10d),
> > + NV_CSPMU_EVENT_ATTR_4(socket, prb_data, 0x111),
> > +
> > + NV_CSPMU_EVENT_ATTR_4(socket, rd_outstanding, 0x115),
> > + NV_CSPMU_EVENT_ATTR_4(socket, dl_outstanding, 0x119),
> > + NV_CSPMU_EVENT_ATTR_4(socket, wb_outstanding, 0x11d),
> > + NV_CSPMU_EVENT_ATTR_4(socket, wr_outstanding, 0x121),
> > + NV_CSPMU_EVENT_ATTR_4(socket, ev_outstanding, 0x125),
> > + NV_CSPMU_EVENT_ATTR_4(socket, prb_outstanding, 0x129),
> > +
> > + NV_CSPMU_EVENT_ATTR_4(socket, rd_access, 0x12d),
> > + NV_CSPMU_EVENT_ATTR_4(socket, dl_access, 0x131),
> > + NV_CSPMU_EVENT_ATTR_4(socket, wb_access, 0x135),
> > + NV_CSPMU_EVENT_ATTR_4(socket, wr_access, 0x139),
> > + NV_CSPMU_EVENT_ATTR_4(socket, ev_access, 0x13d),
> > + NV_CSPMU_EVENT_ATTR_4(socket, prb_access, 0x141),
> > +
> > + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_rd_data, 0x145),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_rd_access, 0x149),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wb_access, 0x14d),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_rd_outstanding, 0x151),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wr_outstanding, 0x155),
> > +
> > + NV_CSPMU_EVENT_ATTR_4(ocu, rem_rd_data, 0x159),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, rem_rd_access, 0x15d),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wb_access, 0x161),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, rem_rd_outstanding, 0x165),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wr_outstanding, 0x169),
> > +
> > + ARM_CSPMU_EVENT_ATTR(gmem_rd_data, 0x16d),
> > + ARM_CSPMU_EVENT_ATTR(gmem_rd_access, 0x16e),
> > + ARM_CSPMU_EVENT_ATTR(gmem_rd_outstanding, 0x16f),
> > + ARM_CSPMU_EVENT_ATTR(gmem_dl_rsp, 0x170),
> > + ARM_CSPMU_EVENT_ATTR(gmem_dl_access, 0x171),
> > + ARM_CSPMU_EVENT_ATTR(gmem_dl_outstanding, 0x172),
> > + ARM_CSPMU_EVENT_ATTR(gmem_wb_data, 0x173),
> > + ARM_CSPMU_EVENT_ATTR(gmem_wb_access, 0x174),
> > + ARM_CSPMU_EVENT_ATTR(gmem_wb_outstanding, 0x175),
> > + ARM_CSPMU_EVENT_ATTR(gmem_ev_rsp, 0x176),
> > + ARM_CSPMU_EVENT_ATTR(gmem_ev_access, 0x177),
> > + ARM_CSPMU_EVENT_ATTR(gmem_ev_outstanding, 0x178),
> > + ARM_CSPMU_EVENT_ATTR(gmem_wr_data, 0x179),
> > + ARM_CSPMU_EVENT_ATTR(gmem_wr_outstanding, 0x17a),
> > + ARM_CSPMU_EVENT_ATTR(gmem_wr_access, 0x17b),
> > +
> > + NV_CSPMU_EVENT_ATTR_4(socket, wr_data, 0x17c),
> > +
> > + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wr_data, 0x180),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wb_data, 0x184),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wr_access, 0x188),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, gmem_wb_outstanding, 0x18c),
> > +
> > + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wr_data, 0x190),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wb_data, 0x194),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wr_access, 0x198),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, rem_wb_outstanding, 0x19c),
> > +
> > + ARM_CSPMU_EVENT_ATTR(gmem_wr_total_bytes, 0x1a0),
> > + ARM_CSPMU_EVENT_ATTR(remote_socket_wr_total_bytes,
> 0x1a1),
> > + ARM_CSPMU_EVENT_ATTR(remote_socket_rd_data, 0x1a2),
> > + ARM_CSPMU_EVENT_ATTR(remote_socket_rd_outstanding,
> 0x1a3),
> > + ARM_CSPMU_EVENT_ATTR(remote_socket_rd_access, 0x1a4),
> > +
> > + ARM_CSPMU_EVENT_ATTR(cmem_rd_data, 0x1a5),
> > + ARM_CSPMU_EVENT_ATTR(cmem_rd_access, 0x1a6),
> > + ARM_CSPMU_EVENT_ATTR(cmem_rd_outstanding, 0x1a7),
> > + ARM_CSPMU_EVENT_ATTR(cmem_dl_rsp, 0x1a8),
> > + ARM_CSPMU_EVENT_ATTR(cmem_dl_access, 0x1a9),
> > + ARM_CSPMU_EVENT_ATTR(cmem_dl_outstanding, 0x1aa),
> > + ARM_CSPMU_EVENT_ATTR(cmem_wb_data, 0x1ab),
> > + ARM_CSPMU_EVENT_ATTR(cmem_wb_access, 0x1ac),
> > + ARM_CSPMU_EVENT_ATTR(cmem_wb_outstanding, 0x1ad),
> > + ARM_CSPMU_EVENT_ATTR(cmem_ev_rsp, 0x1ae),
> > + ARM_CSPMU_EVENT_ATTR(cmem_ev_access, 0x1af),
> > + ARM_CSPMU_EVENT_ATTR(cmem_ev_outstanding, 0x1b0),
> > + ARM_CSPMU_EVENT_ATTR(cmem_wr_data, 0x1b1),
> > + ARM_CSPMU_EVENT_ATTR(cmem_wr_outstanding, 0x1b2),
> > +
> > + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_rd_data, 0x1b3),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_rd_access, 0x1b7),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wb_access, 0x1bb),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_rd_outstanding, 0x1bf),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wr_outstanding, 0x1c3),
> > +
> > + ARM_CSPMU_EVENT_ATTR(ocu_prb_access, 0x1c7),
> > + ARM_CSPMU_EVENT_ATTR(ocu_prb_data, 0x1c8),
> > + ARM_CSPMU_EVENT_ATTR(ocu_prb_outstanding, 0x1c9),
> > +
> > + ARM_CSPMU_EVENT_ATTR(cmem_wr_access, 0x1ca),
> > +
> > + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wr_access, 0x1cb),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wb_data, 0x1cf),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wr_data, 0x1d3),
> > + NV_CSPMU_EVENT_ATTR_4(ocu, cmem_wb_outstanding, 0x1d7),
> > +
> > + ARM_CSPMU_EVENT_ATTR(cmem_wr_total_bytes, 0x1db),
> > +
> > + ARM_CSPMU_EVENT_ATTR(cycles,
> ARM_CSPMU_EVT_CYCLES_DEFAULT),
> > + NULL,
> > +};
> > +
> > +static struct attribute *mcf_pmu_event_attrs[] = {
> > + ARM_CSPMU_EVENT_ATTR(rd_bytes_loc, 0x0),
> > + ARM_CSPMU_EVENT_ATTR(rd_bytes_rem, 0x1),
> > + ARM_CSPMU_EVENT_ATTR(wr_bytes_loc, 0x2),
> > + ARM_CSPMU_EVENT_ATTR(wr_bytes_rem, 0x3),
> > + ARM_CSPMU_EVENT_ATTR(total_bytes_loc, 0x4),
> > + ARM_CSPMU_EVENT_ATTR(total_bytes_rem, 0x5),
> > + ARM_CSPMU_EVENT_ATTR(rd_req_loc, 0x6),
> > + ARM_CSPMU_EVENT_ATTR(rd_req_rem, 0x7),
> > + ARM_CSPMU_EVENT_ATTR(wr_req_loc, 0x8),
> > + ARM_CSPMU_EVENT_ATTR(wr_req_rem, 0x9),
> > + ARM_CSPMU_EVENT_ATTR(total_req_loc, 0xa),
> > + ARM_CSPMU_EVENT_ATTR(total_req_rem, 0xb),
> > + ARM_CSPMU_EVENT_ATTR(rd_cum_outs_loc, 0xc),
> > + ARM_CSPMU_EVENT_ATTR(rd_cum_outs_rem, 0xd),
> > + ARM_CSPMU_EVENT_ATTR(cycles,
> ARM_CSPMU_EVT_CYCLES_DEFAULT),
> > + NULL,
> > +};
> > +
> > +static struct attribute *generic_pmu_event_attrs[] = {
> > + ARM_CSPMU_EVENT_ATTR(cycles,
> ARM_CSPMU_EVT_CYCLES_DEFAULT),
> > + NULL,
> > +};
> > +
> > +static struct attribute *scf_pmu_format_attrs[] = {
> > + ARM_CSPMU_FORMAT_EVENT_ATTR,
> > + NULL,
> > +};
> > +
> > +static struct attribute *mcf_pcie_pmu_format_attrs[] = {
> > + ARM_CSPMU_FORMAT_EVENT_ATTR,
> > + ARM_CSPMU_FORMAT_ATTR(root_port, "config1:0-9"),
> > + NULL,
> > +};
> > +
> > +static struct attribute *mcf_gpu_pmu_format_attrs[] = {
> > + ARM_CSPMU_FORMAT_EVENT_ATTR,
> > + ARM_CSPMU_FORMAT_ATTR(gpu, "config1:0-1"),
> > + NULL,
> > +};
> > +
> > +static struct attribute *mcf_nvlink_pmu_format_attrs[] = {
> > + ARM_CSPMU_FORMAT_EVENT_ATTR,
> > + ARM_CSPMU_FORMAT_ATTR(rem_socket, "config1:0-3"),
> > + NULL,
> > +};
> > +
> > +static struct attribute *generic_pmu_format_attrs[] = {
> > + ARM_CSPMU_FORMAT_EVENT_ATTR,
> > + ARM_CSPMU_FORMAT_FILTER_ATTR,
> > + NULL,
> > +};
> > +
> > +static struct attribute **
> > +nv_cspmu_get_event_attrs(const struct arm_cspmu *cspmu)
> > +{
> > + const struct nv_cspmu_ctx *ctx = to_nv_cspmu_ctx(cspmu);
> > +
> > + return ctx->event_attr;
> > +}
> > +
> > +static struct attribute **
> > +nv_cspmu_get_format_attrs(const struct arm_cspmu *cspmu)
> > +{
> > + const struct nv_cspmu_ctx *ctx = to_nv_cspmu_ctx(cspmu);
> > +
> > + return ctx->format_attr;
> > +}
> > +
> > +static const char *
> > +nv_cspmu_get_name(const struct arm_cspmu *cspmu)
> > +{
> > + const struct nv_cspmu_ctx *ctx = to_nv_cspmu_ctx(cspmu);
> > +
> > + return ctx->name;
> > +}
> > +
> > +static u32 nv_cspmu_event_filter(const struct perf_event *event)
> > +{
> > + const struct nv_cspmu_ctx *ctx =
> > + to_nv_cspmu_ctx(to_arm_cspmu(event->pmu));
> > +
> > + return event->attr.config1 & ctx->filter_mask;
> > +}
> > +
> > +enum nv_cspmu_name_fmt {
> > + NAME_FMT_GENERIC,
> > + NAME_FMT_PROC
> > +};
> > +
> > +struct nv_cspmu_match {
> > + u32 prodid;
> > + u32 prodid_mask;
> > + u64 filter_mask;
> > + const char *name_pattern;
> > + enum nv_cspmu_name_fmt name_fmt;
> > + struct attribute **event_attr;
> > + struct attribute **format_attr;
> > +};
> > +
> > +static const struct nv_cspmu_match nv_cspmu_match[] = {
>
> Similar coding style nit below.
>

Sure, I will update this.

>
> Otherwise,
>
> Acked-by: Suzuki K Poulose <[email protected]>

Thanks!

Unfortunately, we need to update the name of the PMUs and remove
some of the attributes in NVIDIA implementation. This requires a change
in nvidia_cspmu.c and nvidia-pmu.rst. I hope you are fine if I include this
change on v5 patch.

Regards,
Besar

2022-09-28 11:11:07

by Suzuki K Poulose

[permalink] [raw]
Subject: Re: [PATCH v4 2/2] perf: arm_cspmu: Add support for NVIDIA SCF and MCF attribute

On 28/09/2022 02:38, Besar Wicaksono wrote:
>
>
>> -----Original Message-----
>> From: Suzuki K Poulose <[email protected]>
>> Sent: Tuesday, September 27, 2022 6:43 AM
>> To: Besar Wicaksono <[email protected]>; [email protected];
>> [email protected]; [email protected]; [email protected]
>> Cc: [email protected]; [email protected];
>> [email protected]; [email protected];
>> [email protected]; [email protected]; Thierry Reding
>> <[email protected]>; Jonathan Hunter <[email protected]>; Vikram
>> Sethi <[email protected]>; [email protected];
>> [email protected]; [email protected]
>> Subject: Re: [PATCH v4 2/2] perf: arm_cspmu: Add support for NVIDIA SCF
>> and MCF attribute
>>
>> External email: Use caution opening links or attachments
>>
>>
>> On 14/08/2022 19:23, Besar Wicaksono wrote:
>>> Add support for NVIDIA System Cache Fabric (SCF) and Memory Control
>>> Fabric (MCF) PMU attributes for CoreSight PMU implementation in
>>> NVIDIA devices.
>>>
>>> Signed-off-by: Besar Wicaksono <[email protected]>

>>> +struct nv_cspmu_match {
>>> + u32 prodid;
>>> + u32 prodid_mask;
>>> + u64 filter_mask;
>>> + const char *name_pattern;
>>> + enum nv_cspmu_name_fmt name_fmt;
>>> + struct attribute **event_attr;
>>> + struct attribute **format_attr;
>>> +};
>>> +
>>> +static const struct nv_cspmu_match nv_cspmu_match[] = {
>>
>> Similar coding style nit below.
>>
>
> Sure, I will update this.
>
>>
>> Otherwise,
>>
>> Acked-by: Suzuki K Poulose <[email protected]>
>
> Thanks!
>
> Unfortunately, we need to update the name of the PMUs and remove
> some of the attributes in NVIDIA implementation. This requires a change
> in nvidia_cspmu.c and nvidia-pmu.rst. I hope you are fine if I include this
> change on v5 patch.

That should be fine.

Suzuki