2020-07-15 15:34:11

by Ben Levinsky

[permalink] [raw]
Subject: [PATCH v6 0/5] Provide basic driver to control Arm R5 co-processor found on

Currently it is able to start, stop and load elf on to the
processor.

The driver was tested on Xilinx ZynqMP and Versal.

v2:
- remove domain struct as per review from Mathieu
v3:
- add xilinx-related platform mgmt fn's instead of wrapping around
function pointer in xilinx eemi ops struct
- update zynqmp_r5 yaml parsing to not raise warnings for extra
information in children of R5 node. The warning "node has a unit
name, but no reg or ranges property" will still be raised though
as this particular node is needed to describe the
'#address-cells' and '#size-cells' information.
v4:
- add default values for enums
- fix formatting as per checkpatch.pl --strict. Note that 1 warning and 1 check
are still raised as each is due to fixing the warning results in that
particular line going over 80 characters.
- remove warning '/example-0/rpu@ff9a0000/r5@0:
node has a unit name, but no reg or ranges property'
by adding reg to r5 node.
v5:
- update device tree sample and yaml parsing to not raise any warnings
- description for memory-region in yaml parsing
- compatible string in yaml parsing for TCM
- parse_fw change from use of rproc_of_resm_mem_entry_init to rproc_mem_entry_init and use of alloc/release
- var's of type zynqmp_r5_pdata all have same local variable name
- use dev_dbg instead of dev_info
v6:
- adding memory carveouts is handled much more similarly. All mem carveouts are
now described in reserved memory as needed. That is, TCM nodes are not
coupled to remoteproc anymore. This is reflected in the remoteproc R5 driver
and the device tree binding.
- remove mailbox from device tree binding as it is not necessary for elf
loading



Ben Levinsky (5):
firmware: xilinx: Add ZynqMP firmware ioctl enums for RPU
configuration.
firmware: xilinx: Add shutdown/wakeup APIs
firmware: xilinx: Add RPU configuration APIs
dt-bindings: remoteproc: Add documentation for ZynqMP R5 rproc
bindings
remoteproc: Add initial zynqmp R5 remoteproc driver

.../xilinx,zynqmp-r5-remoteproc.yaml | 73 ++
drivers/firmware/xilinx/zynqmp.c | 134 +++
drivers/remoteproc/Kconfig | 10 +
drivers/remoteproc/Makefile | 1 +
drivers/remoteproc/zynqmp_r5_remoteproc.c | 911 ++++++++++++++++++
include/linux/firmware/xlnx-zynqmp.h | 75 ++
6 files changed, 1204 insertions(+)
create mode 100644 Documentation/devicetree/bindings/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml
create mode 100644 drivers/remoteproc/zynqmp_r5_remoteproc.c

--
2.17.1


2020-07-15 15:34:16

by Ben Levinsky

[permalink] [raw]
Subject: [PATCH v6 4/5] dt-bindings: remoteproc: Add documentation for ZynqMP R5 rproc bindings

Add binding for ZynqMP R5 OpenAMP.

Represent the RPU domain resources in one device node. Each RPU
processor is a subnode of the top RPU domain node.

Signed-off-by: Ben Levinsky <[email protected]>
Signed-off-by: Jason Wu <[email protected]>
Signed-off-by: Wendy Liang <[email protected]>
Signed-off-by: Michal Simek <[email protected]>
---
v3:
- update zynqmp_r5 yaml parsing to not raise warnings for extra
information in children of R5 node. The warning "node has a unit
name, but no reg or ranges property" will still be raised though
as this particular node is needed to describe the
'#address-cells' and '#size-cells' information.
v4::
- remove warning '/example-0/rpu@ff9a0000/r5@0:
node has a unit name, but no reg or ranges property'
by adding reg to r5 node.
v5:
- update device tree sample and yaml parsing to not raise any warnings
- description for memory-region in yaml parsing
- compatible string in yaml parsing for TCM
v6:
- remove coupling TCM nodes with remoteproc
- remove mailbox as it is optional not needed
---
.../xilinx,zynqmp-r5-remoteproc.yaml | 73 +++++++++++++++++++
1 file changed, 73 insertions(+)
create mode 100644 Documentation/devicetree/bindings/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml

diff --git a/Documentation/devicetree/bindings/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml b/Documentation/devicetree/bindings/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml
new file mode 100644
index 000000000000..54f916d9b037
--- /dev/null
+++ b/Documentation/devicetree/bindings/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml
@@ -0,0 +1,73 @@
+# SPDX-License-Identifier: (GPL-2.0 OR BSD-2-Clause)
+%YAML 1.2
+---
+$id: "http://devicetree.org/schemas/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml#"
+$schema: "http://devicetree.org/meta-schemas/core.yaml#"
+
+title: Xilinx R5 remote processor controller bindings
+
+description:
+ This document defines the binding for the remoteproc component that loads and
+ boots firmwares on the Xilinx Zynqmp and Versal family chipset.
+
+ Note that the Linux has global addressing view of the R5-related memory (TCM)
+ so the absolute address ranges are provided in TCM reg's.
+maintainers:
+ - Ed Mooring <[email protected]>
+ - Ben Levinsky <[email protected]>
+
+properties:
+ compatible:
+ const: "xlnx,zynqmp-r5-remoteproc-1.0"
+
+ lockstep-mode:
+ description:
+ R5 core configuration (split is 0 or lock-step and 1)
+ maxItems: 1
+
+ interrupts:
+ description:
+ Interrupt mapping for remoteproc IPI. It is required if the
+ user uses the remoteproc driver with the RPMsg kernel driver.
+ maxItems: 6
+
+ memory-region:
+ description:
+ collection of memory carveouts used for elf-loading and inter-processor
+ communication.
+ maxItems: 4
+ minItems: 4
+ pnode-id:
+ maxItems: 1
+ mboxes:
+ maxItems: 2
+ mbox-names:
+ maxItems: 2
+
+examples:
+ - |
+ reserved-memory {
+ #address-cells = <1>;
+ #size-cells = <1>;
+ ranges;
+ elf_load: rproc@3ed000000 {
+ no-map;
+ reg = <0x3ed00000 0x40000>;
+ };
+ };
+ rpu {
+ compatible = "xlnx,zynqmp-r5-remoteproc-1.0";
+ #address-cells = <1>;
+ #size-cells = <1>;
+ ranges;
+ lockstep-mode = <0>;
+ r5_0 {
+ ranges;
+ #address-cells = <1>;
+ #size-cells = <1>;
+ memory-region = <&elf_load>;
+ pnode-id = <0x7>;
+ };
+ };
+
+...
--
2.17.1

2020-07-15 15:34:23

by Ben Levinsky

[permalink] [raw]
Subject: [PATCH v6 1/5] firmware: xilinx: Add ZynqMP firmware ioctl enums for RPU configuration.

Add ZynqMP firmware ioctl enums for RPU configuration.

Signed-off-by: Ben Levinsky <[email protected]>
---
v3:
- add xilinx-related platform mgmt fn's instead of wrapping around
function pointer in xilinx eemi ops struct
v4:
- add default values for enums
---
include/linux/firmware/xlnx-zynqmp.h | 19 +++++++++++++++++++
1 file changed, 19 insertions(+)

diff --git a/include/linux/firmware/xlnx-zynqmp.h b/include/linux/firmware/xlnx-zynqmp.h
index 5968df82b991..bb347dfe4ba4 100644
--- a/include/linux/firmware/xlnx-zynqmp.h
+++ b/include/linux/firmware/xlnx-zynqmp.h
@@ -104,6 +104,10 @@ enum pm_ret_status {
};

enum pm_ioctl_id {
+ IOCTL_GET_RPU_OPER_MODE = 0,
+ IOCTL_SET_RPU_OPER_MODE = 1,
+ IOCTL_RPU_BOOT_ADDR_CONFIG = 2,
+ IOCTL_TCM_COMB_CONFIG = 3,
IOCTL_SD_DLL_RESET = 6,
IOCTL_SET_SD_TAPDELAY,
IOCTL_SET_PLL_FRAC_MODE,
@@ -129,6 +133,21 @@ enum pm_query_id {
PM_QID_CLOCK_GET_MAX_DIVISOR,
};

+enum rpu_oper_mode {
+ PM_RPU_MODE_LOCKSTEP = 0,
+ PM_RPU_MODE_SPLIT = 1,
+};
+
+enum rpu_boot_mem {
+ PM_RPU_BOOTMEM_LOVEC = 0,
+ PM_RPU_BOOTMEM_HIVEC = 1,
+};
+
+enum rpu_tcm_comb {
+ PM_RPU_TCM_SPLIT = 0,
+ PM_RPU_TCM_COMB = 1,
+};
+
enum zynqmp_pm_reset_action {
PM_RESET_ACTION_RELEASE,
PM_RESET_ACTION_ASSERT,
--
2.17.1

2020-07-15 15:35:12

by Ben Levinsky

[permalink] [raw]
Subject: [PATCH v6 5/5] remoteproc: Add initial zynqmp R5 remoteproc driver

R5 is included in Xilinx Zynq UltraScale MPSoC so by adding this
remotproc driver, we can boot the R5 sub-system in different
configurations.

Acked-by: Stefano Stabellini <[email protected]>
Acked-by: Ben Levinsky <[email protected]>
Reviewed-by: Radhey Shyam Pandey <[email protected]>
Signed-off-by: Ben Levinsky <[email protected]>
Signed-off-by: Wendy Liang <[email protected]>
Signed-off-by: Michal Simek <[email protected]>
Signed-off-by: Ed Mooring <[email protected]>
Signed-off-by: Jason Wu <[email protected]>
Tested-by: Ben Levinsky <[email protected]>
---
v2:
- remove domain struct as per review from Mathieu
v3:
- add xilinx-related platform mgmt fn's instead of wrapping around
function pointer in xilinx eemi ops struct
v4:
- add default values for enums
- fix formatting as per checkpatch.pl --strict. Note that 1 warning and 1 check
are still raised as each is due to fixing the warning results in that
particular line going over 80 characters.
v5:
- parse_fw change from use of rproc_of_resm_mem_entry_init to rproc_mem_entry_init and use of alloc/release
- var's of type zynqmp_r5_pdata all have same local variable name
- use dev_dbg instead of dev_info
v6:
- adding memory carveouts is handled much more similarly. All mem carveouts are
now described in reserved memory as needed. That is, TCM nodes are not
coupled to remoteproc anymore. This is reflected in the remoteproc R5 driver
and the device tree binding.
- remove mailbox from device tree binding as it is not necessary for elf
loading
- use lockstep-mode property for configuring RPU
---
drivers/remoteproc/Kconfig | 10 +
drivers/remoteproc/Makefile | 1 +
drivers/remoteproc/zynqmp_r5_remoteproc.c | 911 ++++++++++++++++++++++
3 files changed, 922 insertions(+)
create mode 100644 drivers/remoteproc/zynqmp_r5_remoteproc.c

diff --git a/drivers/remoteproc/Kconfig b/drivers/remoteproc/Kconfig
index c4d1731295eb..342a7e668636 100644
--- a/drivers/remoteproc/Kconfig
+++ b/drivers/remoteproc/Kconfig
@@ -249,6 +249,16 @@ config STM32_RPROC

This can be either built-in or a loadable module.

+config ZYNQMP_R5_REMOTEPROC
+ tristate "ZynqMP_R5 remoteproc support"
+ depends on ARM64 && PM && ARCH_ZYNQMP
+ select RPMSG_VIRTIO
+ select MAILBOX
+ select ZYNQMP_IPI_MBOX
+ help
+ Say y here to support ZynqMP R5 remote processors via the remote
+ processor framework.
+
endif # REMOTEPROC

endmenu
diff --git a/drivers/remoteproc/Makefile b/drivers/remoteproc/Makefile
index e8b886e511f0..04d1c95d06d7 100644
--- a/drivers/remoteproc/Makefile
+++ b/drivers/remoteproc/Makefile
@@ -28,5 +28,6 @@ obj-$(CONFIG_QCOM_WCNSS_PIL) += qcom_wcnss_pil.o
qcom_wcnss_pil-y += qcom_wcnss.o
qcom_wcnss_pil-y += qcom_wcnss_iris.o
obj-$(CONFIG_ST_REMOTEPROC) += st_remoteproc.o
+obj-$(CONFIG_ZYNQMP_R5_REMOTEPROC) += zynqmp_r5_remoteproc.o
obj-$(CONFIG_ST_SLIM_REMOTEPROC) += st_slim_rproc.o
obj-$(CONFIG_STM32_RPROC) += stm32_rproc.o
diff --git a/drivers/remoteproc/zynqmp_r5_remoteproc.c b/drivers/remoteproc/zynqmp_r5_remoteproc.c
new file mode 100644
index 000000000000..b600759e257e
--- /dev/null
+++ b/drivers/remoteproc/zynqmp_r5_remoteproc.c
@@ -0,0 +1,911 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Zynq R5 Remote Processor driver
+ *
+ * Copyright (C) 2019, 2020 Xilinx Inc. Ben Levinsky <[email protected]>
+ * Copyright (C) 2015 - 2018 Xilinx Inc.
+ * Copyright (C) 2015 Jason Wu <[email protected]>
+ *
+ * Based on origin OMAP and Zynq Remote Processor driver
+ *
+ * Copyright (C) 2012 Michal Simek <[email protected]>
+ * Copyright (C) 2012 PetaLogix
+ * Copyright (C) 2011 Texas Instruments, Inc.
+ * Copyright (C) 2011 Google, Inc.
+ */
+
+#include <linux/atomic.h>
+#include <linux/cpu.h>
+#include <linux/dma-mapping.h>
+#include <linux/delay.h>
+#include <linux/err.h>
+#include <linux/firmware/xlnx-zynqmp.h>
+#include <linux/genalloc.h>
+#include <linux/idr.h>
+#include <linux/interrupt.h>
+#include <linux/kernel.h>
+#include <linux/list.h>
+#include <linux/mailbox_client.h>
+#include <linux/mailbox/zynqmp-ipi-message.h>
+#include <linux/module.h>
+#include <linux/of_address.h>
+#include <linux/of_irq.h>
+#include <linux/of_platform.h>
+#include <linux/of_reserved_mem.h>
+#include <linux/pfn.h>
+#include <linux/platform_device.h>
+#include <linux/remoteproc.h>
+#include <linux/skbuff.h>
+#include <linux/slab.h>
+#include <linux/sysfs.h>
+
+#include "remoteproc_internal.h"
+
+#define MAX_RPROCS 2 /* Support up to 2 RPU */
+#define MAX_MEM_PNODES 4 /* Max power nodes for one RPU memory instance */
+
+#define DEFAULT_FIRMWARE_NAME "rproc-rpu-fw"
+
+/* PM proc states */
+#define PM_PROC_STATE_ACTIVE 1U
+
+/* IPI buffer MAX length */
+#define IPI_BUF_LEN_MAX 32U
+/* RX mailbox client buffer max length */
+#define RX_MBOX_CLIENT_BUF_MAX (IPI_BUF_LEN_MAX + \
+ sizeof(struct zynqmp_ipi_message))
+
+#define ZYNQMP_R5_NUM_TCM_BANKS 4
+
+/* lookup table mapping power-node-ID of TCM bank to absolute base address */
+static unsigned long tcm_addr_to_pnode[ZYNQMP_R5_NUM_TCM_BANKS][2] = {
+ {0xFFE00000, 0xF },
+ {0xFFE20000, 0x10},
+ {0xFFE90000, 0x10},
+ {0xFFEB0000, 0x11},
+};
+
+static bool autoboot __read_mostly;
+
+/**
+ * struct zynqmp_r5_mem - zynqmp rpu memory data
+ * @pnode_id: TCM power domain ids
+ * @res: memory resource
+ * @node: list node
+ */
+struct zynqmp_r5_mem {
+ u32 pnode_id[MAX_MEM_PNODES];
+ struct resource res;
+ struct list_head node;
+};
+
+/**
+ * struct zynqmp_r5_pdata - zynqmp rpu remote processor private data
+ * @dev: device of RPU instance
+ * @rproc: rproc handle
+ * @pnode_id: RPU CPU power domain id
+ * @mems: memory resources
+ * @is_r5_mode_set: indicate if r5 operation mode is set
+ * @tx_mc: tx mailbox client
+ * @rx_mc: rx mailbox client
+ * @tx_chan: tx mailbox channel
+ * @rx_chan: rx mailbox channel
+ * @mbox_work: mbox_work for the RPU remoteproc
+ * @tx_mc_skbs: socket buffers for tx mailbox client
+ * @rx_mc_buf: rx mailbox client buffer to save the rx message
+ */
+struct zynqmp_r5_pdata {
+ struct device dev;
+ struct rproc *rproc;
+ u32 pnode_id;
+ struct list_head mems;
+ bool is_r5_mode_set;
+ struct mbox_client tx_mc;
+ struct mbox_client rx_mc;
+ struct mbox_chan *tx_chan;
+ struct mbox_chan *rx_chan;
+ struct work_struct mbox_work;
+ struct sk_buff_head tx_mc_skbs;
+ unsigned char rx_mc_buf[RX_MBOX_CLIENT_BUF_MAX];
+};
+
+/**
+ * table of RPUs
+ */
+struct zynqmp_r5_pdata rpus[MAX_RPROCS];
+/**
+ * RPU core configuration
+ */
+enum rpu_oper_mode rpu_mode;
+
+/*
+ * r5_set_mode - set RPU operation mode
+ * @pdata: Remote processor private data
+ *
+ * set RPU oepration mode
+ *
+ * Return: 0 for success, negative value for failure
+ */
+static int r5_set_mode(struct zynqmp_r5_pdata *pdata)
+{
+ u32 val[PAYLOAD_ARG_CNT] = {0}, expect, tcm_mode;
+ struct device *dev = &pdata->dev;
+ int ret;
+
+ expect = (u32)rpu_mode;
+ ret = zynqmp_pm_get_rpu_mode(pdata->pnode_id, 0, 0, val);
+ if (ret < 0) {
+ dev_err(dev, "failed to get RPU oper mode.\n");
+ return ret;
+ }
+ if (val[0] == expect) {
+ dev_dbg(dev, "RPU mode matches: %x\n", val[0]);
+ } else {
+ ret = zynqmp_pm_set_rpu_mode(pdata->pnode_id,
+ expect, 0, val);
+ if (ret < 0) {
+ dev_err(dev,
+ "failed to set RPU oper mode.\n");
+ return ret;
+ }
+ }
+
+ tcm_mode = (expect == (u32)PM_RPU_MODE_LOCKSTEP) ?
+ PM_RPU_TCM_COMB : PM_RPU_TCM_SPLIT;
+ ret = zynqmp_pm_set_tcm_config(pdata->pnode_id, tcm_mode, 0, val);
+ if (ret < 0) {
+ dev_err(dev, "failed to config TCM to %x.\n",
+ expect);
+ return ret;
+ }
+ pdata->is_r5_mode_set = true;
+ return 0;
+}
+
+/*
+ * ZynqMP R5 remoteproc memory release function
+ */
+static int zynqmp_r5_mem_release(struct rproc *rproc,
+ struct rproc_mem_entry *mem)
+{
+ struct zynqmp_r5_mem *priv;
+ int i, ret;
+ struct device *dev = &rproc->dev;
+
+ priv = mem->priv;
+ if (!priv)
+ return 0;
+ for (i = 0; i < MAX_MEM_PNODES; i++) {
+ if (priv->pnode_id[i]) {
+ dev_dbg(dev, "%s, pnode %d\n",
+ __func__, priv->pnode_id[i]);
+ ret = zynqmp_pm_release_node(priv->pnode_id[i]);
+ if (ret < 0) {
+ dev_err(dev,
+ "failed to release power node: %u\n",
+ priv->pnode_id[i]);
+ return ret;
+ }
+ } else {
+ break;
+ }
+ }
+ return 0;
+}
+
+/*
+ * ZynqMP R5 remoteproc operations
+ */
+static int zynqmp_r5_rproc_start(struct rproc *rproc)
+{
+ struct device *dev = rproc->dev.parent;
+ struct zynqmp_r5_pdata *pdata = rproc->priv;
+ enum rpu_boot_mem bootmem;
+ int ret;
+
+ if ((rproc->bootaddr & 0xF0000000) == 0xF0000000)
+ bootmem = PM_RPU_BOOTMEM_HIVEC;
+ else
+ bootmem = PM_RPU_BOOTMEM_LOVEC;
+ dev_dbg(dev, "RPU boot from %s.",
+ bootmem == PM_RPU_BOOTMEM_HIVEC ? "OCM" : "TCM");
+ ret = zynqmp_pm_request_wakeup(pdata->pnode_id, 1,
+ bootmem, ZYNQMP_PM_REQUEST_ACK_NO);
+ if (ret < 0) {
+ dev_err(dev, "failed to boot R5.\n");
+ return ret;
+ }
+ return 0;
+}
+
+static int zynqmp_r5_rproc_stop(struct rproc *rproc)
+{
+ struct zynqmp_r5_pdata *pdata = rproc->priv;
+ int ret;
+
+ ret = zynqmp_pm_force_powerdown(pdata->pnode_id,
+ ZYNQMP_PM_REQUEST_ACK_BLOCKING);
+ if (ret < 0) {
+ dev_err(&pdata->dev, "failed to shutdown R5.\n");
+ return ret;
+ }
+ return 0;
+}
+
+static int zynqmp_r5_rproc_mem_alloc(struct rproc *rproc,
+ struct rproc_mem_entry *mem)
+{
+ struct device *dev = rproc->dev.parent;
+ void *va;
+
+ dev_dbg(rproc->dev.parent, "map memory: %pa\n", &mem->dma);
+ va = ioremap_wc(mem->dma, mem->len);
+ if (IS_ERR_OR_NULL(va)) {
+ dev_err(dev, "Unable to map memory region: %pa+%lx\n",
+ &mem->dma, mem->len);
+ return -ENOMEM;
+ }
+
+ /* Update memory entry va */
+ mem->va = va;
+
+ return 0;
+}
+
+static int zynqmp_r5_rproc_mem_release(struct rproc *rproc,
+ struct rproc_mem_entry *mem)
+{
+ dev_dbg(rproc->dev.parent, "unmap memory: %pa\n", &mem->dma);
+ iounmap(mem->va);
+
+ return 0;
+}
+
+/*
+ * TCM needs mapping to R5 relative address and xilinx platform mgmt call
+ */
+struct rproc_mem_entry *handle_tcm_parsing(struct device *dev,
+ struct reserved_mem *rmem,
+ struct device_node *node,
+ int lookup_idx)
+{
+ void *va;
+ dma_addr_t dma;
+ resource_size_t size;
+ int ret;
+ u32 pnode_id;
+ struct resource rsc;
+ struct rproc_mem_entry *mem;
+
+ pnode_id = tcm_addr_to_pnode[lookup_idx][1];
+ ret = zynqmp_pm_request_node(pnode_id,
+ ZYNQMP_PM_CAPABILITY_ACCESS, 0,
+ ZYNQMP_PM_REQUEST_ACK_BLOCKING);
+ if (ret < 0) {
+ dev_err(dev, "failed to request power node: %u\n", pnode_id);
+ return -EINVAL;
+ }
+
+ ret = of_address_to_resource(node, 0, &rsc);
+ if (ret < 0) {
+ dev_err(dev, "failed to get resource of memory %s",
+ of_node_full_name(node));
+ return -EINVAL;
+ }
+ size = resource_size(&rsc);
+ va = devm_ioremap_wc(dev, rsc.start, size);
+ if (!va)
+ return -ENOMEM;
+
+ /* zero out tcm base address */
+ if (rsc.start & 0xffe00000) {
+ /* R5 can't see anything past 0xfffff so wipe it */
+ rsc.start &= 0x000fffff;
+ /*
+ * handle tcm banks 1 a and b (0xffe9000 and
+ * 0xffeb0000)
+ */
+ if (rsc.start & 0x80000)
+ rsc.start -= 0x90000;
+ }
+
+ dma = (dma_addr_t)rsc.start;
+ mem = rproc_mem_entry_init(dev, va, dma, (int)size, rsc.start,
+ NULL, zynqmp_r5_mem_release,
+ rsc.name);
+ if (!mem)
+ return -ENOMEM;
+
+ return mem;
+}
+
+static int parse_mem_regions(struct rproc *rproc)
+{
+ int num_mems, i;
+ struct zynqmp_r5_pdata *pdata = rproc->priv;
+ struct device *dev = &pdata->dev;
+ struct device_node *np = dev->of_node;
+ struct rproc_mem_entry *mem;
+
+ num_mems = of_count_phandle_with_args(np, "memory-region", NULL);
+ if (num_mems <= 0)
+ return 0;
+ for (i = 0; i < num_mems; i++) {
+ struct device_node *node;
+ struct reserved_mem *rmem;
+
+ node = of_parse_phandle(np, "memory-region", i);
+ rmem = of_reserved_mem_lookup(node);
+ if (!rmem) {
+ dev_err(dev, "unable to acquire memory-region\n");
+ return -EINVAL;
+ }
+
+ if (strstr(node->name, "vdev0buffer")) {
+ /* Register DMA region */
+ mem = rproc_mem_entry_init(dev, NULL,
+ (dma_addr_t)rmem->base,
+ rmem->size, rmem->base,
+ NULL, NULL,
+ "vdev0buffer");
+ if (!mem) {
+ dev_err(dev, "unable to initialize memory-region %s\n",
+ node->name);
+ return -ENOMEM;
+ }
+ dev_dbg(dev, "parsed %s at %llx\r\n", mem->name,
+ mem->dma);
+ } else if (strstr(node->name, "vdev0vring")) {
+ int vring_id;
+ char name[16];
+
+ /*
+ * can be 1 of multiple vring IDs per IPC channel
+ * e.g. 'vdev0vring0' and 'vdev0vring1'
+ */
+ vring_id = node->name[14] - '0';
+ snprintf(name, sizeof(name), "vdev0vring%d", vring_id);
+ /* Register vring */
+ mem = rproc_mem_entry_init(dev, NULL,
+ (dma_addr_t)rmem->base,
+ rmem->size, rmem->base,
+ zynqmp_r5_rproc_mem_alloc,
+ zynqmp_r5_rproc_mem_release,
+ name);
+ dev_dbg(dev, "parsed %s at %llx\r\n", mem->name,
+ mem->dma);
+ } else {
+ int idx;
+
+ /*
+ * if TCM update address space for R5 and
+ * make xilinx platform mgmt call
+ */
+ for (idx = 0; idx < ZYNQMP_R5_NUM_TCM_BANKS; idx++) {
+ if (tcm_addr_to_pnode[idx][0] == rmem->base)
+ break;
+ }
+
+ if (idx != ZYNQMP_R5_NUM_TCM_BANKS) {
+ mem = handle_tcm_parsing(dev, rmem, node, idx);
+ } else {
+ mem = rproc_mem_entry_init(dev, NULL,
+ (dma_addr_t)rmem->base,
+ rmem->size, rmem->base,
+ zynqmp_r5_rproc_mem_alloc,
+ zynqmp_r5_rproc_mem_release,
+ node->name);
+ }
+
+ if (!mem) {
+ dev_err(dev,
+ "unable to init memory-region %s\n",
+ node->name);
+ return -ENOMEM;
+ }
+ }
+ rproc_add_carveout(rproc, mem);
+ }
+
+ return 0;
+}
+
+static int zynqmp_r5_parse_fw(struct rproc *rproc, const struct firmware *fw)
+{
+ int ret;
+ struct zynqmp_r5_pdata *pdata = rproc->priv;
+ struct device *dev = &pdata->dev;
+
+ ret = parse_mem_regions(rproc);
+ if (ret) {
+ dev_err(dev, "parse_mem_regions failed %x\n", ret);
+ return ret;
+ }
+
+ ret = rproc_elf_load_rsc_table(rproc, fw);
+ if (ret == -EINVAL) {
+ dev_dbg(dev, "no resource table found.\n");
+ ret = 0;
+ }
+ return ret;
+}
+
+/* kick a firmware */
+static void zynqmp_r5_rproc_kick(struct rproc *rproc, int vqid)
+{
+ struct device *dev = rproc->dev.parent;
+ struct zynqmp_r5_pdata *pdata = rproc->priv;
+
+ dev_dbg(dev, "KICK Firmware to start send messages vqid %d\n", vqid);
+
+ if (vqid < 0) {
+ /* If vqid is negative, does not pass the vqid to
+ * mailbox. As vqid is supposed to be 0 or possive.
+ * It also gives a way to just kick instead but
+ * not use the IPI buffer. It is better to provide
+ * a proper way to pass the short message, which will
+ * need to sync to upstream first, for now,
+ * use negative vqid to assume no message will be
+ * passed with IPI buffer, but just raise interrupt.
+ * This will be faster as it doesn't need to copy the
+ * message to the IPI buffer.
+ *
+ * It will ignore the return, as failure is due to
+ * there already kicks in the mailbox queue.
+ */
+ (void)mbox_send_message(pdata->tx_chan, NULL);
+ } else {
+ struct sk_buff *skb;
+ unsigned int skb_len;
+ struct zynqmp_ipi_message *mb_msg;
+ int ret;
+
+ skb_len = (unsigned int)(sizeof(vqid) + sizeof(mb_msg));
+ skb = alloc_skb(skb_len, GFP_ATOMIC);
+ if (!skb) {
+ dev_err(dev,
+ "Failed to allocate skb to kick remote.\n");
+ return;
+ }
+ mb_msg = (struct zynqmp_ipi_message *)skb_put(skb, skb_len);
+ mb_msg->len = sizeof(vqid);
+ memcpy(mb_msg->data, &vqid, sizeof(vqid));
+ skb_queue_tail(&pdata->tx_mc_skbs, skb);
+ ret = mbox_send_message(pdata->tx_chan, mb_msg);
+ if (ret < 0) {
+ dev_warn(dev, "Failed to kick remote.\n");
+ skb_dequeue_tail(&pdata->tx_mc_skbs);
+ kfree_skb(skb);
+ }
+ }
+}
+
+static struct rproc_ops zynqmp_r5_rproc_ops = {
+ .start = zynqmp_r5_rproc_start,
+ .stop = zynqmp_r5_rproc_stop,
+ .load = rproc_elf_load_segments,
+ .parse_fw = zynqmp_r5_parse_fw,
+ .find_loaded_rsc_table = rproc_elf_find_loaded_rsc_table,
+ .sanity_check = rproc_elf_sanity_check,
+ .get_boot_addr = rproc_elf_get_boot_addr,
+ .kick = zynqmp_r5_rproc_kick,
+};
+
+/* zynqmp_r5_mem_probe() - probes RPU TCM memory device
+ * @pdata: pointer to the RPU remoteproc private data
+ * @node: pointer to the memory node
+ *
+ * Function to retrieve resources for RPU TCM memory device.
+ */
+static int zynqmp_r5_mem_probe(struct zynqmp_r5_pdata *pdata,
+ struct device_node *node)
+{
+ struct device *dev;
+ struct zynqmp_r5_mem *mem;
+ int ret;
+ struct property *prop;
+ const __be32 *cur;
+ u32 val;
+ int i;
+
+ dev = &pdata->dev;
+ mem = devm_kzalloc(dev, sizeof(*mem), GFP_KERNEL);
+ if (!mem)
+ return -ENOMEM;
+ ret = of_address_to_resource(node, 0, &mem->res);
+ if (ret < 0) {
+ dev_err(dev, "failed to get resource of memory %s",
+ of_node_full_name(node));
+ return -EINVAL;
+ }
+
+ /* Get the power domain id */
+ i = 0;
+ if (of_find_property(node, "pnode-id", NULL)) {
+ of_property_for_each_u32(node, "pnode-id", prop, cur, val)
+ mem->pnode_id[i++] = val;
+ }
+ list_add_tail(&mem->node, &pdata->mems);
+ return 0;
+}
+
+/**
+ * zynqmp_r5_release() - ZynqMP R5 device release function
+ * @dev: pointer to the device struct of ZynqMP R5
+ *
+ * Function to release ZynqMP R5 device.
+ */
+static void zynqmp_r5_release(struct device *dev)
+{
+ struct zynqmp_r5_pdata *pdata;
+ struct rproc *rproc;
+ struct sk_buff *skb;
+
+ pdata = dev_get_drvdata(dev);
+ rproc = pdata->rproc;
+ if (rproc) {
+ rproc_del(rproc);
+ rproc_free(rproc);
+ }
+ if (pdata->tx_chan)
+ mbox_free_channel(pdata->tx_chan);
+ if (pdata->rx_chan)
+ mbox_free_channel(pdata->rx_chan);
+ /* Discard all SKBs */
+ while (!skb_queue_empty(&pdata->tx_mc_skbs)) {
+ skb = skb_dequeue(&pdata->tx_mc_skbs);
+ kfree_skb(skb);
+ }
+
+ put_device(dev->parent);
+}
+
+/**
+ * event_notified_idr_cb() - event notified idr callback
+ * @id: idr id
+ * @ptr: pointer to idr private data
+ * @data: data passed to idr_for_each callback
+ *
+ * Pass notification to remoteproc virtio
+ *
+ * Return: 0. having return is to satisfy the idr_for_each() function
+ * pointer input argument requirement.
+ **/
+static int event_notified_idr_cb(int id, void *ptr, void *data)
+{
+ struct rproc *rproc = data;
+
+ (void)rproc_vq_interrupt(rproc, id);
+ return 0;
+}
+
+/**
+ * handle_event_notified() - remoteproc notification work funciton
+ * @work: pointer to the work structure
+ *
+ * It checks each registered remoteproc notify IDs.
+ */
+static void handle_event_notified(struct work_struct *work)
+{
+ struct rproc *rproc;
+ struct zynqmp_r5_pdata *pdata;
+
+ pdata = container_of(work, struct zynqmp_r5_pdata, mbox_work);
+
+ (void)mbox_send_message(pdata->rx_chan, NULL);
+ rproc = pdata->rproc;
+ /*
+ * We only use IPI for interrupt. The firmware side may or may
+ * not write the notifyid when it trigger IPI.
+ * And thus, we scan through all the registered notifyids.
+ */
+ idr_for_each(&rproc->notifyids, event_notified_idr_cb, rproc);
+}
+
+/**
+ * zynqmp_r5_mb_rx_cb() - Receive channel mailbox callback
+ * @cl: mailbox client
+ * @mssg: message pointer
+ *
+ * It will schedule the R5 notification work.
+ */
+static void zynqmp_r5_mb_rx_cb(struct mbox_client *cl, void *mssg)
+{
+ struct zynqmp_r5_pdata *pdata;
+
+ pdata = container_of(cl, struct zynqmp_r5_pdata, rx_mc);
+ if (mssg) {
+ struct zynqmp_ipi_message *ipi_msg, *buf_msg;
+ size_t len;
+
+ ipi_msg = (struct zynqmp_ipi_message *)mssg;
+ buf_msg = (struct zynqmp_ipi_message *)pdata->rx_mc_buf;
+ len = (ipi_msg->len >= IPI_BUF_LEN_MAX) ?
+ IPI_BUF_LEN_MAX : ipi_msg->len;
+ buf_msg->len = len;
+ memcpy(buf_msg->data, ipi_msg->data, len);
+ }
+ schedule_work(&pdata->mbox_work);
+}
+
+/**
+ * zynqmp_r5_mb_tx_done() - Request has been sent to the remote
+ * @cl: mailbox client
+ * @mssg: pointer to the message which has been sent
+ * @r: status of last TX - OK or error
+ *
+ * It will be called by the mailbox framework when the last TX has done.
+ */
+static void zynqmp_r5_mb_tx_done(struct mbox_client *cl, void *mssg, int r)
+{
+ struct zynqmp_r5_pdata *pdata;
+ struct sk_buff *skb;
+
+ if (!mssg)
+ return;
+ pdata = container_of(cl, struct zynqmp_r5_pdata, tx_mc);
+ skb = skb_dequeue(&pdata->tx_mc_skbs);
+ kfree_skb(skb);
+}
+
+/**
+ * zynqmp_r5_setup_mbox() - Setup mailboxes
+ *
+ * @pdata: pointer to the ZynqMP R5 processor platform data
+ * @node: pointer of the device node
+ *
+ * Function to setup mailboxes to talk to RPU.
+ *
+ * Return: 0 for success, negative value for failure.
+ */
+static int zynqmp_r5_setup_mbox(struct zynqmp_r5_pdata *pdata,
+ struct device_node *node)
+{
+ struct device *dev = &pdata->dev;
+ struct mbox_client *mclient;
+
+ /* Setup TX mailbox channel client */
+ mclient = &pdata->tx_mc;
+ mclient->dev = dev;
+ mclient->rx_callback = NULL;
+ mclient->tx_block = false;
+ mclient->knows_txdone = false;
+ mclient->tx_done = zynqmp_r5_mb_tx_done;
+
+ /* Setup TX mailbox channel client */
+ mclient = &pdata->rx_mc;
+ mclient->dev = dev;
+ mclient->rx_callback = zynqmp_r5_mb_rx_cb;
+ mclient->tx_block = false;
+ mclient->knows_txdone = false;
+
+ INIT_WORK(&pdata->mbox_work, handle_event_notified);
+
+ /* Request TX and RX channels */
+ pdata->tx_chan = mbox_request_channel_byname(&pdata->tx_mc, "tx");
+ if (IS_ERR(pdata->tx_chan)) {
+ dev_err(dev, "failed to request mbox tx channel.\n");
+ pdata->tx_chan = NULL;
+ return -EINVAL;
+ }
+ pdata->rx_chan = mbox_request_channel_byname(&pdata->rx_mc, "rx");
+ if (IS_ERR(pdata->rx_chan)) {
+ dev_err(dev, "failed to request mbox rx channel.\n");
+ pdata->rx_chan = NULL;
+ return -EINVAL;
+ }
+ skb_queue_head_init(&pdata->tx_mc_skbs);
+ return 0;
+}
+
+/**
+ * zynqmp_r5_probe() - Probes ZynqMP R5 processor device node
+ * @pdata: pointer to the ZynqMP R5 processor platform data
+ * @pdev: parent RPU domain platform device
+ * @node: pointer of the device node
+ *
+ * Function to retrieve the information of the ZynqMP R5 device node.
+ *
+ * Return: 0 for success, negative value for failure.
+ */
+static int zynqmp_r5_probe(struct zynqmp_r5_pdata *pdata,
+ struct platform_device *pdev,
+ struct device_node *node)
+{
+ struct device *dev = &pdata->dev;
+ struct rproc *rproc;
+ struct device_node *nc;
+ int ret;
+
+ /* Create device for ZynqMP R5 device */
+ dev->parent = &pdev->dev;
+ dev->release = zynqmp_r5_release;
+ dev->of_node = node;
+ dev_set_name(dev, "%s", of_node_full_name(node));
+ dev_set_drvdata(dev, pdata);
+ ret = device_register(dev);
+ if (ret) {
+ dev_err(dev, "failed to register device.\n");
+ return ret;
+ }
+ get_device(&pdev->dev);
+
+ /* Allocate remoteproc instance */
+ rproc = rproc_alloc(dev, dev_name(dev), &zynqmp_r5_rproc_ops, NULL, 0);
+ if (!rproc) {
+ dev_err(dev, "rproc allocation failed.\n");
+ ret = -ENOMEM;
+ goto error;
+ }
+ rproc->auto_boot = autoboot;
+ pdata->rproc = rproc;
+ rproc->priv = pdata;
+
+ /*
+ * The device has not been spawned from a device tree, so
+ * arch_setup_dma_ops has not been called, thus leaving
+ * the device with dummy DMA ops.
+ * Fix this by inheriting the parent's DMA ops and mask.
+ */
+ rproc->dev.dma_mask = pdev->dev.dma_mask;
+ set_dma_ops(&rproc->dev, get_dma_ops(&pdev->dev));
+
+ /* Probe R5 memory devices */
+ INIT_LIST_HEAD(&pdata->mems);
+ for_each_available_child_of_node(node, nc) {
+ ret = zynqmp_r5_mem_probe(pdata, nc);
+ if (ret) {
+ dev_err(dev, "failed to probe memory %s.\n",
+ of_node_full_name(nc));
+ goto error;
+ }
+ }
+
+ /* Set up DMA mask */
+ ret = dma_set_coherent_mask(dev, DMA_BIT_MASK(32));
+ if (ret) {
+ dev_warn(dev, "dma_set_coherent_mask failed: %d\n", ret);
+ /* If DMA is not configured yet, try to configure it. */
+ ret = of_dma_configure(dev, node, true);
+ if (ret) {
+ dev_err(dev, "failed to configure DMA.\n");
+ goto error;
+ }
+ }
+
+ /* Get R5 power domain node */
+ ret = of_property_read_u32(node, "pnode-id", &pdata->pnode_id);
+ if (ret) {
+ dev_err(dev, "failed to get power node id.\n");
+ goto error;
+ }
+
+ /* TODO Check if R5 is running */
+
+ /* Set up R5 if not already setup */
+ ret = pdata->is_r5_mode_set ? 0 : r5_set_mode(pdata);
+ if (ret) {
+ dev_err(dev, "failed to set R5 operation mode.\n");
+ return ret;
+ }
+
+ if (!of_get_property(dev->of_node, "mboxes", NULL)) {
+ dev_dbg(dev, "no mailboxes.\n");
+ goto error;
+ } else {
+ ret = zynqmp_r5_setup_mbox(pdata, node);
+ if (ret < 0)
+ goto error;
+ }
+
+ /* Add R5 remoteproc */
+ ret = rproc_add(rproc);
+ if (ret) {
+ dev_err(dev, "rproc registration failed\n");
+ goto error;
+ }
+ return 0;
+error:
+ if (pdata->rproc)
+ rproc_free(pdata->rproc);
+ pdata->rproc = NULL;
+ device_unregister(dev);
+ put_device(&pdev->dev);
+ return ret;
+}
+
+static int zynqmp_r5_remoteproc_probe(struct platform_device *pdev)
+{
+ int ret, i = 0;
+ u32 *lockstep_mode;
+ struct device *dev = &pdev->dev;
+ struct device_node *nc;
+ struct zynqmp_r5_pdata *pdata;
+
+ pdata = devm_kzalloc(dev, sizeof(*pdata), GFP_KERNEL);
+ lockstep_mode = devm_kzalloc(dev, sizeof(u32 *), GFP_KERNEL);
+ if (!pdata || !lockstep_mode)
+ return -ENOMEM;
+
+ platform_set_drvdata(pdev, pdata);
+
+ of_property_read_u32(dev->of_node, "lockstep-mode", lockstep_mode);
+
+ if (!(*lockstep_mode)) {
+ rpu_mode = PM_RPU_MODE_SPLIT;
+ } else if (*lockstep_mode == 1) {
+ rpu_mode = PM_RPU_MODE_LOCKSTEP;
+ } else {
+ dev_err(dev,
+ "Invalid lockstep-mode mode provided - %x %d\n",
+ *lockstep_mode, rpu_mode);
+ return -EINVAL;
+ }
+ dev_dbg(dev, "RPU configuration: %s\r\n",
+ (*lockstep_mode) ? "lockstep" : "split");
+
+ for_each_available_child_of_node(dev->of_node, nc) {
+ ret = zynqmp_r5_probe(&rpus[i], pdev, nc);
+ if (ret) {
+ dev_err(dev, "failed to probe rpu %s.\n",
+ of_node_full_name(nc));
+ return ret;
+ }
+ i++;
+ }
+
+ return 0;
+}
+
+static int zynqmp_r5_remoteproc_remove(struct platform_device *pdev)
+{
+ int i;
+
+ for (i = 0; i < MAX_RPROCS; i++) {
+ struct zynqmp_r5_pdata *pdata = &rpus[i];
+ struct rproc *rproc;
+
+ rproc = pdata->rproc;
+ if (rproc) {
+ rproc_del(rproc);
+ rproc_free(rproc);
+ pdata->rproc = NULL;
+ }
+ if (pdata->tx_chan) {
+ mbox_free_channel(pdata->tx_chan);
+ pdata->tx_chan = NULL;
+ }
+ if (pdata->rx_chan) {
+ mbox_free_channel(pdata->rx_chan);
+ pdata->rx_chan = NULL;
+ }
+
+ device_unregister(&pdata->dev);
+ }
+
+ return 0;
+}
+
+/* Match table for OF platform binding */
+static const struct of_device_id zynqmp_r5_remoteproc_match[] = {
+ { .compatible = "xlnx,zynqmp-r5-remoteproc-1.0", },
+ { /* end of list */ },
+};
+MODULE_DEVICE_TABLE(of, zynqmp_r5_remoteproc_match);
+
+static struct platform_driver zynqmp_r5_remoteproc_driver = {
+ .probe = zynqmp_r5_remoteproc_probe,
+ .remove = zynqmp_r5_remoteproc_remove,
+ .driver = {
+ .name = "zynqmp_r5_remoteproc",
+ .of_match_table = zynqmp_r5_remoteproc_match,
+ },
+};
+module_platform_driver(zynqmp_r5_remoteproc_driver);
+
+module_param_named(autoboot, autoboot, bool, 0444);
+MODULE_PARM_DESC(autoboot,
+ "enable | disable autoboot. (default: false)");
+
+MODULE_AUTHOR("Ben Levinsky <[email protected]>");
+MODULE_LICENSE("GPL v2");
--
2.17.1

2020-07-15 15:36:24

by Ben Levinsky

[permalink] [raw]
Subject: [PATCH v6 2/5] firmware: xilinx: Add shutdown/wakeup APIs

Add shutdown/wakeup a resource eemi operations to shutdown
or bringup a resource.

Signed-off-by: Ben Levinsky <[email protected]>
---
v3:
- add xilinx-related platform mgmt fn's instead of wrapping around
function pointer in xilinx eemi ops struct
- fix formatting
v4:
- add default values for enum
---
drivers/firmware/xilinx/zynqmp.c | 35 ++++++++++++++++++++++++++++
include/linux/firmware/xlnx-zynqmp.h | 22 +++++++++++++++++
2 files changed, 57 insertions(+)

diff --git a/drivers/firmware/xilinx/zynqmp.c b/drivers/firmware/xilinx/zynqmp.c
index 8d1ff2454e2e..f1a0bd35deeb 100644
--- a/drivers/firmware/xilinx/zynqmp.c
+++ b/drivers/firmware/xilinx/zynqmp.c
@@ -846,6 +846,41 @@ int zynqmp_pm_release_node(const u32 node)
}
EXPORT_SYMBOL_GPL(zynqmp_pm_release_node);

+/**
+ * zynqmp_pm_force_powerdown - PM call to request for another PU or subsystem to
+ * be powered down forcefully
+ * @target: Node ID of the targeted PU or subsystem
+ * @ack: Flag to specify whether acknowledge is requested
+ *
+ * Return: status, either success or error+reason
+ */
+int zynqmp_pm_force_powerdown(const u32 target,
+ const enum zynqmp_pm_request_ack ack)
+{
+ return zynqmp_pm_invoke_fn(PM_FORCE_POWERDOWN, target, ack, 0, 0, NULL);
+}
+EXPORT_SYMBOL_GPL(zynqmp_pm_force_powerdown);
+
+/**
+ * zynqmp_pm_request_wakeup - PM call to wake up selected master or subsystem
+ * @node: Node ID of the master or subsystem
+ * @set_addr: Specifies whether the address argument is relevant
+ * @address: Address from which to resume when woken up
+ * @ack: Flag to specify whether acknowledge requested
+ *
+ * Return: status, either success or error+reason
+ */
+int zynqmp_pm_request_wakeup(const u32 node,
+ const bool set_addr,
+ const u64 address,
+ const enum zynqmp_pm_request_ack ack)
+{
+ /* set_addr flag is encoded into 1st bit of address */
+ return zynqmp_pm_invoke_fn(PM_REQUEST_WAKEUP, node, address | set_addr,
+ address >> 32, ack, NULL);
+}
+EXPORT_SYMBOL_GPL(zynqmp_pm_request_wakeup);
+
/**
* zynqmp_pm_set_requirement() - PM call to set requirement for PM slaves
* @node: Node ID of the slave
diff --git a/include/linux/firmware/xlnx-zynqmp.h b/include/linux/firmware/xlnx-zynqmp.h
index bb347dfe4ba4..4d83a7d69c4c 100644
--- a/include/linux/firmware/xlnx-zynqmp.h
+++ b/include/linux/firmware/xlnx-zynqmp.h
@@ -64,6 +64,8 @@

enum pm_api_id {
PM_GET_API_VERSION = 1,
+ PM_FORCE_POWERDOWN = 8,
+ PM_REQUEST_WAKEUP = 10,
PM_SYSTEM_SHUTDOWN = 12,
PM_REQUEST_NODE = 13,
PM_RELEASE_NODE,
@@ -376,6 +378,12 @@ int zynqmp_pm_write_pggs(u32 index, u32 value);
int zynqmp_pm_read_pggs(u32 index, u32 *value);
int zynqmp_pm_system_shutdown(const u32 type, const u32 subtype);
int zynqmp_pm_set_boot_health_status(u32 value);
+int zynqmp_pm_force_powerdown(const u32 target,
+ const enum zynqmp_pm_request_ack ack);
+int zynqmp_pm_request_wakeup(const u32 node,
+ const bool set_addr,
+ const u64 address,
+ const enum zynqmp_pm_request_ack ack);
#else
static inline struct zynqmp_eemi_ops *zynqmp_pm_get_eemi_ops(void)
{
@@ -526,6 +534,20 @@ static inline int zynqmp_pm_set_boot_health_status(u32 value)
{
return -ENODEV;
}
+
+static inline int zynqmp_pm_force_powerdown(const u32 target,
+ const enum zynqmp_pm_request_ack ack)
+{
+ return -ENODEV;
+}
+
+static inline int zynqmp_pm_request_wakeup(const u32 node,
+ const bool set_addr,
+ const u64 address,
+ const enum zynqmp_pm_request_ack ack)
+{
+ return -ENODEV;
+}
#endif

#endif /* __FIRMWARE_ZYNQMP_H__ */
--
2.17.1

2020-07-15 15:36:37

by Ben Levinsky

[permalink] [raw]
Subject: [PATCH v6 3/5] firmware: xilinx: Add RPU configuration APIs

This patch adds APIs to provide access and a configuration interface
to the current power state of a sub-system on Zynqmp sub-system.

Signed-off-by: Ben Levinsky <[email protected]>
---
v3:
- add xilinx-related platform mgmt fn's instead of wrapping around
function pointer in xilinx eemi ops struct
v4:
- add default values for enums
---
drivers/firmware/xilinx/zynqmp.c | 99 ++++++++++++++++++++++++++++
include/linux/firmware/xlnx-zynqmp.h | 34 ++++++++++
2 files changed, 133 insertions(+)

diff --git a/drivers/firmware/xilinx/zynqmp.c b/drivers/firmware/xilinx/zynqmp.c
index f1a0bd35deeb..fdd61d258e55 100644
--- a/drivers/firmware/xilinx/zynqmp.c
+++ b/drivers/firmware/xilinx/zynqmp.c
@@ -846,6 +846,61 @@ int zynqmp_pm_release_node(const u32 node)
}
EXPORT_SYMBOL_GPL(zynqmp_pm_release_node);

+/**
+ * zynqmp_pm_get_rpu_mode() - Get RPU mode
+ * @node_id: Node ID of the device
+ * @arg1: Argument 1 to requested IOCTL call
+ * @arg2: Argument 2 to requested IOCTL call
+ * @out: Returned output value
+ *
+ * Return: RPU mode
+ */
+int zynqmp_pm_get_rpu_mode(u32 node_id,
+ u32 arg1, u32 arg2, u32 *out)
+{
+ return zynqmp_pm_invoke_fn(PM_IOCTL, node_id,
+ IOCTL_GET_RPU_OPER_MODE, 0, 0, out);
+}
+EXPORT_SYMBOL_GPL(zynqmp_pm_get_rpu_mode);
+
+/**
+ * zynqmp_pm_set_rpu_mode() - Set RPU mode
+ * @node_id: Node ID of the device
+ * @ioctl_id: ID of the requested IOCTL
+ * @arg2: Argument 2 to requested IOCTL call
+ * @out: Returned output value
+ *
+ * This function is used to set RPU mode.
+ *
+ * Return: Returns status, either success or error+reason
+ */
+int zynqmp_pm_set_rpu_mode(u32 node_id,
+ u32 arg1, u32 arg2, u32 *out)
+{
+ return zynqmp_pm_invoke_fn(PM_IOCTL, node_id,
+ IOCTL_SET_RPU_OPER_MODE, 0, 0, out);
+}
+EXPORT_SYMBOL_GPL(zynqmp_pm_set_rpu_mode);
+
+/**
+ * zynqmp_pm_tcm_comb_config - configure TCM
+ * @node_id: Node ID of the device
+ * @arg1: Argument 1 to requested IOCTL call
+ * @arg2: Argument 2 to requested IOCTL call
+ * @out: Returned output value
+ *
+ * This function is used to set RPU mode.
+ *
+ * Return: Returns status, either success or error+reason
+ */
+int zynqmp_pm_set_tcm_config(u32 node_id,
+ u32 arg1, u32 arg2, u32 *out)
+{
+ return zynqmp_pm_invoke_fn(PM_IOCTL, node_id,
+ IOCTL_TCM_COMB_CONFIG, 0, 0, out);
+}
+EXPORT_SYMBOL_GPL(zynqmp_pm_set_tcm_config);
+
/**
* zynqmp_pm_force_powerdown - PM call to request for another PU or subsystem to
* be powered down forcefully
@@ -881,6 +936,50 @@ int zynqmp_pm_request_wakeup(const u32 node,
}
EXPORT_SYMBOL_GPL(zynqmp_pm_request_wakeup);

+/**
+ * zynqmp_pm_get_node_status - PM call to request a node's current power state
+ * @node: ID of the component or sub-system in question
+ * @status: Current operating state of the requested node
+ * @requirements: Current requirements asserted on the node,
+ * used for slave nodes only.
+ * @usage: Usage information, used for slave nodes only:
+ * PM_USAGE_NO_MASTER - No master is currently using
+ * the node
+ * PM_USAGE_CURRENT_MASTER - Only requesting master is
+ * currently using the node
+ * PM_USAGE_OTHER_MASTER - Only other masters are
+ * currently using the node
+ * PM_USAGE_BOTH_MASTERS - Both the current and at least
+ * one other master is currently
+ * using the node
+ *
+ * Return: status, either success or error+reason
+ */
+int zynqmp_pm_get_node_status(const u32 node,
+ u32 *status, u32 *requirements,
+ u32 *usage)
+
+{
+ u32 ret_payload[PAYLOAD_ARG_CNT];
+ int ret;
+
+ if (!status)
+ return -EINVAL;
+
+ ret = zynqmp_pm_invoke_fn(PM_GET_NODE_STATUS, node, 0, 0, 0,
+ ret_payload);
+ if (ret_payload[0] == XST_PM_SUCCESS) {
+ *status = ret_payload[1];
+ if (requirements)
+ *requirements = ret_payload[2];
+ if (usage)
+ *usage = ret_payload[3];
+ }
+
+ return ret;
+}
+EXPORT_SYMBOL_GPL(zynqmp_pm_get_node_status);
+
/**
* zynqmp_pm_set_requirement() - PM call to set requirement for PM slaves
* @node: Node ID of the slave
diff --git a/include/linux/firmware/xlnx-zynqmp.h b/include/linux/firmware/xlnx-zynqmp.h
index 4d83a7d69c4c..0caca9e2223a 100644
--- a/include/linux/firmware/xlnx-zynqmp.h
+++ b/include/linux/firmware/xlnx-zynqmp.h
@@ -64,6 +64,7 @@

enum pm_api_id {
PM_GET_API_VERSION = 1,
+ PM_GET_NODE_STATUS = 3,
PM_FORCE_POWERDOWN = 8,
PM_REQUEST_WAKEUP = 10,
PM_SYSTEM_SHUTDOWN = 12,
@@ -384,6 +385,14 @@ int zynqmp_pm_request_wakeup(const u32 node,
const bool set_addr,
const u64 address,
const enum zynqmp_pm_request_ack ack);
+int zynqmp_pm_get_node_status(const u32 node, u32 *status,
+ u32 *requirements, u32 *usage);
+int zynqmp_pm_get_rpu_mode(u32 node_id,
+ u32 arg1, u32 arg2, u32 *out);
+int zynqmp_pm_set_rpu_mode(u32 node_id,
+ u32 arg1, u32 arg2, u32 *out);
+int zynqmp_pm_set_tcm_config(u32 node_id,
+ u32 arg1, u32 arg2, u32 *out);
#else
static inline struct zynqmp_eemi_ops *zynqmp_pm_get_eemi_ops(void)
{
@@ -548,6 +557,31 @@ static inline int zynqmp_pm_request_wakeup(const u32 node,
{
return -ENODEV;
}
+
+static inline int zynqmp_pm_get_node_status(const u32 node,
+ u32 *status, u32 *requirements,
+ u32 *usage)
+{
+ return -ENODEV;
+}
+
+static inline int zynqmp_pm_get_rpu_mode(u32 node_id,
+ u32 arg1, u32 arg2, u32 *out)
+{
+ return -ENODEV;
+}
+
+static inline int zynqmp_pm_set_rpu_mode(u32 node_id,
+ u32 arg1, u32 arg2, u32 *out)
+{
+ return -ENODEV;
+}
+
+static inline int zynqmp_pm_set_tcm_config(u32 node_id,
+ u32 arg1, u32 arg2, u32 *out)
+{
+ return -ENODEV;
+}
#endif

#endif /* __FIRMWARE_ZYNQMP_H__ */
--
2.17.1

2020-07-27 23:03:14

by Mathieu Poirier

[permalink] [raw]
Subject: Re: [PATCH v6 5/5] remoteproc: Add initial zynqmp R5 remoteproc driver

On Wed, Jul 15, 2020 at 08:33:17AM -0700, Ben Levinsky wrote:
> R5 is included in Xilinx Zynq UltraScale MPSoC so by adding this
> remotproc driver, we can boot the R5 sub-system in different
> configurations.
>
> Acked-by: Stefano Stabellini <[email protected]>
> Acked-by: Ben Levinsky <[email protected]>
> Reviewed-by: Radhey Shyam Pandey <[email protected]>
> Signed-off-by: Ben Levinsky <[email protected]>
> Signed-off-by: Wendy Liang <[email protected]>
> Signed-off-by: Michal Simek <[email protected]>
> Signed-off-by: Ed Mooring <[email protected]>
> Signed-off-by: Jason Wu <[email protected]>
> Tested-by: Ben Levinsky <[email protected]>
> ---
> v2:
> - remove domain struct as per review from Mathieu
> v3:
> - add xilinx-related platform mgmt fn's instead of wrapping around
> function pointer in xilinx eemi ops struct
> v4:
> - add default values for enums
> - fix formatting as per checkpatch.pl --strict. Note that 1 warning and 1 check
> are still raised as each is due to fixing the warning results in that
> particular line going over 80 characters.
> v5:
> - parse_fw change from use of rproc_of_resm_mem_entry_init to rproc_mem_entry_init and use of alloc/release
> - var's of type zynqmp_r5_pdata all have same local variable name
> - use dev_dbg instead of dev_info
> v6:
> - adding memory carveouts is handled much more similarly. All mem carveouts are
> now described in reserved memory as needed. That is, TCM nodes are not
> coupled to remoteproc anymore. This is reflected in the remoteproc R5 driver
> and the device tree binding.
> - remove mailbox from device tree binding as it is not necessary for elf
> loading
> - use lockstep-mode property for configuring RPU
> ---
> drivers/remoteproc/Kconfig | 10 +
> drivers/remoteproc/Makefile | 1 +
> drivers/remoteproc/zynqmp_r5_remoteproc.c | 911 ++++++++++++++++++++++
> 3 files changed, 922 insertions(+)
> create mode 100644 drivers/remoteproc/zynqmp_r5_remoteproc.c
>
> diff --git a/drivers/remoteproc/Kconfig b/drivers/remoteproc/Kconfig
> index c4d1731295eb..342a7e668636 100644
> --- a/drivers/remoteproc/Kconfig
> +++ b/drivers/remoteproc/Kconfig
> @@ -249,6 +249,16 @@ config STM32_RPROC
>
> This can be either built-in or a loadable module.
>
> +config ZYNQMP_R5_REMOTEPROC
> + tristate "ZynqMP_R5 remoteproc support"
> + depends on ARM64 && PM && ARCH_ZYNQMP
> + select RPMSG_VIRTIO
> + select MAILBOX
> + select ZYNQMP_IPI_MBOX
> + help
> + Say y here to support ZynqMP R5 remote processors via the remote
> + processor framework.
> +
> endif # REMOTEPROC
>
> endmenu
> diff --git a/drivers/remoteproc/Makefile b/drivers/remoteproc/Makefile
> index e8b886e511f0..04d1c95d06d7 100644
> --- a/drivers/remoteproc/Makefile
> +++ b/drivers/remoteproc/Makefile
> @@ -28,5 +28,6 @@ obj-$(CONFIG_QCOM_WCNSS_PIL) += qcom_wcnss_pil.o
> qcom_wcnss_pil-y += qcom_wcnss.o
> qcom_wcnss_pil-y += qcom_wcnss_iris.o
> obj-$(CONFIG_ST_REMOTEPROC) += st_remoteproc.o
> +obj-$(CONFIG_ZYNQMP_R5_REMOTEPROC) += zynqmp_r5_remoteproc.o
> obj-$(CONFIG_ST_SLIM_REMOTEPROC) += st_slim_rproc.o
> obj-$(CONFIG_STM32_RPROC) += stm32_rproc.o
> diff --git a/drivers/remoteproc/zynqmp_r5_remoteproc.c b/drivers/remoteproc/zynqmp_r5_remoteproc.c
> new file mode 100644
> index 000000000000..b600759e257e
> --- /dev/null
> +++ b/drivers/remoteproc/zynqmp_r5_remoteproc.c
> @@ -0,0 +1,911 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Zynq R5 Remote Processor driver
> + *
> + * Copyright (C) 2019, 2020 Xilinx Inc. Ben Levinsky <[email protected]>
> + * Copyright (C) 2015 - 2018 Xilinx Inc.
> + * Copyright (C) 2015 Jason Wu <[email protected]>
> + *
> + * Based on origin OMAP and Zynq Remote Processor driver
> + *
> + * Copyright (C) 2012 Michal Simek <[email protected]>
> + * Copyright (C) 2012 PetaLogix
> + * Copyright (C) 2011 Texas Instruments, Inc.
> + * Copyright (C) 2011 Google, Inc.
> + */
> +
> +#include <linux/atomic.h>

Unused

> +#include <linux/cpu.h>
> +#include <linux/dma-mapping.h>
> +#include <linux/delay.h>

Unused

> +#include <linux/err.h>
> +#include <linux/firmware/xlnx-zynqmp.h>
> +#include <linux/genalloc.h>

Unused

> +#include <linux/idr.h>
> +#include <linux/interrupt.h>
> +#include <linux/kernel.h>
> +#include <linux/list.h>
> +#include <linux/mailbox_client.h>
> +#include <linux/mailbox/zynqmp-ipi-message.h>
> +#include <linux/module.h>
> +#include <linux/of_address.h>
> +#include <linux/of_irq.h>
> +#include <linux/of_platform.h>
> +#include <linux/of_reserved_mem.h>
> +#include <linux/pfn.h>

Unused

> +#include <linux/platform_device.h>
> +#include <linux/remoteproc.h>
> +#include <linux/skbuff.h>
> +#include <linux/slab.h>
> +#include <linux/sysfs.h>
> +
> +#include "remoteproc_internal.h"
> +
> +#define MAX_RPROCS 2 /* Support up to 2 RPU */
> +#define MAX_MEM_PNODES 4 /* Max power nodes for one RPU memory instance */
> +
> +#define DEFAULT_FIRMWARE_NAME "rproc-rpu-fw"
> +
> +/* PM proc states */
> +#define PM_PROC_STATE_ACTIVE 1U

Unused

> +
> +/* IPI buffer MAX length */
> +#define IPI_BUF_LEN_MAX 32U
> +/* RX mailbox client buffer max length */
> +#define RX_MBOX_CLIENT_BUF_MAX (IPI_BUF_LEN_MAX + \
> + sizeof(struct zynqmp_ipi_message))
> +
> +#define ZYNQMP_R5_NUM_TCM_BANKS 4
> +
> +/* lookup table mapping power-node-ID of TCM bank to absolute base address */
> +static unsigned long tcm_addr_to_pnode[ZYNQMP_R5_NUM_TCM_BANKS][2] = {
> + {0xFFE00000, 0xF },
> + {0xFFE20000, 0x10},
> + {0xFFE90000, 0x10},
> + {0xFFEB0000, 0x11},
> +};
> +
> +static bool autoboot __read_mostly;
> +
> +/**
> + * struct zynqmp_r5_mem - zynqmp rpu memory data
> + * @pnode_id: TCM power domain ids
> + * @res: memory resource
> + * @node: list node
> + */
> +struct zynqmp_r5_mem {
> + u32 pnode_id[MAX_MEM_PNODES];
> + struct resource res;
> + struct list_head node;
> +};
> +
> +/**
> + * struct zynqmp_r5_pdata - zynqmp rpu remote processor private data
> + * @dev: device of RPU instance
> + * @rproc: rproc handle
> + * @pnode_id: RPU CPU power domain id
> + * @mems: memory resources
> + * @is_r5_mode_set: indicate if r5 operation mode is set
> + * @tx_mc: tx mailbox client
> + * @rx_mc: rx mailbox client
> + * @tx_chan: tx mailbox channel
> + * @rx_chan: rx mailbox channel
> + * @mbox_work: mbox_work for the RPU remoteproc
> + * @tx_mc_skbs: socket buffers for tx mailbox client
> + * @rx_mc_buf: rx mailbox client buffer to save the rx message
> + */
> +struct zynqmp_r5_pdata {
> + struct device dev;
> + struct rproc *rproc;
> + u32 pnode_id;
> + struct list_head mems;
> + bool is_r5_mode_set;
> + struct mbox_client tx_mc;
> + struct mbox_client rx_mc;
> + struct mbox_chan *tx_chan;
> + struct mbox_chan *rx_chan;
> + struct work_struct mbox_work;
> + struct sk_buff_head tx_mc_skbs;
> + unsigned char rx_mc_buf[RX_MBOX_CLIENT_BUF_MAX];
> +};
> +
> +/**
> + * table of RPUs
> + */
> +struct zynqmp_r5_pdata rpus[MAX_RPROCS];
> +/**
> + * RPU core configuration
> + */
> +enum rpu_oper_mode rpu_mode;
> +
> +/*
> + * r5_set_mode - set RPU operation mode
> + * @pdata: Remote processor private data
> + *
> + * set RPU oepration mode
> + *
> + * Return: 0 for success, negative value for failure
> + */
> +static int r5_set_mode(struct zynqmp_r5_pdata *pdata)
> +{
> + u32 val[PAYLOAD_ARG_CNT] = {0}, expect, tcm_mode;
> + struct device *dev = &pdata->dev;
> + int ret;
> +
> + expect = (u32)rpu_mode;
> + ret = zynqmp_pm_get_rpu_mode(pdata->pnode_id, 0, 0, val);
> + if (ret < 0) {
> + dev_err(dev, "failed to get RPU oper mode.\n");
> + return ret;
> + }
> + if (val[0] == expect) {
> + dev_dbg(dev, "RPU mode matches: %x\n", val[0]);
> + } else {
> + ret = zynqmp_pm_set_rpu_mode(pdata->pnode_id,
> + expect, 0, val);
> + if (ret < 0) {
> + dev_err(dev,
> + "failed to set RPU oper mode.\n");
> + return ret;
> + }
> + }
> +
> + tcm_mode = (expect == (u32)PM_RPU_MODE_LOCKSTEP) ?
> + PM_RPU_TCM_COMB : PM_RPU_TCM_SPLIT;
> + ret = zynqmp_pm_set_tcm_config(pdata->pnode_id, tcm_mode, 0, val);
> + if (ret < 0) {
> + dev_err(dev, "failed to config TCM to %x.\n",
> + expect);
> + return ret;
> + }
> + pdata->is_r5_mode_set = true;
> + return 0;
> +}
> +
> +/*
> + * ZynqMP R5 remoteproc memory release function
> + */
> +static int zynqmp_r5_mem_release(struct rproc *rproc,
> + struct rproc_mem_entry *mem)
> +{
> + struct zynqmp_r5_mem *priv;
> + int i, ret;
> + struct device *dev = &rproc->dev;
> +
> + priv = mem->priv;
> + if (!priv)
> + return 0;
> + for (i = 0; i < MAX_MEM_PNODES; i++) {
> + if (priv->pnode_id[i]) {
> + dev_dbg(dev, "%s, pnode %d\n",
> + __func__, priv->pnode_id[i]);
> + ret = zynqmp_pm_release_node(priv->pnode_id[i]);
> + if (ret < 0) {
> + dev_err(dev,
> + "failed to release power node: %u\n",
> + priv->pnode_id[i]);
> + return ret;
> + }
> + } else {
> + break;
> + }
> + }
> + return 0;
> +}
> +
> +/*
> + * ZynqMP R5 remoteproc operations
> + */
> +static int zynqmp_r5_rproc_start(struct rproc *rproc)
> +{
> + struct device *dev = rproc->dev.parent;
> + struct zynqmp_r5_pdata *pdata = rproc->priv;
> + enum rpu_boot_mem bootmem;
> + int ret;
> +
> + if ((rproc->bootaddr & 0xF0000000) == 0xF0000000)
> + bootmem = PM_RPU_BOOTMEM_HIVEC;
> + else
> + bootmem = PM_RPU_BOOTMEM_LOVEC;
> + dev_dbg(dev, "RPU boot from %s.",
> + bootmem == PM_RPU_BOOTMEM_HIVEC ? "OCM" : "TCM");
> + ret = zynqmp_pm_request_wakeup(pdata->pnode_id, 1,
> + bootmem, ZYNQMP_PM_REQUEST_ACK_NO);
> + if (ret < 0) {
> + dev_err(dev, "failed to boot R5.\n");
> + return ret;
> + }
> + return 0;
> +}
> +
> +static int zynqmp_r5_rproc_stop(struct rproc *rproc)
> +{
> + struct zynqmp_r5_pdata *pdata = rproc->priv;
> + int ret;
> +
> + ret = zynqmp_pm_force_powerdown(pdata->pnode_id,
> + ZYNQMP_PM_REQUEST_ACK_BLOCKING);
> + if (ret < 0) {
> + dev_err(&pdata->dev, "failed to shutdown R5.\n");
> + return ret;
> + }
> + return 0;
> +}
> +
> +static int zynqmp_r5_rproc_mem_alloc(struct rproc *rproc,
> + struct rproc_mem_entry *mem)
> +{
> + struct device *dev = rproc->dev.parent;
> + void *va;
> +
> + dev_dbg(rproc->dev.parent, "map memory: %pa\n", &mem->dma);
> + va = ioremap_wc(mem->dma, mem->len);
> + if (IS_ERR_OR_NULL(va)) {
> + dev_err(dev, "Unable to map memory region: %pa+%lx\n",
> + &mem->dma, mem->len);
> + return -ENOMEM;
> + }
> +
> + /* Update memory entry va */
> + mem->va = va;
> +
> + return 0;
> +}
> +
> +static int zynqmp_r5_rproc_mem_release(struct rproc *rproc,
> + struct rproc_mem_entry *mem)
> +{
> + dev_dbg(rproc->dev.parent, "unmap memory: %pa\n", &mem->dma);
> + iounmap(mem->va);
> +
> + return 0;
> +}
> +
> +/*
> + * TCM needs mapping to R5 relative address and xilinx platform mgmt call
> + */
> +struct rproc_mem_entry *handle_tcm_parsing(struct device *dev,
> + struct reserved_mem *rmem,
> + struct device_node *node,
> + int lookup_idx)
> +{
> + void *va;
> + dma_addr_t dma;
> + resource_size_t size;
> + int ret;
> + u32 pnode_id;
> + struct resource rsc;
> + struct rproc_mem_entry *mem;
> +
> + pnode_id = tcm_addr_to_pnode[lookup_idx][1];
> + ret = zynqmp_pm_request_node(pnode_id,
> + ZYNQMP_PM_CAPABILITY_ACCESS, 0,
> + ZYNQMP_PM_REQUEST_ACK_BLOCKING);
> + if (ret < 0) {
> + dev_err(dev, "failed to request power node: %u\n", pnode_id);
> + return -EINVAL;
> + }
> +
> + ret = of_address_to_resource(node, 0, &rsc);
> + if (ret < 0) {
> + dev_err(dev, "failed to get resource of memory %s",
> + of_node_full_name(node));
> + return -EINVAL;
> + }
> + size = resource_size(&rsc);
> + va = devm_ioremap_wc(dev, rsc.start, size);
> + if (!va)
> + return -ENOMEM;
> +
> + /* zero out tcm base address */
> + if (rsc.start & 0xffe00000) {
> + /* R5 can't see anything past 0xfffff so wipe it */
> + rsc.start &= 0x000fffff;
> + /*
> + * handle tcm banks 1 a and b (0xffe9000 and
> + * 0xffeb0000)
> + */
> + if (rsc.start & 0x80000)
> + rsc.start -= 0x90000;
> + }
> +
> + dma = (dma_addr_t)rsc.start;
> + mem = rproc_mem_entry_init(dev, va, dma, (int)size, rsc.start,
> + NULL, zynqmp_r5_mem_release,
> + rsc.name);
> + if (!mem)
> + return -ENOMEM;
> +
> + return mem;
> +}
> +
> +static int parse_mem_regions(struct rproc *rproc)
> +{
> + int num_mems, i;
> + struct zynqmp_r5_pdata *pdata = rproc->priv;
> + struct device *dev = &pdata->dev;
> + struct device_node *np = dev->of_node;
> + struct rproc_mem_entry *mem;
> +
> + num_mems = of_count_phandle_with_args(np, "memory-region", NULL);
> + if (num_mems <= 0)
> + return 0;
> + for (i = 0; i < num_mems; i++) {
> + struct device_node *node;
> + struct reserved_mem *rmem;
> +
> + node = of_parse_phandle(np, "memory-region", i);
> + rmem = of_reserved_mem_lookup(node);
> + if (!rmem) {
> + dev_err(dev, "unable to acquire memory-region\n");
> + return -EINVAL;
> + }
> +
> + if (strstr(node->name, "vdev0buffer")) {
> + /* Register DMA region */
> + mem = rproc_mem_entry_init(dev, NULL,
> + (dma_addr_t)rmem->base,
> + rmem->size, rmem->base,
> + NULL, NULL,
> + "vdev0buffer");
> + if (!mem) {
> + dev_err(dev, "unable to initialize memory-region %s\n",
> + node->name);
> + return -ENOMEM;
> + }
> + dev_dbg(dev, "parsed %s at %llx\r\n", mem->name,
> + mem->dma);
> + } else if (strstr(node->name, "vdev0vring")) {
> + int vring_id;
> + char name[16];
> +
> + /*
> + * can be 1 of multiple vring IDs per IPC channel
> + * e.g. 'vdev0vring0' and 'vdev0vring1'
> + */
> + vring_id = node->name[14] - '0';
> + snprintf(name, sizeof(name), "vdev0vring%d", vring_id);
> + /* Register vring */
> + mem = rproc_mem_entry_init(dev, NULL,
> + (dma_addr_t)rmem->base,
> + rmem->size, rmem->base,
> + zynqmp_r5_rproc_mem_alloc,
> + zynqmp_r5_rproc_mem_release,
> + name);
> + dev_dbg(dev, "parsed %s at %llx\r\n", mem->name,
> + mem->dma);
> + } else {
> + int idx;
> +
> + /*
> + * if TCM update address space for R5 and
> + * make xilinx platform mgmt call
> + */
> + for (idx = 0; idx < ZYNQMP_R5_NUM_TCM_BANKS; idx++) {
> + if (tcm_addr_to_pnode[idx][0] == rmem->base)
> + break;
> + }
> +
> + if (idx != ZYNQMP_R5_NUM_TCM_BANKS) {
> + mem = handle_tcm_parsing(dev, rmem, node, idx);
> + } else {
> + mem = rproc_mem_entry_init(dev, NULL,
> + (dma_addr_t)rmem->base,
> + rmem->size, rmem->base,
> + zynqmp_r5_rproc_mem_alloc,
> + zynqmp_r5_rproc_mem_release,
> + node->name);
> + }
> +
> + if (!mem) {
> + dev_err(dev,
> + "unable to init memory-region %s\n",
> + node->name);
> + return -ENOMEM;
> + }
> + }
> + rproc_add_carveout(rproc, mem);
> + }
> +
> + return 0;
> +}
> +
> +static int zynqmp_r5_parse_fw(struct rproc *rproc, const struct firmware *fw)
> +{
> + int ret;
> + struct zynqmp_r5_pdata *pdata = rproc->priv;
> + struct device *dev = &pdata->dev;
> +
> + ret = parse_mem_regions(rproc);
> + if (ret) {
> + dev_err(dev, "parse_mem_regions failed %x\n", ret);
> + return ret;
> + }
> +
> + ret = rproc_elf_load_rsc_table(rproc, fw);
> + if (ret == -EINVAL) {
> + dev_dbg(dev, "no resource table found.\n");
> + ret = 0;
> + }
> + return ret;
> +}
> +
> +/* kick a firmware */
> +static void zynqmp_r5_rproc_kick(struct rproc *rproc, int vqid)
> +{
> + struct device *dev = rproc->dev.parent;
> + struct zynqmp_r5_pdata *pdata = rproc->priv;
> +
> + dev_dbg(dev, "KICK Firmware to start send messages vqid %d\n", vqid);
> +
> + if (vqid < 0) {
> + /* If vqid is negative, does not pass the vqid to
> + * mailbox. As vqid is supposed to be 0 or possive.
> + * It also gives a way to just kick instead but
> + * not use the IPI buffer. It is better to provide
> + * a proper way to pass the short message, which will
> + * need to sync to upstream first, for now,
> + * use negative vqid to assume no message will be
> + * passed with IPI buffer, but just raise interrupt.
> + * This will be faster as it doesn't need to copy the
> + * message to the IPI buffer.
> + *
> + * It will ignore the return, as failure is due to
> + * there already kicks in the mailbox queue.
> + */
> + (void)mbox_send_message(pdata->tx_chan, NULL);
> + } else {
> + struct sk_buff *skb;
> + unsigned int skb_len;
> + struct zynqmp_ipi_message *mb_msg;
> + int ret;
> +
> + skb_len = (unsigned int)(sizeof(vqid) + sizeof(mb_msg));
> + skb = alloc_skb(skb_len, GFP_ATOMIC);
> + if (!skb) {
> + dev_err(dev,
> + "Failed to allocate skb to kick remote.\n");
> + return;
> + }
> + mb_msg = (struct zynqmp_ipi_message *)skb_put(skb, skb_len);
> + mb_msg->len = sizeof(vqid);
> + memcpy(mb_msg->data, &vqid, sizeof(vqid));
> + skb_queue_tail(&pdata->tx_mc_skbs, skb);
> + ret = mbox_send_message(pdata->tx_chan, mb_msg);
> + if (ret < 0) {
> + dev_warn(dev, "Failed to kick remote.\n");
> + skb_dequeue_tail(&pdata->tx_mc_skbs);
> + kfree_skb(skb);
> + }
> + }
> +}
> +
> +static struct rproc_ops zynqmp_r5_rproc_ops = {
> + .start = zynqmp_r5_rproc_start,
> + .stop = zynqmp_r5_rproc_stop,
> + .load = rproc_elf_load_segments,
> + .parse_fw = zynqmp_r5_parse_fw,
> + .find_loaded_rsc_table = rproc_elf_find_loaded_rsc_table,
> + .sanity_check = rproc_elf_sanity_check,
> + .get_boot_addr = rproc_elf_get_boot_addr,
> + .kick = zynqmp_r5_rproc_kick,
> +};
> +
> +/* zynqmp_r5_mem_probe() - probes RPU TCM memory device
> + * @pdata: pointer to the RPU remoteproc private data
> + * @node: pointer to the memory node
> + *
> + * Function to retrieve resources for RPU TCM memory device.
> + */
> +static int zynqmp_r5_mem_probe(struct zynqmp_r5_pdata *pdata,
> + struct device_node *node)
> +{
> + struct device *dev;
> + struct zynqmp_r5_mem *mem;
> + int ret;
> + struct property *prop;
> + const __be32 *cur;
> + u32 val;
> + int i;
> +
> + dev = &pdata->dev;
> + mem = devm_kzalloc(dev, sizeof(*mem), GFP_KERNEL);
> + if (!mem)
> + return -ENOMEM;
> + ret = of_address_to_resource(node, 0, &mem->res);
> + if (ret < 0) {
> + dev_err(dev, "failed to get resource of memory %s",
> + of_node_full_name(node));
> + return -EINVAL;
> + }
> +
> + /* Get the power domain id */
> + i = 0;
> + if (of_find_property(node, "pnode-id", NULL)) {
> + of_property_for_each_u32(node, "pnode-id", prop, cur, val)
> + mem->pnode_id[i++] = val;
> + }
> + list_add_tail(&mem->node, &pdata->mems);
> + return 0;
> +}
> +
> +/**
> + * zynqmp_r5_release() - ZynqMP R5 device release function
> + * @dev: pointer to the device struct of ZynqMP R5
> + *
> + * Function to release ZynqMP R5 device.
> + */
> +static void zynqmp_r5_release(struct device *dev)
> +{
> + struct zynqmp_r5_pdata *pdata;
> + struct rproc *rproc;
> + struct sk_buff *skb;
> +
> + pdata = dev_get_drvdata(dev);
> + rproc = pdata->rproc;
> + if (rproc) {
> + rproc_del(rproc);
> + rproc_free(rproc);
> + }
> + if (pdata->tx_chan)
> + mbox_free_channel(pdata->tx_chan);
> + if (pdata->rx_chan)
> + mbox_free_channel(pdata->rx_chan);
> + /* Discard all SKBs */
> + while (!skb_queue_empty(&pdata->tx_mc_skbs)) {
> + skb = skb_dequeue(&pdata->tx_mc_skbs);
> + kfree_skb(skb);
> + }
> +
> + put_device(dev->parent);
> +}
> +
> +/**
> + * event_notified_idr_cb() - event notified idr callback
> + * @id: idr id
> + * @ptr: pointer to idr private data
> + * @data: data passed to idr_for_each callback
> + *
> + * Pass notification to remoteproc virtio
> + *
> + * Return: 0. having return is to satisfy the idr_for_each() function
> + * pointer input argument requirement.
> + **/
> +static int event_notified_idr_cb(int id, void *ptr, void *data)
> +{
> + struct rproc *rproc = data;
> +
> + (void)rproc_vq_interrupt(rproc, id);
> + return 0;
> +}
> +
> +/**
> + * handle_event_notified() - remoteproc notification work funciton
> + * @work: pointer to the work structure
> + *
> + * It checks each registered remoteproc notify IDs.
> + */
> +static void handle_event_notified(struct work_struct *work)
> +{
> + struct rproc *rproc;
> + struct zynqmp_r5_pdata *pdata;
> +
> + pdata = container_of(work, struct zynqmp_r5_pdata, mbox_work);
> +
> + (void)mbox_send_message(pdata->rx_chan, NULL);
> + rproc = pdata->rproc;
> + /*
> + * We only use IPI for interrupt. The firmware side may or may
> + * not write the notifyid when it trigger IPI.
> + * And thus, we scan through all the registered notifyids.
> + */
> + idr_for_each(&rproc->notifyids, event_notified_idr_cb, rproc);
> +}
> +
> +/**
> + * zynqmp_r5_mb_rx_cb() - Receive channel mailbox callback
> + * @cl: mailbox client
> + * @mssg: message pointer
> + *
> + * It will schedule the R5 notification work.
> + */
> +static void zynqmp_r5_mb_rx_cb(struct mbox_client *cl, void *mssg)
> +{
> + struct zynqmp_r5_pdata *pdata;
> +
> + pdata = container_of(cl, struct zynqmp_r5_pdata, rx_mc);
> + if (mssg) {
> + struct zynqmp_ipi_message *ipi_msg, *buf_msg;
> + size_t len;
> +
> + ipi_msg = (struct zynqmp_ipi_message *)mssg;
> + buf_msg = (struct zynqmp_ipi_message *)pdata->rx_mc_buf;
> + len = (ipi_msg->len >= IPI_BUF_LEN_MAX) ?
> + IPI_BUF_LEN_MAX : ipi_msg->len;
> + buf_msg->len = len;
> + memcpy(buf_msg->data, ipi_msg->data, len);
> + }
> + schedule_work(&pdata->mbox_work);
> +}
> +
> +/**
> + * zynqmp_r5_mb_tx_done() - Request has been sent to the remote
> + * @cl: mailbox client
> + * @mssg: pointer to the message which has been sent
> + * @r: status of last TX - OK or error
> + *
> + * It will be called by the mailbox framework when the last TX has done.
> + */
> +static void zynqmp_r5_mb_tx_done(struct mbox_client *cl, void *mssg, int r)
> +{
> + struct zynqmp_r5_pdata *pdata;
> + struct sk_buff *skb;
> +
> + if (!mssg)
> + return;
> + pdata = container_of(cl, struct zynqmp_r5_pdata, tx_mc);
> + skb = skb_dequeue(&pdata->tx_mc_skbs);
> + kfree_skb(skb);
> +}
> +
> +/**
> + * zynqmp_r5_setup_mbox() - Setup mailboxes
> + *
> + * @pdata: pointer to the ZynqMP R5 processor platform data
> + * @node: pointer of the device node
> + *
> + * Function to setup mailboxes to talk to RPU.
> + *
> + * Return: 0 for success, negative value for failure.
> + */
> +static int zynqmp_r5_setup_mbox(struct zynqmp_r5_pdata *pdata,
> + struct device_node *node)
> +{
> + struct device *dev = &pdata->dev;
> + struct mbox_client *mclient;
> +
> + /* Setup TX mailbox channel client */
> + mclient = &pdata->tx_mc;
> + mclient->dev = dev;
> + mclient->rx_callback = NULL;
> + mclient->tx_block = false;
> + mclient->knows_txdone = false;
> + mclient->tx_done = zynqmp_r5_mb_tx_done;
> +
> + /* Setup TX mailbox channel client */
> + mclient = &pdata->rx_mc;
> + mclient->dev = dev;
> + mclient->rx_callback = zynqmp_r5_mb_rx_cb;
> + mclient->tx_block = false;
> + mclient->knows_txdone = false;
> +
> + INIT_WORK(&pdata->mbox_work, handle_event_notified);
> +
> + /* Request TX and RX channels */
> + pdata->tx_chan = mbox_request_channel_byname(&pdata->tx_mc, "tx");
> + if (IS_ERR(pdata->tx_chan)) {
> + dev_err(dev, "failed to request mbox tx channel.\n");
> + pdata->tx_chan = NULL;
> + return -EINVAL;
> + }
> + pdata->rx_chan = mbox_request_channel_byname(&pdata->rx_mc, "rx");
> + if (IS_ERR(pdata->rx_chan)) {
> + dev_err(dev, "failed to request mbox rx channel.\n");
> + pdata->rx_chan = NULL;
> + return -EINVAL;
> + }
> + skb_queue_head_init(&pdata->tx_mc_skbs);
> + return 0;
> +}
> +
> +/**
> + * zynqmp_r5_probe() - Probes ZynqMP R5 processor device node
> + * @pdata: pointer to the ZynqMP R5 processor platform data
> + * @pdev: parent RPU domain platform device
> + * @node: pointer of the device node
> + *
> + * Function to retrieve the information of the ZynqMP R5 device node.
> + *
> + * Return: 0 for success, negative value for failure.
> + */
> +static int zynqmp_r5_probe(struct zynqmp_r5_pdata *pdata,
> + struct platform_device *pdev,
> + struct device_node *node)
> +{
> + struct device *dev = &pdata->dev;
> + struct rproc *rproc;
> + struct device_node *nc;
> + int ret;
> +
> + /* Create device for ZynqMP R5 device */
> + dev->parent = &pdev->dev;
> + dev->release = zynqmp_r5_release;
> + dev->of_node = node;
> + dev_set_name(dev, "%s", of_node_full_name(node));
> + dev_set_drvdata(dev, pdata);
> + ret = device_register(dev);
> + if (ret) {
> + dev_err(dev, "failed to register device.\n");
> + return ret;
> + }
> + get_device(&pdev->dev);
> +
> + /* Allocate remoteproc instance */
> + rproc = rproc_alloc(dev, dev_name(dev), &zynqmp_r5_rproc_ops, NULL, 0);
> + if (!rproc) {
> + dev_err(dev, "rproc allocation failed.\n");
> + ret = -ENOMEM;
> + goto error;
> + }
> + rproc->auto_boot = autoboot;
> + pdata->rproc = rproc;
> + rproc->priv = pdata;
> +
> + /*
> + * The device has not been spawned from a device tree, so
> + * arch_setup_dma_ops has not been called, thus leaving
> + * the device with dummy DMA ops.
> + * Fix this by inheriting the parent's DMA ops and mask.
> + */
> + rproc->dev.dma_mask = pdev->dev.dma_mask;
> + set_dma_ops(&rproc->dev, get_dma_ops(&pdev->dev));
> +
> + /* Probe R5 memory devices */
> + INIT_LIST_HEAD(&pdata->mems);
> + for_each_available_child_of_node(node, nc) {
> + ret = zynqmp_r5_mem_probe(pdata, nc);
> + if (ret) {
> + dev_err(dev, "failed to probe memory %s.\n",
> + of_node_full_name(nc));
> + goto error;
> + }
> + }
> +
> + /* Set up DMA mask */
> + ret = dma_set_coherent_mask(dev, DMA_BIT_MASK(32));
> + if (ret) {
> + dev_warn(dev, "dma_set_coherent_mask failed: %d\n", ret);
> + /* If DMA is not configured yet, try to configure it. */
> + ret = of_dma_configure(dev, node, true);
> + if (ret) {
> + dev_err(dev, "failed to configure DMA.\n");
> + goto error;
> + }
> + }
> +
> + /* Get R5 power domain node */
> + ret = of_property_read_u32(node, "pnode-id", &pdata->pnode_id);
> + if (ret) {
> + dev_err(dev, "failed to get power node id.\n");
> + goto error;
> + }
> +
> + /* TODO Check if R5 is running */
> +
> + /* Set up R5 if not already setup */
> + ret = pdata->is_r5_mode_set ? 0 : r5_set_mode(pdata);
> + if (ret) {
> + dev_err(dev, "failed to set R5 operation mode.\n");
> + return ret;
> + }
> +
> + if (!of_get_property(dev->of_node, "mboxes", NULL)) {
> + dev_dbg(dev, "no mailboxes.\n");
> + goto error;
> + } else {
> + ret = zynqmp_r5_setup_mbox(pdata, node);
> + if (ret < 0)
> + goto error;
> + }
> +
> + /* Add R5 remoteproc */
> + ret = rproc_add(rproc);
> + if (ret) {
> + dev_err(dev, "rproc registration failed\n");
> + goto error;
> + }
> + return 0;
> +error:
> + if (pdata->rproc)
> + rproc_free(pdata->rproc);
> + pdata->rproc = NULL;
> + device_unregister(dev);
> + put_device(&pdev->dev);
> + return ret;
> +}
> +
> +static int zynqmp_r5_remoteproc_probe(struct platform_device *pdev)
> +{
> + int ret, i = 0;
> + u32 *lockstep_mode;

Can this just be a regular variable?

> + struct device *dev = &pdev->dev;
> + struct device_node *nc;
> + struct zynqmp_r5_pdata *pdata;
> +
> + pdata = devm_kzalloc(dev, sizeof(*pdata), GFP_KERNEL);
> + lockstep_mode = devm_kzalloc(dev, sizeof(u32 *), GFP_KERNEL);
> + if (!pdata || !lockstep_mode)
> + return -ENOMEM;
> +
> + platform_set_drvdata(pdev, pdata);
> +
> + of_property_read_u32(dev->of_node, "lockstep-mode", lockstep_mode);

To make Rob's life a little easier this should probably be "xilinx,cluster-mode"
so that it follows what TI has done.

I am out of time for today, I will continue tomorrow.

Thanks,
Mathieu

> +
> + if (!(*lockstep_mode)) {
> + rpu_mode = PM_RPU_MODE_SPLIT;
> + } else if (*lockstep_mode == 1) {
> + rpu_mode = PM_RPU_MODE_LOCKSTEP;
> + } else {
> + dev_err(dev,
> + "Invalid lockstep-mode mode provided - %x %d\n",
> + *lockstep_mode, rpu_mode);
> + return -EINVAL;
> + }
> + dev_dbg(dev, "RPU configuration: %s\r\n",
> + (*lockstep_mode) ? "lockstep" : "split");
> +
> + for_each_available_child_of_node(dev->of_node, nc) {
> + ret = zynqmp_r5_probe(&rpus[i], pdev, nc);
> + if (ret) {
> + dev_err(dev, "failed to probe rpu %s.\n",
> + of_node_full_name(nc));
> + return ret;
> + }
> + i++;
> + }
> +
> + return 0;
> +}
> +
> +static int zynqmp_r5_remoteproc_remove(struct platform_device *pdev)
> +{
> + int i;
> +
> + for (i = 0; i < MAX_RPROCS; i++) {
> + struct zynqmp_r5_pdata *pdata = &rpus[i];
> + struct rproc *rproc;
> +
> + rproc = pdata->rproc;
> + if (rproc) {
> + rproc_del(rproc);
> + rproc_free(rproc);
> + pdata->rproc = NULL;
> + }
> + if (pdata->tx_chan) {
> + mbox_free_channel(pdata->tx_chan);
> + pdata->tx_chan = NULL;
> + }
> + if (pdata->rx_chan) {
> + mbox_free_channel(pdata->rx_chan);
> + pdata->rx_chan = NULL;
> + }
> +
> + device_unregister(&pdata->dev);
> + }
> +
> + return 0;
> +}
> +
> +/* Match table for OF platform binding */
> +static const struct of_device_id zynqmp_r5_remoteproc_match[] = {
> + { .compatible = "xlnx,zynqmp-r5-remoteproc-1.0", },
> + { /* end of list */ },
> +};
> +MODULE_DEVICE_TABLE(of, zynqmp_r5_remoteproc_match);
> +
> +static struct platform_driver zynqmp_r5_remoteproc_driver = {
> + .probe = zynqmp_r5_remoteproc_probe,
> + .remove = zynqmp_r5_remoteproc_remove,
> + .driver = {
> + .name = "zynqmp_r5_remoteproc",
> + .of_match_table = zynqmp_r5_remoteproc_match,
> + },
> +};
> +module_platform_driver(zynqmp_r5_remoteproc_driver);
> +
> +module_param_named(autoboot, autoboot, bool, 0444);
> +MODULE_PARM_DESC(autoboot,
> + "enable | disable autoboot. (default: false)");
> +
> +MODULE_AUTHOR("Ben Levinsky <[email protected]>");
> +MODULE_LICENSE("GPL v2");
> --
> 2.17.1
>

2020-07-28 05:10:17

by Michal Simek

[permalink] [raw]
Subject: Re: [PATCH v6 5/5] remoteproc: Add initial zynqmp R5 remoteproc driver



On 28. 07. 20 0:59, Mathieu Poirier wrote:
> On Wed, Jul 15, 2020 at 08:33:17AM -0700, Ben Levinsky wrote:
>> R5 is included in Xilinx Zynq UltraScale MPSoC so by adding this
>> remotproc driver, we can boot the R5 sub-system in different
>> configurations.
>>
>> Acked-by: Stefano Stabellini <[email protected]>
>> Acked-by: Ben Levinsky <[email protected]>
>> Reviewed-by: Radhey Shyam Pandey <[email protected]>
>> Signed-off-by: Ben Levinsky <[email protected]>
>> Signed-off-by: Wendy Liang <[email protected]>
>> Signed-off-by: Michal Simek <[email protected]>
>> Signed-off-by: Ed Mooring <[email protected]>
>> Signed-off-by: Jason Wu <[email protected]>
>> Tested-by: Ben Levinsky <[email protected]>
>> ---
>> v2:
>> - remove domain struct as per review from Mathieu
>> v3:
>> - add xilinx-related platform mgmt fn's instead of wrapping around
>> function pointer in xilinx eemi ops struct
>> v4:
>> - add default values for enums
>> - fix formatting as per checkpatch.pl --strict. Note that 1 warning and 1 check
>> are still raised as each is due to fixing the warning results in that
>> particular line going over 80 characters.
>> v5:
>> - parse_fw change from use of rproc_of_resm_mem_entry_init to rproc_mem_entry_init and use of alloc/release
>> - var's of type zynqmp_r5_pdata all have same local variable name
>> - use dev_dbg instead of dev_info
>> v6:
>> - adding memory carveouts is handled much more similarly. All mem carveouts are
>> now described in reserved memory as needed. That is, TCM nodes are not
>> coupled to remoteproc anymore. This is reflected in the remoteproc R5 driver
>> and the device tree binding.
>> - remove mailbox from device tree binding as it is not necessary for elf
>> loading
>> - use lockstep-mode property for configuring RPU
>> ---
>> drivers/remoteproc/Kconfig | 10 +
>> drivers/remoteproc/Makefile | 1 +
>> drivers/remoteproc/zynqmp_r5_remoteproc.c | 911 ++++++++++++++++++++++
>> 3 files changed, 922 insertions(+)
>> create mode 100644 drivers/remoteproc/zynqmp_r5_remoteproc.c
>>
>> diff --git a/drivers/remoteproc/Kconfig b/drivers/remoteproc/Kconfig
>> index c4d1731295eb..342a7e668636 100644
>> --- a/drivers/remoteproc/Kconfig
>> +++ b/drivers/remoteproc/Kconfig
>> @@ -249,6 +249,16 @@ config STM32_RPROC
>>
>> This can be either built-in or a loadable module.
>>
>> +config ZYNQMP_R5_REMOTEPROC
>> + tristate "ZynqMP_R5 remoteproc support"
>> + depends on ARM64 && PM && ARCH_ZYNQMP
>> + select RPMSG_VIRTIO
>> + select MAILBOX
>> + select ZYNQMP_IPI_MBOX
>> + help
>> + Say y here to support ZynqMP R5 remote processors via the remote
>> + processor framework.
>> +
>> endif # REMOTEPROC
>>
>> endmenu
>> diff --git a/drivers/remoteproc/Makefile b/drivers/remoteproc/Makefile
>> index e8b886e511f0..04d1c95d06d7 100644
>> --- a/drivers/remoteproc/Makefile
>> +++ b/drivers/remoteproc/Makefile
>> @@ -28,5 +28,6 @@ obj-$(CONFIG_QCOM_WCNSS_PIL) += qcom_wcnss_pil.o
>> qcom_wcnss_pil-y += qcom_wcnss.o
>> qcom_wcnss_pil-y += qcom_wcnss_iris.o
>> obj-$(CONFIG_ST_REMOTEPROC) += st_remoteproc.o
>> +obj-$(CONFIG_ZYNQMP_R5_REMOTEPROC) += zynqmp_r5_remoteproc.o
>> obj-$(CONFIG_ST_SLIM_REMOTEPROC) += st_slim_rproc.o
>> obj-$(CONFIG_STM32_RPROC) += stm32_rproc.o
>> diff --git a/drivers/remoteproc/zynqmp_r5_remoteproc.c b/drivers/remoteproc/zynqmp_r5_remoteproc.c
>> new file mode 100644
>> index 000000000000..b600759e257e
>> --- /dev/null
>> +++ b/drivers/remoteproc/zynqmp_r5_remoteproc.c
>> @@ -0,0 +1,911 @@
>> +// SPDX-License-Identifier: GPL-2.0
>> +/*
>> + * Zynq R5 Remote Processor driver
>> + *
>> + * Copyright (C) 2019, 2020 Xilinx Inc. Ben Levinsky <[email protected]>
>> + * Copyright (C) 2015 - 2018 Xilinx Inc.
>> + * Copyright (C) 2015 Jason Wu <[email protected]>
>> + *
>> + * Based on origin OMAP and Zynq Remote Processor driver
>> + *
>> + * Copyright (C) 2012 Michal Simek <[email protected]>
>> + * Copyright (C) 2012 PetaLogix
>> + * Copyright (C) 2011 Texas Instruments, Inc.
>> + * Copyright (C) 2011 Google, Inc.
>> + */
>> +
>> +#include <linux/atomic.h>
>
> Unused
>
>> +#include <linux/cpu.h>
>> +#include <linux/dma-mapping.h>
>> +#include <linux/delay.h>
>
> Unused
>
>> +#include <linux/err.h>
>> +#include <linux/firmware/xlnx-zynqmp.h>
>> +#include <linux/genalloc.h>
>
> Unused
>
>> +#include <linux/idr.h>
>> +#include <linux/interrupt.h>
>> +#include <linux/kernel.h>
>> +#include <linux/list.h>
>> +#include <linux/mailbox_client.h>
>> +#include <linux/mailbox/zynqmp-ipi-message.h>
>> +#include <linux/module.h>
>> +#include <linux/of_address.h>
>> +#include <linux/of_irq.h>
>> +#include <linux/of_platform.h>
>> +#include <linux/of_reserved_mem.h>
>> +#include <linux/pfn.h>
>
> Unused
>
>> +#include <linux/platform_device.h>
>> +#include <linux/remoteproc.h>
>> +#include <linux/skbuff.h>
>> +#include <linux/slab.h>
>> +#include <linux/sysfs.h>
>> +
>> +#include "remoteproc_internal.h"
>> +
>> +#define MAX_RPROCS 2 /* Support up to 2 RPU */
>> +#define MAX_MEM_PNODES 4 /* Max power nodes for one RPU memory instance */
>> +
>> +#define DEFAULT_FIRMWARE_NAME "rproc-rpu-fw"
>> +
>> +/* PM proc states */
>> +#define PM_PROC_STATE_ACTIVE 1U
>
> Unused
>
>> +
>> +/* IPI buffer MAX length */
>> +#define IPI_BUF_LEN_MAX 32U
>> +/* RX mailbox client buffer max length */
>> +#define RX_MBOX_CLIENT_BUF_MAX (IPI_BUF_LEN_MAX + \
>> + sizeof(struct zynqmp_ipi_message))
>> +
>> +#define ZYNQMP_R5_NUM_TCM_BANKS 4
>> +
>> +/* lookup table mapping power-node-ID of TCM bank to absolute base address */
>> +static unsigned long tcm_addr_to_pnode[ZYNQMP_R5_NUM_TCM_BANKS][2] = {
>> + {0xFFE00000, 0xF },
>> + {0xFFE20000, 0x10},
>> + {0xFFE90000, 0x10},
>> + {0xFFEB0000, 0x11},
>> +};
>> +
>> +static bool autoboot __read_mostly;
>> +
>> +/**
>> + * struct zynqmp_r5_mem - zynqmp rpu memory data
>> + * @pnode_id: TCM power domain ids
>> + * @res: memory resource
>> + * @node: list node
>> + */
>> +struct zynqmp_r5_mem {
>> + u32 pnode_id[MAX_MEM_PNODES];
>> + struct resource res;
>> + struct list_head node;
>> +};
>> +
>> +/**
>> + * struct zynqmp_r5_pdata - zynqmp rpu remote processor private data
>> + * @dev: device of RPU instance
>> + * @rproc: rproc handle
>> + * @pnode_id: RPU CPU power domain id
>> + * @mems: memory resources
>> + * @is_r5_mode_set: indicate if r5 operation mode is set
>> + * @tx_mc: tx mailbox client
>> + * @rx_mc: rx mailbox client
>> + * @tx_chan: tx mailbox channel
>> + * @rx_chan: rx mailbox channel
>> + * @mbox_work: mbox_work for the RPU remoteproc
>> + * @tx_mc_skbs: socket buffers for tx mailbox client
>> + * @rx_mc_buf: rx mailbox client buffer to save the rx message
>> + */
>> +struct zynqmp_r5_pdata {
>> + struct device dev;
>> + struct rproc *rproc;
>> + u32 pnode_id;
>> + struct list_head mems;
>> + bool is_r5_mode_set;
>> + struct mbox_client tx_mc;
>> + struct mbox_client rx_mc;
>> + struct mbox_chan *tx_chan;
>> + struct mbox_chan *rx_chan;
>> + struct work_struct mbox_work;
>> + struct sk_buff_head tx_mc_skbs;
>> + unsigned char rx_mc_buf[RX_MBOX_CLIENT_BUF_MAX];
>> +};
>> +
>> +/**
>> + * table of RPUs
>> + */
>> +struct zynqmp_r5_pdata rpus[MAX_RPROCS];
>> +/**
>> + * RPU core configuration
>> + */
>> +enum rpu_oper_mode rpu_mode;
>> +
>> +/*
>> + * r5_set_mode - set RPU operation mode
>> + * @pdata: Remote processor private data
>> + *
>> + * set RPU oepration mode
>> + *
>> + * Return: 0 for success, negative value for failure
>> + */
>> +static int r5_set_mode(struct zynqmp_r5_pdata *pdata)
>> +{
>> + u32 val[PAYLOAD_ARG_CNT] = {0}, expect, tcm_mode;
>> + struct device *dev = &pdata->dev;
>> + int ret;
>> +
>> + expect = (u32)rpu_mode;
>> + ret = zynqmp_pm_get_rpu_mode(pdata->pnode_id, 0, 0, val);
>> + if (ret < 0) {
>> + dev_err(dev, "failed to get RPU oper mode.\n");
>> + return ret;
>> + }
>> + if (val[0] == expect) {
>> + dev_dbg(dev, "RPU mode matches: %x\n", val[0]);
>> + } else {
>> + ret = zynqmp_pm_set_rpu_mode(pdata->pnode_id,
>> + expect, 0, val);
>> + if (ret < 0) {
>> + dev_err(dev,
>> + "failed to set RPU oper mode.\n");
>> + return ret;
>> + }
>> + }
>> +
>> + tcm_mode = (expect == (u32)PM_RPU_MODE_LOCKSTEP) ?
>> + PM_RPU_TCM_COMB : PM_RPU_TCM_SPLIT;
>> + ret = zynqmp_pm_set_tcm_config(pdata->pnode_id, tcm_mode, 0, val);
>> + if (ret < 0) {
>> + dev_err(dev, "failed to config TCM to %x.\n",
>> + expect);
>> + return ret;
>> + }
>> + pdata->is_r5_mode_set = true;
>> + return 0;
>> +}
>> +
>> +/*
>> + * ZynqMP R5 remoteproc memory release function
>> + */
>> +static int zynqmp_r5_mem_release(struct rproc *rproc,
>> + struct rproc_mem_entry *mem)
>> +{
>> + struct zynqmp_r5_mem *priv;
>> + int i, ret;
>> + struct device *dev = &rproc->dev;
>> +
>> + priv = mem->priv;
>> + if (!priv)
>> + return 0;
>> + for (i = 0; i < MAX_MEM_PNODES; i++) {
>> + if (priv->pnode_id[i]) {
>> + dev_dbg(dev, "%s, pnode %d\n",
>> + __func__, priv->pnode_id[i]);
>> + ret = zynqmp_pm_release_node(priv->pnode_id[i]);
>> + if (ret < 0) {
>> + dev_err(dev,
>> + "failed to release power node: %u\n",
>> + priv->pnode_id[i]);
>> + return ret;
>> + }
>> + } else {
>> + break;
>> + }
>> + }
>> + return 0;
>> +}
>> +
>> +/*
>> + * ZynqMP R5 remoteproc operations
>> + */
>> +static int zynqmp_r5_rproc_start(struct rproc *rproc)
>> +{
>> + struct device *dev = rproc->dev.parent;
>> + struct zynqmp_r5_pdata *pdata = rproc->priv;
>> + enum rpu_boot_mem bootmem;
>> + int ret;
>> +
>> + if ((rproc->bootaddr & 0xF0000000) == 0xF0000000)
>> + bootmem = PM_RPU_BOOTMEM_HIVEC;
>> + else
>> + bootmem = PM_RPU_BOOTMEM_LOVEC;
>> + dev_dbg(dev, "RPU boot from %s.",
>> + bootmem == PM_RPU_BOOTMEM_HIVEC ? "OCM" : "TCM");
>> + ret = zynqmp_pm_request_wakeup(pdata->pnode_id, 1,
>> + bootmem, ZYNQMP_PM_REQUEST_ACK_NO);
>> + if (ret < 0) {
>> + dev_err(dev, "failed to boot R5.\n");
>> + return ret;
>> + }
>> + return 0;
>> +}
>> +
>> +static int zynqmp_r5_rproc_stop(struct rproc *rproc)
>> +{
>> + struct zynqmp_r5_pdata *pdata = rproc->priv;
>> + int ret;
>> +
>> + ret = zynqmp_pm_force_powerdown(pdata->pnode_id,
>> + ZYNQMP_PM_REQUEST_ACK_BLOCKING);
>> + if (ret < 0) {
>> + dev_err(&pdata->dev, "failed to shutdown R5.\n");
>> + return ret;
>> + }
>> + return 0;
>> +}
>> +
>> +static int zynqmp_r5_rproc_mem_alloc(struct rproc *rproc,
>> + struct rproc_mem_entry *mem)
>> +{
>> + struct device *dev = rproc->dev.parent;
>> + void *va;
>> +
>> + dev_dbg(rproc->dev.parent, "map memory: %pa\n", &mem->dma);
>> + va = ioremap_wc(mem->dma, mem->len);
>> + if (IS_ERR_OR_NULL(va)) {
>> + dev_err(dev, "Unable to map memory region: %pa+%lx\n",
>> + &mem->dma, mem->len);
>> + return -ENOMEM;
>> + }
>> +
>> + /* Update memory entry va */
>> + mem->va = va;
>> +
>> + return 0;
>> +}
>> +
>> +static int zynqmp_r5_rproc_mem_release(struct rproc *rproc,
>> + struct rproc_mem_entry *mem)
>> +{
>> + dev_dbg(rproc->dev.parent, "unmap memory: %pa\n", &mem->dma);
>> + iounmap(mem->va);
>> +
>> + return 0;
>> +}
>> +
>> +/*
>> + * TCM needs mapping to R5 relative address and xilinx platform mgmt call
>> + */
>> +struct rproc_mem_entry *handle_tcm_parsing(struct device *dev,
>> + struct reserved_mem *rmem,
>> + struct device_node *node,
>> + int lookup_idx)
>> +{
>> + void *va;
>> + dma_addr_t dma;
>> + resource_size_t size;
>> + int ret;
>> + u32 pnode_id;
>> + struct resource rsc;
>> + struct rproc_mem_entry *mem;
>> +
>> + pnode_id = tcm_addr_to_pnode[lookup_idx][1];
>> + ret = zynqmp_pm_request_node(pnode_id,
>> + ZYNQMP_PM_CAPABILITY_ACCESS, 0,
>> + ZYNQMP_PM_REQUEST_ACK_BLOCKING);
>> + if (ret < 0) {
>> + dev_err(dev, "failed to request power node: %u\n", pnode_id);
>> + return -EINVAL;
>> + }
>> +
>> + ret = of_address_to_resource(node, 0, &rsc);
>> + if (ret < 0) {
>> + dev_err(dev, "failed to get resource of memory %s",
>> + of_node_full_name(node));
>> + return -EINVAL;
>> + }
>> + size = resource_size(&rsc);
>> + va = devm_ioremap_wc(dev, rsc.start, size);
>> + if (!va)
>> + return -ENOMEM;
>> +
>> + /* zero out tcm base address */
>> + if (rsc.start & 0xffe00000) {
>> + /* R5 can't see anything past 0xfffff so wipe it */
>> + rsc.start &= 0x000fffff;
>> + /*
>> + * handle tcm banks 1 a and b (0xffe9000 and
>> + * 0xffeb0000)
>> + */
>> + if (rsc.start & 0x80000)
>> + rsc.start -= 0x90000;
>> + }
>> +
>> + dma = (dma_addr_t)rsc.start;
>> + mem = rproc_mem_entry_init(dev, va, dma, (int)size, rsc.start,
>> + NULL, zynqmp_r5_mem_release,
>> + rsc.name);
>> + if (!mem)
>> + return -ENOMEM;
>> +
>> + return mem;
>> +}
>> +
>> +static int parse_mem_regions(struct rproc *rproc)
>> +{
>> + int num_mems, i;
>> + struct zynqmp_r5_pdata *pdata = rproc->priv;
>> + struct device *dev = &pdata->dev;
>> + struct device_node *np = dev->of_node;
>> + struct rproc_mem_entry *mem;
>> +
>> + num_mems = of_count_phandle_with_args(np, "memory-region", NULL);
>> + if (num_mems <= 0)
>> + return 0;
>> + for (i = 0; i < num_mems; i++) {
>> + struct device_node *node;
>> + struct reserved_mem *rmem;
>> +
>> + node = of_parse_phandle(np, "memory-region", i);
>> + rmem = of_reserved_mem_lookup(node);
>> + if (!rmem) {
>> + dev_err(dev, "unable to acquire memory-region\n");
>> + return -EINVAL;
>> + }
>> +
>> + if (strstr(node->name, "vdev0buffer")) {
>> + /* Register DMA region */
>> + mem = rproc_mem_entry_init(dev, NULL,
>> + (dma_addr_t)rmem->base,
>> + rmem->size, rmem->base,
>> + NULL, NULL,
>> + "vdev0buffer");
>> + if (!mem) {
>> + dev_err(dev, "unable to initialize memory-region %s\n",
>> + node->name);
>> + return -ENOMEM;
>> + }
>> + dev_dbg(dev, "parsed %s at %llx\r\n", mem->name,
>> + mem->dma);
>> + } else if (strstr(node->name, "vdev0vring")) {
>> + int vring_id;
>> + char name[16];
>> +
>> + /*
>> + * can be 1 of multiple vring IDs per IPC channel
>> + * e.g. 'vdev0vring0' and 'vdev0vring1'
>> + */
>> + vring_id = node->name[14] - '0';
>> + snprintf(name, sizeof(name), "vdev0vring%d", vring_id);
>> + /* Register vring */
>> + mem = rproc_mem_entry_init(dev, NULL,
>> + (dma_addr_t)rmem->base,
>> + rmem->size, rmem->base,
>> + zynqmp_r5_rproc_mem_alloc,
>> + zynqmp_r5_rproc_mem_release,
>> + name);
>> + dev_dbg(dev, "parsed %s at %llx\r\n", mem->name,
>> + mem->dma);
>> + } else {
>> + int idx;
>> +
>> + /*
>> + * if TCM update address space for R5 and
>> + * make xilinx platform mgmt call
>> + */
>> + for (idx = 0; idx < ZYNQMP_R5_NUM_TCM_BANKS; idx++) {
>> + if (tcm_addr_to_pnode[idx][0] == rmem->base)
>> + break;
>> + }
>> +
>> + if (idx != ZYNQMP_R5_NUM_TCM_BANKS) {
>> + mem = handle_tcm_parsing(dev, rmem, node, idx);
>> + } else {
>> + mem = rproc_mem_entry_init(dev, NULL,
>> + (dma_addr_t)rmem->base,
>> + rmem->size, rmem->base,
>> + zynqmp_r5_rproc_mem_alloc,
>> + zynqmp_r5_rproc_mem_release,
>> + node->name);
>> + }
>> +
>> + if (!mem) {
>> + dev_err(dev,
>> + "unable to init memory-region %s\n",
>> + node->name);
>> + return -ENOMEM;
>> + }
>> + }
>> + rproc_add_carveout(rproc, mem);
>> + }
>> +
>> + return 0;
>> +}
>> +
>> +static int zynqmp_r5_parse_fw(struct rproc *rproc, const struct firmware *fw)
>> +{
>> + int ret;
>> + struct zynqmp_r5_pdata *pdata = rproc->priv;
>> + struct device *dev = &pdata->dev;
>> +
>> + ret = parse_mem_regions(rproc);
>> + if (ret) {
>> + dev_err(dev, "parse_mem_regions failed %x\n", ret);
>> + return ret;
>> + }
>> +
>> + ret = rproc_elf_load_rsc_table(rproc, fw);
>> + if (ret == -EINVAL) {
>> + dev_dbg(dev, "no resource table found.\n");
>> + ret = 0;
>> + }
>> + return ret;
>> +}
>> +
>> +/* kick a firmware */
>> +static void zynqmp_r5_rproc_kick(struct rproc *rproc, int vqid)
>> +{
>> + struct device *dev = rproc->dev.parent;
>> + struct zynqmp_r5_pdata *pdata = rproc->priv;
>> +
>> + dev_dbg(dev, "KICK Firmware to start send messages vqid %d\n", vqid);
>> +
>> + if (vqid < 0) {
>> + /* If vqid is negative, does not pass the vqid to
>> + * mailbox. As vqid is supposed to be 0 or possive.
>> + * It also gives a way to just kick instead but
>> + * not use the IPI buffer. It is better to provide
>> + * a proper way to pass the short message, which will
>> + * need to sync to upstream first, for now,
>> + * use negative vqid to assume no message will be
>> + * passed with IPI buffer, but just raise interrupt.
>> + * This will be faster as it doesn't need to copy the
>> + * message to the IPI buffer.
>> + *
>> + * It will ignore the return, as failure is due to
>> + * there already kicks in the mailbox queue.
>> + */
>> + (void)mbox_send_message(pdata->tx_chan, NULL);
>> + } else {
>> + struct sk_buff *skb;
>> + unsigned int skb_len;
>> + struct zynqmp_ipi_message *mb_msg;
>> + int ret;
>> +
>> + skb_len = (unsigned int)(sizeof(vqid) + sizeof(mb_msg));
>> + skb = alloc_skb(skb_len, GFP_ATOMIC);
>> + if (!skb) {
>> + dev_err(dev,
>> + "Failed to allocate skb to kick remote.\n");
>> + return;
>> + }
>> + mb_msg = (struct zynqmp_ipi_message *)skb_put(skb, skb_len);
>> + mb_msg->len = sizeof(vqid);
>> + memcpy(mb_msg->data, &vqid, sizeof(vqid));
>> + skb_queue_tail(&pdata->tx_mc_skbs, skb);
>> + ret = mbox_send_message(pdata->tx_chan, mb_msg);
>> + if (ret < 0) {
>> + dev_warn(dev, "Failed to kick remote.\n");
>> + skb_dequeue_tail(&pdata->tx_mc_skbs);
>> + kfree_skb(skb);
>> + }
>> + }
>> +}
>> +
>> +static struct rproc_ops zynqmp_r5_rproc_ops = {
>> + .start = zynqmp_r5_rproc_start,
>> + .stop = zynqmp_r5_rproc_stop,
>> + .load = rproc_elf_load_segments,
>> + .parse_fw = zynqmp_r5_parse_fw,
>> + .find_loaded_rsc_table = rproc_elf_find_loaded_rsc_table,
>> + .sanity_check = rproc_elf_sanity_check,
>> + .get_boot_addr = rproc_elf_get_boot_addr,
>> + .kick = zynqmp_r5_rproc_kick,
>> +};
>> +
>> +/* zynqmp_r5_mem_probe() - probes RPU TCM memory device
>> + * @pdata: pointer to the RPU remoteproc private data
>> + * @node: pointer to the memory node
>> + *
>> + * Function to retrieve resources for RPU TCM memory device.
>> + */
>> +static int zynqmp_r5_mem_probe(struct zynqmp_r5_pdata *pdata,
>> + struct device_node *node)
>> +{
>> + struct device *dev;
>> + struct zynqmp_r5_mem *mem;
>> + int ret;
>> + struct property *prop;
>> + const __be32 *cur;
>> + u32 val;
>> + int i;
>> +
>> + dev = &pdata->dev;
>> + mem = devm_kzalloc(dev, sizeof(*mem), GFP_KERNEL);
>> + if (!mem)
>> + return -ENOMEM;
>> + ret = of_address_to_resource(node, 0, &mem->res);
>> + if (ret < 0) {
>> + dev_err(dev, "failed to get resource of memory %s",
>> + of_node_full_name(node));
>> + return -EINVAL;
>> + }
>> +
>> + /* Get the power domain id */
>> + i = 0;
>> + if (of_find_property(node, "pnode-id", NULL)) {
>> + of_property_for_each_u32(node, "pnode-id", prop, cur, val)
>> + mem->pnode_id[i++] = val;
>> + }
>> + list_add_tail(&mem->node, &pdata->mems);
>> + return 0;
>> +}
>> +
>> +/**
>> + * zynqmp_r5_release() - ZynqMP R5 device release function
>> + * @dev: pointer to the device struct of ZynqMP R5
>> + *
>> + * Function to release ZynqMP R5 device.
>> + */
>> +static void zynqmp_r5_release(struct device *dev)
>> +{
>> + struct zynqmp_r5_pdata *pdata;
>> + struct rproc *rproc;
>> + struct sk_buff *skb;
>> +
>> + pdata = dev_get_drvdata(dev);
>> + rproc = pdata->rproc;
>> + if (rproc) {
>> + rproc_del(rproc);
>> + rproc_free(rproc);
>> + }
>> + if (pdata->tx_chan)
>> + mbox_free_channel(pdata->tx_chan);
>> + if (pdata->rx_chan)
>> + mbox_free_channel(pdata->rx_chan);
>> + /* Discard all SKBs */
>> + while (!skb_queue_empty(&pdata->tx_mc_skbs)) {
>> + skb = skb_dequeue(&pdata->tx_mc_skbs);
>> + kfree_skb(skb);
>> + }
>> +
>> + put_device(dev->parent);
>> +}
>> +
>> +/**
>> + * event_notified_idr_cb() - event notified idr callback
>> + * @id: idr id
>> + * @ptr: pointer to idr private data
>> + * @data: data passed to idr_for_each callback
>> + *
>> + * Pass notification to remoteproc virtio
>> + *
>> + * Return: 0. having return is to satisfy the idr_for_each() function
>> + * pointer input argument requirement.
>> + **/
>> +static int event_notified_idr_cb(int id, void *ptr, void *data)
>> +{
>> + struct rproc *rproc = data;
>> +
>> + (void)rproc_vq_interrupt(rproc, id);
>> + return 0;
>> +}
>> +
>> +/**
>> + * handle_event_notified() - remoteproc notification work funciton
>> + * @work: pointer to the work structure
>> + *
>> + * It checks each registered remoteproc notify IDs.
>> + */
>> +static void handle_event_notified(struct work_struct *work)
>> +{
>> + struct rproc *rproc;
>> + struct zynqmp_r5_pdata *pdata;
>> +
>> + pdata = container_of(work, struct zynqmp_r5_pdata, mbox_work);
>> +
>> + (void)mbox_send_message(pdata->rx_chan, NULL);
>> + rproc = pdata->rproc;
>> + /*
>> + * We only use IPI for interrupt. The firmware side may or may
>> + * not write the notifyid when it trigger IPI.
>> + * And thus, we scan through all the registered notifyids.
>> + */
>> + idr_for_each(&rproc->notifyids, event_notified_idr_cb, rproc);
>> +}
>> +
>> +/**
>> + * zynqmp_r5_mb_rx_cb() - Receive channel mailbox callback
>> + * @cl: mailbox client
>> + * @mssg: message pointer
>> + *
>> + * It will schedule the R5 notification work.
>> + */
>> +static void zynqmp_r5_mb_rx_cb(struct mbox_client *cl, void *mssg)
>> +{
>> + struct zynqmp_r5_pdata *pdata;
>> +
>> + pdata = container_of(cl, struct zynqmp_r5_pdata, rx_mc);
>> + if (mssg) {
>> + struct zynqmp_ipi_message *ipi_msg, *buf_msg;
>> + size_t len;
>> +
>> + ipi_msg = (struct zynqmp_ipi_message *)mssg;
>> + buf_msg = (struct zynqmp_ipi_message *)pdata->rx_mc_buf;
>> + len = (ipi_msg->len >= IPI_BUF_LEN_MAX) ?
>> + IPI_BUF_LEN_MAX : ipi_msg->len;
>> + buf_msg->len = len;
>> + memcpy(buf_msg->data, ipi_msg->data, len);
>> + }
>> + schedule_work(&pdata->mbox_work);
>> +}
>> +
>> +/**
>> + * zynqmp_r5_mb_tx_done() - Request has been sent to the remote
>> + * @cl: mailbox client
>> + * @mssg: pointer to the message which has been sent
>> + * @r: status of last TX - OK or error
>> + *
>> + * It will be called by the mailbox framework when the last TX has done.
>> + */
>> +static void zynqmp_r5_mb_tx_done(struct mbox_client *cl, void *mssg, int r)
>> +{
>> + struct zynqmp_r5_pdata *pdata;
>> + struct sk_buff *skb;
>> +
>> + if (!mssg)
>> + return;
>> + pdata = container_of(cl, struct zynqmp_r5_pdata, tx_mc);
>> + skb = skb_dequeue(&pdata->tx_mc_skbs);
>> + kfree_skb(skb);
>> +}
>> +
>> +/**
>> + * zynqmp_r5_setup_mbox() - Setup mailboxes
>> + *
>> + * @pdata: pointer to the ZynqMP R5 processor platform data
>> + * @node: pointer of the device node
>> + *
>> + * Function to setup mailboxes to talk to RPU.
>> + *
>> + * Return: 0 for success, negative value for failure.
>> + */
>> +static int zynqmp_r5_setup_mbox(struct zynqmp_r5_pdata *pdata,
>> + struct device_node *node)
>> +{
>> + struct device *dev = &pdata->dev;
>> + struct mbox_client *mclient;
>> +
>> + /* Setup TX mailbox channel client */
>> + mclient = &pdata->tx_mc;
>> + mclient->dev = dev;
>> + mclient->rx_callback = NULL;
>> + mclient->tx_block = false;
>> + mclient->knows_txdone = false;
>> + mclient->tx_done = zynqmp_r5_mb_tx_done;
>> +
>> + /* Setup TX mailbox channel client */
>> + mclient = &pdata->rx_mc;
>> + mclient->dev = dev;
>> + mclient->rx_callback = zynqmp_r5_mb_rx_cb;
>> + mclient->tx_block = false;
>> + mclient->knows_txdone = false;
>> +
>> + INIT_WORK(&pdata->mbox_work, handle_event_notified);
>> +
>> + /* Request TX and RX channels */
>> + pdata->tx_chan = mbox_request_channel_byname(&pdata->tx_mc, "tx");
>> + if (IS_ERR(pdata->tx_chan)) {
>> + dev_err(dev, "failed to request mbox tx channel.\n");
>> + pdata->tx_chan = NULL;
>> + return -EINVAL;
>> + }
>> + pdata->rx_chan = mbox_request_channel_byname(&pdata->rx_mc, "rx");
>> + if (IS_ERR(pdata->rx_chan)) {
>> + dev_err(dev, "failed to request mbox rx channel.\n");
>> + pdata->rx_chan = NULL;
>> + return -EINVAL;
>> + }
>> + skb_queue_head_init(&pdata->tx_mc_skbs);
>> + return 0;
>> +}
>> +
>> +/**
>> + * zynqmp_r5_probe() - Probes ZynqMP R5 processor device node
>> + * @pdata: pointer to the ZynqMP R5 processor platform data
>> + * @pdev: parent RPU domain platform device
>> + * @node: pointer of the device node
>> + *
>> + * Function to retrieve the information of the ZynqMP R5 device node.
>> + *
>> + * Return: 0 for success, negative value for failure.
>> + */
>> +static int zynqmp_r5_probe(struct zynqmp_r5_pdata *pdata,
>> + struct platform_device *pdev,
>> + struct device_node *node)
>> +{
>> + struct device *dev = &pdata->dev;
>> + struct rproc *rproc;
>> + struct device_node *nc;
>> + int ret;
>> +
>> + /* Create device for ZynqMP R5 device */
>> + dev->parent = &pdev->dev;
>> + dev->release = zynqmp_r5_release;
>> + dev->of_node = node;
>> + dev_set_name(dev, "%s", of_node_full_name(node));
>> + dev_set_drvdata(dev, pdata);
>> + ret = device_register(dev);
>> + if (ret) {
>> + dev_err(dev, "failed to register device.\n");
>> + return ret;
>> + }
>> + get_device(&pdev->dev);
>> +
>> + /* Allocate remoteproc instance */
>> + rproc = rproc_alloc(dev, dev_name(dev), &zynqmp_r5_rproc_ops, NULL, 0);
>> + if (!rproc) {
>> + dev_err(dev, "rproc allocation failed.\n");
>> + ret = -ENOMEM;
>> + goto error;
>> + }
>> + rproc->auto_boot = autoboot;
>> + pdata->rproc = rproc;
>> + rproc->priv = pdata;
>> +
>> + /*
>> + * The device has not been spawned from a device tree, so
>> + * arch_setup_dma_ops has not been called, thus leaving
>> + * the device with dummy DMA ops.
>> + * Fix this by inheriting the parent's DMA ops and mask.
>> + */
>> + rproc->dev.dma_mask = pdev->dev.dma_mask;
>> + set_dma_ops(&rproc->dev, get_dma_ops(&pdev->dev));
>> +
>> + /* Probe R5 memory devices */
>> + INIT_LIST_HEAD(&pdata->mems);
>> + for_each_available_child_of_node(node, nc) {
>> + ret = zynqmp_r5_mem_probe(pdata, nc);
>> + if (ret) {
>> + dev_err(dev, "failed to probe memory %s.\n",
>> + of_node_full_name(nc));
>> + goto error;
>> + }
>> + }
>> +
>> + /* Set up DMA mask */
>> + ret = dma_set_coherent_mask(dev, DMA_BIT_MASK(32));
>> + if (ret) {
>> + dev_warn(dev, "dma_set_coherent_mask failed: %d\n", ret);
>> + /* If DMA is not configured yet, try to configure it. */
>> + ret = of_dma_configure(dev, node, true);
>> + if (ret) {
>> + dev_err(dev, "failed to configure DMA.\n");
>> + goto error;
>> + }
>> + }
>> +
>> + /* Get R5 power domain node */
>> + ret = of_property_read_u32(node, "pnode-id", &pdata->pnode_id);
>> + if (ret) {
>> + dev_err(dev, "failed to get power node id.\n");
>> + goto error;
>> + }
>> +
>> + /* TODO Check if R5 is running */
>> +
>> + /* Set up R5 if not already setup */
>> + ret = pdata->is_r5_mode_set ? 0 : r5_set_mode(pdata);
>> + if (ret) {
>> + dev_err(dev, "failed to set R5 operation mode.\n");
>> + return ret;
>> + }
>> +
>> + if (!of_get_property(dev->of_node, "mboxes", NULL)) {
>> + dev_dbg(dev, "no mailboxes.\n");
>> + goto error;
>> + } else {
>> + ret = zynqmp_r5_setup_mbox(pdata, node);
>> + if (ret < 0)
>> + goto error;
>> + }
>> +
>> + /* Add R5 remoteproc */
>> + ret = rproc_add(rproc);
>> + if (ret) {
>> + dev_err(dev, "rproc registration failed\n");
>> + goto error;
>> + }
>> + return 0;
>> +error:
>> + if (pdata->rproc)
>> + rproc_free(pdata->rproc);
>> + pdata->rproc = NULL;
>> + device_unregister(dev);
>> + put_device(&pdev->dev);
>> + return ret;
>> +}
>> +
>> +static int zynqmp_r5_remoteproc_probe(struct platform_device *pdev)
>> +{
>> + int ret, i = 0;
>> + u32 *lockstep_mode;
>
> Can this just be a regular variable?
>
>> + struct device *dev = &pdev->dev;
>> + struct device_node *nc;
>> + struct zynqmp_r5_pdata *pdata;
>> +
>> + pdata = devm_kzalloc(dev, sizeof(*pdata), GFP_KERNEL);
>> + lockstep_mode = devm_kzalloc(dev, sizeof(u32 *), GFP_KERNEL);
>> + if (!pdata || !lockstep_mode)
>> + return -ENOMEM;
>> +
>> + platform_set_drvdata(pdev, pdata);
>> +
>> + of_property_read_u32(dev->of_node, "lockstep-mode", lockstep_mode);
>
> To make Rob's life a little easier this should probably be "xilinx,cluster-mode"
> so that it follows what TI has done.
>

Note: If this is required change then you need to use xlnx,cluster-mode.

Thanks,
Michal

2020-07-28 21:01:14

by Mathieu Poirier

[permalink] [raw]
Subject: Re: [PATCH v6 5/5] remoteproc: Add initial zynqmp R5 remoteproc driver

On Wed, Jul 15, 2020 at 08:33:17AM -0700, Ben Levinsky wrote:
> R5 is included in Xilinx Zynq UltraScale MPSoC so by adding this
> remotproc driver, we can boot the R5 sub-system in different
> configurations.
>
> Acked-by: Stefano Stabellini <[email protected]>
> Acked-by: Ben Levinsky <[email protected]>
> Reviewed-by: Radhey Shyam Pandey <[email protected]>
> Signed-off-by: Ben Levinsky <[email protected]>
> Signed-off-by: Wendy Liang <[email protected]>
> Signed-off-by: Michal Simek <[email protected]>
> Signed-off-by: Ed Mooring <[email protected]>
> Signed-off-by: Jason Wu <[email protected]>
> Tested-by: Ben Levinsky <[email protected]>
> ---
> v2:
> - remove domain struct as per review from Mathieu
> v3:
> - add xilinx-related platform mgmt fn's instead of wrapping around
> function pointer in xilinx eemi ops struct
> v4:
> - add default values for enums
> - fix formatting as per checkpatch.pl --strict. Note that 1 warning and 1 check
> are still raised as each is due to fixing the warning results in that
> particular line going over 80 characters.
> v5:
> - parse_fw change from use of rproc_of_resm_mem_entry_init to rproc_mem_entry_init and use of alloc/release
> - var's of type zynqmp_r5_pdata all have same local variable name
> - use dev_dbg instead of dev_info
> v6:
> - adding memory carveouts is handled much more similarly. All mem carveouts are
> now described in reserved memory as needed. That is, TCM nodes are not
> coupled to remoteproc anymore. This is reflected in the remoteproc R5 driver
> and the device tree binding.
> - remove mailbox from device tree binding as it is not necessary for elf
> loading
> - use lockstep-mode property for configuring RPU
> ---
> drivers/remoteproc/Kconfig | 10 +
> drivers/remoteproc/Makefile | 1 +
> drivers/remoteproc/zynqmp_r5_remoteproc.c | 911 ++++++++++++++++++++++
> 3 files changed, 922 insertions(+)
> create mode 100644 drivers/remoteproc/zynqmp_r5_remoteproc.c
>
> diff --git a/drivers/remoteproc/Kconfig b/drivers/remoteproc/Kconfig
> index c4d1731295eb..342a7e668636 100644
> --- a/drivers/remoteproc/Kconfig
> +++ b/drivers/remoteproc/Kconfig
> @@ -249,6 +249,16 @@ config STM32_RPROC
>
> This can be either built-in or a loadable module.
>
> +config ZYNQMP_R5_REMOTEPROC
> + tristate "ZynqMP_R5 remoteproc support"
> + depends on ARM64 && PM && ARCH_ZYNQMP
> + select RPMSG_VIRTIO
> + select MAILBOX
> + select ZYNQMP_IPI_MBOX
> + help
> + Say y here to support ZynqMP R5 remote processors via the remote
> + processor framework.
> +
> endif # REMOTEPROC
>
> endmenu
> diff --git a/drivers/remoteproc/Makefile b/drivers/remoteproc/Makefile
> index e8b886e511f0..04d1c95d06d7 100644
> --- a/drivers/remoteproc/Makefile
> +++ b/drivers/remoteproc/Makefile
> @@ -28,5 +28,6 @@ obj-$(CONFIG_QCOM_WCNSS_PIL) += qcom_wcnss_pil.o
> qcom_wcnss_pil-y += qcom_wcnss.o
> qcom_wcnss_pil-y += qcom_wcnss_iris.o
> obj-$(CONFIG_ST_REMOTEPROC) += st_remoteproc.o
> +obj-$(CONFIG_ZYNQMP_R5_REMOTEPROC) += zynqmp_r5_remoteproc.o
> obj-$(CONFIG_ST_SLIM_REMOTEPROC) += st_slim_rproc.o
> obj-$(CONFIG_STM32_RPROC) += stm32_rproc.o
> diff --git a/drivers/remoteproc/zynqmp_r5_remoteproc.c b/drivers/remoteproc/zynqmp_r5_remoteproc.c
> new file mode 100644
> index 000000000000..b600759e257e
> --- /dev/null
> +++ b/drivers/remoteproc/zynqmp_r5_remoteproc.c
> @@ -0,0 +1,911 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Zynq R5 Remote Processor driver
> + *
> + * Copyright (C) 2019, 2020 Xilinx Inc. Ben Levinsky <[email protected]>
> + * Copyright (C) 2015 - 2018 Xilinx Inc.
> + * Copyright (C) 2015 Jason Wu <[email protected]>
> + *
> + * Based on origin OMAP and Zynq Remote Processor driver
> + *
> + * Copyright (C) 2012 Michal Simek <[email protected]>
> + * Copyright (C) 2012 PetaLogix
> + * Copyright (C) 2011 Texas Instruments, Inc.
> + * Copyright (C) 2011 Google, Inc.
> + */
> +
> +#include <linux/atomic.h>
> +#include <linux/cpu.h>
> +#include <linux/dma-mapping.h>
> +#include <linux/delay.h>
> +#include <linux/err.h>
> +#include <linux/firmware/xlnx-zynqmp.h>
> +#include <linux/genalloc.h>
> +#include <linux/idr.h>
> +#include <linux/interrupt.h>
> +#include <linux/kernel.h>
> +#include <linux/list.h>
> +#include <linux/mailbox_client.h>
> +#include <linux/mailbox/zynqmp-ipi-message.h>
> +#include <linux/module.h>
> +#include <linux/of_address.h>
> +#include <linux/of_irq.h>
> +#include <linux/of_platform.h>
> +#include <linux/of_reserved_mem.h>
> +#include <linux/pfn.h>
> +#include <linux/platform_device.h>
> +#include <linux/remoteproc.h>
> +#include <linux/skbuff.h>
> +#include <linux/slab.h>
> +#include <linux/sysfs.h>
> +
> +#include "remoteproc_internal.h"
> +
> +#define MAX_RPROCS 2 /* Support up to 2 RPU */
> +#define MAX_MEM_PNODES 4 /* Max power nodes for one RPU memory instance */
> +
> +#define DEFAULT_FIRMWARE_NAME "rproc-rpu-fw"
> +
> +/* PM proc states */
> +#define PM_PROC_STATE_ACTIVE 1U
> +
> +/* IPI buffer MAX length */
> +#define IPI_BUF_LEN_MAX 32U
> +/* RX mailbox client buffer max length */
> +#define RX_MBOX_CLIENT_BUF_MAX (IPI_BUF_LEN_MAX + \
> + sizeof(struct zynqmp_ipi_message))
> +
> +#define ZYNQMP_R5_NUM_TCM_BANKS 4
> +
> +/* lookup table mapping power-node-ID of TCM bank to absolute base address */
> +static unsigned long tcm_addr_to_pnode[ZYNQMP_R5_NUM_TCM_BANKS][2] = {
> + {0xFFE00000, 0xF },
> + {0xFFE20000, 0x10},
> + {0xFFE90000, 0x10},
> + {0xFFEB0000, 0x11},
> +};
> +
> +static bool autoboot __read_mostly;
> +
> +/**
> + * struct zynqmp_r5_mem - zynqmp rpu memory data
> + * @pnode_id: TCM power domain ids
> + * @res: memory resource
> + * @node: list node
> + */
> +struct zynqmp_r5_mem {
> + u32 pnode_id[MAX_MEM_PNODES];
> + struct resource res;
> + struct list_head node;
> +};
> +
> +/**
> + * struct zynqmp_r5_pdata - zynqmp rpu remote processor private data
> + * @dev: device of RPU instance
> + * @rproc: rproc handle
> + * @pnode_id: RPU CPU power domain id
> + * @mems: memory resources
> + * @is_r5_mode_set: indicate if r5 operation mode is set
> + * @tx_mc: tx mailbox client
> + * @rx_mc: rx mailbox client
> + * @tx_chan: tx mailbox channel
> + * @rx_chan: rx mailbox channel
> + * @mbox_work: mbox_work for the RPU remoteproc
> + * @tx_mc_skbs: socket buffers for tx mailbox client
> + * @rx_mc_buf: rx mailbox client buffer to save the rx message
> + */
> +struct zynqmp_r5_pdata {
> + struct device dev;
> + struct rproc *rproc;
> + u32 pnode_id;
> + struct list_head mems;
> + bool is_r5_mode_set;
> + struct mbox_client tx_mc;
> + struct mbox_client rx_mc;
> + struct mbox_chan *tx_chan;
> + struct mbox_chan *rx_chan;
> + struct work_struct mbox_work;
> + struct sk_buff_head tx_mc_skbs;
> + unsigned char rx_mc_buf[RX_MBOX_CLIENT_BUF_MAX];
> +};
> +
> +/**
> + * table of RPUs
> + */
> +struct zynqmp_r5_pdata rpus[MAX_RPROCS];
> +/**
> + * RPU core configuration
> + */
> +enum rpu_oper_mode rpu_mode;
> +
> +/*
> + * r5_set_mode - set RPU operation mode
> + * @pdata: Remote processor private data
> + *
> + * set RPU oepration mode
> + *
> + * Return: 0 for success, negative value for failure
> + */
> +static int r5_set_mode(struct zynqmp_r5_pdata *pdata)
> +{
> + u32 val[PAYLOAD_ARG_CNT] = {0}, expect, tcm_mode;
> + struct device *dev = &pdata->dev;
> + int ret;
> +
> + expect = (u32)rpu_mode;
> + ret = zynqmp_pm_get_rpu_mode(pdata->pnode_id, 0, 0, val);
> + if (ret < 0) {
> + dev_err(dev, "failed to get RPU oper mode.\n");
> + return ret;
> + }
> + if (val[0] == expect) {
> + dev_dbg(dev, "RPU mode matches: %x\n", val[0]);
> + } else {
> + ret = zynqmp_pm_set_rpu_mode(pdata->pnode_id,
> + expect, 0, val);
> + if (ret < 0) {
> + dev_err(dev,
> + "failed to set RPU oper mode.\n");
> + return ret;
> + }
> + }
> +
> + tcm_mode = (expect == (u32)PM_RPU_MODE_LOCKSTEP) ?
> + PM_RPU_TCM_COMB : PM_RPU_TCM_SPLIT;
> + ret = zynqmp_pm_set_tcm_config(pdata->pnode_id, tcm_mode, 0, val);
> + if (ret < 0) {
> + dev_err(dev, "failed to config TCM to %x.\n",
> + expect);
> + return ret;
> + }
> + pdata->is_r5_mode_set = true;
> + return 0;
> +}
> +
> +/*
> + * ZynqMP R5 remoteproc memory release function
> + */
> +static int zynqmp_r5_mem_release(struct rproc *rproc,
> + struct rproc_mem_entry *mem)
> +{
> + struct zynqmp_r5_mem *priv;
> + int i, ret;
> + struct device *dev = &rproc->dev;
> +
> + priv = mem->priv;
> + if (!priv)
> + return 0;
> + for (i = 0; i < MAX_MEM_PNODES; i++) {
> + if (priv->pnode_id[i]) {
> + dev_dbg(dev, "%s, pnode %d\n",
> + __func__, priv->pnode_id[i]);
> + ret = zynqmp_pm_release_node(priv->pnode_id[i]);
> + if (ret < 0) {
> + dev_err(dev,
> + "failed to release power node: %u\n",
> + priv->pnode_id[i]);
> + return ret;
> + }
> + } else {
> + break;
> + }
> + }
> + return 0;
> +}
> +
> +/*
> + * ZynqMP R5 remoteproc operations
> + */
> +static int zynqmp_r5_rproc_start(struct rproc *rproc)
> +{
> + struct device *dev = rproc->dev.parent;
> + struct zynqmp_r5_pdata *pdata = rproc->priv;
> + enum rpu_boot_mem bootmem;
> + int ret;
> +
> + if ((rproc->bootaddr & 0xF0000000) == 0xF0000000)
> + bootmem = PM_RPU_BOOTMEM_HIVEC;
> + else
> + bootmem = PM_RPU_BOOTMEM_LOVEC;
> + dev_dbg(dev, "RPU boot from %s.",
> + bootmem == PM_RPU_BOOTMEM_HIVEC ? "OCM" : "TCM");
> + ret = zynqmp_pm_request_wakeup(pdata->pnode_id, 1,
> + bootmem, ZYNQMP_PM_REQUEST_ACK_NO);
> + if (ret < 0) {
> + dev_err(dev, "failed to boot R5.\n");
> + return ret;
> + }
> + return 0;
> +}
> +
> +static int zynqmp_r5_rproc_stop(struct rproc *rproc)
> +{
> + struct zynqmp_r5_pdata *pdata = rproc->priv;
> + int ret;
> +
> + ret = zynqmp_pm_force_powerdown(pdata->pnode_id,
> + ZYNQMP_PM_REQUEST_ACK_BLOCKING);
> + if (ret < 0) {
> + dev_err(&pdata->dev, "failed to shutdown R5.\n");
> + return ret;
> + }
> + return 0;
> +}
> +
> +static int zynqmp_r5_rproc_mem_alloc(struct rproc *rproc,
> + struct rproc_mem_entry *mem)
> +{
> + struct device *dev = rproc->dev.parent;
> + void *va;
> +
> + dev_dbg(rproc->dev.parent, "map memory: %pa\n", &mem->dma);
> + va = ioremap_wc(mem->dma, mem->len);
> + if (IS_ERR_OR_NULL(va)) {
> + dev_err(dev, "Unable to map memory region: %pa+%lx\n",
> + &mem->dma, mem->len);
> + return -ENOMEM;
> + }
> +
> + /* Update memory entry va */
> + mem->va = va;
> +
> + return 0;
> +}
> +
> +static int zynqmp_r5_rproc_mem_release(struct rproc *rproc,
> + struct rproc_mem_entry *mem)
> +{
> + dev_dbg(rproc->dev.parent, "unmap memory: %pa\n", &mem->dma);
> + iounmap(mem->va);
> +
> + return 0;
> +}
> +
> +/*
> + * TCM needs mapping to R5 relative address and xilinx platform mgmt call
> + */
> +struct rproc_mem_entry *handle_tcm_parsing(struct device *dev,
> + struct reserved_mem *rmem,
> + struct device_node *node,
> + int lookup_idx)
> +{
> + void *va;
> + dma_addr_t dma;
> + resource_size_t size;
> + int ret;
> + u32 pnode_id;
> + struct resource rsc;
> + struct rproc_mem_entry *mem;
> +
> + pnode_id = tcm_addr_to_pnode[lookup_idx][1];
> + ret = zynqmp_pm_request_node(pnode_id,
> + ZYNQMP_PM_CAPABILITY_ACCESS, 0,
> + ZYNQMP_PM_REQUEST_ACK_BLOCKING);
> + if (ret < 0) {
> + dev_err(dev, "failed to request power node: %u\n", pnode_id);
> + return -EINVAL;
> + }
> +
> + ret = of_address_to_resource(node, 0, &rsc);
> + if (ret < 0) {
> + dev_err(dev, "failed to get resource of memory %s",
> + of_node_full_name(node));
> + return -EINVAL;
> + }
> + size = resource_size(&rsc);
> + va = devm_ioremap_wc(dev, rsc.start, size);
> + if (!va)
> + return -ENOMEM;
> +
> + /* zero out tcm base address */
> + if (rsc.start & 0xffe00000) {
> + /* R5 can't see anything past 0xfffff so wipe it */
> + rsc.start &= 0x000fffff;
> + /*
> + * handle tcm banks 1 a and b (0xffe9000 and
> + * 0xffeb0000)
> + */
> + if (rsc.start & 0x80000)
> + rsc.start -= 0x90000;
> + }
> +
> + dma = (dma_addr_t)rsc.start;
> + mem = rproc_mem_entry_init(dev, va, dma, (int)size, rsc.start,
> + NULL, zynqmp_r5_mem_release,
> + rsc.name);
> + if (!mem)
> + return -ENOMEM;
> +
> + return mem;
> +}
> +
> +static int parse_mem_regions(struct rproc *rproc)
> +{
> + int num_mems, i;
> + struct zynqmp_r5_pdata *pdata = rproc->priv;
> + struct device *dev = &pdata->dev;
> + struct device_node *np = dev->of_node;
> + struct rproc_mem_entry *mem;
> +
> + num_mems = of_count_phandle_with_args(np, "memory-region", NULL);
> + if (num_mems <= 0)
> + return 0;
> + for (i = 0; i < num_mems; i++) {
> + struct device_node *node;
> + struct reserved_mem *rmem;
> +
> + node = of_parse_phandle(np, "memory-region", i);
> + rmem = of_reserved_mem_lookup(node);
> + if (!rmem) {
> + dev_err(dev, "unable to acquire memory-region\n");
> + return -EINVAL;
> + }
> +
> + if (strstr(node->name, "vdev0buffer")) {
> + /* Register DMA region */
> + mem = rproc_mem_entry_init(dev, NULL,
> + (dma_addr_t)rmem->base,
> + rmem->size, rmem->base,
> + NULL, NULL,
> + "vdev0buffer");
> + if (!mem) {
> + dev_err(dev, "unable to initialize memory-region %s\n",
> + node->name);
> + return -ENOMEM;
> + }
> + dev_dbg(dev, "parsed %s at %llx\r\n", mem->name,
> + mem->dma);
> + } else if (strstr(node->name, "vdev0vring")) {
> + int vring_id;
> + char name[16];
> +
> + /*
> + * can be 1 of multiple vring IDs per IPC channel
> + * e.g. 'vdev0vring0' and 'vdev0vring1'
> + */
> + vring_id = node->name[14] - '0';
> + snprintf(name, sizeof(name), "vdev0vring%d", vring_id);
> + /* Register vring */
> + mem = rproc_mem_entry_init(dev, NULL,
> + (dma_addr_t)rmem->base,
> + rmem->size, rmem->base,
> + zynqmp_r5_rproc_mem_alloc,
> + zynqmp_r5_rproc_mem_release,
> + name);
> + dev_dbg(dev, "parsed %s at %llx\r\n", mem->name,
> + mem->dma);
> + } else {
> + int idx;
> +
> + /*
> + * if TCM update address space for R5 and
> + * make xilinx platform mgmt call
> + */
> + for (idx = 0; idx < ZYNQMP_R5_NUM_TCM_BANKS; idx++) {
> + if (tcm_addr_to_pnode[idx][0] == rmem->base)
> + break;
> + }
> +
> + if (idx != ZYNQMP_R5_NUM_TCM_BANKS) {
> + mem = handle_tcm_parsing(dev, rmem, node, idx);
> + } else {
> + mem = rproc_mem_entry_init(dev, NULL,
> + (dma_addr_t)rmem->base,
> + rmem->size, rmem->base,
> + zynqmp_r5_rproc_mem_alloc,
> + zynqmp_r5_rproc_mem_release,
> + node->name);
> + }
> +
> + if (!mem) {
> + dev_err(dev,
> + "unable to init memory-region %s\n",
> + node->name);
> + return -ENOMEM;
> + }
> + }
> + rproc_add_carveout(rproc, mem);
> + }
> +
> + return 0;
> +}
> +
> +static int zynqmp_r5_parse_fw(struct rproc *rproc, const struct firmware *fw)
> +{
> + int ret;
> + struct zynqmp_r5_pdata *pdata = rproc->priv;
> + struct device *dev = &pdata->dev;
> +
> + ret = parse_mem_regions(rproc);
> + if (ret) {
> + dev_err(dev, "parse_mem_regions failed %x\n", ret);
> + return ret;
> + }
> +
> + ret = rproc_elf_load_rsc_table(rproc, fw);
> + if (ret == -EINVAL) {
> + dev_dbg(dev, "no resource table found.\n");
> + ret = 0;
> + }
> + return ret;
> +}
> +
> +/* kick a firmware */
> +static void zynqmp_r5_rproc_kick(struct rproc *rproc, int vqid)
> +{
> + struct device *dev = rproc->dev.parent;
> + struct zynqmp_r5_pdata *pdata = rproc->priv;
> +
> + dev_dbg(dev, "KICK Firmware to start send messages vqid %d\n", vqid);
> +
> + if (vqid < 0) {
> + /* If vqid is negative, does not pass the vqid to
> + * mailbox. As vqid is supposed to be 0 or possive.
> + * It also gives a way to just kick instead but
> + * not use the IPI buffer. It is better to provide
> + * a proper way to pass the short message, which will
> + * need to sync to upstream first, for now,
> + * use negative vqid to assume no message will be
> + * passed with IPI buffer, but just raise interrupt.
> + * This will be faster as it doesn't need to copy the
> + * message to the IPI buffer.
> + *
> + * It will ignore the return, as failure is due to
> + * there already kicks in the mailbox queue.
> + */
> + (void)mbox_send_message(pdata->tx_chan, NULL);
> + } else {
> + struct sk_buff *skb;
> + unsigned int skb_len;
> + struct zynqmp_ipi_message *mb_msg;
> + int ret;
> +
> + skb_len = (unsigned int)(sizeof(vqid) + sizeof(mb_msg));
> + skb = alloc_skb(skb_len, GFP_ATOMIC);
> + if (!skb) {
> + dev_err(dev,
> + "Failed to allocate skb to kick remote.\n");
> + return;
> + }
> + mb_msg = (struct zynqmp_ipi_message *)skb_put(skb, skb_len);
> + mb_msg->len = sizeof(vqid);
> + memcpy(mb_msg->data, &vqid, sizeof(vqid));
> + skb_queue_tail(&pdata->tx_mc_skbs, skb);
> + ret = mbox_send_message(pdata->tx_chan, mb_msg);
> + if (ret < 0) {
> + dev_warn(dev, "Failed to kick remote.\n");
> + skb_dequeue_tail(&pdata->tx_mc_skbs);
> + kfree_skb(skb);
> + }
> + }
> +}
> +
> +static struct rproc_ops zynqmp_r5_rproc_ops = {
> + .start = zynqmp_r5_rproc_start,
> + .stop = zynqmp_r5_rproc_stop,
> + .load = rproc_elf_load_segments,
> + .parse_fw = zynqmp_r5_parse_fw,
> + .find_loaded_rsc_table = rproc_elf_find_loaded_rsc_table,
> + .sanity_check = rproc_elf_sanity_check,
> + .get_boot_addr = rproc_elf_get_boot_addr,
> + .kick = zynqmp_r5_rproc_kick,
> +};
> +
> +/* zynqmp_r5_mem_probe() - probes RPU TCM memory device
> + * @pdata: pointer to the RPU remoteproc private data
> + * @node: pointer to the memory node
> + *
> + * Function to retrieve resources for RPU TCM memory device.
> + */
> +static int zynqmp_r5_mem_probe(struct zynqmp_r5_pdata *pdata,
> + struct device_node *node)
> +{
> + struct device *dev;
> + struct zynqmp_r5_mem *mem;
> + int ret;

int ret, i = 0;

> + struct property *prop;
> + const __be32 *cur;
> + u32 val;
> + int i;
> +
> + dev = &pdata->dev;
> + mem = devm_kzalloc(dev, sizeof(*mem), GFP_KERNEL);
> + if (!mem)
> + return -ENOMEM;
> + ret = of_address_to_resource(node, 0, &mem->res);
> + if (ret < 0) {
> + dev_err(dev, "failed to get resource of memory %s",
> + of_node_full_name(node));
> + return -EINVAL;
> + }
> +
> + /* Get the power domain id */
> + i = 0;

In version 5 of this set initialisation was done at declaration time - why was
it moved here? The high fluctuation of changes in the code makes reviewing your
work very difficult.

> + if (of_find_property(node, "pnode-id", NULL)) {
> + of_property_for_each_u32(node, "pnode-id", prop, cur, val)
> + mem->pnode_id[i++] = val;
> + }
> + list_add_tail(&mem->node, &pdata->mems);
> + return 0;
> +}
> +
> +/**
> + * zynqmp_r5_release() - ZynqMP R5 device release function
> + * @dev: pointer to the device struct of ZynqMP R5
> + *
> + * Function to release ZynqMP R5 device.
> + */
> +static void zynqmp_r5_release(struct device *dev)
> +{
> + struct zynqmp_r5_pdata *pdata;
> + struct rproc *rproc;
> + struct sk_buff *skb;
> +
> + pdata = dev_get_drvdata(dev);
> + rproc = pdata->rproc;
> + if (rproc) {
> + rproc_del(rproc);
> + rproc_free(rproc);
> + }
> + if (pdata->tx_chan)
> + mbox_free_channel(pdata->tx_chan);
> + if (pdata->rx_chan)
> + mbox_free_channel(pdata->rx_chan);
> + /* Discard all SKBs */
> + while (!skb_queue_empty(&pdata->tx_mc_skbs)) {
> + skb = skb_dequeue(&pdata->tx_mc_skbs);
> + kfree_skb(skb);
> + }
> +
> + put_device(dev->parent);
> +}
> +
> +/**
> + * event_notified_idr_cb() - event notified idr callback
> + * @id: idr id
> + * @ptr: pointer to idr private data
> + * @data: data passed to idr_for_each callback
> + *
> + * Pass notification to remoteproc virtio
> + *
> + * Return: 0. having return is to satisfy the idr_for_each() function
> + * pointer input argument requirement.
> + **/
> +static int event_notified_idr_cb(int id, void *ptr, void *data)
> +{
> + struct rproc *rproc = data;
> +
> + (void)rproc_vq_interrupt(rproc, id);
> + return 0;
> +}
> +
> +/**
> + * handle_event_notified() - remoteproc notification work funciton
> + * @work: pointer to the work structure
> + *
> + * It checks each registered remoteproc notify IDs.
> + */
> +static void handle_event_notified(struct work_struct *work)
> +{
> + struct rproc *rproc;
> + struct zynqmp_r5_pdata *pdata;
> +
> + pdata = container_of(work, struct zynqmp_r5_pdata, mbox_work);
> +
> + (void)mbox_send_message(pdata->rx_chan, NULL);
> + rproc = pdata->rproc;
> + /*
> + * We only use IPI for interrupt. The firmware side may or may
> + * not write the notifyid when it trigger IPI.
> + * And thus, we scan through all the registered notifyids.
> + */
> + idr_for_each(&rproc->notifyids, event_notified_idr_cb, rproc);
> +}
> +
> +/**
> + * zynqmp_r5_mb_rx_cb() - Receive channel mailbox callback
> + * @cl: mailbox client
> + * @mssg: message pointer
> + *
> + * It will schedule the R5 notification work.
> + */
> +static void zynqmp_r5_mb_rx_cb(struct mbox_client *cl, void *mssg)
> +{
> + struct zynqmp_r5_pdata *pdata;
> +
> + pdata = container_of(cl, struct zynqmp_r5_pdata, rx_mc);
> + if (mssg) {
> + struct zynqmp_ipi_message *ipi_msg, *buf_msg;
> + size_t len;
> +
> + ipi_msg = (struct zynqmp_ipi_message *)mssg;
> + buf_msg = (struct zynqmp_ipi_message *)pdata->rx_mc_buf;
> + len = (ipi_msg->len >= IPI_BUF_LEN_MAX) ?
> + IPI_BUF_LEN_MAX : ipi_msg->len;
> + buf_msg->len = len;
> + memcpy(buf_msg->data, ipi_msg->data, len);
> + }
> + schedule_work(&pdata->mbox_work);
> +}
> +
> +/**
> + * zynqmp_r5_mb_tx_done() - Request has been sent to the remote
> + * @cl: mailbox client
> + * @mssg: pointer to the message which has been sent
> + * @r: status of last TX - OK or error
> + *
> + * It will be called by the mailbox framework when the last TX has done.
> + */
> +static void zynqmp_r5_mb_tx_done(struct mbox_client *cl, void *mssg, int r)
> +{
> + struct zynqmp_r5_pdata *pdata;
> + struct sk_buff *skb;
> +
> + if (!mssg)
> + return;
> + pdata = container_of(cl, struct zynqmp_r5_pdata, tx_mc);
> + skb = skb_dequeue(&pdata->tx_mc_skbs);
> + kfree_skb(skb);
> +}
> +
> +/**
> + * zynqmp_r5_setup_mbox() - Setup mailboxes
> + *
> + * @pdata: pointer to the ZynqMP R5 processor platform data
> + * @node: pointer of the device node
> + *
> + * Function to setup mailboxes to talk to RPU.
> + *
> + * Return: 0 for success, negative value for failure.
> + */
> +static int zynqmp_r5_setup_mbox(struct zynqmp_r5_pdata *pdata,
> + struct device_node *node)
> +{
> + struct device *dev = &pdata->dev;
> + struct mbox_client *mclient;
> +
> + /* Setup TX mailbox channel client */
> + mclient = &pdata->tx_mc;
> + mclient->dev = dev;
> + mclient->rx_callback = NULL;
> + mclient->tx_block = false;
> + mclient->knows_txdone = false;
> + mclient->tx_done = zynqmp_r5_mb_tx_done;
> +
> + /* Setup TX mailbox channel client */
> + mclient = &pdata->rx_mc;
> + mclient->dev = dev;
> + mclient->rx_callback = zynqmp_r5_mb_rx_cb;
> + mclient->tx_block = false;
> + mclient->knows_txdone = false;
> +
> + INIT_WORK(&pdata->mbox_work, handle_event_notified);
> +
> + /* Request TX and RX channels */
> + pdata->tx_chan = mbox_request_channel_byname(&pdata->tx_mc, "tx");
> + if (IS_ERR(pdata->tx_chan)) {
> + dev_err(dev, "failed to request mbox tx channel.\n");
> + pdata->tx_chan = NULL;
> + return -EINVAL;
> + }
> + pdata->rx_chan = mbox_request_channel_byname(&pdata->rx_mc, "rx");
> + if (IS_ERR(pdata->rx_chan)) {
> + dev_err(dev, "failed to request mbox rx channel.\n");
> + pdata->rx_chan = NULL;
> + return -EINVAL;
> + }
> + skb_queue_head_init(&pdata->tx_mc_skbs);
> + return 0;
> +}
> +
> +/**
> + * zynqmp_r5_probe() - Probes ZynqMP R5 processor device node
> + * @pdata: pointer to the ZynqMP R5 processor platform data
> + * @pdev: parent RPU domain platform device
> + * @node: pointer of the device node
> + *
> + * Function to retrieve the information of the ZynqMP R5 device node.
> + *
> + * Return: 0 for success, negative value for failure.
> + */
> +static int zynqmp_r5_probe(struct zynqmp_r5_pdata *pdata,
> + struct platform_device *pdev,
> + struct device_node *node)
> +{
> + struct device *dev = &pdata->dev;
> + struct rproc *rproc;
> + struct device_node *nc;
> + int ret;
> +
> + /* Create device for ZynqMP R5 device */
> + dev->parent = &pdev->dev;
> + dev->release = zynqmp_r5_release;
> + dev->of_node = node;
> + dev_set_name(dev, "%s", of_node_full_name(node));
> + dev_set_drvdata(dev, pdata);
> + ret = device_register(dev);
> + if (ret) {
> + dev_err(dev, "failed to register device.\n");
> + return ret;
> + }
> + get_device(&pdev->dev);
> +
> + /* Allocate remoteproc instance */
> + rproc = rproc_alloc(dev, dev_name(dev), &zynqmp_r5_rproc_ops, NULL, 0);
> + if (!rproc) {
> + dev_err(dev, "rproc allocation failed.\n");
> + ret = -ENOMEM;
> + goto error;
> + }
> + rproc->auto_boot = autoboot;
> + pdata->rproc = rproc;
> + rproc->priv = pdata;
> +
> + /*
> + * The device has not been spawned from a device tree, so
> + * arch_setup_dma_ops has not been called, thus leaving
> + * the device with dummy DMA ops.
> + * Fix this by inheriting the parent's DMA ops and mask.
> + */

This _might_ be fixed by calling devm_of_platform_populate().

> + rproc->dev.dma_mask = pdev->dev.dma_mask;
> + set_dma_ops(&rproc->dev, get_dma_ops(&pdev->dev));
> +
> + /* Probe R5 memory devices */
> + INIT_LIST_HEAD(&pdata->mems);
> + for_each_available_child_of_node(node, nc) {
> + ret = zynqmp_r5_mem_probe(pdata, nc);

Contrary to version 5 of this series, the example in the bindings does not
include TCM nodes. As such I have to wonder if TCMs are needed to boot the R5s
- if so the bindings are wrong and if not, this loop needs to be removed.

> + if (ret) {
> + dev_err(dev, "failed to probe memory %s.\n",
> + of_node_full_name(nc));
> + goto error;
> + }
> + }
> +
> + /* Set up DMA mask */
> + ret = dma_set_coherent_mask(dev, DMA_BIT_MASK(32));
> + if (ret) {
> + dev_warn(dev, "dma_set_coherent_mask failed: %d\n", ret);
> + /* If DMA is not configured yet, try to configure it. */
> + ret = of_dma_configure(dev, node, true);
> + if (ret) {
> + dev_err(dev, "failed to configure DMA.\n");
> + goto error;
> + }
> + }
> +
> + /* Get R5 power domain node */
> + ret = of_property_read_u32(node, "pnode-id", &pdata->pnode_id);
> + if (ret) {
> + dev_err(dev, "failed to get power node id.\n");
> + goto error;
> + }
> +
> + /* TODO Check if R5 is running */
> +
> + /* Set up R5 if not already setup */
> + ret = pdata->is_r5_mode_set ? 0 : r5_set_mode(pdata);

Have you encountered a case where ->is_r5_mode_set is not '0'? As far as I can
tell @pdata is &rpu[i] from zynqmp_r5_remoteproc_probe() and is global. As such
it should be located in the bss section and initialised to 0 by default.

> + if (ret) {
> + dev_err(dev, "failed to set R5 operation mode.\n");
> + return ret;
> + }
> +
> + if (!of_get_property(dev->of_node, "mboxes", NULL)) {
> + dev_dbg(dev, "no mailboxes.\n");
> + goto error;

The changelog in patch 4 specifically mentions that mailboxes are not needed
and yet the code does otherwise.

> + } else {
> + ret = zynqmp_r5_setup_mbox(pdata, node);
> + if (ret < 0)
> + goto error;
> + }
> +
> + /* Add R5 remoteproc */
> + ret = rproc_add(rproc);
> + if (ret) {
> + dev_err(dev, "rproc registration failed\n");
> + goto error;
> + }
> + return 0;
> +error:
> + if (pdata->rproc)
> + rproc_free(pdata->rproc);
> + pdata->rproc = NULL;
> + device_unregister(dev);
> + put_device(&pdev->dev);
> + return ret;
> +}
> +
> +static int zynqmp_r5_remoteproc_probe(struct platform_device *pdev)
> +{
> + int ret, i = 0;
> + u32 *lockstep_mode;
> + struct device *dev = &pdev->dev;
> + struct device_node *nc;
> + struct zynqmp_r5_pdata *pdata;
> +
> + pdata = devm_kzalloc(dev, sizeof(*pdata), GFP_KERNEL);
> + lockstep_mode = devm_kzalloc(dev, sizeof(u32 *), GFP_KERNEL);
> + if (!pdata || !lockstep_mode)
> + return -ENOMEM;
> +
> + platform_set_drvdata(pdev, pdata);

As far as I can tell the above, along with allocating memory for @pdata, is not
needed since zynqmp_r5_remoteproc_remove() uses rpus[].

I have only reviewed the _probe() function and already encountered a fair amount
of fundemantal errors. As such I will stop my review here. I will need to see a
reviewed-by tag (on the mailing list) by Stephano or Michal before reviewing the
next set.

Thanks,
Mathieu

> +
> + of_property_read_u32(dev->of_node, "lockstep-mode", lockstep_mode);
> +
> + if (!(*lockstep_mode)) {
> + rpu_mode = PM_RPU_MODE_SPLIT;
> + } else if (*lockstep_mode == 1) {
> + rpu_mode = PM_RPU_MODE_LOCKSTEP;
> + } else {
> + dev_err(dev,
> + "Invalid lockstep-mode mode provided - %x %d\n",
> + *lockstep_mode, rpu_mode);
> + return -EINVAL;
> + }
> + dev_dbg(dev, "RPU configuration: %s\r\n",
> + (*lockstep_mode) ? "lockstep" : "split");
> +
> + for_each_available_child_of_node(dev->of_node, nc) {
> + ret = zynqmp_r5_probe(&rpus[i], pdev, nc);
> + if (ret) {
> + dev_err(dev, "failed to probe rpu %s.\n",
> + of_node_full_name(nc));
> + return ret;
> + }
> + i++;
> + }
> +
> + return 0;
> +}
> +
> +static int zynqmp_r5_remoteproc_remove(struct platform_device *pdev)
> +{
> + int i;
> +
> + for (i = 0; i < MAX_RPROCS; i++) {
> + struct zynqmp_r5_pdata *pdata = &rpus[i];
> + struct rproc *rproc;
> +
> + rproc = pdata->rproc;
> + if (rproc) {
> + rproc_del(rproc);
> + rproc_free(rproc);
> + pdata->rproc = NULL;
> + }
> + if (pdata->tx_chan) {
> + mbox_free_channel(pdata->tx_chan);
> + pdata->tx_chan = NULL;
> + }
> + if (pdata->rx_chan) {
> + mbox_free_channel(pdata->rx_chan);
> + pdata->rx_chan = NULL;
> + }
> +
> + device_unregister(&pdata->dev);
> + }
> +
> + return 0;
> +}
> +
> +/* Match table for OF platform binding */
> +static const struct of_device_id zynqmp_r5_remoteproc_match[] = {
> + { .compatible = "xlnx,zynqmp-r5-remoteproc-1.0", },
> + { /* end of list */ },
> +};
> +MODULE_DEVICE_TABLE(of, zynqmp_r5_remoteproc_match);
> +
> +static struct platform_driver zynqmp_r5_remoteproc_driver = {
> + .probe = zynqmp_r5_remoteproc_probe,
> + .remove = zynqmp_r5_remoteproc_remove,
> + .driver = {
> + .name = "zynqmp_r5_remoteproc",
> + .of_match_table = zynqmp_r5_remoteproc_match,
> + },
> +};
> +module_platform_driver(zynqmp_r5_remoteproc_driver);
> +
> +module_param_named(autoboot, autoboot, bool, 0444);
> +MODULE_PARM_DESC(autoboot,
> + "enable | disable autoboot. (default: false)");
> +
> +MODULE_AUTHOR("Ben Levinsky <[email protected]>");
> +MODULE_LICENSE("GPL v2");
> --
> 2.17.1
>

2020-07-31 23:04:26

by Stefano Stabellini

[permalink] [raw]
Subject: Re: [PATCH v6 5/5] remoteproc: Add initial zynqmp R5 remoteproc driver

On Tue, 28 Jul 2020, Mathieu Poirier wrote:
> On Wed, Jul 15, 2020 at 08:33:17AM -0700, Ben Levinsky wrote:
> > R5 is included in Xilinx Zynq UltraScale MPSoC so by adding this
> > remotproc driver, we can boot the R5 sub-system in different
> > configurations.
> >
> > Acked-by: Stefano Stabellini <[email protected]>
> > Acked-by: Ben Levinsky <[email protected]>
> > Reviewed-by: Radhey Shyam Pandey <[email protected]>
> > Signed-off-by: Ben Levinsky <[email protected]>
> > Signed-off-by: Wendy Liang <[email protected]>
> > Signed-off-by: Michal Simek <[email protected]>
> > Signed-off-by: Ed Mooring <[email protected]>
> > Signed-off-by: Jason Wu <[email protected]>
> > Tested-by: Ben Levinsky <[email protected]>

[...]

> > +static int zynqmp_r5_remoteproc_probe(struct platform_device *pdev)
> > +{
> > + int ret, i = 0;
> > + u32 *lockstep_mode;
> > + struct device *dev = &pdev->dev;
> > + struct device_node *nc;
> > + struct zynqmp_r5_pdata *pdata;
> > +
> > + pdata = devm_kzalloc(dev, sizeof(*pdata), GFP_KERNEL);
> > + lockstep_mode = devm_kzalloc(dev, sizeof(u32 *), GFP_KERNEL);
> > + if (!pdata || !lockstep_mode)
> > + return -ENOMEM;
> > +
> > + platform_set_drvdata(pdev, pdata);
>
> As far as I can tell the above, along with allocating memory for @pdata, is not
> needed since zynqmp_r5_remoteproc_remove() uses rpus[].
>
> I have only reviewed the _probe() function and already encountered a fair amount
> of fundemantal errors. As such I will stop my review here. I will need to see a
> reviewed-by tag (on the mailing list) by Stephano or Michal before reviewing the
> next set.

Let me take this opportunity to say that my Acked-by on this version of
the series was an unintentional miscommunication: I didn't give my
Acked-by as I haven't even read the patches yet.

I'll circle back with Michal and we'll make sure for either of us to do
a round of public reviews on the next version.