Received: by 2002:a05:6a10:af89:0:0:0:0 with SMTP id iu9csp1191566pxb; Fri, 21 Jan 2022 11:54:46 -0800 (PST) X-Google-Smtp-Source: ABdhPJy2g12DSxFtCi8g8Z89rKJczipv5UfjaL4SPF7T3kN4sKtDzqnWoKLST4njmCM/2zRm50Fq X-Received: by 2002:a05:6a00:be4:b0:4be:f488:8dbd with SMTP id x36-20020a056a000be400b004bef4888dbdmr5143336pfu.27.1642794886000; Fri, 21 Jan 2022 11:54:46 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1642794885; cv=none; d=google.com; s=arc-20160816; b=wEjuiNxyS1segVX+m3DxwTNr2GnSLKXT7MG0Wc6w2SidiPG43nWw4Z7uFHg/x+Lr/q oBtpZx1wp9EaNopQg6hU1t40zGvXnzIezXJJzs10mPppGPtr+7ZlmaErEjgOLUqBaY7/ pkvpU9kwRzu6uxAXGWrHmrlDGlPLq7EKbRiZWRh8JJNp+nvi4hZBP06ruzkiuNVdX0zj 5QHgc/2dPJCDLrTye7xzVc/gVIo9U8YHW8+FptJPWMWhAKHq7OcT6aqgnuGdtWUypLst P6dVpXVuyAMR2G9qmuvTIAsRmOLRPYjguzMvldXUx1JyCUutIcEpibqHq3uBbXM8tSUM 7XSA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=DQ8qY4OyOkFvaykGyjwGKL6pcDR4EejouSObBAhGz24=; b=ZXnBF6ZLKH94hrKmIlfC6aPPb6DzmnVPPxbDpXfLdY1+Qgv8Sd355KpmsYSx5u2Bg9 IdyCErC6m7paGE73f+V8ZsRTMTkrmrQl05UzTTNM5ESX90W9B6kDTrgl/PgTmtcKBUvr NYyRWh62Tz7aS2IEPycCNN3YsNurAByJ5gwn5YYWm5B+R9QzxXb3zr9jWnygfJVs17PG CtkWuexcBpeijScUyYn1ATZhgm4Nh3B4pz6Sm2Z8A25iK5ZfwoFT+9MQpNXW9LJEJFIE pel8VraxVgF0fZPnsidH/mrrq+2XZkoFBTwFq2gWawGxZEBayGgf3kwaNsYuT2Ydvzrk yowQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=auF8Wjon; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id t32si2250173pfg.98.2022.01.21.11.54.33; Fri, 21 Jan 2022 11:54:45 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=auF8Wjon; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1356626AbiASSET (ORCPT + 99 others); Wed, 19 Jan 2022 13:04:19 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:38204 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1356620AbiASSES (ORCPT ); Wed, 19 Jan 2022 13:04:18 -0500 Received: from mail-pj1-x102d.google.com (mail-pj1-x102d.google.com [IPv6:2607:f8b0:4864:20::102d]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id F1666C061574 for ; Wed, 19 Jan 2022 10:04:17 -0800 (PST) Received: by mail-pj1-x102d.google.com with SMTP id n16-20020a17090a091000b001b46196d572so3326643pjn.5 for ; Wed, 19 Jan 2022 10:04:17 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to; bh=DQ8qY4OyOkFvaykGyjwGKL6pcDR4EejouSObBAhGz24=; b=auF8WjonDn3tEyolyKygoNil+sWIDVBM3Jz2kvvsd5VGc6ahPn9M9S9fvSnmBb1cZM B45mmEYDAnmh1Z7VlN2JporKbxsES+MciXVQgbmwgXRj8ozsnzdak8PklbNoQWy4B9+a d24F0nsRs4ZpSbdqA2/V11XOScJr2XCbZQf+XqtZwovX9oKxZ9hQuLUuJczM+EOYBA0B r2A4iOVjNy6q3qZKTs8Uwe0qKJE0nlP4Ghi0CkROotR41SRTJqDhVjYgAhQmJWsq2z/M MJl5ijyxGrdUAB1/01ugOtihqBx6amNsIj2WRuBJnKOXqjU16brOyyayQvQAF98DLo/s AuzA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=DQ8qY4OyOkFvaykGyjwGKL6pcDR4EejouSObBAhGz24=; b=zrtuMJUN+SyS/QZCqoIbKdYsQENq5ps1XbPS+15TcE5Hnb/sUDraq9Erl02s+GT7Ca urPZOhJEfnIyyh+JRIpx/KTFn3foIb6T6NV5Ee4/5xO6B0m6oSqNkPiRSuI7R3an+EIh /I8ROFqCKf+A/qnkyVxFqHORG+aBRaNiQPB2jgC4DMyd5fAHXNByDrA5oUWLrEll4k6o ptRBTn7dfa5DymotpQZ14r3mFowlmcGsr6wmp6lf54Svy8xpakbU6vpG3rv8yu76J46d KoSbeJA0N3Me/3v7Fp+WtDn+jsLfFGjKoxpwj+N1YERmbFLPKLdlaI7THd4xBuHWFKYv aWaQ== X-Gm-Message-State: AOAM532/fz69N3f/qT+jOX+eiaoABkmMe+x5HYkDYe8111Fexl/ZKc2g PNN3VwKsIt1RR9kniQMKFPZRfw== X-Received: by 2002:a17:90b:4d8a:: with SMTP id oj10mr5693610pjb.226.1642615457370; Wed, 19 Jan 2022 10:04:17 -0800 (PST) Received: from p14s (S0106889e681aac74.cg.shawcable.net. [68.147.0.187]) by smtp.gmail.com with ESMTPSA id d32sm345607pgb.46.2022.01.19.10.04.15 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 19 Jan 2022 10:04:16 -0800 (PST) Date: Wed, 19 Jan 2022 11:04:13 -0700 From: Mathieu Poirier To: Peng Fan Cc: "Peng Fan (OSS)" , "bjorn.andersson@linaro.org" , "shawnguo@kernel.org" , "s.hauer@pengutronix.de" , "kernel@pengutronix.de" , "festevam@gmail.com" , dl-linux-imx , "linux-remoteproc@vger.kernel.org" , "linux-arm-kernel@lists.infradead.org" , "linux-kernel@vger.kernel.org" Subject: Re: [PATCH 6/9] remoteproc: imx_rproc: support attaching to i.MX8QXP M4 Message-ID: <20220119180413.GB1282028@p14s> References: <20220111033333.403448-1-peng.fan@oss.nxp.com> <20220111033333.403448-9-peng.fan@oss.nxp.com> <20220118185729.GH1119324@p14s> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, Jan 19, 2022 at 02:28:31AM +0000, Peng Fan wrote: > > Subject: Re: [PATCH 6/9] remoteproc: imx_rproc: support attaching to > > i.MX8QXP M4 > > > > On Tue, Jan 11, 2022 at 11:33:30AM +0800, Peng Fan (OSS) wrote: > > > From: Peng Fan > > > > > > When M4 is kicked by SCFW, M4 runs in its own hardware partition, > > > Linux could only do IPC with M4, it could not start, stop, update image. > > > > > > When M4 crash reboot, it could notify Linux, so Linux could prepare to > > > reattach to M4 after M4 recovery. > > > > > > Signed-off-by: Peng Fan > > > --- > > > drivers/remoteproc/imx_rproc.c | 96 > > > ++++++++++++++++++++++++++++++++++ > > > 1 file changed, 96 insertions(+) > > > > > > diff --git a/drivers/remoteproc/imx_rproc.c > > > b/drivers/remoteproc/imx_rproc.c index 0e99a3ca6fbc..5f04aea2f6a1 > > > 100644 > > > --- a/drivers/remoteproc/imx_rproc.c > > > +++ b/drivers/remoteproc/imx_rproc.c > > > @@ -6,6 +6,7 @@ > > > #include > > > #include > > > #include > > > +#include > > > #include > > > #include > > > #include > > > @@ -59,6 +60,8 @@ > > > #define IMX_SIP_RPROC_STARTED 0x01 > > > #define IMX_SIP_RPROC_STOP 0x02 > > > > > > +#define IMX_SC_IRQ_GROUP_REBOOTED 5 > > > + > > > /** > > > * struct imx_rproc_mem - slim internal memory structure > > > * @cpu_addr: MPU virtual address of the memory region @@ -90,6 > > > +93,23 @@ struct imx_rproc { > > > struct workqueue_struct *workqueue; > > > void __iomem *rsc_table; > > > bool has_clk; > > > + struct imx_sc_ipc *ipc_handle; > > > + struct notifier_block proc_nb; > > > + u32 rproc_pt; > > > + u32 rsrc; > > > > There is no documentation for the above two fields and I have to guess what > > they do. > > Fix in V2. > > > > > > +}; > > > + > > > +static const struct imx_rproc_att imx_rproc_att_imx8qxp[] = { > > > + /* dev addr , sys addr , size , flags */ > > > + { 0x08000000, 0x08000000, 0x10000000, 0}, > > > + /* TCML/U */ > > > + { 0x1FFE0000, 0x34FE0000, 0x00040000, ATT_OWN | ATT_IOMEM }, > > > + /* OCRAM(Low 96KB) */ > > > + { 0x21000000, 0x00100000, 0x00018000, 0}, > > > + /* OCRAM */ > > > + { 0x21100000, 0x00100000, 0x00040000, 0}, > > > + /* DDR (Data) */ > > > + { 0x80000000, 0x80000000, 0x60000000, 0 }, > > > }; > > > > > > static const struct imx_rproc_att imx_rproc_att_imx8mn[] = { @@ > > > -236,6 +256,12 @@ static const struct imx_rproc_dcfg > > imx_rproc_cfg_imx8ulp = { > > > .method = IMX_RPROC_NONE, > > > }; > > > > > > +static const struct imx_rproc_dcfg imx_rproc_cfg_imx8qxp = { > > > + .att = imx_rproc_att_imx8qxp, > > > + .att_size = ARRAY_SIZE(imx_rproc_att_imx8qxp), > > > + .method = IMX_RPROC_SCU_API, > > > +}; > > > + > > > static const struct imx_rproc_dcfg imx_rproc_cfg_imx7ulp = { > > > .att = imx_rproc_att_imx7ulp, > > > .att_size = ARRAY_SIZE(imx_rproc_att_imx7ulp), > > > @@ -491,6 +517,11 @@ static int imx_rproc_attach(struct rproc *rproc) > > > return 0; > > > } > > > > > > +static int imx_rproc_detach(struct rproc *rproc) { > > > + return 0; > > > > Is it possible to detach the remote processor from the application core? If > > not please write a comment that says so. > > No from my understanding. > > And shouldn't this return some > > kind of error so that users don't think the operation was carried out > > successfully? > > No. This is to match patch 3/9 to support M-core self recovery. After M-core > crash reboot, remoteproc framework just detach and re-attach. Right, but by returning 0 in imx_rproc_detach() issueing something like: # echo detach > /sys/class/remoteproc/remoteproc0/state will return without an error message, leading users to believe the remote processor has been detached. Returning an error code like -EINVAL would tell users the operation was not completed successfully. More comments to come shortly... > > > > > I am out of time for today and as such will continue with this set tomorrow. > > Thanks for you reviewing the patchset. > > Thanks, > Peng. > > > > > Thanks, > > Mathieu > > > > > +} > > > + > > > static struct resource_table *imx_rproc_get_loaded_rsc_table(struct > > > rproc *rproc, size_t *table_sz) { > > > struct imx_rproc *priv = rproc->priv; @@ -525,6 +556,7 @@ > > > imx_rproc_elf_find_loaded_rsc_table(struct rproc *rproc, const struct > > > firmware * static const struct rproc_ops imx_rproc_ops = { > > > .prepare = imx_rproc_prepare, > > > .attach = imx_rproc_attach, > > > + .detach = imx_rproc_detach, > > > .start = imx_rproc_start, > > > .stop = imx_rproc_stop, > > > .kick = imx_rproc_kick, > > > @@ -671,6 +703,22 @@ static void imx_rproc_free_mbox(struct rproc > > *rproc) > > > mbox_free_channel(priv->rx_ch); > > > } > > > > > > +static int imx_rproc_partition_notify(struct notifier_block *nb, > > > + unsigned long event, void *group) { > > > + struct imx_rproc *priv = container_of(nb, struct imx_rproc, > > > +proc_nb); > > > + > > > + /* Ignore other irqs */ > > > + if (!((event & BIT(priv->rproc_pt)) && (*(u8 *)group == > > IMX_SC_IRQ_GROUP_REBOOTED))) > > > + return 0; > > > + > > > + rproc_report_crash(priv->rproc, RPROC_WATCHDOG); > > > + > > > + pr_info("Patition%d reset!\n", priv->rproc_pt); > > > + > > > + return 0; > > > +} > > > + > > > static int imx_rproc_detect_mode(struct imx_rproc *priv) { > > > struct regmap_config config = { .name = "imx-rproc" }; @@ -680,6 > > > +728,7 @@ static int imx_rproc_detect_mode(struct imx_rproc *priv) > > > struct arm_smccc_res res; > > > int ret; > > > u32 val; > > > + u8 pt; > > > > > > switch (dcfg->method) { > > > case IMX_RPROC_NONE: > > > @@ -690,6 +739,52 @@ static int imx_rproc_detect_mode(struct > > imx_rproc *priv) > > > if (res.a0) > > > priv->rproc->state = RPROC_DETACHED; > > > return 0; > > > + case IMX_RPROC_SCU_API: > > > + ret = imx_scu_get_handle(&priv->ipc_handle); > > > + if (ret) > > > + return ret; > > > + ret = of_property_read_u32(dev->of_node, "rsrc-id", &priv->rsrc); > > > + if (ret) { > > > + dev_err(dev, "no rsrc-id\n"); > > > + return ret; > > > + } > > > + > > > + /* > > > + * If Mcore resource is not owned by Acore partition, It is kicked by > > ROM, > > > + * and Linux could only do IPC with Mcore and nothing else. > > > + */ > > > + if (!imx_sc_rm_is_resource_owned(priv->ipc_handle, priv->rsrc)) { > > > + > > > + priv->has_clk = false; > > > + priv->rproc->self_recovery = true; > > > + priv->rproc->state = RPROC_DETACHED; > > > + > > > + /* Get partition id and enable irq in SCFW */ > > > + ret = imx_sc_rm_get_resource_owner(priv->ipc_handle, > > priv->rsrc, &pt); > > > + if (ret) { > > > + dev_err(dev, "not able to get resource owner\n"); > > > + return ret; > > > + } > > > + > > > + priv->rproc_pt = pt; > > > + priv->proc_nb.notifier_call = imx_rproc_partition_notify; > > > + > > > + ret = imx_scu_irq_register_notifier(&priv->proc_nb); > > > + if (ret) { > > > + dev_warn(dev, "register scu notifier failed.\n"); > > > + return ret; > > > + } > > > + > > > + ret = > > imx_scu_irq_group_enable(IMX_SC_IRQ_GROUP_REBOOTED, > > > + BIT(priv->rproc_pt), true); > > > + if (ret) { > > > + imx_scu_irq_unregister_notifier(&priv->proc_nb); > > > + dev_warn(dev, "Enable irq failed.\n"); > > > + return ret; > > > + } > > > + } > > > + > > > + return 0; > > > default: > > > break; > > > } > > > @@ -847,6 +942,7 @@ static const struct of_device_id > > imx_rproc_of_match[] = { > > > { .compatible = "fsl,imx8mm-cm4", .data = &imx_rproc_cfg_imx8mq }, > > > { .compatible = "fsl,imx8mn-cm7", .data = &imx_rproc_cfg_imx8mn }, > > > { .compatible = "fsl,imx8mp-cm7", .data = &imx_rproc_cfg_imx8mn }, > > > + { .compatible = "fsl,imx8qxp-cm4", .data = &imx_rproc_cfg_imx8qxp }, > > > { .compatible = "fsl,imx8ulp-cm33", .data = &imx_rproc_cfg_imx8ulp }, > > > {}, > > > }; > > > -- > > > 2.25.1 > > >