Received: by 2002:a25:683:0:0:0:0:0 with SMTP id 125csp4560688ybg; Mon, 8 Jun 2020 10:54:46 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxB66LD+2O1Vr9rdA1yFGznae2IZl8mpqAZU9Y10TvA+V3RL/hPGG4gbFzFv9Hq6cKLEx4k X-Received: by 2002:a05:6402:1247:: with SMTP id l7mr21932382edw.61.1591638886290; Mon, 08 Jun 2020 10:54:46 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1591638886; cv=none; d=google.com; s=arc-20160816; b=H0+SWEYF0ksrfVFPL5oDfxzicm/oGa8HAS607avKWwHrwKLEOrvpO7dwrbXZWbfC/l 8D2Aok7UWPCWZiQoteFkQRN8yCOt2A2QbTXy2qFK1lCRytnc0IAas41ngvahHtDGGQVI x96ToUJ3dslIqfSTRL3k61Yo4da8czwue50gICN+voWMTiV4/ckBnpvC4dftR+qALlcJ tqn/ZfgM3GLEhQl0CQis/LG8vlyflKYGfIMVtU01tUcyA8bsMAEqu/fpyuxPKKjZ7cbf t6fH/uKxyrOMRGlz7R3EaivndKdAQ6easloOY6fnZgJoeRvyZiL7LZIZoTd/UMQnepam dpAA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:mime-version:message-id:date:references :in-reply-to:subject:cc:to:from; bh=lUN/1YRK42XYdEpAZ5Ey8+TdhnMosets714rBqsuyYs=; b=ZkXZ3qYc6mDFqU1uRKDS3AMlowLzP6CzJCRWTmewFnZaZYzwjk0x9cy97FRog66C5F rdiJCjzdRLxQo2kp9r6QNOck4RER+LflhCjccp8F7JAic3LSE+61K7tSioYthlliFBCW q1ltWCLkJRiFefFesBfVxqSDqwPjR2HrJug9tUYFJuyB7LOu4F4cYG//Q+xZt2RndduO Yp7OXIETv9OyqfPbVfWO/BFHSUAY1UVxGWy7zz3l6T/jWL6mj3+3GJKs7rtTboSBxh24 R5VQYXyeBIlYICHQOFC9OSictJR4gbyuiTGIaca7EK4kLCx/IWM+AAt2Yk7DWWjlTzjV 57rA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=ibm.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id l2si9879146edf.333.2020.06.08.10.54.23; Mon, 08 Jun 2020 10:54:46 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=ibm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1730084AbgFHRs2 (ORCPT + 99 others); Mon, 8 Jun 2020 13:48:28 -0400 Received: from mx0b-001b2d01.pphosted.com ([148.163.158.5]:22834 "EHLO mx0a-001b2d01.pphosted.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1726097AbgFHRs2 (ORCPT ); Mon, 8 Jun 2020 13:48:28 -0400 Received: from pps.filterd (m0098413.ppops.net [127.0.0.1]) by mx0b-001b2d01.pphosted.com (8.16.0.42/8.16.0.42) with SMTP id 058HilPd110622; Mon, 8 Jun 2020 13:47:46 -0400 Received: from pps.reinject (localhost [127.0.0.1]) by mx0b-001b2d01.pphosted.com with ESMTP id 31g59rmccj-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Mon, 08 Jun 2020 13:47:46 -0400 Received: from m0098413.ppops.net (m0098413.ppops.net [127.0.0.1]) by pps.reinject (8.16.0.36/8.16.0.36) with SMTP id 058HikbA110537; Mon, 8 Jun 2020 13:47:45 -0400 Received: from ppma01fra.de.ibm.com (46.49.7a9f.ip4.static.sl-reverse.com [159.122.73.70]) by mx0b-001b2d01.pphosted.com with ESMTP id 31g59rmcbv-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Mon, 08 Jun 2020 13:47:45 -0400 Received: from pps.filterd (ppma01fra.de.ibm.com [127.0.0.1]) by ppma01fra.de.ibm.com (8.16.0.42/8.16.0.42) with SMTP id 058HfM6T001927; Mon, 8 Jun 2020 17:47:44 GMT Received: from b06cxnps3075.portsmouth.uk.ibm.com (d06relay10.portsmouth.uk.ibm.com [9.149.109.195]) by ppma01fra.de.ibm.com with ESMTP id 31g2s7st5s-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Mon, 08 Jun 2020 17:47:43 +0000 Received: from b06wcsmtp001.portsmouth.uk.ibm.com (b06wcsmtp001.portsmouth.uk.ibm.com [9.149.105.160]) by b06cxnps3075.portsmouth.uk.ibm.com (8.14.9/8.14.9/NCO v10.0) with ESMTP id 058HlfZI49938666 (version=TLSv1/SSLv3 cipher=DHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Mon, 8 Jun 2020 17:47:41 GMT Received: from b06wcsmtp001.portsmouth.uk.ibm.com (unknown [127.0.0.1]) by IMSVA (Postfix) with ESMTP id 3B6E6A4067; Mon, 8 Jun 2020 17:47:41 +0000 (GMT) Received: from b06wcsmtp001.portsmouth.uk.ibm.com (unknown [127.0.0.1]) by IMSVA (Postfix) with ESMTP id 4F256A4062; Mon, 8 Jun 2020 17:47:37 +0000 (GMT) Received: from vajain21-in-ibm-com (unknown [9.85.74.153]) by b06wcsmtp001.portsmouth.uk.ibm.com (Postfix) with SMTP; Mon, 8 Jun 2020 17:47:36 +0000 (GMT) Received: by vajain21-in-ibm-com (sSMTP sendmail emulation); Mon, 08 Jun 2020 23:17:35 +0530 From: Vaibhav Jain To: linuxppc-dev@lists.ozlabs.org, linux-nvdimm@lists.01.org, linux-kernel@vger.kernel.org, Ira Weiny Cc: Dan Williams , "Aneesh Kumar K . V" , Michael Ellerman , "Oliver O'Halloran" , Santosh Sivaraj , Steven Rostedt Subject: Re: [PATCH v11 3/6] powerpc/papr_scm: Fetch nvdimm health information from PHYP In-Reply-To: <20200607131339.476036-4-vaibhav@linux.ibm.com> References: <20200607131339.476036-1-vaibhav@linux.ibm.com> <20200607131339.476036-4-vaibhav@linux.ibm.com> Date: Mon, 08 Jun 2020 23:17:35 +0530 Message-ID: <87wo4hwitk.fsf@linux.ibm.com> MIME-Version: 1.0 Content-Type: text/plain X-TM-AS-GCONF: 00 X-Proofpoint-Virus-Version: vendor=fsecure engine=2.50.10434:6.0.216,18.0.687 definitions=2020-06-08_17:2020-06-08,2020-06-08 signatures=0 X-Proofpoint-Spam-Details: rule=outbound_notspam policy=outbound score=0 lowpriorityscore=0 mlxscore=0 impostorscore=0 clxscore=1015 adultscore=0 spamscore=0 phishscore=0 malwarescore=0 mlxlogscore=999 cotscore=-2147483648 bulkscore=0 suspectscore=0 priorityscore=1501 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2004280000 definitions=main-2006080121 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Ira, During v9 you had provided your ack to this patch [1] and also had made a review comment in a later patch regarding an avoidable 'goto' statement. I have since updated the patch addressing that review comment. Can you please provide your ack to this patch too. [1] https://lore.kernel.org/linux-nvdimm/20200603231814.GK1505637@iweiny-DESK2.sc.intel.com/T/#m668d7b35a2394104f11afdae5951e420a8ccffe6 [2] "I missed this... probably did not need the goto in the first patch?" https://lore.kernel.org/linux-nvdimm/20200603231814.GK1505637@iweiny-DESK2.sc.intel.com/T/#m1ebdd309ac0cb6f47d3b574b8d05374b21ff75df Thanks, ~ Vaibhav Vaibhav Jain writes: > Implement support for fetching nvdimm health information via > H_SCM_HEALTH hcall as documented in Ref[1]. The hcall returns a pair > of 64-bit bitmap, bitwise-and of which is then stored in > 'struct papr_scm_priv' and subsequently partially exposed to > user-space via newly introduced dimm specific attribute > 'papr/flags'. Since the hcall is costly, the health information is > cached and only re-queried, 60s after the previous successful hcall. > > The patch also adds a documentation text describing flags reported by > the the new sysfs attribute 'papr/flags' is also introduced at > Documentation/ABI/testing/sysfs-bus-papr-pmem. > > [1] commit 58b278f568f0 ("powerpc: Provide initial documentation for > PAPR hcalls") > > Cc: "Aneesh Kumar K . V" > Cc: Dan Williams > Cc: Michael Ellerman > Cc: Ira Weiny > Signed-off-by: Vaibhav Jain > --- > Changelog: > > v10..v11: > * None > > v9..v10: > * Removed an avoidable 'goto' in __drc_pmem_query_health. [ Ira ]. > > Resend: > * Added ack from Aneesh. > > v8..v9: > * Rename some variables and defines to reduce usage of term SCM > replacing it with PMEM [Dan Williams, Aneesh] > * s/PAPR_SCM_DIMM/PAPR_PMEM/g > * s/papr_scm_nd_attributes/papr_nd_attributes/g > * s/papr_scm_nd_attribute_group/papr_nd_attribute_group/g > * s/papr_scm_dimm_attr_groups/papr_nd_attribute_groups/g > * Renamed file sysfs-bus-papr-scm to sysfs-bus-papr-pmem > > v7..v8: > * Update type of variable 'rc' in __drc_pmem_query_health() and > drc_pmem_query_health() to long and int respectively. [ Ira ] > * Updated the patch description to s/64 bit Big Endian Number/64-bit > bitmap/ [ Ira, Aneesh ]. > > Resend: > * None > > v6..v7 : > * Used the exported buf_seq_printf() function to generate content for > 'papr/flags' > * Moved the PAPR_SCM_DIMM_* bit-flags macro definitions to papr_scm.c > and removed the papr_scm.h file [Mpe] > * Some minor consistency issued in sysfs-bus-papr-scm > documentation. [Mpe] > * s/dimm_mutex/health_mutex/g [Mpe] > * Split drc_pmem_query_health() into two function one of which takes > care of caching and locking. [Mpe] > * Fixed a local copy creation of dimm health information using > READ_ONCE(). [Mpe] > > v5..v6 : > * Change the flags sysfs attribute from 'papr_flags' to 'papr/flags' > [Dan Williams] > * Include documentation for 'papr/flags' attr [Dan Williams] > * Change flag 'save_fail' to 'flush_fail' [Dan Williams] > * Caching of health bitmap to reduce expensive hcalls [Dan Williams] > * Removed usage of PPC_BIT from 'papr-scm.h' header [Mpe] > * Replaced two __be64 integers from papr_scm_priv to a single u64 > integer [Mpe] > * Updated patch description to reflect the changes made in this > version. > * Removed avoidable usage of 'papr_scm_priv.dimm_mutex' from > flags_show() [Dan Williams] > > v4..v5 : > * None > > v3..v4 : > * None > > v2..v3 : > * Removed PAPR_SCM_DIMM_HEALTH_NON_CRITICAL as a condition for > NVDIMM unarmed [Aneesh] > > v1..v2 : > * New patch in the series. > --- > Documentation/ABI/testing/sysfs-bus-papr-pmem | 27 +++ > arch/powerpc/platforms/pseries/papr_scm.c | 168 +++++++++++++++++- > 2 files changed, 193 insertions(+), 2 deletions(-) > create mode 100644 Documentation/ABI/testing/sysfs-bus-papr-pmem > > diff --git a/Documentation/ABI/testing/sysfs-bus-papr-pmem b/Documentation/ABI/testing/sysfs-bus-papr-pmem > new file mode 100644 > index 000000000000..5b10d036a8d4 > --- /dev/null > +++ b/Documentation/ABI/testing/sysfs-bus-papr-pmem > @@ -0,0 +1,27 @@ > +What: /sys/bus/nd/devices/nmemX/papr/flags > +Date: Apr, 2020 > +KernelVersion: v5.8 > +Contact: linuxppc-dev , linux-nvdimm@lists.01.org, > +Description: > + (RO) Report flags indicating various states of a > + papr-pmem NVDIMM device. Each flag maps to a one or > + more bits set in the dimm-health-bitmap retrieved in > + response to H_SCM_HEALTH hcall. The details of the bit > + flags returned in response to this hcall is available > + at 'Documentation/powerpc/papr_hcalls.rst' . Below are > + the flags reported in this sysfs file: > + > + * "not_armed" : Indicates that NVDIMM contents will not > + survive a power cycle. > + * "flush_fail" : Indicates that NVDIMM contents > + couldn't be flushed during last > + shut-down event. > + * "restore_fail": Indicates that NVDIMM contents > + couldn't be restored during NVDIMM > + initialization. > + * "encrypted" : NVDIMM contents are encrypted. > + * "smart_notify": There is health event for the NVDIMM. > + * "scrubbed" : Indicating that contents of the > + NVDIMM have been scrubbed. > + * "locked" : Indicating that NVDIMM contents cant > + be modified until next power cycle. > diff --git a/arch/powerpc/platforms/pseries/papr_scm.c b/arch/powerpc/platforms/pseries/papr_scm.c > index f35592423380..0c091622b15e 100644 > --- a/arch/powerpc/platforms/pseries/papr_scm.c > +++ b/arch/powerpc/platforms/pseries/papr_scm.c > @@ -12,6 +12,7 @@ > #include > #include > #include > +#include > > #include > > @@ -22,6 +23,44 @@ > (1ul << ND_CMD_GET_CONFIG_DATA) | \ > (1ul << ND_CMD_SET_CONFIG_DATA)) > > +/* DIMM health bitmap bitmap indicators */ > +/* SCM device is unable to persist memory contents */ > +#define PAPR_PMEM_UNARMED (1ULL << (63 - 0)) > +/* SCM device failed to persist memory contents */ > +#define PAPR_PMEM_SHUTDOWN_DIRTY (1ULL << (63 - 1)) > +/* SCM device contents are persisted from previous IPL */ > +#define PAPR_PMEM_SHUTDOWN_CLEAN (1ULL << (63 - 2)) > +/* SCM device contents are not persisted from previous IPL */ > +#define PAPR_PMEM_EMPTY (1ULL << (63 - 3)) > +/* SCM device memory life remaining is critically low */ > +#define PAPR_PMEM_HEALTH_CRITICAL (1ULL << (63 - 4)) > +/* SCM device will be garded off next IPL due to failure */ > +#define PAPR_PMEM_HEALTH_FATAL (1ULL << (63 - 5)) > +/* SCM contents cannot persist due to current platform health status */ > +#define PAPR_PMEM_HEALTH_UNHEALTHY (1ULL << (63 - 6)) > +/* SCM device is unable to persist memory contents in certain conditions */ > +#define PAPR_PMEM_HEALTH_NON_CRITICAL (1ULL << (63 - 7)) > +/* SCM device is encrypted */ > +#define PAPR_PMEM_ENCRYPTED (1ULL << (63 - 8)) > +/* SCM device has been scrubbed and locked */ > +#define PAPR_PMEM_SCRUBBED_AND_LOCKED (1ULL << (63 - 9)) > + > +/* Bits status indicators for health bitmap indicating unarmed dimm */ > +#define PAPR_PMEM_UNARMED_MASK (PAPR_PMEM_UNARMED | \ > + PAPR_PMEM_HEALTH_UNHEALTHY) > + > +/* Bits status indicators for health bitmap indicating unflushed dimm */ > +#define PAPR_PMEM_BAD_SHUTDOWN_MASK (PAPR_PMEM_SHUTDOWN_DIRTY) > + > +/* Bits status indicators for health bitmap indicating unrestored dimm */ > +#define PAPR_PMEM_BAD_RESTORE_MASK (PAPR_PMEM_EMPTY) > + > +/* Bit status indicators for smart event notification */ > +#define PAPR_PMEM_SMART_EVENT_MASK (PAPR_PMEM_HEALTH_CRITICAL | \ > + PAPR_PMEM_HEALTH_FATAL | \ > + PAPR_PMEM_HEALTH_UNHEALTHY) > + > +/* private struct associated with each region */ > struct papr_scm_priv { > struct platform_device *pdev; > struct device_node *dn; > @@ -39,6 +78,15 @@ struct papr_scm_priv { > struct resource res; > struct nd_region *region; > struct nd_interleave_set nd_set; > + > + /* Protect dimm health data from concurrent read/writes */ > + struct mutex health_mutex; > + > + /* Last time the health information of the dimm was updated */ > + unsigned long lasthealth_jiffies; > + > + /* Health information for the dimm */ > + u64 health_bitmap; > }; > > static int drc_pmem_bind(struct papr_scm_priv *p) > @@ -144,6 +192,61 @@ static int drc_pmem_query_n_bind(struct papr_scm_priv *p) > return drc_pmem_bind(p); > } > > +/* > + * Issue hcall to retrieve dimm health info and populate papr_scm_priv with the > + * health information. > + */ > +static int __drc_pmem_query_health(struct papr_scm_priv *p) > +{ > + unsigned long ret[PLPAR_HCALL_BUFSIZE]; > + long rc; > + > + /* issue the hcall */ > + rc = plpar_hcall(H_SCM_HEALTH, ret, p->drc_index); > + if (rc != H_SUCCESS) { > + dev_err(&p->pdev->dev, > + "Failed to query health information, Err:%ld\n", rc); > + return -ENXIO; > + } > + > + p->lasthealth_jiffies = jiffies; > + p->health_bitmap = ret[0] & ret[1]; > + > + dev_dbg(&p->pdev->dev, > + "Queried dimm health info. Bitmap:0x%016lx Mask:0x%016lx\n", > + ret[0], ret[1]); > + > + return 0; > +} > + > +/* Min interval in seconds for assuming stable dimm health */ > +#define MIN_HEALTH_QUERY_INTERVAL 60 > + > +/* Query cached health info and if needed call drc_pmem_query_health */ > +static int drc_pmem_query_health(struct papr_scm_priv *p) > +{ > + unsigned long cache_timeout; > + int rc; > + > + /* Protect concurrent modifications to papr_scm_priv */ > + rc = mutex_lock_interruptible(&p->health_mutex); > + if (rc) > + return rc; > + > + /* Jiffies offset for which the health data is assumed to be same */ > + cache_timeout = p->lasthealth_jiffies + > + msecs_to_jiffies(MIN_HEALTH_QUERY_INTERVAL * 1000); > + > + /* Fetch new health info is its older than MIN_HEALTH_QUERY_INTERVAL */ > + if (time_after(jiffies, cache_timeout)) > + rc = __drc_pmem_query_health(p); > + else > + /* Assume cached health data is valid */ > + rc = 0; > + > + mutex_unlock(&p->health_mutex); > + return rc; > +} > > static int papr_scm_meta_get(struct papr_scm_priv *p, > struct nd_cmd_get_config_data_hdr *hdr) > @@ -286,6 +389,64 @@ static int papr_scm_ndctl(struct nvdimm_bus_descriptor *nd_desc, > return 0; > } > > +static ssize_t flags_show(struct device *dev, > + struct device_attribute *attr, char *buf) > +{ > + struct nvdimm *dimm = to_nvdimm(dev); > + struct papr_scm_priv *p = nvdimm_provider_data(dimm); > + struct seq_buf s; > + u64 health; > + int rc; > + > + rc = drc_pmem_query_health(p); > + if (rc) > + return rc; > + > + /* Copy health_bitmap locally, check masks & update out buffer */ > + health = READ_ONCE(p->health_bitmap); > + > + seq_buf_init(&s, buf, PAGE_SIZE); > + if (health & PAPR_PMEM_UNARMED_MASK) > + seq_buf_printf(&s, "not_armed "); > + > + if (health & PAPR_PMEM_BAD_SHUTDOWN_MASK) > + seq_buf_printf(&s, "flush_fail "); > + > + if (health & PAPR_PMEM_BAD_RESTORE_MASK) > + seq_buf_printf(&s, "restore_fail "); > + > + if (health & PAPR_PMEM_ENCRYPTED) > + seq_buf_printf(&s, "encrypted "); > + > + if (health & PAPR_PMEM_SMART_EVENT_MASK) > + seq_buf_printf(&s, "smart_notify "); > + > + if (health & PAPR_PMEM_SCRUBBED_AND_LOCKED) > + seq_buf_printf(&s, "scrubbed locked "); > + > + if (seq_buf_used(&s)) > + seq_buf_printf(&s, "\n"); > + > + return seq_buf_used(&s); > +} > +DEVICE_ATTR_RO(flags); > + > +/* papr_scm specific dimm attributes */ > +static struct attribute *papr_nd_attributes[] = { > + &dev_attr_flags.attr, > + NULL, > +}; > + > +static struct attribute_group papr_nd_attribute_group = { > + .name = "papr", > + .attrs = papr_nd_attributes, > +}; > + > +static const struct attribute_group *papr_nd_attr_groups[] = { > + &papr_nd_attribute_group, > + NULL, > +}; > + > static int papr_scm_nvdimm_init(struct papr_scm_priv *p) > { > struct device *dev = &p->pdev->dev; > @@ -312,8 +473,8 @@ static int papr_scm_nvdimm_init(struct papr_scm_priv *p) > dimm_flags = 0; > set_bit(NDD_LABELING, &dimm_flags); > > - p->nvdimm = nvdimm_create(p->bus, p, NULL, dimm_flags, > - PAPR_SCM_DIMM_CMD_MASK, 0, NULL); > + p->nvdimm = nvdimm_create(p->bus, p, papr_nd_attr_groups, > + dimm_flags, PAPR_SCM_DIMM_CMD_MASK, 0, NULL); > if (!p->nvdimm) { > dev_err(dev, "Error creating DIMM object for %pOF\n", p->dn); > goto err; > @@ -399,6 +560,9 @@ static int papr_scm_probe(struct platform_device *pdev) > if (!p) > return -ENOMEM; > > + /* Initialize the dimm mutex */ > + mutex_init(&p->health_mutex); > + > /* optional DT properties */ > of_property_read_u32(dn, "ibm,metadata-size", &metadata_size); > > -- > 2.26.2 > -- Cheers ~ Vaibhav