Received: by 2002:a05:7412:40d:b0:e2:908c:2ebd with SMTP id 13csp1089600rdf; Wed, 22 Nov 2023 05:30:09 -0800 (PST) X-Google-Smtp-Source: AGHT+IH1NoXMnUzE9uYFycYUKhHNFJOWbtWC+drNpuWhiYYWG2Vi9KWhgZKTLYxzeULhXuFQRdjO X-Received: by 2002:aca:90a:0:b0:3b6:4e53:e7ac with SMTP id 10-20020aca090a000000b003b64e53e7acmr2139605oij.56.1700659808966; Wed, 22 Nov 2023 05:30:08 -0800 (PST) ARC-Seal: i=2; a=rsa-sha256; t=1700659808; cv=pass; d=google.com; s=arc-20160816; b=HBK9DOr3tXOZBXqUyMByDPakwKrgCk5gZ9laNBb5eZK0n+UkH5d64MzZKwXDVWg+GQ 8l1fAvkUjBTuzJMlzq138JpJwt6c3IrBfDKHu6QG8D2miDJVmkSbcUoMC+4WWw6Etyv0 nNPxhGxAg9sBeE2aNg3gzKh6mR4M/aU5rU/sgO0gEPAhyJiAWauDISUmkphCa1+XsQzZ /gwNhRPuOzzX+qAnWxZmTR5nQsfVTwBXZcbZE7wcyx4m5rQ8IHkzFJPWnwOl0x5m0sw1 yPZREZKL0Jb81xGTc4Ek0jnXLCNl1SVS0gmosC15uPHQoUB01V2Cnrv+hfzK1vq5IVKa uc9g== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:user-agent:in-reply-to:content-disposition :mime-version:references:mail-followup-to:message-id:subject:cc:to :from:date:dkim-signature; bh=D/f7S2Biejbu76wbA2JL6WsfsOKW1/+8APoduISDCV8=; fh=MdAOXgR7LTF8cdQqD9Fd7aj+8Geara3jgDEczizEc2w=; b=cUuMVTZyawTDYpXV4ql9M9sA7sPVz41z0xSj2jiFHtYuaNq9VVonbuYuNJC4u+BLuQ rKQGi3wPgur8IBYxfNZ7bR1Pdfnxzr3NsTj/7oEtQSjOCwfBiw8+2hor1anvBaeYwpC1 xI9UHq+8SSk6tGS2QIjQXaOw94SJZrOLR/OLZAgtt+XXCeh0gl4TSPXUxi9dcErky2q/ l+seckaT4tcfl1bcgadM8wVsb09M3nevu0cpOmI1SUqa6K1UkviaIKW0p2blXN9FCjSs y9+219AsF0if9Ap8nlyOsw7DgwdtbvIvB/+/AarUosaZ+55uf2N7W1YqF+wrYzRLYHaw z7Qw== ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@stgolabs.net header.s=dreamhost header.b=VLL7TTFD; arc=pass (i=1); spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.33 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from lipwig.vger.email (lipwig.vger.email. [23.128.96.33]) by mx.google.com with ESMTPS id o18-20020a63e352000000b005bf77518dc6si12500988pgj.286.2023.11.22.05.30.08 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 22 Nov 2023 05:30:08 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.33 as permitted sender) client-ip=23.128.96.33; Authentication-Results: mx.google.com; dkim=pass header.i=@stgolabs.net header.s=dreamhost header.b=VLL7TTFD; arc=pass (i=1); spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.33 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by lipwig.vger.email (Postfix) with ESMTP id 6C1028183EC3; Tue, 21 Nov 2023 16:22:46 -0800 (PST) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.11 at lipwig.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S234955AbjKVAVz (ORCPT + 99 others); Tue, 21 Nov 2023 19:21:55 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:49310 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234949AbjKVAVx (ORCPT ); Tue, 21 Nov 2023 19:21:53 -0500 Received: from bee.birch.relay.mailchannels.net (bee.birch.relay.mailchannels.net [23.83.209.14]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 866D59D; Tue, 21 Nov 2023 16:21:48 -0800 (PST) X-Sender-Id: dreamhost|x-authsender|dave@stgolabs.net Received: from relay.mailchannels.net (localhost [127.0.0.1]) by relay.mailchannels.net (Postfix) with ESMTP id DF4A72C2B8B; Wed, 22 Nov 2023 00:21:47 +0000 (UTC) Received: from pdx1-sub0-mail-a211.dreamhost.com (unknown [127.0.0.6]) (Authenticated sender: dreamhost) by relay.mailchannels.net (Postfix) with ESMTPA id 3F5172C2A57; Wed, 22 Nov 2023 00:21:47 +0000 (UTC) ARC-Seal: i=1; s=arc-2022; d=mailchannels.net; t=1700612507; a=rsa-sha256; cv=none; b=0Y7OXV7ufO92usIQMaBz5sYMlsc3zCUKoexPFu8wAgCwZeG6vciBVA29lJ4LmWKuyj8Ug5 HyIc+RizCJVwHoFIrOPidAPBl2xHBhvcuasG/yPaNpRLyWoPOsLK9b79nZwcuacd7iDwiw d0TnL1Xjk1hei26xq6xl1Rr+8fPztiKdjIyLXgeDNjGK3O2XjZ2nOL0EdVAIpMz7BFrv6X f97XBCIkX6Ym/4P7EikaKIkPVeQl9gP94YMUpLYcStVH+VYTdM9cD2eZhImKvjai/wQMN3 oSqt/bfAgQA8U/86wZkUFRrhohF4dfsgzPqlw7uaWzaM6b3HRGnkObnc7u+HDw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=mailchannels.net; s=arc-2022; t=1700612507; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references:dkim-signature; bh=D/f7S2Biejbu76wbA2JL6WsfsOKW1/+8APoduISDCV8=; b=r/5cvNG+v+boT7oKZNTA8YlMemTtt4Z3UQcgNe8+Yuc2IqRuEnohwKOec6j4P6BtZwmmpY fuObaSzOHy8iRo+zl+TfC6brV42HibBMP2mdBn80CcReAvzVJw1YJRRKwQpnHREQMW9rIb MXsaSAPyI8cr5ARuuZSlguAmjwLxBGNZ8+XkLubWA7rVwEIbnKTZcfm3DwYt0svB9Ehh+f 43OZbH33uzRtHRmLLLl4zMoAjepSiRk7hjLEjrWu+ssYMT3Hjea2jkBjzlOCaxgVEQebrk M6wZc4U6ojOE0L3FkPQnHNAXx0Tpv1MnKUMGqIDzrBOEtpnFPvnv0J+TatLtiw== ARC-Authentication-Results: i=1; rspamd-5f5ccc4f7-vhtb8; auth=pass smtp.auth=dreamhost smtp.mailfrom=dave@stgolabs.net X-Sender-Id: dreamhost|x-authsender|dave@stgolabs.net X-MC-Relay: Neutral X-MailChannels-SenderId: dreamhost|x-authsender|dave@stgolabs.net X-MailChannels-Auth-Id: dreamhost X-Decisive-Occur: 25c0e4b066611bc5_1700612507746_541272267 X-MC-Loop-Signature: 1700612507746:2811177988 X-MC-Ingress-Time: 1700612507746 Received: from pdx1-sub0-mail-a211.dreamhost.com (pop.dreamhost.com [64.90.62.162]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384) by 100.110.238.199 (trex/6.9.2); Wed, 22 Nov 2023 00:21:47 +0000 Received: from offworld (unknown [104.36.30.214]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange ECDHE (P-256) server-signature RSA-PSS (2048 bits) server-digest SHA256) (No client certificate requested) (Authenticated sender: dave@stgolabs.net) by pdx1-sub0-mail-a211.dreamhost.com (Postfix) with ESMTPSA id 4SZhkc4gdDz64; Tue, 21 Nov 2023 16:21:44 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=stgolabs.net; s=dreamhost; t=1700612507; bh=D/f7S2Biejbu76wbA2JL6WsfsOKW1/+8APoduISDCV8=; h=Date:From:To:Cc:Subject:Content-Type; b=VLL7TTFDTijyxxUPESw0u9NjkNP+e3WTLntH0+VYxD2SM7WUQ2A6u4QGx+bvdUKC0 rdm6JcSNKLXcL+9I0WwPtgj8pVozQPv34FwgVLX69rzb7hhacvQZuGMqJFdoz9pWPL SYyaOILq0TYLkXfi5MQfz2CzGoMVH9QrZi4QiSGxHQEazvAmhDH7y+bonRxRvkNpxV tjUe87DfxZ/lA0+W+SVlVrW4k8LsioDHYEiParV0hsuljJOL5zROFrQ7/3DGbKyexC DQfnmKjKjzwNo0Mirp2E9zL7+oRBgXPkCcojmJEbkS1bey9SFR80IlR8NGLWnpDmle GaYH5PSh5s2mQ== Date: Tue, 21 Nov 2023 16:21:40 -0800 From: Davidlohr Bueso To: shiju.jose@huawei.com Cc: linux-cxl@vger.kernel.org, linux-mm@kvack.org, jonathan.cameron@huawei.com, dave.jiang@intel.com, alison.schofield@intel.com, vishal.l.verma@intel.com, ira.weiny@intel.com, dan.j.williams@intel.com, linux-acpi@vger.kernel.org, linux-kernel@vger.kernel.org, david@redhat.com, Vilas.Sridharan@amd.com, leo.duran@amd.com, Yazen.Ghannam@amd.com, rientjes@google.com, jiaqiyan@google.com, tony.luck@intel.com, Jon.Grimm@amd.com, dave.hansen@linux.intel.com, rafael@kernel.org, lenb@kernel.org, naoya.horiguchi@nec.com, james.morse@arm.com, jthoughton@google.com, somasundaram.a@hpe.com, erdemaktas@google.com, pgonda@google.com, duenwen@google.com, mike.malvestuto@intel.com, gthelen@google.com, wschwartz@amperecomputing.com, dferguson@amperecomputing.com, tanxiaofei@huawei.com, prime.zeng@hisilicon.com, kangkang.shen@futurewei.com, wanghuiqiang@huawei.com, linuxarm@huawei.com, fan.ni@samsung.com, a.manzanares@samsung.com Subject: Re: [PATCH v2 04/10] cxl/memscrub: Add CXL device patrol scrub control feature Message-ID: Mail-Followup-To: shiju.jose@huawei.com, linux-cxl@vger.kernel.org, linux-mm@kvack.org, jonathan.cameron@huawei.com, dave.jiang@intel.com, alison.schofield@intel.com, vishal.l.verma@intel.com, ira.weiny@intel.com, dan.j.williams@intel.com, linux-acpi@vger.kernel.org, linux-kernel@vger.kernel.org, david@redhat.com, Vilas.Sridharan@amd.com, leo.duran@amd.com, Yazen.Ghannam@amd.com, rientjes@google.com, jiaqiyan@google.com, tony.luck@intel.com, Jon.Grimm@amd.com, dave.hansen@linux.intel.com, rafael@kernel.org, lenb@kernel.org, naoya.horiguchi@nec.com, james.morse@arm.com, jthoughton@google.com, somasundaram.a@hpe.com, erdemaktas@google.com, pgonda@google.com, duenwen@google.com, mike.malvestuto@intel.com, gthelen@google.com, wschwartz@amperecomputing.com, dferguson@amperecomputing.com, tanxiaofei@huawei.com, prime.zeng@hisilicon.com, kangkang.shen@futurewei.com, wanghuiqiang@huawei.com, linuxarm@huawei.com, fan.ni@samsung.com, a.manzanares@samsung.com References: <20231121101844.1161-1-shiju.jose@huawei.com> <20231121101844.1161-5-shiju.jose@huawei.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii; format=flowed Content-Disposition: inline In-Reply-To: <20231121101844.1161-5-shiju.jose@huawei.com> User-Agent: NeoMutt/20231006 X-Spam-Status: No, score=-0.9 required=5.0 tests=DKIM_SIGNED,DKIM_VALID, DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lipwig.vger.email Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (lipwig.vger.email [0.0.0.0]); Tue, 21 Nov 2023 16:22:46 -0800 (PST) On Tue, 21 Nov 2023, shiju.jose@huawei.com wrote: >From: Shiju Jose > >CXL spec 3.1 section 8.2.9.9.11.1 describes the device patrol scrub control >feature. The device patrol scrub proactively locates and makes corrections >to errors in regular cycle. The patrol scrub control allows the request to >configure patrol scrub input configurations. > >The patrol scrub control allows the requester to specify the number of >hours for which the patrol scrub cycles must be completed, provided that >the requested number is not less than the minimum number of hours for the >patrol scrub cycle that the device is capable of. In addition, the patrol >scrub controls allow the host to disable and enable the feature in case >disabling of the feature is needed for other purposes such as >performance-aware operations which require the background operations to be >turned off. > >Signed-off-by: Shiju Jose >--- > drivers/cxl/Kconfig | 17 +++ > drivers/cxl/core/Makefile | 1 + > drivers/cxl/core/memscrub.c | 272 ++++++++++++++++++++++++++++++++++++ > drivers/cxl/cxlmem.h | 8 ++ > drivers/cxl/pci.c | 6 + > 5 files changed, 304 insertions(+) > create mode 100644 drivers/cxl/core/memscrub.c > >diff --git a/drivers/cxl/Kconfig b/drivers/cxl/Kconfig >index 8ea1d340e438..e91f5acc94f2 100644 >--- a/drivers/cxl/Kconfig >+++ b/drivers/cxl/Kconfig >@@ -154,4 +154,21 @@ config CXL_PMU > monitoring units and provide standard perf based interfaces. > > If unsure say 'm'. >+ >+config CXL_SCRUB >+ tristate "CXL: Memory scrub feature" >+ depends on CXL_PCI >+ depends on CXL_MEM >+ help >+ The CXL memory scrub control is an optional feature allows host to >+ control the scrub configurations of CXL Type 3 devices, which >+ support patrol scrub and/or DDR5 ECS(Error Check Scrub). >+ >+ Say 'y/m' to enable the CXL memory scrub driver that will attach to >+ CXL.mem devices for memory scrub control feature. See sections >+ 8.2.9.9.11.1 and 8.2.9.9.11.2 in the CXL 3.1 specification for a >+ detailed description of CXL memory scrub control features. >+ >+ If unsure say 'm'. Does this really warrant yet another kconfig option? >+ > endif >diff --git a/drivers/cxl/core/Makefile b/drivers/cxl/core/Makefile >index 1f66b5d4d935..99e3202f868f 100644 >--- a/drivers/cxl/core/Makefile >+++ b/drivers/cxl/core/Makefile >@@ -15,3 +15,4 @@ cxl_core-y += hdm.o > cxl_core-y += pmu.o > cxl_core-$(CONFIG_TRACING) += trace.o > cxl_core-$(CONFIG_CXL_REGION) += region.o >+cxl_core-$(CONFIG_CXL_SCRUB) += memscrub.o >diff --git a/drivers/cxl/core/memscrub.c b/drivers/cxl/core/memscrub.c >new file mode 100644 >index 000000000000..7d98a1593abe >--- /dev/null >+++ b/drivers/cxl/core/memscrub.c >@@ -0,0 +1,272 @@ >+// SPDX-License-Identifier: GPL-2.0-or-later >+/* >+ * cxl_memscrub.c - CXL memory scrub driver >+ * >+ * Copyright (c) 2023 HiSilicon Limited. >+ * >+ * - Provides functions to configure patrol scrub >+ * feature of the CXL memory devices. >+ */ >+ >+#define pr_fmt(fmt) "CXL_MEM_SCRUB: " fmt >+ >+#include >+ >+/* CXL memory scrub feature common definitions */ >+#define CXL_SCRUB_MAX_ATTRB_RANGE_LENGTH 128 >+ >+static int cxl_mem_get_supported_feature_entry(struct cxl_memdev *cxlmd, const uuid_t *feat_uuid, >+ struct cxl_mbox_supp_feat_entry *feat_entry_out) It would be nicer if the feat_entry_out pointer was returned instead of passed by parameter. >+{ >+ struct cxl_mbox_get_supp_feats_out *feats_out __free(kvfree) = NULL; >+ struct cxl_mbox_supp_feat_entry *feat_entry; >+ struct cxl_dev_state *cxlds = cxlmd->cxlds; >+ struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlds); >+ struct cxl_mbox_get_supp_feats_in pi; >+ bool is_support_feature = false; >+ int feat_index, count; >+ int nentries; >+ int ret; >+ >+ feat_index = 0; >+ pi.count = sizeof(struct cxl_mbox_get_supp_feats_out) + >+ sizeof(struct cxl_mbox_supp_feat_entry); >+ feats_out = kvmalloc(pi.count, GFP_KERNEL); >+ if (!feats_out) >+ return -ENOMEM; >+ >+ do { >+ pi.start_index = feat_index; >+ memset(feats_out, 0, pi.count); >+ ret = cxl_get_supported_features(mds, &pi, feats_out); >+ if (ret) >+ return ret; >+ >+ nentries = feats_out->entries; >+ if (!nentries) >+ break; >+ >+ /* Check CXL memdev supports the feature */ >+ feat_entry = (void *)feats_out->feat_entries; >+ for (count = 0; count < nentries; count++, feat_entry++) { >+ if (uuid_equal(&feat_entry->uuid, feat_uuid)) { >+ is_support_feature = true; >+ memcpy(feat_entry_out, feat_entry, sizeof(*feat_entry_out)); >+ break; Just return 0 in this case, and then you can do without the is_supported_feature flag. >+ } >+ } >+ if (is_support_feature) >+ break; >+ feat_index += nentries; >+ } while (nentries); >+ >+ if (!is_support_feature) >+ return -ENOTSUPP; >+ >+ return 0; >+} >+ >+/* CXL memory patrol scrub control definitions */ >+#define CXL_MEMDEV_PS_GET_FEAT_VERSION 0x01 >+#define CXL_MEMDEV_PS_SET_FEAT_VERSION 0x01 >+ >+static const uuid_t cxl_patrol_scrub_uuid = >+ UUID_INIT(0x96dad7d6, 0xfde8, 0x482b, 0xa7, 0x33, 0x75, 0x77, 0x4e, \ >+ 0x06, 0xdb, 0x8a); >+ >+/* CXL memory patrol scrub control functions */ >+struct cxl_patrol_scrub_context { >+ struct device *dev; >+ u16 get_feat_size; >+ u16 set_feat_size; >+ bool scrub_cycle_changable; s/changable/changeable >+}; >+ >+/** >+ * struct cxl_memdev_ps_params - CXL memory patrol scrub parameter data structure. >+ * @enable: [IN] enable(1)/disable(0) patrol scrub. >+ * @scrub_cycle_changable: [OUT] scrub cycle attribute of patrol scrub is changeable. >+ * @speed: [IN] Requested patrol scrub cycle in hours. >+ * [OUT] Current patrol scrub cycle in hours. >+ * @min_speed:[OUT] minimum patrol scrub cycle, in hours, supported. >+ * @speed_avail:[OUT] Supported patrol scrub cycle in hours. >+ */ >+struct cxl_memdev_ps_params { >+ bool enable; >+ bool scrub_cycle_changable; >+ u16 speed; >+ u16 min_speed; >+ char speed_avail[CXL_SCRUB_MAX_ATTRB_RANGE_LENGTH]; >+}; >+ >+enum { >+ CXL_MEMDEV_PS_PARAM_ENABLE = 0, >+ CXL_MEMDEV_PS_PARAM_SPEED, >+}; >+ >+#define CXL_MEMDEV_PS_SCRUB_CYCLE_CHANGE_CAP_MASK BIT(0) >+#define CXL_MEMDEV_PS_SCRUB_CYCLE_REALTIME_REPORT_CAP_MASK BIT(1) >+#define CXL_MEMDEV_PS_CUR_SCRUB_CYCLE_MASK GENMASK(7, 0) >+#define CXL_MEMDEV_PS_MIN_SCRUB_CYCLE_MASK GENMASK(15, 8) >+#define CXL_MEMDEV_PS_FLAG_ENABLED_MASK BIT(0) >+ >+struct cxl_memdev_ps_feat_read_attrbs { >+ u8 scrub_cycle_cap; >+ __le16 scrub_cycle; >+ u8 scrub_flags; >+} __packed; >+ >+struct cxl_memdev_ps_set_feat_pi { >+ struct cxl_mbox_set_feat_in pi; >+ u8 scrub_cycle_hr; >+ u8 scrub_flags; >+} __packed; >+ >+static int cxl_mem_ps_get_attrbs(struct device *dev, >+ struct cxl_memdev_ps_params *params) >+{ >+ struct cxl_memdev_ps_feat_read_attrbs *rd_attrbs __free(kvfree) = NULL; >+ struct cxl_mbox_get_feat_in pi = { >+ .uuid = cxl_patrol_scrub_uuid, >+ .offset = 0, >+ .count = sizeof(struct cxl_memdev_ps_feat_read_attrbs), >+ .selection = CXL_GET_FEAT_SEL_CURRENT_VALUE, >+ }; >+ struct cxl_memdev *cxlmd = to_cxl_memdev(dev); >+ struct cxl_dev_state *cxlds = cxlmd->cxlds; >+ struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlds); >+ int ret; >+ >+ if (!mds) >+ return -EFAULT; >+ >+ rd_attrbs = kvmalloc(pi.count, GFP_KERNEL); >+ if (!rd_attrbs) >+ return -ENOMEM; >+ >+ ret = cxl_get_feature(mds, &pi, rd_attrbs); >+ if (ret) { >+ params->enable = 0; >+ params->speed = 0; >+ snprintf(params->speed_avail, CXL_SCRUB_MAX_ATTRB_RANGE_LENGTH, >+ "Unavailable"); >+ return ret; >+ } >+ params->scrub_cycle_changable = FIELD_GET(CXL_MEMDEV_PS_SCRUB_CYCLE_CHANGE_CAP_MASK, >+ rd_attrbs->scrub_cycle_cap); >+ params->enable = FIELD_GET(CXL_MEMDEV_PS_FLAG_ENABLED_MASK, >+ rd_attrbs->scrub_flags); >+ params->speed = FIELD_GET(CXL_MEMDEV_PS_CUR_SCRUB_CYCLE_MASK, >+ rd_attrbs->scrub_cycle); >+ params->min_speed = FIELD_GET(CXL_MEMDEV_PS_MIN_SCRUB_CYCLE_MASK, >+ rd_attrbs->scrub_cycle); >+ snprintf(params->speed_avail, CXL_SCRUB_MAX_ATTRB_RANGE_LENGTH, >+ "Minimum scrub cycle = %d hour", params->min_speed); >+ >+ return 0; >+} >+ >+static int cxl_mem_ps_set_attrbs(struct device *dev, >+ struct cxl_memdev_ps_params *params, u8 param_type) >+{ >+ struct cxl_memdev_ps_set_feat_pi set_pi = { >+ .pi.uuid = cxl_patrol_scrub_uuid, >+ .pi.flags = CXL_SET_FEAT_FLAG_MOD_VALUE_SAVED_ACROSS_RESET | >+ CXL_SET_FEAT_FLAG_FULL_DATA_TRANSFER, >+ .pi.offset = 0, >+ .pi.version = CXL_MEMDEV_PS_SET_FEAT_VERSION, >+ }; >+ struct cxl_memdev *cxlmd = to_cxl_memdev(dev); >+ struct cxl_dev_state *cxlds = cxlmd->cxlds; >+ struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlds); >+ struct cxl_memdev_ps_params rd_params; >+ int ret; >+ >+ if (!mds) >+ return -EFAULT; >+ >+ ret = cxl_mem_ps_get_attrbs(dev, &rd_params); >+ if (ret) { >+ dev_err(dev, "Get cxlmemdev patrol scrub params fail ret=%d\n", >+ ret); >+ return ret; >+ } >+ >+ switch (param_type) { >+ case CXL_MEMDEV_PS_PARAM_ENABLE: >+ set_pi.scrub_flags = FIELD_PREP(CXL_MEMDEV_PS_FLAG_ENABLED_MASK, >+ params->enable); >+ set_pi.scrub_cycle_hr = FIELD_PREP(CXL_MEMDEV_PS_CUR_SCRUB_CYCLE_MASK, >+ rd_params.speed); >+ break; >+ case CXL_MEMDEV_PS_PARAM_SPEED: >+ if (params->speed < rd_params.min_speed) { >+ dev_err(dev, "Invalid CXL patrol scrub cycle(%d) to set\n", >+ params->speed); >+ dev_err(dev, "Minimum supported CXL patrol scrub cycle in hour %d\n", >+ params->min_speed); >+ return -EINVAL; >+ } >+ set_pi.scrub_cycle_hr = FIELD_PREP(CXL_MEMDEV_PS_CUR_SCRUB_CYCLE_MASK, >+ params->speed); >+ set_pi.scrub_flags = FIELD_PREP(CXL_MEMDEV_PS_FLAG_ENABLED_MASK, >+ rd_params.enable); >+ break; >+ default: >+ dev_err(dev, "Invalid CXL patrol scrub parameter to set\n"); >+ return -EINVAL; >+ } >+ >+ ret = cxl_set_feature(mds, &set_pi, sizeof(set_pi)); >+ if (ret) { >+ dev_err(dev, "CXL patrol scrub set feature fail ret=%d\n", >+ ret); >+ return ret; >+ } >+ >+ /* Verify attribute set successfully */ >+ if (param_type == CXL_MEMDEV_PS_PARAM_SPEED) { >+ ret = cxl_mem_ps_get_attrbs(dev, &rd_params); >+ if (ret) { >+ dev_err(dev, "Get cxlmemdev patrol scrub params fail ret=%d\n", ret); >+ return ret; >+ } >+ if (rd_params.speed != params->speed) >+ return -EFAULT; >+ } >+ >+ return 0; >+} >+ >+int cxl_mem_patrol_scrub_init(struct cxl_memdev *cxlmd) >+{ >+ struct cxl_patrol_scrub_context *cxl_ps_ctx; >+ struct cxl_mbox_supp_feat_entry feat_entry; >+ struct cxl_memdev_ps_params params; >+ int ret; >+ >+ ret = cxl_mem_get_supported_feature_entry(cxlmd, &cxl_patrol_scrub_uuid, >+ &feat_entry); >+ if (ret < 0) >+ return ret; >+ >+ if (!(feat_entry.attrb_flags & CXL_FEAT_ENTRY_FLAG_CHANGABLE)) >+ return -ENOTSUPP; >+ >+ cxl_ps_ctx = devm_kzalloc(&cxlmd->dev, sizeof(*cxl_ps_ctx), GFP_KERNEL); >+ if (!cxl_ps_ctx) >+ return -ENOMEM; >+ >+ cxl_ps_ctx->get_feat_size = feat_entry.get_feat_size; >+ cxl_ps_ctx->set_feat_size = feat_entry.set_feat_size; >+ ret = cxl_mem_ps_get_attrbs(&cxlmd->dev, ¶ms); >+ if (ret) { >+ dev_err(&cxlmd->dev, "Get CXL patrol scrub params fail ret=%d\n", >+ ret); >+ return ret; >+ } >+ cxl_ps_ctx->scrub_cycle_changable = params.scrub_cycle_changable; >+ >+ return 0; >+} >+EXPORT_SYMBOL_NS_GPL(cxl_mem_patrol_scrub_init, CXL); >diff --git a/drivers/cxl/cxlmem.h b/drivers/cxl/cxlmem.h >index 46131dcd0900..25c46e72af16 100644 >--- a/drivers/cxl/cxlmem.h >+++ b/drivers/cxl/cxlmem.h >@@ -983,6 +983,14 @@ int cxl_trigger_poison_list(struct cxl_memdev *cxlmd); > int cxl_inject_poison(struct cxl_memdev *cxlmd, u64 dpa); > int cxl_clear_poison(struct cxl_memdev *cxlmd, u64 dpa); > >+/* cxl memory scrub functions */ >+#ifdef CONFIG_CXL_SCRUB >+int cxl_mem_patrol_scrub_init(struct cxl_memdev *cxlmd); >+#else >+static inline int cxl_mem_patrol_scrub_init(struct cxl_memdev *cxlmd) >+{ return -ENOTSUPP; } >+#endif >+ > #ifdef CONFIG_CXL_SUSPEND > void cxl_mem_active_inc(void); > void cxl_mem_active_dec(void); >diff --git a/drivers/cxl/pci.c b/drivers/cxl/pci.c >index 0155fb66b580..86bba8794bb4 100644 >--- a/drivers/cxl/pci.c >+++ b/drivers/cxl/pci.c >@@ -881,6 +881,12 @@ static int cxl_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id) > if (rc) > return rc; > >+ /* >+ * Initialize optional CXL scrub features >+ */ >+ if (cxl_mem_patrol_scrub_init(cxlmd)) >+ dev_dbg(&pdev->dev, "cxl_mem_patrol_scrub_init failed\n"); Unnecessary debug msg. >+ > rc = devm_cxl_sanitize_setup_notifier(&pdev->dev, cxlmd); > if (rc) > return rc; >-- >2.34.1 >