Received: by 2002:a05:6a10:9848:0:0:0:0 with SMTP id x8csp3146090pxf; Mon, 15 Mar 2021 02:38:03 -0700 (PDT) X-Google-Smtp-Source: ABdhPJylcQvgY4m10vPZc5UhyoIpU4f8Tm6OdOv4TkD0cR8daq6GM4ZuyyrnVFBFNL/+piK63ILo X-Received: by 2002:aa7:d503:: with SMTP id y3mr28634946edq.142.1615801083349; Mon, 15 Mar 2021 02:38:03 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1615801083; cv=none; d=google.com; s=arc-20160816; b=xqQfw63OUMNfS23tnrWeGGivobC0TKaCAqjdEKBepO9oAKAdtb7ha/0z2c24Ujdnrt g+/O93kXsbx8fKGIJpSPK56O/wpOZToTzERe7mwSQncnpM4DO11QBKGeI98XMLGM6U88 nUjAT3bNSAoxBiCni3v0FsZ7D467M3wU+ePMdd4qMGS6Mdy0mv5psC6dZiPN00YZBRO8 oVeUGDuThKIEPPzN54xrHpLPRMM5QdcTOw3fhjYpDLXYSaCFbHx7LEqMY1Yrq0YIwQ+X 8beTOEO97LHGk6D0nMkPArjWfy7BPLhdtKWRTgWFYGbrK+xP+T4ZMWQg/s+ZyNU7Jyyc 7J3g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:user-agent:message-id:references:in-reply-to :subject:cc:to:from:date:content-transfer-encoding:mime-version :sender:dkim-signature; bh=lLf6xLv6j6kjmGnZxe7LXmYbE7Y8IzEhAXO3Hf+8ZBM=; b=bzi1Gar+nhSfA8HU7417BsraiJm3UDqNliea1gI1g7jIhZzvp0q+5/mEVAUB3DbBuf j7zo1OhlxP4EhYI5yTvPUVTe2IRbNaK5NVQaxnp2zVhz++7dTAa4QCeRJhnFO2mEgLvu yglibfA/XRH9X9gP9BcdQh/AsH5IqWomABsDIAYB4EpKfGWW6PQEpophH5uNqNrjxakd YM0ImtnyLv1arwGPAB3GIQALJKYFoFQcuw+OeZGGMYzVmCS4oMkXN0RUQjkf2xhZTzaq f6j75p7eoS4LKPSYDWJkZqDCvGz9Ek/UgXsY3rW600qmtScefJmjgbwERXNjOSv+fAj6 Opcg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@mg.codeaurora.org header.s=smtp header.b=LMEMIcHA; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id o11si10660206eju.241.2021.03.15.02.37.40; Mon, 15 Mar 2021 02:38:03 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@mg.codeaurora.org header.s=smtp header.b=LMEMIcHA; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229574AbhCOJgp (ORCPT + 99 others); Mon, 15 Mar 2021 05:36:45 -0400 Received: from z11.mailgun.us ([104.130.96.11]:44394 "EHLO z11.mailgun.us" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229553AbhCOJgm (ORCPT ); Mon, 15 Mar 2021 05:36:42 -0400 DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mg.codeaurora.org; q=dns/txt; s=smtp; t=1615801002; h=Message-ID: References: In-Reply-To: Subject: Cc: To: From: Date: Content-Transfer-Encoding: Content-Type: MIME-Version: Sender; bh=lLf6xLv6j6kjmGnZxe7LXmYbE7Y8IzEhAXO3Hf+8ZBM=; b=LMEMIcHA+262a69F2LTMNx2darHLtMHPsRVt2BPOWaDlSSMrGkrHUKwxeZZLJpxFVhWDzFGK DmtBNnzoFy9uqP4ONdWqstCad2ZTBDCTxIF2B2cfUCe2er9pV9I31IfmDNnv6rrjroqKe14i ufuF2ro4E+FADWO8oOz4+SN9weg= X-Mailgun-Sending-Ip: 104.130.96.11 X-Mailgun-Sid: WyI0MWYwYSIsICJsaW51eC1rZXJuZWxAdmdlci5rZXJuZWwub3JnIiwgImJlOWU0YSJd Received: from smtp.codeaurora.org (ec2-35-166-182-171.us-west-2.compute.amazonaws.com [35.166.182.171]) by smtp-out-n04.prod.us-west-2.postgun.com with SMTP id 604f2a9f3f267701a4762bf9 (version=TLS1.2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256); Mon, 15 Mar 2021 09:36:31 GMT Sender: cang=codeaurora.org@mg.codeaurora.org Received: by smtp.codeaurora.org (Postfix, from userid 1001) id D6DC4C433C6; Mon, 15 Mar 2021 09:36:31 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-caf-mail-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-2.9 required=2.0 tests=ALL_TRUSTED,BAYES_00 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.codeaurora.org (localhost.localdomain [127.0.0.1]) (using TLSv1 with cipher ECDHE-RSA-AES256-SHA (256/256 bits)) (No client certificate requested) (Authenticated sender: cang) by smtp.codeaurora.org (Postfix) with ESMTPSA id 0D9A4C433CA; Mon, 15 Mar 2021 09:36:28 +0000 (UTC) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit Date: Mon, 15 Mar 2021 17:36:28 +0800 From: Can Guo To: Avri Altman Cc: "James E . J . Bottomley" , "Martin K . Petersen" , linux-scsi@vger.kernel.org, linux-kernel@vger.kernel.org, gregkh@linuxfoundation.org, Bart Van Assche , yongmyung lee , Daejun Park , alim.akhtar@samsung.com, asutoshd@codeaurora.org, Zang Leigang , Avi Shchislowski , Bean Huo , stanley.chu@mediatek.com Subject: Re: [PATCH v5 07/10] scsi: ufshpb: Add "Cold" regions timer In-Reply-To: <20210302132503.224670-8-avri.altman@wdc.com> References: <20210302132503.224670-1-avri.altman@wdc.com> <20210302132503.224670-8-avri.altman@wdc.com> Message-ID: X-Sender: cang@codeaurora.org User-Agent: Roundcube Webmail/1.3.9 Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2021-03-02 21:25, Avri Altman wrote: > In order not to hang on to “cold” regions, we shall inactivate a > region that has no READ access for a predefined amount of time - > READ_TO_MS. For that purpose we shall monitor the active regions list, > polling it on every POLLING_INTERVAL_MS. On timeout expiry we shall add > the region to the "to-be-inactivated" list, unless it is clean and did > not exhaust its READ_TO_EXPIRIES - another parameter. > > All this does not apply to pinned regions. > > Signed-off-by: Avri Altman > --- > drivers/scsi/ufs/ufshpb.c | 65 +++++++++++++++++++++++++++++++++++++++ > drivers/scsi/ufs/ufshpb.h | 6 ++++ > 2 files changed, 71 insertions(+) > > diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c > index 0034fa03fdc6..89a930e72cff 100644 > --- a/drivers/scsi/ufs/ufshpb.c > +++ b/drivers/scsi/ufs/ufshpb.c > @@ -18,6 +18,9 @@ > > #define ACTIVATION_THRESHOLD 4 /* 4 IOs */ > #define EVICTION_THRESHOLD (ACTIVATION_THRESHOLD << 6) /* 256 IOs */ > +#define READ_TO_MS 1000 > +#define READ_TO_EXPIRIES 100 > +#define POLLING_INTERVAL_MS 200 > > /* memory management */ > static struct kmem_cache *ufshpb_mctx_cache; > @@ -1024,12 +1027,61 @@ static int > ufshpb_check_srgns_issue_state(struct ufshpb_lu *hpb, > return 0; > } > > +static void ufshpb_read_to_handler(struct work_struct *work) > +{ > + struct delayed_work *dwork = to_delayed_work(work); > + struct ufshpb_lu *hpb; > + struct victim_select_info *lru_info; > + struct ufshpb_region *rgn; > + unsigned long flags; > + LIST_HEAD(expired_list); > + > + hpb = container_of(dwork, struct ufshpb_lu, ufshpb_read_to_work); > + > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); > + > + lru_info = &hpb->lru_info; > + > + list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn) { > + bool timedout = ktime_after(ktime_get(), rgn->read_timeout); > + > + if (timedout) { > + rgn->read_timeout_expiries--; > + if (is_rgn_dirty(rgn) || > + rgn->read_timeout_expiries == 0) > + list_add(&rgn->list_expired_rgn, &expired_list); > + else > + rgn->read_timeout = ktime_add_ms(ktime_get(), > + READ_TO_MS); > + } > + } > + > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); > + > + list_for_each_entry(rgn, &expired_list, list_expired_rgn) { Here can be problematic - since you don't have the native expired_list initialized before use, if above loop did not insert anything to expired_list, it shall become a dead loop here. And, which lock is protecting rgn->list_expired_rgn? If two read_to_handler works are running in parallel, one can be inserting it to its expired_list while another can be deleting it. Can Guo. > + list_del_init(&rgn->list_expired_rgn); > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); > + ufshpb_update_inactive_info(hpb, rgn->rgn_idx); > + hpb->stats.rb_inactive_cnt++; > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); > + } > + > + ufshpb_kick_map_work(hpb); > + > + schedule_delayed_work(&hpb->ufshpb_read_to_work, > + msecs_to_jiffies(POLLING_INTERVAL_MS)); > +} > + > static void ufshpb_add_lru_info(struct victim_select_info *lru_info, > struct ufshpb_region *rgn) > { > rgn->rgn_state = HPB_RGN_ACTIVE; > list_add_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); > atomic_inc(&lru_info->active_cnt); > + if (rgn->hpb->is_hcm) { > + rgn->read_timeout = ktime_add_ms(ktime_get(), READ_TO_MS); > + rgn->read_timeout_expiries = READ_TO_EXPIRIES; > + } > } > > static void ufshpb_hit_lru_info(struct victim_select_info *lru_info, > @@ -1813,6 +1865,7 @@ static int ufshpb_alloc_region_tbl(struct > ufs_hba *hba, struct ufshpb_lu *hpb) > > INIT_LIST_HEAD(&rgn->list_inact_rgn); > INIT_LIST_HEAD(&rgn->list_lru_rgn); > + INIT_LIST_HEAD(&rgn->list_expired_rgn); > > if (rgn_idx == hpb->rgns_per_lu - 1) { > srgn_cnt = ((hpb->srgns_per_lu - 1) % > @@ -1834,6 +1887,7 @@ static int ufshpb_alloc_region_tbl(struct > ufs_hba *hba, struct ufshpb_lu *hpb) > } > > rgn->rgn_flags = 0; > + rgn->hpb = hpb; > } > > return 0; > @@ -2053,6 +2107,8 @@ static int ufshpb_lu_hpb_init(struct ufs_hba > *hba, struct ufshpb_lu *hpb) > ufshpb_normalization_work_handler); > INIT_WORK(&hpb->ufshpb_lun_reset_work, > ufshpb_reset_work_handler); > + INIT_DELAYED_WORK(&hpb->ufshpb_read_to_work, > + ufshpb_read_to_handler); > } > > hpb->map_req_cache = kmem_cache_create("ufshpb_req_cache", > @@ -2087,6 +2143,10 @@ static int ufshpb_lu_hpb_init(struct ufs_hba > *hba, struct ufshpb_lu *hpb) > ufshpb_stat_init(hpb); > ufshpb_param_init(hpb); > > + if (hpb->is_hcm) > + schedule_delayed_work(&hpb->ufshpb_read_to_work, > + msecs_to_jiffies(POLLING_INTERVAL_MS)); > + > return 0; > > release_pre_req_mempool: > @@ -2154,6 +2214,7 @@ static void ufshpb_discard_rsp_lists(struct > ufshpb_lu *hpb) > static void ufshpb_cancel_jobs(struct ufshpb_lu *hpb) > { > if (hpb->is_hcm) { > + cancel_delayed_work_sync(&hpb->ufshpb_read_to_work); > cancel_work_sync(&hpb->ufshpb_lun_reset_work); > cancel_work_sync(&hpb->ufshpb_normalization_work); > } > @@ -2264,6 +2325,10 @@ void ufshpb_resume(struct ufs_hba *hba) > continue; > ufshpb_set_state(hpb, HPB_PRESENT); > ufshpb_kick_map_work(hpb); > + if (hpb->is_hcm) > + schedule_delayed_work(&hpb->ufshpb_read_to_work, > + msecs_to_jiffies(POLLING_INTERVAL_MS)); > + > } > } > > diff --git a/drivers/scsi/ufs/ufshpb.h b/drivers/scsi/ufs/ufshpb.h > index 37c1b0ea0c0a..b49e9a34267f 100644 > --- a/drivers/scsi/ufs/ufshpb.h > +++ b/drivers/scsi/ufs/ufshpb.h > @@ -109,6 +109,7 @@ struct ufshpb_subregion { > }; > > struct ufshpb_region { > + struct ufshpb_lu *hpb; > struct ufshpb_subregion *srgn_tbl; > enum HPB_RGN_STATE rgn_state; > int rgn_idx; > @@ -126,6 +127,10 @@ struct ufshpb_region { > /* region reads - for host mode */ > spinlock_t rgn_lock; > unsigned int reads; > + /* region "cold" timer - for host mode */ > + ktime_t read_timeout; > + unsigned int read_timeout_expiries; > + struct list_head list_expired_rgn; > }; > > #define for_each_sub_region(rgn, i, srgn) \ > @@ -219,6 +224,7 @@ struct ufshpb_lu { > struct victim_select_info lru_info; > struct work_struct ufshpb_normalization_work; > struct work_struct ufshpb_lun_reset_work; > + struct delayed_work ufshpb_read_to_work; > > /* pinned region information */ > u32 lu_pinned_start;