Received: by 2002:a05:6a10:9848:0:0:0:0 with SMTP id x8csp200390pxf; Wed, 31 Mar 2021 00:42:51 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxrRX5s6Ssoq8p4zZdsOzLTjnccmBk6eHxoKUTBo+7v/Wat1QAdkXWwGlkgfw4I2Y0Qkc+R X-Received: by 2002:a05:6402:1d39:: with SMTP id dh25mr2138114edb.282.1617176570870; Wed, 31 Mar 2021 00:42:50 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1617176570; cv=none; d=google.com; s=arc-20160816; b=PFawO/PTCrnMAWX7qVJ2nyvr/pUnhJ83d/pPFxvVd1xr+EQDcrR7wIhxnofD+Gj6gd imsOfrhuq5Bw6sv38lriiI7nzcm4c2qmMdeTn7TD5ARCgE0CFi5TIkk9Su2R2HT7D3sK csRCS26iFBCx3Eetcl8PQIqxr5G8GLjQ8vKQrnyLie6tNG4YeKkfCUmqMQfNfNaGx4Tm zUs9cHJVgxPgL3348u/BzoRrrb+T+GlZd0h32+p8i1P5VUKJGaNib/gDKEwtjAjL3rXj tjYZHWKWKekXYiC4K+svZpx4l6rdok6pr8lF2DYWo1nS0ZvEBG7z4tmtakQK452pLeN4 bEcw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :wdcironportexception:ironport-sdr:ironport-sdr:ironport-sdr :dkim-signature; bh=tDMvOXDbrWfj03PuMQzHD3t2PB8ReIxnCdNlX1g3yvs=; b=jhUKRb3Y8wTxjPe33nHiUT2CJjQm67eMzAOQdE8AMDerF5yDhVj8R2ve8upwUxVJ+7 1BORYJr7mINRKcYgSUjN3PSAQcH8d4q8F0i4eDQF3V6UBRFgWokq9xVnh0hAlAwt8weK tZAQYD5gYtKQ4Gb+373Tlfvu24ZQ46n2XpakQDoX9LH1LVIUiY7Eb/lcAgxmshKssv4O 9XgZQ7Cj33WqTfJXHHT9e7CPDnCTTENcjT9fdZISdaAflYBcTsvbPs4zgsijHqa7JtQM Qb5L9Ty9TUsfwJb1IgrGEfNaynNs3PsGk8bjOwf6Yv3jtan3ayEwuymcjl+JvWwsmwcL udRA== ARC-Authentication-Results: i=1; mx.google.com; dkim=fail header.i=@wdc.com header.s=dkim.wdc.com header.b=A8Ve6qes; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=wdc.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id rp14si994776ejb.435.2021.03.31.00.42.25; Wed, 31 Mar 2021 00:42:50 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=fail header.i=@wdc.com header.s=dkim.wdc.com header.b=A8Ve6qes; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=wdc.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S234210AbhCaHk6 (ORCPT + 99 others); Wed, 31 Mar 2021 03:40:58 -0400 Received: from esa1.hgst.iphmx.com ([68.232.141.245]:8968 "EHLO esa1.hgst.iphmx.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234088AbhCaHke (ORCPT ); Wed, 31 Mar 2021 03:40:34 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1617176434; x=1648712434; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=BI+q1+vMMk859Iz5KZFdJyLrCicWlgpuK1Jmh18UGMI=; b=A8Ve6qese1mBEfcXs/0znzknK7UYSKOU+JE0/pyrHzYO/4hW4oADnMZX wGy2i3Fj2v/iselHO4j2ubhp2JoDK/7oPfXllj47v/dtQH0RRWOApai2W 4GmVH9WP+l+1llyjxJc8Lwv+rVPLY/bXHAEln0DZwGW2MYH1n+dYw513r 9iPh+vawxp7/He+glwjmNMaPbzEzpw2NC5CcaI9MIQYrMZ63HfSSHkbP9 xm6bevclLgs5tnxpzmNjF3VccsV9COlbvpV3DnOK0O+MV2bjQfr7QLwYV fCXmQ6wFughowlcVLvCmwK3B7Gbsv5y4C7G/cffKNcTm61BcFeK3lJtJb w==; IronPort-SDR: vwAGrBkkonW4EywBlH4N+jSS6sLvlIHjgrwvAyAw0gRJ144B1idpy+l0JqSqKIeLk6lVoVopdE Qm98hyADJTK2zC8/iER0N6/+oAYQIAQAchJN6z1Ebe+Wo9UUTN3hscyVk+ZtB2LS8di1BzPMci M+o24LLwKMCayLTI0GOIyQ1/9RgWLpA+5jt9jC7fTUFh6Da2lRk7lwZrTXOLSSBwIUNgMG/4qd 3hu4F2o0JTRJ1i3Gp6Q0jPu4uXvCJThLpGL0g/7Fiv23vYsKmN66PUU4BZkgmqI1bh6voQTN7t ohA= X-IronPort-AV: E=Sophos;i="5.81,293,1610380800"; d="scan'208";a="274239166" Received: from uls-op-cesaip01.wdc.com (HELO uls-op-cesaep01.wdc.com) ([199.255.45.14]) by ob1.hgst.iphmx.com with ESMTP; 31 Mar 2021 15:40:33 +0800 IronPort-SDR: 4ewX9uMG4glTTBy5hjWA2RN9Dd4AoJLn11vpMQbmTJzCVqsl1beAC4vPwS/HBkhvT3AEp9wFFK vsp+7We3BYKiGSqMvSqs4Zsvl9g5F8v/GCRyEjZfGHt5FiIF0LKKMsLkhHVIrKyrhQquTr7XYL vLIOTYMVUWbrlGjxelYYAQewhlt4bDdUgLnkZ3zEjhSiYpqUJ0rahtD6zo2CUeql1Vr/wpeKz5 8Rvv9aPB8kgO2C2quTIKTFWNf3Iw0dWmxzsPe737Dfs3bMxPLt63jB24PhknGhLUvvHgVwxAjY d3tdIoEiXJT/siz9R+LCu89g Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep01.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 31 Mar 2021 00:22:17 -0700 IronPort-SDR: blLlkpBFbwk7b28R9+YxKUyv1VhtkEthqjwtGy1JepW/Rra6q4LPZL5TBah4V1r8tCmRn63Zc7 VZDEkz7rFTlC9W2PIIIChswjGPLHLz4wNzh1SGCiOnDM94bkZ4pWpzNNRJ32R09KNz87qangve pSB4lvB191FDblIOEYDg1NGv4bHx6TgPKj4x84/ICkDKUFqoNlVwtb+IOqv4NdSNWmpqhP+hCH HtMQ4Ejc+g4RndX1hYuBSltA+WBOGQlSOGWGqR9Oh8k9XmFziVY1pAiuwtnhHksaAghizd8zUD b2s= WDCIronportException: Internal Received: from bxygm33.sdcorp.global.sandisk.com ([10.0.231.247]) by uls-op-cesaip01.wdc.com with ESMTP; 31 Mar 2021 00:40:29 -0700 From: Avri Altman To: "James E . J . Bottomley" , "Martin K . Petersen" , linux-scsi@vger.kernel.org, linux-kernel@vger.kernel.org Cc: gregkh@linuxfoundation.org, Bart Van Assche , yongmyung lee , Daejun Park , alim.akhtar@samsung.com, asutoshd@codeaurora.org, Zang Leigang , Avi Shchislowski , Bean Huo , cang@codeaurora.org, stanley.chu@mediatek.com, Avri Altman Subject: [PATCH v7 04/11] scsi: ufshpb: Add reads counter Date: Wed, 31 Mar 2021 10:39:45 +0300 Message-Id: <20210331073952.102162-5-avri.altman@wdc.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20210331073952.102162-1-avri.altman@wdc.com> References: <20210331073952.102162-1-avri.altman@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org In host control mode, reads are the major source of activation trials. Keep track of those reads counters, for both active as well inactive regions. We reset the read counter upon write - we are only interested in "clean" reads. Keep those counters normalized, as we are using those reads as a comparative score, to make various decisions. If during consecutive normalizations an active region has exhaust its reads - inactivate it. while at it, protect the {active,inactive}_count stats by adding them into the applicable handler. Signed-off-by: Avri Altman --- drivers/scsi/ufs/ufshpb.c | 94 ++++++++++++++++++++++++++++++++++++--- drivers/scsi/ufs/ufshpb.h | 9 ++++ 2 files changed, 97 insertions(+), 6 deletions(-) diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c index 252fcfb48862..3ab66421dc00 100644 --- a/drivers/scsi/ufs/ufshpb.c +++ b/drivers/scsi/ufs/ufshpb.c @@ -16,6 +16,8 @@ #include "ufshpb.h" #include "../sd.h" +#define ACTIVATION_THRESHOLD 8 /* 8 IOs */ + /* memory management */ static struct kmem_cache *ufshpb_mctx_cache; static mempool_t *ufshpb_mctx_pool; @@ -26,6 +28,9 @@ static int tot_active_srgn_pages; static struct workqueue_struct *ufshpb_wq; +static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int rgn_idx, + int srgn_idx); + bool ufshpb_is_allowed(struct ufs_hba *hba) { return !(hba->ufshpb_dev.hpb_disabled); @@ -148,7 +153,7 @@ static void ufshpb_iterate_rgn(struct ufshpb_lu *hpb, int rgn_idx, int srgn_idx, int srgn_offset, int cnt, bool set_dirty) { struct ufshpb_region *rgn; - struct ufshpb_subregion *srgn; + struct ufshpb_subregion *srgn, *prev_srgn = NULL; int set_bit_len; int bitmap_len; unsigned long flags; @@ -167,15 +172,39 @@ static void ufshpb_iterate_rgn(struct ufshpb_lu *hpb, int rgn_idx, int srgn_idx, else set_bit_len = cnt; - if (set_dirty) - set_bit(RGN_FLAG_DIRTY, &rgn->rgn_flags); - spin_lock_irqsave(&hpb->rgn_state_lock, flags); if (set_dirty && rgn->rgn_state != HPB_RGN_INACTIVE && srgn->srgn_state == HPB_SRGN_VALID) bitmap_set(srgn->mctx->ppn_dirty, srgn_offset, set_bit_len); spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); + if (hpb->is_hcm && prev_srgn != srgn) { + bool activate = false; + + spin_lock(&rgn->rgn_lock); + if (set_dirty) { + rgn->reads -= srgn->reads; + srgn->reads = 0; + set_bit(RGN_FLAG_DIRTY, &rgn->rgn_flags); + } else { + srgn->reads++; + rgn->reads++; + if (srgn->reads == ACTIVATION_THRESHOLD) + activate = true; + } + spin_unlock(&rgn->rgn_lock); + + if (activate) { + spin_lock_irqsave(&hpb->rsp_list_lock, flags); + ufshpb_update_active_info(hpb, rgn_idx, srgn_idx); + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, + "activate region %d-%d\n", rgn_idx, srgn_idx); + } + + prev_srgn = srgn; + } + srgn_offset = 0; if (++srgn_idx == hpb->srgns_per_rgn) { srgn_idx = 0; @@ -604,6 +633,19 @@ int ufshpb_prep(struct ufs_hba *hba, struct ufshcd_lrb *lrbp) if (!ufshpb_is_support_chunk(hpb, transfer_len)) return 0; + if (hpb->is_hcm) { + /* + * in host control mode, reads are the main source for + * activation trials. + */ + ufshpb_iterate_rgn(hpb, rgn_idx, srgn_idx, srgn_offset, + transfer_len, false); + + /* keep those counters normalized */ + if (rgn->reads > hpb->entries_per_srgn) + schedule_work(&hpb->ufshpb_normalization_work); + } + spin_lock_irqsave(&hpb->rgn_state_lock, flags); if (ufshpb_test_ppn_dirty(hpb, rgn_idx, srgn_idx, srgn_offset, transfer_len)) { @@ -755,6 +797,8 @@ static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int rgn_idx, if (list_empty(&srgn->list_act_srgn)) list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); + + hpb->stats.rb_active_cnt++; } static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int rgn_idx) @@ -770,6 +814,8 @@ static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int rgn_idx) if (list_empty(&rgn->list_inact_rgn)) list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn); + + hpb->stats.rb_inactive_cnt++; } static void ufshpb_activate_subregion(struct ufshpb_lu *hpb, @@ -1090,6 +1136,7 @@ static int ufshpb_evict_region(struct ufshpb_lu *hpb, struct ufshpb_region *rgn) rgn->rgn_idx); goto out; } + if (!list_empty(&rgn->list_lru_rgn)) { if (ufshpb_check_srgns_issue_state(hpb, rgn)) { ret = -EBUSY; @@ -1284,7 +1331,6 @@ static void ufshpb_rsp_req_region_update(struct ufshpb_lu *hpb, if (srgn->srgn_state == HPB_SRGN_VALID) srgn->srgn_state = HPB_SRGN_INVALID; spin_unlock(&hpb->rgn_state_lock); - hpb->stats.rb_active_cnt++; } if (hpb->is_hcm) { @@ -1316,7 +1362,6 @@ static void ufshpb_rsp_req_region_update(struct ufshpb_lu *hpb, } spin_unlock(&hpb->rgn_state_lock); - hpb->stats.rb_inactive_cnt++; } out: @@ -1515,6 +1560,36 @@ static void ufshpb_run_inactive_region_list(struct ufshpb_lu *hpb) spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); } +static void ufshpb_normalization_work_handler(struct work_struct *work) +{ + struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu, + ufshpb_normalization_work); + int rgn_idx; + + for (rgn_idx = 0; rgn_idx < hpb->rgns_per_lu; rgn_idx++) { + struct ufshpb_region *rgn = hpb->rgn_tbl + rgn_idx; + int srgn_idx; + + spin_lock(&rgn->rgn_lock); + rgn->reads = 0; + for (srgn_idx = 0; srgn_idx < hpb->srgns_per_rgn; srgn_idx++) { + struct ufshpb_subregion *srgn = rgn->srgn_tbl + srgn_idx; + + srgn->reads >>= 1; + rgn->reads += srgn->reads; + } + spin_unlock(&rgn->rgn_lock); + + if (rgn->rgn_state != HPB_RGN_ACTIVE || rgn->reads) + continue; + + /* if region is active but has no reads - inactivate it */ + spin_lock(&hpb->rsp_list_lock); + ufshpb_update_inactive_info(hpb, rgn->rgn_idx); + spin_unlock(&hpb->rsp_list_lock); + } +} + static void ufshpb_map_work_handler(struct work_struct *work) { struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu, map_work); @@ -1674,6 +1749,8 @@ static int ufshpb_alloc_region_tbl(struct ufs_hba *hba, struct ufshpb_lu *hpb) rgn = rgn_table + rgn_idx; rgn->rgn_idx = rgn_idx; + spin_lock_init(&rgn->rgn_lock); + INIT_LIST_HEAD(&rgn->list_inact_rgn); INIT_LIST_HEAD(&rgn->list_lru_rgn); @@ -1915,6 +1992,9 @@ static int ufshpb_lu_hpb_init(struct ufs_hba *hba, struct ufshpb_lu *hpb) INIT_LIST_HEAD(&hpb->list_hpb_lu); INIT_WORK(&hpb->map_work, ufshpb_map_work_handler); + if (hpb->is_hcm) + INIT_WORK(&hpb->ufshpb_normalization_work, + ufshpb_normalization_work_handler); hpb->map_req_cache = kmem_cache_create("ufshpb_req_cache", sizeof(struct ufshpb_req), 0, 0, NULL); @@ -2014,6 +2094,8 @@ static void ufshpb_discard_rsp_lists(struct ufshpb_lu *hpb) static void ufshpb_cancel_jobs(struct ufshpb_lu *hpb) { + if (hpb->is_hcm) + cancel_work_sync(&hpb->ufshpb_normalization_work); cancel_work_sync(&hpb->map_work); } diff --git a/drivers/scsi/ufs/ufshpb.h b/drivers/scsi/ufs/ufshpb.h index 032672114881..87495e59fcf1 100644 --- a/drivers/scsi/ufs/ufshpb.h +++ b/drivers/scsi/ufs/ufshpb.h @@ -106,6 +106,10 @@ struct ufshpb_subregion { int rgn_idx; int srgn_idx; bool is_last; + + /* subregion reads - for host mode */ + unsigned int reads; + /* below information is used by rsp_list */ struct list_head list_act_srgn; }; @@ -123,6 +127,10 @@ struct ufshpb_region { struct list_head list_lru_rgn; unsigned long rgn_flags; #define RGN_FLAG_DIRTY 0 + + /* region reads - for host mode */ + spinlock_t rgn_lock; + unsigned int reads; }; #define for_each_sub_region(rgn, i, srgn) \ @@ -212,6 +220,7 @@ struct ufshpb_lu { /* for selecting victim */ struct victim_select_info lru_info; + struct work_struct ufshpb_normalization_work; /* pinned region information */ u32 lu_pinned_start; -- 2.25.1