Received: by 2002:a05:6a10:f3d0:0:0:0:0 with SMTP id a16csp2865809pxv; Mon, 12 Jul 2021 03:57:27 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxR6ida5eHnZo5/q0t3EcBTG6m2RL1JJUKPlw+unMXbV3oHx4ogKGfbrpTmO2pZR7mNK2Ho X-Received: by 2002:a05:6e02:12a8:: with SMTP id f8mr13454116ilr.82.1626087447778; Mon, 12 Jul 2021 03:57:27 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1626087447; cv=none; d=google.com; s=arc-20160816; b=fRH3MlYsN4s90GjkWN+YvMQBAZXUFDtV6VE5Xn4CAab2wrXMt+pjDXKic3ji67OhOy i5yRwx/oFYuq2Q+flHWanKs/TnBPO0bf7QU0/pvaT4VYOU6idEVM2BgOuofxVRszjzJ6 +rxECmwyeW6VqT8eOYTGEcrUT+VP5t0LN3PCIBXWdrUa1jUHi96h69to2HsH0MXwiKY5 w5u8VbOAr2vwyl3Rh+zn1IMmuXHr+MvtYVqf7KOcxKfiXm2rDEJ2xJ/EEWmQUlHvuoqN FFddMa1iuDrkvp2VaVovohacxcc8B4XfRUrDPiwtEARtRp6BEcQMM8qmg9ZJ7JvOYFv3 lPxw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :wdcironportexception:ironport-sdr:ironport-sdr:ironport-sdr :dkim-signature; bh=j/g+zWhACdZd0MOWKVfPFq7U5P9/d8pkZKa8GYp6Ol8=; b=xGlUarCb0hvk6oYz+EgsNDLEXh0KM2lHRAMCX5NJJSU3MrFZO8JYba1DNGJhJCfyWY XIJUs3qjCPFXjHC9NKwYNPLfZH+Msh45pGWa/gHLf7ffXKY2xu8HDly+HarC6H0KX7Kd DkZUCr5TJu6kecdI1jMoUQ13L3PRS0Sh9wZXSZ3ZgQcD1v6z0AQd+cDf0xwsd0uIvGlH 9LpIxec9rIB4HGZngx90vAmHXNG7wyVF6U0roKYYhrPj1UAu0NAuHSMhlH0tza2IHJPF S0KFJZZGPNuKOfUArdXQlC7Bm1U7FUDHWlPaavtcgPCw5Uf2+2LfvNE4vHr+7AnnoLmJ yQyg== ARC-Authentication-Results: i=1; mx.google.com; dkim=fail header.i=@wdc.com header.s=dkim.wdc.com header.b=qY5vqLtw; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=wdc.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id t3si18309053iof.33.2021.07.12.03.57.16; Mon, 12 Jul 2021 03:57:27 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=fail header.i=@wdc.com header.s=dkim.wdc.com header.b=qY5vqLtw; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=wdc.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S241208AbhGLJye (ORCPT + 99 others); Mon, 12 Jul 2021 05:54:34 -0400 Received: from esa4.hgst.iphmx.com ([216.71.154.42]:4475 "EHLO esa4.hgst.iphmx.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S241166AbhGLJy2 (ORCPT ); Mon, 12 Jul 2021 05:54:28 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1626083500; x=1657619500; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=2mE6HnixVQ0nNDwBgw5qlos26G/asVJ4KERSAzfQHOY=; b=qY5vqLtwviDxeOeJIufv/LjXyZT6477eVuMfRRLais6Sz1mAg1LOTcKQ OAALsDxNtlD6LDiZpkUmqHG/tB80snaD0YGU0QyI3UooN3V8fKqkpbaHe O0/7UwC3wHQusOrKIn96Mn1dpX0A8AsWnU0w1oNplmTA5YMMuegoTA7uM x9GqcVXBP+QaCnFWefRutPfXAl+8il0JyLWLVrg5tsUbKlPFlLWeLqUcr yAXxYni9oGd+q/HIf45qGrQmxN9lErcqwRs7mF0Z2lQehrpj1TnLMvJVb db2S3+Ci18rbjcl1UEVz3AtvsMTj/oECl1TwG3sxNFKTBL9x+lQ3Ggo1I g==; IronPort-SDR: vMbXU+5KCCex2Yw/ota5N+8TshTRQeUhfueit0JmjbaeN62UZZz6JTMsmLircC8vaCJxCdBqEP /a70cD227xV2jHGpO4vXiIZVwFNVt4uq+R2+Lboq+rccXfgDVm2lfen1KA6S9mxFRMoqxZq71P m+U5puHU1B8QPK8qh1J4Kv0Pdwx3j37ScN9pH880J/D0a04vNAHXBO058ETI/aQddiLm/22Yhb 2mbWp3TBCZXw7i7IyJRHI2Qpaebqy7Ki+b12cqvjNQphOyb5AlajJJ4GDVd0mE4bARQM6Mf6Gy 7sg= X-IronPort-AV: E=Sophos;i="5.84,232,1620662400"; d="scan'208";a="173586801" Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 12 Jul 2021 17:51:40 +0800 IronPort-SDR: HcoQhn/MGu6dqQ2au5wPmMYKDFpPzIRyF0DG4t3zOetND2g0XEvYhiidIrLw8ZLb/GwkdXjGIi aR3iofZEenr9Jd2QVxBSV4BglH3jELPQGvWCZugFkjdesif/DrgMC+f2j9wmBdu2O7HTtTZ+Va QjzSh+VqmS465ArCKfAtxirpx6nFRl3Fn22NodQsMmo2ZGAKvCnzLa/j+nqr6qwvbWsNH7MWOZ SGcHZfATcmFjx7zbbTslmh+TFJoANsDDwQ+rLaS3fCOqsYq7uJBKAjx9S8gHc5KgIQBzkSs1RP Gfx+ZDn6eBOK9uQFcfGComUn Received: from uls-op-cesaip02.wdc.com ([10.248.3.37]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 12 Jul 2021 02:28:25 -0700 IronPort-SDR: A3KuAPb39dMfnQnL/KpZNdd0sU1ssBM+KoNeolY5F9RgA1YbfJTx5ZuUy3T/uSgUAw+IroHw5m ZqJHKxA5wQtsT711jHR2Xyxim/QHEpKrvuGNQ3mDdT9Hgrlf1O/UhnQBdBPBJeloq7Q2yqiyba gH3jjsNVXW3wCWLLzUQxmVjymL16xApInS0NbtMwAzDP6l9skfarbNHt0KG6cyGtEjEJuNmwd+ L8JuIozuA0StVz7dXginFTdj1vCfotRlGEpmymhHEvVY53YuBHWlAA/ah3d5NsVznt9RJFP2yZ G2M= WDCIronportException: Internal Received: from bxygm33.sdcorp.global.sandisk.com ([10.0.231.247]) by uls-op-cesaip02.wdc.com with ESMTP; 12 Jul 2021 02:51:36 -0700 From: Avri Altman To: "James E . J . Bottomley" , "Martin K . Petersen" , linux-scsi@vger.kernel.org, linux-kernel@vger.kernel.org Cc: gregkh@linuxfoundation.org, Bart Van Assche , yongmyung lee , Daejun Park , alim.akhtar@samsung.com, asutoshd@codeaurora.org, Zang Leigang , Avi Shchislowski , Bean Huo , cang@codeaurora.org, stanley.chu@mediatek.com, Avri Altman Subject: [PATCH v12 04/12] scsi: ufshpb: Add reads counter Date: Mon, 12 Jul 2021 12:50:31 +0300 Message-Id: <20210712095039.8093-5-avri.altman@wdc.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20210712095039.8093-1-avri.altman@wdc.com> References: <20210712095039.8093-1-avri.altman@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org In host control mode, reads are the major source of activation trials. Keep track of those reads counters, for both active as well inactive regions. We reset the read counter upon write - we are only interested in "clean" reads. Keep those counters normalized, as we are using those reads as a comparative score, to make various decisions. If during consecutive normalizations an active region has exhaust its reads - inactivate it. while at it, protect the {active,inactive}_count stats by adding them into the applicable handler. Signed-off-by: Avri Altman Reviewed-by: Daejun Park --- drivers/scsi/ufs/ufshpb.c | 94 ++++++++++++++++++++++++++++++++++++--- drivers/scsi/ufs/ufshpb.h | 9 ++++ 2 files changed, 97 insertions(+), 6 deletions(-) diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c index 3663f427dff6..d46235527fba 100644 --- a/drivers/scsi/ufs/ufshpb.c +++ b/drivers/scsi/ufs/ufshpb.c @@ -16,6 +16,8 @@ #include "ufshpb.h" #include "../sd.h" +#define ACTIVATION_THRESHOLD 8 /* 8 IOs */ + /* memory management */ static struct kmem_cache *ufshpb_mctx_cache; static mempool_t *ufshpb_mctx_pool; @@ -26,6 +28,9 @@ static int tot_active_srgn_pages; static struct workqueue_struct *ufshpb_wq; +static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int rgn_idx, + int srgn_idx); + bool ufshpb_is_allowed(struct ufs_hba *hba) { return !(hba->ufshpb_dev.hpb_disabled); @@ -149,7 +154,7 @@ static void ufshpb_iterate_rgn(struct ufshpb_lu *hpb, int rgn_idx, int srgn_idx, int srgn_offset, int cnt, bool set_dirty) { struct ufshpb_region *rgn; - struct ufshpb_subregion *srgn; + struct ufshpb_subregion *srgn, *prev_srgn = NULL; int set_bit_len; int bitmap_len; unsigned long flags; @@ -168,15 +173,39 @@ static void ufshpb_iterate_rgn(struct ufshpb_lu *hpb, int rgn_idx, int srgn_idx, else set_bit_len = cnt; - if (set_dirty) - set_bit(RGN_FLAG_DIRTY, &rgn->rgn_flags); - spin_lock_irqsave(&hpb->rgn_state_lock, flags); if (set_dirty && rgn->rgn_state != HPB_RGN_INACTIVE && srgn->srgn_state == HPB_SRGN_VALID) bitmap_set(srgn->mctx->ppn_dirty, srgn_offset, set_bit_len); spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); + if (hpb->is_hcm && prev_srgn != srgn) { + bool activate = false; + + spin_lock(&rgn->rgn_lock); + if (set_dirty) { + rgn->reads -= srgn->reads; + srgn->reads = 0; + set_bit(RGN_FLAG_DIRTY, &rgn->rgn_flags); + } else { + srgn->reads++; + rgn->reads++; + if (srgn->reads == ACTIVATION_THRESHOLD) + activate = true; + } + spin_unlock(&rgn->rgn_lock); + + if (activate) { + spin_lock_irqsave(&hpb->rsp_list_lock, flags); + ufshpb_update_active_info(hpb, rgn_idx, srgn_idx); + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, + "activate region %d-%d\n", rgn_idx, srgn_idx); + } + + prev_srgn = srgn; + } + srgn_offset = 0; if (++srgn_idx == hpb->srgns_per_rgn) { srgn_idx = 0; @@ -605,6 +634,19 @@ int ufshpb_prep(struct ufs_hba *hba, struct ufshcd_lrb *lrbp) WARN_ON_ONCE(transfer_len > HPB_MULTI_CHUNK_HIGH); + if (hpb->is_hcm) { + /* + * in host control mode, reads are the main source for + * activation trials. + */ + ufshpb_iterate_rgn(hpb, rgn_idx, srgn_idx, srgn_offset, + transfer_len, false); + + /* keep those counters normalized */ + if (rgn->reads > hpb->entries_per_srgn) + schedule_work(&hpb->ufshpb_normalization_work); + } + spin_lock_irqsave(&hpb->rgn_state_lock, flags); if (ufshpb_test_ppn_dirty(hpb, rgn_idx, srgn_idx, srgn_offset, transfer_len)) { @@ -756,6 +798,8 @@ static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int rgn_idx, if (list_empty(&srgn->list_act_srgn)) list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); + + hpb->stats.rb_active_cnt++; } static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int rgn_idx) @@ -771,6 +815,8 @@ static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int rgn_idx) if (list_empty(&rgn->list_inact_rgn)) list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn); + + hpb->stats.rb_inactive_cnt++; } static void ufshpb_activate_subregion(struct ufshpb_lu *hpb, @@ -1084,6 +1130,7 @@ static int ufshpb_evict_region(struct ufshpb_lu *hpb, struct ufshpb_region *rgn) rgn->rgn_idx); goto out; } + if (!list_empty(&rgn->list_lru_rgn)) { if (ufshpb_check_srgns_issue_state(hpb, rgn)) { ret = -EBUSY; @@ -1278,7 +1325,6 @@ static void ufshpb_rsp_req_region_update(struct ufshpb_lu *hpb, if (srgn->srgn_state == HPB_SRGN_VALID) srgn->srgn_state = HPB_SRGN_INVALID; spin_unlock(&hpb->rgn_state_lock); - hpb->stats.rb_active_cnt++; } if (hpb->is_hcm) { @@ -1310,7 +1356,6 @@ static void ufshpb_rsp_req_region_update(struct ufshpb_lu *hpb, } spin_unlock(&hpb->rgn_state_lock); - hpb->stats.rb_inactive_cnt++; } out: @@ -1509,6 +1554,36 @@ static void ufshpb_run_inactive_region_list(struct ufshpb_lu *hpb) spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); } +static void ufshpb_normalization_work_handler(struct work_struct *work) +{ + struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu, + ufshpb_normalization_work); + int rgn_idx; + + for (rgn_idx = 0; rgn_idx < hpb->rgns_per_lu; rgn_idx++) { + struct ufshpb_region *rgn = hpb->rgn_tbl + rgn_idx; + int srgn_idx; + + spin_lock(&rgn->rgn_lock); + rgn->reads = 0; + for (srgn_idx = 0; srgn_idx < hpb->srgns_per_rgn; srgn_idx++) { + struct ufshpb_subregion *srgn = rgn->srgn_tbl + srgn_idx; + + srgn->reads >>= 1; + rgn->reads += srgn->reads; + } + spin_unlock(&rgn->rgn_lock); + + if (rgn->rgn_state != HPB_RGN_ACTIVE || rgn->reads) + continue; + + /* if region is active but has no reads - inactivate it */ + spin_lock(&hpb->rsp_list_lock); + ufshpb_update_inactive_info(hpb, rgn->rgn_idx); + spin_unlock(&hpb->rsp_list_lock); + } +} + static void ufshpb_map_work_handler(struct work_struct *work) { struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu, map_work); @@ -1667,6 +1742,8 @@ static int ufshpb_alloc_region_tbl(struct ufs_hba *hba, struct ufshpb_lu *hpb) rgn = rgn_table + rgn_idx; rgn->rgn_idx = rgn_idx; + spin_lock_init(&rgn->rgn_lock); + INIT_LIST_HEAD(&rgn->list_inact_rgn); INIT_LIST_HEAD(&rgn->list_lru_rgn); @@ -1908,6 +1985,9 @@ static int ufshpb_lu_hpb_init(struct ufs_hba *hba, struct ufshpb_lu *hpb) INIT_LIST_HEAD(&hpb->list_hpb_lu); INIT_WORK(&hpb->map_work, ufshpb_map_work_handler); + if (hpb->is_hcm) + INIT_WORK(&hpb->ufshpb_normalization_work, + ufshpb_normalization_work_handler); hpb->map_req_cache = kmem_cache_create("ufshpb_req_cache", sizeof(struct ufshpb_req), 0, 0, NULL); @@ -2007,6 +2087,8 @@ static void ufshpb_discard_rsp_lists(struct ufshpb_lu *hpb) static void ufshpb_cancel_jobs(struct ufshpb_lu *hpb) { + if (hpb->is_hcm) + cancel_work_sync(&hpb->ufshpb_normalization_work); cancel_work_sync(&hpb->map_work); } diff --git a/drivers/scsi/ufs/ufshpb.h b/drivers/scsi/ufs/ufshpb.h index 9ab502f82835..33d163e76d41 100644 --- a/drivers/scsi/ufs/ufshpb.h +++ b/drivers/scsi/ufs/ufshpb.h @@ -106,6 +106,10 @@ struct ufshpb_subregion { int rgn_idx; int srgn_idx; bool is_last; + + /* subregion reads - for host mode */ + unsigned int reads; + /* below information is used by rsp_list */ struct list_head list_act_srgn; }; @@ -123,6 +127,10 @@ struct ufshpb_region { struct list_head list_lru_rgn; unsigned long rgn_flags; #define RGN_FLAG_DIRTY 0 + + /* region reads - for host mode */ + spinlock_t rgn_lock; + unsigned int reads; }; #define for_each_sub_region(rgn, i, srgn) \ @@ -212,6 +220,7 @@ struct ufshpb_lu { /* for selecting victim */ struct victim_select_info lru_info; + struct work_struct ufshpb_normalization_work; /* pinned region information */ u32 lu_pinned_start; -- 2.25.1