From: Can Guo <cang@codeaurora.org>
To: Avri Altman <avri.altman@wdc.com>
Cc: "James E . J . Bottomley" <jejb@linux.vnet.ibm.com>,
"Martin K . Petersen" <martin.petersen@oracle.com>,
linux-scsi@vger.kernel.org, linux-kernel@vger.kernel.org,
gregkh@linuxfoundation.org, Bart Van Assche <bvanassche@acm.org>,
yongmyung lee <ymhungry.lee@samsung.com>,
Daejun Park <daejun7.park@samsung.com>,
alim.akhtar@samsung.com, asutoshd@codeaurora.org,
Zang Leigang <zangleigang@hisilicon.com>,
Avi Shchislowski <avi.shchislowski@wdc.com>,
Bean Huo <beanhuo@micron.com>,
stanley.chu@mediatek.com
Subject: Re: [PATCH v6 03/10] scsi: ufshpb: Add region's reads counter
Date: Wed, 24 Mar 2021 17:25:25 +0800 [thread overview]
Message-ID: <48758404e172e8faca07c3fab8a3bd5c@codeaurora.org> (raw)
In-Reply-To: <20210322081044.62003-4-avri.altman@wdc.com>
On 2021-03-22 16:10, Avri Altman wrote:
> In host control mode, reads are the major source of activation trials.
> Keep track of those reads counters, for both active as well inactive
> regions.
>
> We reset the read counter upon write - we are only interested in
> "clean"
> reads.
>
> Keep those counters normalized, as we are using those reads as a
> comparative score, to make various decisions.
> If during consecutive normalizations an active region has exhaust its
> reads - inactivate it.
>
> while at it, protect the {active,inactive}_count stats by adding them
> into the applicable handler.
>
> Signed-off-by: Avri Altman <avri.altman@wdc.com>
> ---
> drivers/scsi/ufs/ufshpb.c | 100 +++++++++++++++++++++++++++++++-------
> drivers/scsi/ufs/ufshpb.h | 5 ++
> 2 files changed, 88 insertions(+), 17 deletions(-)
>
> diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c
> index d4f0bb6d8fa1..a1519cbb4ce0 100644
> --- a/drivers/scsi/ufs/ufshpb.c
> +++ b/drivers/scsi/ufs/ufshpb.c
> @@ -16,6 +16,8 @@
> #include "ufshpb.h"
> #include "../sd.h"
>
> +#define ACTIVATION_THRESHOLD 8 /* 8 IOs */
> +
> /* memory management */
> static struct kmem_cache *ufshpb_mctx_cache;
> static mempool_t *ufshpb_mctx_pool;
> @@ -546,6 +548,23 @@ static int ufshpb_issue_pre_req(struct ufshpb_lu
> *hpb, struct scsi_cmnd *cmd,
> return ret;
> }
>
> +static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int
> rgn_idx,
> + int srgn_idx)
> +{
> + struct ufshpb_region *rgn;
> + struct ufshpb_subregion *srgn;
> +
> + rgn = hpb->rgn_tbl + rgn_idx;
> + srgn = rgn->srgn_tbl + srgn_idx;
> +
> + list_del_init(&rgn->list_inact_rgn);
> +
> + if (list_empty(&srgn->list_act_srgn))
> + list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn);
> +
> + hpb->stats.rb_active_cnt++;
> +}
> +
> /*
> * This function will set up HPB read command using host-side L2P map
> data.
> */
> @@ -596,12 +615,43 @@ int ufshpb_prep(struct ufs_hba *hba, struct
> ufshcd_lrb *lrbp)
> ufshpb_set_ppn_dirty(hpb, rgn_idx, srgn_idx, srgn_offset,
> transfer_len);
> spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
> +
> + if (hpb->is_hcm) {
> + spin_lock(&rgn->rgn_lock);
> + rgn->reads = 0;
> + spin_unlock(&rgn->rgn_lock);
> + }
> +
> return 0;
> }
>
> if (!ufshpb_is_support_chunk(hpb, transfer_len))
> return 0;
>
> + if (hpb->is_hcm) {
> + bool activate = false;
> + /*
> + * in host control mode, reads are the main source for
> + * activation trials.
> + */
> + spin_lock(&rgn->rgn_lock);
> + rgn->reads++;
> + if (rgn->reads == ACTIVATION_THRESHOLD)
> + activate = true;
> + spin_unlock(&rgn->rgn_lock);
> + if (activate) {
> + spin_lock_irqsave(&hpb->rsp_list_lock, flags);
> + ufshpb_update_active_info(hpb, rgn_idx, srgn_idx);
If a transfer_len (possible with HPB2.0) sits accross two
regions/sub-regions,
here it only updates active info of the first region/sub-region.
Thanks,
Can Guo.
> + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
> + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev,
> + "activate region %d-%d\n", rgn_idx, srgn_idx);
> + }
> +
> + /* keep those counters normalized */
> + if (rgn->reads > hpb->entries_per_srgn)
> + schedule_work(&hpb->ufshpb_normalization_work);
> + }
> +
> spin_lock_irqsave(&hpb->rgn_state_lock, flags);
> if (ufshpb_test_ppn_dirty(hpb, rgn_idx, srgn_idx, srgn_offset,
> transfer_len)) {
> @@ -741,21 +791,6 @@ static int ufshpb_clear_dirty_bitmap(struct
> ufshpb_lu *hpb,
> return 0;
> }
>
> -static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int
> rgn_idx,
> - int srgn_idx)
> -{
> - struct ufshpb_region *rgn;
> - struct ufshpb_subregion *srgn;
> -
> - rgn = hpb->rgn_tbl + rgn_idx;
> - srgn = rgn->srgn_tbl + srgn_idx;
> -
> - list_del_init(&rgn->list_inact_rgn);
> -
> - if (list_empty(&srgn->list_act_srgn))
> - list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn);
> -}
> -
> static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int
> rgn_idx)
> {
> struct ufshpb_region *rgn;
> @@ -769,6 +804,8 @@ static void ufshpb_update_inactive_info(struct
> ufshpb_lu *hpb, int rgn_idx)
>
> if (list_empty(&rgn->list_inact_rgn))
> list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn);
> +
> + hpb->stats.rb_inactive_cnt++;
> }
>
> static void ufshpb_activate_subregion(struct ufshpb_lu *hpb,
> @@ -1089,6 +1126,7 @@ static int ufshpb_evict_region(struct ufshpb_lu
> *hpb, struct ufshpb_region *rgn)
> rgn->rgn_idx);
> goto out;
> }
> +
> if (!list_empty(&rgn->list_lru_rgn)) {
> if (ufshpb_check_srgns_issue_state(hpb, rgn)) {
> ret = -EBUSY;
> @@ -1283,7 +1321,6 @@ static void ufshpb_rsp_req_region_update(struct
> ufshpb_lu *hpb,
> if (srgn->srgn_state == HPB_SRGN_VALID)
> srgn->srgn_state = HPB_SRGN_INVALID;
> spin_unlock(&hpb->rgn_state_lock);
> - hpb->stats.rb_active_cnt++;
> }
>
> if (hpb->is_hcm) {
> @@ -1315,7 +1352,6 @@ static void ufshpb_rsp_req_region_update(struct
> ufshpb_lu *hpb,
> }
> spin_unlock(&hpb->rgn_state_lock);
>
> - hpb->stats.rb_inactive_cnt++;
> }
>
> out:
> @@ -1514,6 +1550,29 @@ static void
> ufshpb_run_inactive_region_list(struct ufshpb_lu *hpb)
> spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
> }
>
> +static void ufshpb_normalization_work_handler(struct work_struct
> *work)
> +{
> + struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu,
> + ufshpb_normalization_work);
> + int rgn_idx;
> +
> + for (rgn_idx = 0; rgn_idx < hpb->rgns_per_lu; rgn_idx++) {
> + struct ufshpb_region *rgn = hpb->rgn_tbl + rgn_idx;
> +
> + spin_lock(&rgn->rgn_lock);
> + rgn->reads = (rgn->reads >> 1);
> + spin_unlock(&rgn->rgn_lock);
> +
> + if (rgn->rgn_state != HPB_RGN_ACTIVE || rgn->reads)
> + continue;
> +
> + /* if region is active but has no reads - inactivate it */
> + spin_lock(&hpb->rsp_list_lock);
> + ufshpb_update_inactive_info(hpb, rgn->rgn_idx);
> + spin_unlock(&hpb->rsp_list_lock);
> + }
> +}
> +
> static void ufshpb_map_work_handler(struct work_struct *work)
> {
> struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu,
> map_work);
> @@ -1673,6 +1732,8 @@ static int ufshpb_alloc_region_tbl(struct
> ufs_hba *hba, struct ufshpb_lu *hpb)
> rgn = rgn_table + rgn_idx;
> rgn->rgn_idx = rgn_idx;
>
> + spin_lock_init(&rgn->rgn_lock);
> +
> INIT_LIST_HEAD(&rgn->list_inact_rgn);
> INIT_LIST_HEAD(&rgn->list_lru_rgn);
>
> @@ -1914,6 +1975,9 @@ static int ufshpb_lu_hpb_init(struct ufs_hba
> *hba, struct ufshpb_lu *hpb)
> INIT_LIST_HEAD(&hpb->list_hpb_lu);
>
> INIT_WORK(&hpb->map_work, ufshpb_map_work_handler);
> + if (hpb->is_hcm)
> + INIT_WORK(&hpb->ufshpb_normalization_work,
> + ufshpb_normalization_work_handler);
>
> hpb->map_req_cache = kmem_cache_create("ufshpb_req_cache",
> sizeof(struct ufshpb_req), 0, 0, NULL);
> @@ -2013,6 +2077,8 @@ static void ufshpb_discard_rsp_lists(struct
> ufshpb_lu *hpb)
>
> static void ufshpb_cancel_jobs(struct ufshpb_lu *hpb)
> {
> + if (hpb->is_hcm)
> + cancel_work_sync(&hpb->ufshpb_normalization_work);
> cancel_work_sync(&hpb->map_work);
> }
>
> diff --git a/drivers/scsi/ufs/ufshpb.h b/drivers/scsi/ufs/ufshpb.h
> index 032672114881..32d72c46c57a 100644
> --- a/drivers/scsi/ufs/ufshpb.h
> +++ b/drivers/scsi/ufs/ufshpb.h
> @@ -123,6 +123,10 @@ struct ufshpb_region {
> struct list_head list_lru_rgn;
> unsigned long rgn_flags;
> #define RGN_FLAG_DIRTY 0
> +
> + /* region reads - for host mode */
> + spinlock_t rgn_lock;
> + unsigned int reads;
> };
>
> #define for_each_sub_region(rgn, i, srgn) \
> @@ -212,6 +216,7 @@ struct ufshpb_lu {
>
> /* for selecting victim */
> struct victim_select_info lru_info;
> + struct work_struct ufshpb_normalization_work;
>
> /* pinned region information */
> u32 lu_pinned_start;
next prev parent reply other threads:[~2021-03-24 9:26 UTC|newest]
Thread overview: 19+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-03-22 8:10 [PATCH v6 00/10] Add Host control mode to HPB Avri Altman
2021-03-22 8:10 ` [PATCH v6 01/10] scsi: ufshpb: Cache HPB Control mode on init Avri Altman
2021-03-22 8:10 ` [PATCH v6 02/10] scsi: ufshpb: Add host control mode support to rsp_upiu Avri Altman
2021-03-24 3:31 ` Zang Leigang
2021-03-24 3:59 ` Can Guo
2021-03-24 12:44 ` Avri Altman
2021-03-22 8:10 ` [PATCH v6 03/10] scsi: ufshpb: Add region's reads counter Avri Altman
2021-03-24 9:25 ` Can Guo [this message]
2021-03-24 11:28 ` Avri Altman
2021-03-27 8:02 ` Avri Altman
2021-03-22 8:10 ` [PATCH v6 04/10] scsi: ufshpb: Make eviction depends on region's reads Avri Altman
2021-03-25 5:34 ` Can Guo
2021-03-25 5:56 ` Avri Altman
2021-03-22 8:10 ` [PATCH v6 05/10] scsi: ufshpb: Region inactivation in host mode Avri Altman
2021-03-22 8:10 ` [PATCH v6 06/10] scsi: ufshpb: Add hpb dev reset response Avri Altman
2021-03-22 8:10 ` [PATCH v6 07/10] scsi: ufshpb: Add "Cold" regions timer Avri Altman
2021-03-22 8:10 ` [PATCH v6 08/10] scsi: ufshpb: Limit the number of inflight map requests Avri Altman
2021-03-22 8:10 ` [PATCH v6 09/10] scsi: ufshpb: Add support for host control mode Avri Altman
2021-03-22 8:10 ` [PATCH v6 10/10] scsi: ufshpb: Make host mode parameters configurable Avri Altman
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=48758404e172e8faca07c3fab8a3bd5c@codeaurora.org \
--to=cang@codeaurora.org \
--cc=alim.akhtar@samsung.com \
--cc=asutoshd@codeaurora.org \
--cc=avi.shchislowski@wdc.com \
--cc=avri.altman@wdc.com \
--cc=beanhuo@micron.com \
--cc=bvanassche@acm.org \
--cc=daejun7.park@samsung.com \
--cc=gregkh@linuxfoundation.org \
--cc=jejb@linux.vnet.ibm.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-scsi@vger.kernel.org \
--cc=martin.petersen@oracle.com \
--cc=stanley.chu@mediatek.com \
--cc=ymhungry.lee@samsung.com \
--cc=zangleigang@hisilicon.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox