From: Konrad Dybcio <konrad.dybcio@linaro.org>
To: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>,
Rob Clark <robdclark@gmail.com>, Sean Paul <sean@poorly.run>,
Abhinav Kumar <quic_abhinavk@quicinc.com>
Cc: Marijn Suijten <marijn.suijten@somainline.org>,
Stephen Boyd <swboyd@chromium.org>,
David Airlie <airlied@gmail.com>, Daniel Vetter <daniel@ffwll.ch>,
Bjorn Andersson <andersson@kernel.org>,
linux-arm-msm@vger.kernel.org, dri-devel@lists.freedesktop.org,
freedreno@lists.freedesktop.org
Subject: Re: [PATCH 6/8] drm/msm/dpu: use dpu_perf_cfg in DPU core_perf code
Date: Tue, 20 Jun 2023 12:55:16 +0200 [thread overview]
Message-ID: <bc192a59-f8a1-2786-670c-cd56737dfdba@linaro.org> (raw)
In-Reply-To: <20230620000846.946925-7-dmitry.baryshkov@linaro.org>
On 20.06.2023 02:08, Dmitry Baryshkov wrote:
> Simplify dpu_core_perf code by using only dpu_perf_cfg instead of using
> full-featured catalog data.
>
> Signed-off-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
> ---
Acked-by: Konrad Dybcio <konrad.dybcio@linaro.org>
Check below.
> drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c | 52 ++++++++-----------
> drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.h | 8 +--
> drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c | 2 +-
> 3 files changed, 27 insertions(+), 35 deletions(-)
>
> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c b/drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c
> index 773e641eab28..78a7e3ea27a4 100644
> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c
> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c
> @@ -19,11 +19,11 @@
>
> /**
> * _dpu_core_perf_calc_bw() - to calculate BW per crtc
> - * @kms: pointer to the dpu_kms
> + * @perf_cfg: performance configuration
> * @crtc: pointer to a crtc
> * Return: returns aggregated BW for all planes in crtc.
> */
> -static u64 _dpu_core_perf_calc_bw(struct dpu_kms *kms,
> +static u64 _dpu_core_perf_calc_bw(const struct dpu_perf_cfg *perf_cfg,
> struct drm_crtc *crtc)
> {
> struct drm_plane *plane;
> @@ -39,7 +39,7 @@ static u64 _dpu_core_perf_calc_bw(struct dpu_kms *kms,
> crtc_plane_bw += pstate->plane_fetch_bw;
> }
>
> - bw_factor = kms->catalog->perf->bw_inefficiency_factor;
> + bw_factor = perf_cfg->bw_inefficiency_factor;
It's set to 120 for all SoCs.. and it sounds very much like some kind of a
hack.
The 105 on the other inefficiency factor is easy to spot:
(1024/1000)^2 = 1.048576 =~= 1.05 = 105%
It comes from a MiB-MB-MHz conversion that Qcom splattered all over
downstream as due to ancient tragical design decisions in msmbus
(which leak to the downstream interconnect a bit):
The logic needs to get some input that corresponds to a clock rate
of a bus clock (19.2, 200, 300 Mhz etc.) but the APIs expect a Kbps
value. So at one point they invented a MHZ_TO_MBPS macro which did this
conversion the other way around and probably had to account for it.
I think they tried to make it make more sense, but it ended up being
even more spaghetti :/
Not yet sure how it's done on RPMh icc, but with SMD RPM, passing e.g.
opp-peak-kBps = <(200 * 8 * 1000)>; # 200 MHz * 8-wide * KHz-to-MHz
results in a "correct" end rate.
Konrad
> if (bw_factor) {
> crtc_plane_bw *= bw_factor;
> do_div(crtc_plane_bw, 100);
> @@ -50,12 +50,12 @@ static u64 _dpu_core_perf_calc_bw(struct dpu_kms *kms,
>
> /**
> * _dpu_core_perf_calc_clk() - to calculate clock per crtc
> - * @kms: pointer to the dpu_kms
> + * @perf_cfg: performance configuration
> * @crtc: pointer to a crtc
> * @state: pointer to a crtc state
> * Return: returns max clk for all planes in crtc.
> */
> -static u64 _dpu_core_perf_calc_clk(struct dpu_kms *kms,
> +static u64 _dpu_core_perf_calc_clk(const struct dpu_perf_cfg *perf_cfg,
> struct drm_crtc *crtc, struct drm_crtc_state *state)
> {
> struct drm_plane *plane;
> @@ -76,7 +76,7 @@ static u64 _dpu_core_perf_calc_clk(struct dpu_kms *kms,
> crtc_clk = max(pstate->plane_clk, crtc_clk);
> }
>
> - clk_factor = kms->catalog->perf->clk_inefficiency_factor;
> + clk_factor = perf_cfg->clk_inefficiency_factor;
> if (clk_factor) {
> crtc_clk *= clk_factor;
> do_div(crtc_clk, 100);
> @@ -92,20 +92,20 @@ static struct dpu_kms *_dpu_crtc_get_kms(struct drm_crtc *crtc)
> return to_dpu_kms(priv->kms);
> }
>
> -static void _dpu_core_perf_calc_crtc(struct dpu_kms *kms,
> +static void _dpu_core_perf_calc_crtc(const struct dpu_perf_cfg *perf_cfg,
> struct drm_crtc *crtc,
> struct drm_crtc_state *state,
> struct dpu_core_perf_params *perf)
> {
> - if (!kms || !kms->catalog || !crtc || !state || !perf) {
> + if (!perf_cfg || !crtc || !state || !perf) {
> DPU_ERROR("invalid parameters\n");
> return;
> }
>
> memset(perf, 0, sizeof(struct dpu_core_perf_params));
>
> - perf->bw_ctl = _dpu_core_perf_calc_bw(kms, crtc);
> - perf->core_clk_rate = _dpu_core_perf_calc_clk(kms, crtc, state);
> + perf->bw_ctl = _dpu_core_perf_calc_bw(perf_cfg, crtc);
> + perf->core_clk_rate = _dpu_core_perf_calc_clk(perf_cfg, crtc, state);
>
> DRM_DEBUG_ATOMIC(
> "crtc=%d clk_rate=%llu core_ab=%llu\n",
> @@ -122,6 +122,7 @@ int dpu_core_perf_crtc_check(struct drm_crtc *crtc,
> struct dpu_crtc_state *dpu_cstate;
> struct drm_crtc *tmp_crtc;
> struct dpu_kms *kms;
> + const struct dpu_perf_cfg *perf_cfg;
>
> if (!crtc || !state) {
> DPU_ERROR("invalid crtc\n");
> @@ -129,10 +130,7 @@ int dpu_core_perf_crtc_check(struct drm_crtc *crtc,
> }
>
> kms = _dpu_crtc_get_kms(crtc);
> - if (!kms->catalog) {
> - DPU_ERROR("invalid parameters\n");
> - return 0;
> - }
> + perf_cfg = kms->perf.perf_cfg;
>
> /* we only need bandwidth check on real-time clients (interfaces) */
> if (dpu_crtc_get_client_type(crtc) == NRT_CLIENT)
> @@ -141,7 +139,7 @@ int dpu_core_perf_crtc_check(struct drm_crtc *crtc,
> dpu_cstate = to_dpu_crtc_state(state);
>
> /* obtain new values */
> - _dpu_core_perf_calc_crtc(kms, crtc, state, &dpu_cstate->new_perf);
> + _dpu_core_perf_calc_crtc(perf_cfg, crtc, state, &dpu_cstate->new_perf);
>
> bw_sum_of_intfs = dpu_cstate->new_perf.bw_ctl;
> curr_client_type = dpu_crtc_get_client_type(crtc);
> @@ -164,7 +162,7 @@ int dpu_core_perf_crtc_check(struct drm_crtc *crtc,
> bw = DIV_ROUND_UP_ULL(bw_sum_of_intfs, 1000);
> DRM_DEBUG_ATOMIC("calculated bandwidth=%uk\n", bw);
>
> - threshold = kms->catalog->perf->max_bw_high;
> + threshold = perf_cfg->max_bw_high;
>
> DRM_DEBUG_ATOMIC("final threshold bw limit = %d\n", threshold);
>
> @@ -212,7 +210,7 @@ static int _dpu_core_perf_crtc_update_bus(struct dpu_kms *kms,
> do_div(avg_bw, (kms->num_paths * 1000)); /*Bps_to_icc*/
>
> for (i = 0; i < kms->num_paths; i++)
> - icc_set_bw(kms->path[i], avg_bw, kms->catalog->perf->min_dram_ib);
> + icc_set_bw(kms->path[i], avg_bw, kms->perf.perf_cfg->min_dram_ib);
>
> return ret;
> }
> @@ -293,10 +291,6 @@ int dpu_core_perf_crtc_update(struct drm_crtc *crtc,
> }
>
> kms = _dpu_crtc_get_kms(crtc);
> - if (!kms->catalog) {
> - DPU_ERROR("invalid kms\n");
> - return -EINVAL;
> - }
>
> dpu_crtc = to_dpu_crtc(crtc);
> dpu_cstate = to_dpu_crtc_state(crtc->state);
> @@ -375,7 +369,6 @@ int dpu_core_perf_crtc_update(struct drm_crtc *crtc,
> int dpu_core_perf_debugfs_init(struct dpu_kms *dpu_kms, struct dentry *parent)
> {
> struct dpu_core_perf *perf = &dpu_kms->perf;
> - const struct dpu_mdss_cfg *catalog = perf->catalog;
> struct dentry *entry;
>
> entry = debugfs_create_dir("core_perf", parent);
> @@ -387,15 +380,15 @@ int dpu_core_perf_debugfs_init(struct dpu_kms *dpu_kms, struct dentry *parent)
> debugfs_create_u32("enable_bw_release", 0600, entry,
> (u32 *)&perf->enable_bw_release);
> debugfs_create_u32("threshold_low", 0600, entry,
> - (u32 *)&catalog->perf->max_bw_low);
> + (u32 *)&perf->perf_cfg->max_bw_low);
> debugfs_create_u32("threshold_high", 0600, entry,
> - (u32 *)&catalog->perf->max_bw_high);
> + (u32 *)&perf->perf_cfg->max_bw_high);
> debugfs_create_u32("min_core_ib", 0600, entry,
> - (u32 *)&catalog->perf->min_core_ib);
> + (u32 *)&perf->perf_cfg->min_core_ib);
> debugfs_create_u32("min_llcc_ib", 0600, entry,
> - (u32 *)&catalog->perf->min_llcc_ib);
> + (u32 *)&perf->perf_cfg->min_llcc_ib);
> debugfs_create_u32("min_dram_ib", 0600, entry,
> - (u32 *)&catalog->perf->min_dram_ib);
> + (u32 *)&perf->perf_cfg->min_dram_ib);
>
> return 0;
> }
> @@ -410,17 +403,16 @@ void dpu_core_perf_destroy(struct dpu_core_perf *perf)
>
> perf->max_core_clk_rate = 0;
> perf->core_clk = NULL;
> - perf->catalog = NULL;
> perf->dev = NULL;
> }
>
> int dpu_core_perf_init(struct dpu_core_perf *perf,
> struct drm_device *dev,
> - const struct dpu_mdss_cfg *catalog,
> + const struct dpu_perf_cfg *perf_cfg,
> struct clk *core_clk)
> {
> perf->dev = dev;
> - perf->catalog = catalog;
> + perf->perf_cfg = perf_cfg;
> perf->core_clk = core_clk;
>
> perf->max_core_clk_rate = clk_get_rate(core_clk);
> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.h b/drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.h
> index c29ec72984b8..e8a7916b6f71 100644
> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.h
> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.h
> @@ -28,7 +28,7 @@ struct dpu_core_perf_params {
> * struct dpu_core_perf - definition of core performance context
> * @dev: Pointer to drm device
> * @debugfs_root: top level debug folder
> - * @catalog: Pointer to catalog configuration
> + * @perf_cfg: Platform-specific performance configuration
> * @core_clk: Pointer to the core clock
> * @core_clk_rate: current core clock rate
> * @max_core_clk_rate: maximum allowable core clock rate
> @@ -38,7 +38,7 @@ struct dpu_core_perf_params {
> struct dpu_core_perf {
> struct drm_device *dev;
> struct dentry *debugfs_root;
> - const struct dpu_mdss_cfg *catalog;
> + const struct dpu_perf_cfg *perf_cfg;
> struct clk *core_clk;
> u64 core_clk_rate;
> u64 max_core_clk_rate;
> @@ -79,12 +79,12 @@ void dpu_core_perf_destroy(struct dpu_core_perf *perf);
> * dpu_core_perf_init - initialize the given core performance context
> * @perf: Pointer to core performance context
> * @dev: Pointer to drm device
> - * @catalog: Pointer to catalog
> + * @perf_cfg: Pointer to platform performance configuration
> * @core_clk: pointer to core clock
> */
> int dpu_core_perf_init(struct dpu_core_perf *perf,
> struct drm_device *dev,
> - const struct dpu_mdss_cfg *catalog,
> + const struct dpu_perf_cfg *perf_cfg,
> struct clk *core_clk);
>
> struct dpu_kms;
> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
> index aa8499de1b9f..6e62606e32de 100644
> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
> @@ -1115,7 +1115,7 @@ static int dpu_kms_hw_init(struct msm_kms *kms)
> dpu_kms->hw_vbif[vbif->id] = hw;
> }
>
> - rc = dpu_core_perf_init(&dpu_kms->perf, dev, dpu_kms->catalog,
> + rc = dpu_core_perf_init(&dpu_kms->perf, dev, dpu_kms->catalog->perf,
> msm_clk_bulk_get_clock(dpu_kms->clocks, dpu_kms->num_clocks, "core"));
> if (rc) {
> DPU_ERROR("failed to init perf %d\n", rc);
next prev parent reply other threads:[~2023-06-20 10:55 UTC|newest]
Thread overview: 39+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-06-20 0:08 [PATCH 0/8] drm/msm/dpu: cleanup dpu_core_perf module Dmitry Baryshkov
2023-06-20 0:08 ` [PATCH 1/8] drm/msm/dpu: drop enum dpu_core_perf_data_bus_id Dmitry Baryshkov
2023-06-20 10:35 ` Konrad Dybcio
2023-07-03 21:30 ` Abhinav Kumar
2023-06-20 0:08 ` [PATCH 2/8] drm/msm/dpu: drop performance tuning modes Dmitry Baryshkov
2023-06-20 10:45 ` Konrad Dybcio
2023-07-03 21:40 ` Abhinav Kumar
2023-07-03 22:20 ` Dmitry Baryshkov
2023-07-03 22:26 ` Abhinav Kumar
2023-07-03 22:36 ` Dmitry Baryshkov
2023-06-20 0:08 ` [PATCH 3/8] drm/msm/dpu: drop dpu_core_perf_params::max_per_pipe_ib Dmitry Baryshkov
2023-06-20 10:46 ` Konrad Dybcio
2023-06-20 10:53 ` Dmitry Baryshkov
2023-07-03 21:53 ` Abhinav Kumar
2023-06-20 0:08 ` [PATCH 4/8] drm/msm/dpu: rework indentation in dpu_core_perf Dmitry Baryshkov
2023-06-20 10:47 ` Konrad Dybcio
2023-07-03 22:01 ` Abhinav Kumar
2023-06-20 0:08 ` [PATCH 5/8] drm/msm/dpu: drop the dpu_core_perf_crtc_update()'s stop_req param Dmitry Baryshkov
2023-07-03 22:37 ` Abhinav Kumar
2023-07-03 22:53 ` Dmitry Baryshkov
2023-07-03 22:55 ` Abhinav Kumar
2023-07-03 23:01 ` Dmitry Baryshkov
2023-07-03 23:16 ` Abhinav Kumar
2023-07-04 0:28 ` Dmitry Baryshkov
2023-07-04 0:31 ` Abhinav Kumar
2023-06-20 0:08 ` [PATCH 6/8] drm/msm/dpu: use dpu_perf_cfg in DPU core_perf code Dmitry Baryshkov
2023-06-20 10:55 ` Konrad Dybcio [this message]
2023-06-20 11:18 ` Dmitry Baryshkov
2023-06-20 11:31 ` Konrad Dybcio
2023-07-04 0:46 ` Abhinav Kumar
2023-06-20 0:08 ` [PATCH 7/8] drm/msm/dpu: drop dpu_core_perf_destroy() Dmitry Baryshkov
2023-06-20 10:56 ` Konrad Dybcio
2023-07-03 22:57 ` Abhinav Kumar
2023-07-03 22:59 ` Dmitry Baryshkov
2023-07-04 0:19 ` Abhinav Kumar
2023-07-04 14:31 ` Dmitry Baryshkov
2023-06-20 0:08 ` [PATCH 8/8] drm/msm/dpu: remove unused fields from struct dpu_core_perf Dmitry Baryshkov
2023-06-20 10:56 ` Konrad Dybcio
2023-07-04 0:25 ` Abhinav Kumar
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=bc192a59-f8a1-2786-670c-cd56737dfdba@linaro.org \
--to=konrad.dybcio@linaro.org \
--cc=airlied@gmail.com \
--cc=andersson@kernel.org \
--cc=daniel@ffwll.ch \
--cc=dmitry.baryshkov@linaro.org \
--cc=dri-devel@lists.freedesktop.org \
--cc=freedreno@lists.freedesktop.org \
--cc=linux-arm-msm@vger.kernel.org \
--cc=marijn.suijten@somainline.org \
--cc=quic_abhinavk@quicinc.com \
--cc=robdclark@gmail.com \
--cc=sean@poorly.run \
--cc=swboyd@chromium.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox