From: Imre Deak <imre.deak@intel.com>
To: Jani Nikula <jani.nikula@intel.com>
Cc: intel-gfx@lists.freedesktop.org, intel-xe@lists.freedesktop.org
Subject: Re: [PATCH 14/14] drm/i915/mst: Convert intel_dp_mtp_tu_compute_config() to .4 format
Date: Fri, 31 Jan 2025 17:46:50 +0200 [thread overview]
Message-ID: <Z5zwauaji-K48_uc@ideak-desk.fi.intel.com> (raw)
In-Reply-To: <887306a47ce4550226f5d54178f667a52840a11c.1738327620.git.jani.nikula@intel.com>
On Fri, Jan 31, 2025 at 02:50:07PM +0200, Jani Nikula wrote:
> Move towards always using the fxp q4 or .4 fixed point format for
> compressed bpp. We'll need to pass the more accurate bpp to this
> function later on.
>
> Always use _x16 naming for variables that are in .4 fixed point for
> clarity.
>
> Signed-off-by: Jani Nikula <jani.nikula@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
> ---
> drivers/gpu/drm/i915/display/intel_dp.c | 4 +-
> drivers/gpu/drm/i915/display/intel_dp_mst.c | 52 ++++++++++++---------
> drivers/gpu/drm/i915/display/intel_dp_mst.h | 2 +-
> 3 files changed, 33 insertions(+), 25 deletions(-)
>
> diff --git a/drivers/gpu/drm/i915/display/intel_dp.c b/drivers/gpu/drm/i915/display/intel_dp.c
> index 25160a5d12eb..80f550a59bcb 100644
> --- a/drivers/gpu/drm/i915/display/intel_dp.c
> +++ b/drivers/gpu/drm/i915/display/intel_dp.c
> @@ -2616,8 +2616,8 @@ intel_dp_compute_link_config(struct intel_encoder *encoder,
> ret = intel_dp_mtp_tu_compute_config(intel_dp,
> pipe_config,
> conn_state,
> - pipe_config->pipe_bpp,
> - pipe_config->pipe_bpp,
> + fxp_q4_from_int(pipe_config->pipe_bpp),
> + fxp_q4_from_int(pipe_config->pipe_bpp),
> 0, false);
> if (ret)
> dsc_needed = true;
> diff --git a/drivers/gpu/drm/i915/display/intel_dp_mst.c b/drivers/gpu/drm/i915/display/intel_dp_mst.c
> index 868d0948ca27..b729e27cdde2 100644
> --- a/drivers/gpu/drm/i915/display/intel_dp_mst.c
> +++ b/drivers/gpu/drm/i915/display/intel_dp_mst.c
> @@ -212,7 +212,7 @@ static int intel_dp_mst_dsc_get_slice_count(const struct intel_connector *connec
> int intel_dp_mtp_tu_compute_config(struct intel_dp *intel_dp,
> struct intel_crtc_state *crtc_state,
> struct drm_connector_state *conn_state,
> - int min_bpp, int max_bpp, int step, bool dsc)
> + int min_bpp_x16, int max_bpp_x16, int bpp_step_x16, bool dsc)
> {
> struct intel_display *display = to_intel_display(intel_dp);
> struct drm_atomic_state *state = crtc_state->uapi.state;
> @@ -222,9 +222,14 @@ int intel_dp_mtp_tu_compute_config(struct intel_dp *intel_dp,
> const struct drm_display_mode *adjusted_mode =
> &crtc_state->hw.adjusted_mode;
> bool is_mst = intel_dp->is_mst;
> - int bpp, slots = -EINVAL;
> + int bpp_x16, slots = -EINVAL;
> int dsc_slice_count = 0;
> - int max_dpt_bpp;
> + int max_dpt_bpp_x16;
> +
> + /* shouldn't happen, sanity check */
> + drm_WARN_ON(display->drm, !dsc && (fxp_q4_to_frac(min_bpp_x16) ||
> + fxp_q4_to_frac(max_bpp_x16) ||
> + fxp_q4_to_frac(bpp_step_x16)));
>
> if (is_mst) {
> mst_state = drm_atomic_get_mst_topology_state(state, &intel_dp->mst_mgr);
> @@ -242,15 +247,15 @@ int intel_dp_mtp_tu_compute_config(struct intel_dp *intel_dp,
> crtc_state->fec_enable = !intel_dp_is_uhbr(crtc_state);
> }
>
> - max_dpt_bpp = intel_dp_mst_max_dpt_bpp(crtc_state, dsc);
> - if (max_bpp > max_dpt_bpp) {
> - drm_dbg_kms(display->drm, "Limiting bpp to max DPT bpp (%d -> %d)\n",
> - max_bpp, max_dpt_bpp);
> - max_bpp = max_dpt_bpp;
> + max_dpt_bpp_x16 = fxp_q4_from_int(intel_dp_mst_max_dpt_bpp(crtc_state, dsc));
> + if (max_bpp_x16 > max_dpt_bpp_x16) {
> + drm_dbg_kms(display->drm, "Limiting bpp to max DPT bpp (" FXP_Q4_FMT " -> " FXP_Q4_FMT ")\n",
> + FXP_Q4_ARGS(max_bpp_x16), FXP_Q4_ARGS(max_dpt_bpp_x16));
> + max_bpp_x16 = max_dpt_bpp_x16;
> }
>
> - drm_dbg_kms(display->drm, "Looking for slots in range min bpp %d max bpp %d\n",
> - min_bpp, max_bpp);
> + drm_dbg_kms(display->drm, "Looking for slots in range min bpp " FXP_Q4_FMT " max bpp " FXP_Q4_FMT "\n",
> + FXP_Q4_ARGS(min_bpp_x16), FXP_Q4_ARGS(max_bpp_x16));
>
> if (dsc) {
> dsc_slice_count = intel_dp_mst_dsc_get_slice_count(connector, crtc_state);
> @@ -261,14 +266,15 @@ int intel_dp_mtp_tu_compute_config(struct intel_dp *intel_dp,
> }
> }
>
> - for (bpp = max_bpp; bpp >= min_bpp; bpp -= step) {
> + for (bpp_x16 = max_bpp_x16; bpp_x16 >= min_bpp_x16; bpp_x16 -= bpp_step_x16) {
> int local_bw_overhead;
> int link_bpp_x16;
>
> - drm_dbg_kms(display->drm, "Trying bpp %d\n", bpp);
> + drm_dbg_kms(display->drm, "Trying bpp " FXP_Q4_FMT "\n", FXP_Q4_ARGS(bpp_x16));
>
> - link_bpp_x16 = fxp_q4_from_int(dsc ? bpp :
> - intel_dp_output_bpp(crtc_state->output_format, bpp));
> + link_bpp_x16 = dsc ? bpp_x16 :
> + fxp_q4_from_int(intel_dp_output_bpp(crtc_state->output_format,
> + fxp_q4_to_int(bpp_x16)));
>
> local_bw_overhead = intel_dp_mst_bw_overhead(crtc_state,
> false, dsc_slice_count, link_bpp_x16);
> @@ -356,12 +362,12 @@ int intel_dp_mtp_tu_compute_config(struct intel_dp *intel_dp,
> }
>
> if (!dsc)
> - crtc_state->pipe_bpp = bpp;
> + crtc_state->pipe_bpp = fxp_q4_to_int(bpp_x16);
> else
> - crtc_state->dsc.compressed_bpp_x16 = fxp_q4_from_int(bpp);
> + crtc_state->dsc.compressed_bpp_x16 = bpp_x16;
>
> - drm_dbg_kms(display->drm, "Got %d slots for pipe bpp %d dsc %d\n",
> - slots, bpp, dsc);
> + drm_dbg_kms(display->drm, "Got %d slots for pipe bpp " FXP_Q4_FMT " dsc %d\n",
> + slots, FXP_Q4_ARGS(bpp_x16), dsc);
>
> return 0;
> }
> @@ -379,9 +385,9 @@ static int mst_stream_compute_link_config(struct intel_dp *intel_dp,
> * YUV420 is only half of the pipe bpp value.
> */
> return intel_dp_mtp_tu_compute_config(intel_dp, crtc_state, conn_state,
> - fxp_q4_to_int(limits->link.min_bpp_x16),
> - fxp_q4_to_int(limits->link.max_bpp_x16),
> - 2 * 3, false);
> + limits->link.min_bpp_x16,
> + limits->link.max_bpp_x16,
> + fxp_q4_from_int(2 * 3), false);
> }
>
> static int mst_stream_dsc_compute_link_config(struct intel_dp *intel_dp,
> @@ -435,7 +441,9 @@ static int mst_stream_dsc_compute_link_config(struct intel_dp *intel_dp,
> crtc_state->port_clock = limits->max_rate;
>
> return intel_dp_mtp_tu_compute_config(intel_dp, crtc_state, conn_state,
> - min_compressed_bpp, max_compressed_bpp, 1, true);
> + fxp_q4_from_int(min_compressed_bpp),
> + fxp_q4_from_int(max_compressed_bpp),
> + fxp_q4_from_int(1), true);
> }
>
> static int mst_stream_update_slots(struct intel_dp *intel_dp,
> diff --git a/drivers/gpu/drm/i915/display/intel_dp_mst.h b/drivers/gpu/drm/i915/display/intel_dp_mst.h
> index a713a1c10154..c1bbfeb02ca9 100644
> --- a/drivers/gpu/drm/i915/display/intel_dp_mst.h
> +++ b/drivers/gpu/drm/i915/display/intel_dp_mst.h
> @@ -34,6 +34,6 @@ bool intel_dp_mst_verify_dpcd_state(struct intel_dp *intel_dp);
> int intel_dp_mtp_tu_compute_config(struct intel_dp *intel_dp,
> struct intel_crtc_state *crtc_state,
> struct drm_connector_state *conn_state,
> - int min_bpp, int max_bpp, int step, bool dsc);
> + int min_bpp_x16, int max_bpp_x16, int bpp_step_x16, bool dsc);
>
> #endif /* __INTEL_DP_MST_H__ */
> --
> 2.39.5
>
next prev parent reply other threads:[~2025-01-31 15:45 UTC|newest]
Thread overview: 43+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-01-31 12:49 [PATCH 00/14] drm/i915/dp: dsc fix, refactoring and cleanups Jani Nikula
2025-01-31 12:49 ` [PATCH 01/14] drm/i915/dp: Iterate DSC BPP from high to low on all platforms Jani Nikula
2025-01-31 13:32 ` Imre Deak
2025-02-03 14:46 ` Jani Nikula
2025-01-31 16:13 ` Nautiyal, Ankit K
2025-01-31 12:49 ` [PATCH 02/14] drm/i915/dp: Add intel_dp_dsc_bpp_step_x16() helper to get DSC BPP precision Jani Nikula
2025-01-31 13:45 ` Imre Deak
2025-01-31 14:06 ` Jani Nikula
2025-01-31 23:28 ` [PATCH v2] " Jani Nikula
2025-01-31 12:49 ` [PATCH 03/14] drm/i915/dp: Rename some variables in xelpd_dsc_compute_link_config() Jani Nikula
2025-01-31 13:57 ` Imre Deak
2025-01-31 12:49 ` [PATCH 04/14] drm/i915/dp: Pass .4 BPP values to {icl, xelpd}_dsc_compute_link_config() Jani Nikula
2025-01-31 14:05 ` [PATCH 04/14] drm/i915/dp: Pass .4 BPP values to {icl,xelpd}_dsc_compute_link_config() Imre Deak
2025-01-31 12:49 ` [PATCH 05/14] drm/i915/dp: Move max DSC BPP reduction one level higher Jani Nikula
2025-01-31 14:26 ` Imre Deak
2025-01-31 12:49 ` [PATCH 06/14] drm/i915/dp: Change icl_dsc_compute_link_config() DSC BPP iteration Jani Nikula
2025-01-31 14:30 ` Imre Deak
2025-01-31 12:50 ` [PATCH 07/14] drm/i915/dp: Move force_dsc_fractional_bpp_en check to intel_dp_dsc_valid_bpp() Jani Nikula
2025-01-31 14:32 ` Imre Deak
2025-01-31 12:50 ` [PATCH 08/14] drm/i915/dp: Unify DSC link config functions Jani Nikula
2025-01-31 14:35 ` Imre Deak
2025-01-31 12:50 ` [PATCH 09/14] drm/i915/dp: Inline do_dsc_compute_compressed_bpp() Jani Nikula
2025-01-31 14:48 ` Imre Deak
2025-01-31 12:50 ` [PATCH 10/14] drm/i915/dp: Simplify input BPP checks in intel_dp_dsc_compute_pipe_bpp() Jani Nikula
2025-01-31 14:52 ` Imre Deak
2025-01-31 12:50 ` [PATCH 11/14] drm/i915/dp: Use int for compressed BPP in dsc_compute_link_config() Jani Nikula
2025-01-31 15:08 ` Imre Deak
2025-01-31 15:27 ` Imre Deak
2025-01-31 12:50 ` [PATCH 12/14] drm/i915/dp: Drop compute_pipe_bpp parameter from intel_dp_dsc_compute_config() Jani Nikula
2025-01-31 15:10 ` Imre Deak
2025-01-31 12:50 ` [PATCH 13/14] drm/i915/dp: Pass connector state all the way to dsc_compute_link_config() Jani Nikula
2025-01-31 15:38 ` Imre Deak
2025-01-31 12:50 ` [PATCH 14/14] drm/i915/mst: Convert intel_dp_mtp_tu_compute_config() to .4 format Jani Nikula
2025-01-31 15:46 ` Imre Deak [this message]
2025-01-31 12:57 ` ✓ CI.Patch_applied: success for drm/i915/dp: dsc fix, refactoring and cleanups Patchwork
2025-01-31 12:57 ` ✗ CI.checkpatch: warning " Patchwork
2025-01-31 12:58 ` ✓ CI.KUnit: success " Patchwork
2025-01-31 13:15 ` ✓ CI.Build: " Patchwork
2025-01-31 13:17 ` ✓ CI.Hooks: " Patchwork
2025-01-31 13:18 ` ✗ CI.checksparse: warning " Patchwork
2025-01-31 13:38 ` ✓ Xe.CI.BAT: success " Patchwork
2025-01-31 17:34 ` ✗ Xe.CI.Full: failure " Patchwork
2025-02-01 0:24 ` ✗ CI.Patch_applied: failure for drm/i915/dp: dsc fix, refactoring and cleanups (rev2) Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=Z5zwauaji-K48_uc@ideak-desk.fi.intel.com \
--to=imre.deak@intel.com \
--cc=intel-gfx@lists.freedesktop.org \
--cc=intel-xe@lists.freedesktop.org \
--cc=jani.nikula@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox