From: Jani Nikula <jani.nikula@linux.intel.com>
To: Manasi Navare <manasi.d.navare@intel.com>
Cc: Daniel Vetter <daniel.vetter@intel.com>,
intel-gfx@lists.freedesktop.org, dri-devel@lists.freedesktop.org
Subject: Re: [PATCH 2/4] drm/i915: Compute sink's max lane count/link BW at Hotplug
Date: Tue, 13 Dec 2016 16:28:13 +0200 [thread overview]
Message-ID: <87y3zjsy2q.fsf@intel.com> (raw)
In-Reply-To: <20161208214831.GB26051@intel.com>
On Thu, 08 Dec 2016, Manasi Navare <manasi.d.navare@intel.com> wrote:
> Daniel, can we merge this patch?
Pushed this one to dinq, thanks for the patch.
BR,
Jani.
> It has no dependency on other link train patches,
> it is just a clean up for the existing driver code that
> uses max link rate and lane count values.
> Other link train patches have dependency on this thats why
> it was part of the series.
> But it would be great if this gets merged so that i dont have to
> rebase it tremendously after 3 weeks when I am back and plus
> its good for the driver to start using this clean up. This kind
> of clean up was long due.
>
> Regards
> Manasi
>
> On Thu, Dec 08, 2016 at 11:23:39PM +0200, Jani Nikula wrote:
>> On Tue, 06 Dec 2016, Manasi Navare <manasi.d.navare@intel.com> wrote:
>> > Sink's capabilities are advertised through DPCD registers and get
>> > updated only on hotplug. So they should be computed only once in the
>> > long pulse handler and saved off in intel_dp structure for the use
>> > later. For this reason two new fields max_sink_lane_count and
>> > max_sink_link_bw are added to intel_dp structure.
>> >
>> > This also simplifies the fallback link rate/lane count logic
>> > to handle link training failure. In that case, the max_sink_link_bw
>> > and max_sink_lane_count can be reccomputed to match the fallback
>> > values lowering the sink capabilities due to link train failure.
>> >
>> > Cc: Ville Syrjala <ville.syrjala@linux.intel.com>
>> > Cc: Jani Nikula <jani.nikula@linux.intel.com>
>> > Cc: Daniel Vetter <daniel.vetter@intel.com>
>> > Signed-off-by: Manasi Navare <manasi.d.navare@intel.com>
>>
>> Reviewed-by: Jani Nikula <jani.nikula@intel.com>
>>
>> Eventually we may want to call the fields *link* rates, because that's
>> what they'll effectively be. Transient values that don't reflect the
>> sink or source capabilities, but the link capabilities.
>>
>> > ---
>> > drivers/gpu/drm/i915/intel_dp.c | 10 ++++++++--
>> > drivers/gpu/drm/i915/intel_drv.h | 4 ++++
>> > 2 files changed, 12 insertions(+), 2 deletions(-)
>> >
>> > diff --git a/drivers/gpu/drm/i915/intel_dp.c b/drivers/gpu/drm/i915/intel_dp.c
>> > index db75bb9..434dc7d 100644
>> > --- a/drivers/gpu/drm/i915/intel_dp.c
>> > +++ b/drivers/gpu/drm/i915/intel_dp.c
>> > @@ -156,7 +156,7 @@ static u8 intel_dp_max_lane_count(struct intel_dp *intel_dp)
>> > u8 source_max, sink_max;
>> >
>> > source_max = intel_dig_port->max_lanes;
>> > - sink_max = drm_dp_max_lane_count(intel_dp->dpcd);
>> > + sink_max = intel_dp->max_sink_lane_count;
>> >
>> > return min(source_max, sink_max);
>> > }
>> > @@ -213,7 +213,7 @@ static u8 intel_dp_max_lane_count(struct intel_dp *intel_dp)
>> >
>> > *sink_rates = default_rates;
>> >
>> > - return (intel_dp_max_link_bw(intel_dp) >> 3) + 1;
>> > + return (intel_dp->max_sink_link_bw >> 3) + 1;
>> > }
>> >
>> > static int
>> > @@ -4395,6 +4395,12 @@ static bool intel_digital_port_connected(struct drm_i915_private *dev_priv,
>> > yesno(intel_dp_source_supports_hbr2(intel_dp)),
>> > yesno(drm_dp_tps3_supported(intel_dp->dpcd)));
>> >
>> > + /* Set the max lane count for sink */
>> > + intel_dp->max_sink_lane_count = drm_dp_max_lane_count(intel_dp->dpcd);
>> > +
>> > + /* Set the max link BW for sink */
>> > + intel_dp->max_sink_link_bw = intel_dp_max_link_bw(intel_dp);
>> > +
>> > intel_dp_print_rates(intel_dp);
>> >
>> > intel_dp_read_desc(intel_dp);
>> > diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h
>> > index fd77a3b..b6526ad 100644
>> > --- a/drivers/gpu/drm/i915/intel_drv.h
>> > +++ b/drivers/gpu/drm/i915/intel_drv.h
>> > @@ -906,6 +906,10 @@ struct intel_dp {
>> > /* sink rates as reported by DP_SUPPORTED_LINK_RATES */
>> > uint8_t num_sink_rates;
>> > int sink_rates[DP_MAX_SUPPORTED_RATES];
>> > + /* Max lane count for the sink as per DPCD registers */
>> > + uint8_t max_sink_lane_count;
>> > + /* Max link BW for the sink as per DPCD registers */
>> > + int max_sink_link_bw;
>> > /* sink or branch descriptor */
>> > struct intel_dp_desc desc;
>> > struct drm_dp_aux aux;
>>
>> --
>> Jani Nikula, Intel Open Source Technology Center
--
Jani Nikula, Intel Open Source Technology Center
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx
next prev parent reply other threads:[~2016-12-13 14:28 UTC|newest]
Thread overview: 33+ messages / expand[flat|nested] mbox.gz Atom feed top
2016-12-06 0:27 [PATCH 0/4] Link Training failure handling by sending Hotplug Uevent Manasi Navare
2016-12-06 0:27 ` [PATCH 1/4] drm: Add a new connector atomic property for link status Manasi Navare
2016-12-06 7:23 ` [Intel-gfx] " Daniel Vetter
2016-12-06 15:56 ` Manasi Navare
2016-12-06 16:07 ` [PATCH v4 " Manasi Navare
2016-12-08 15:05 ` Jani Nikula
2016-12-08 15:28 ` [Intel-gfx] " Daniel Vetter
2016-12-08 19:04 ` [PATCH v5 " Manasi Navare
2016-12-08 19:36 ` Sean Paul
2016-12-08 19:48 ` Manasi Navare
2016-12-08 19:47 ` [PATCH v6 " Manasi Navare
2016-12-06 0:27 ` [PATCH 2/4] drm/i915: Compute sink's max lane count/link BW at Hotplug Manasi Navare
2016-12-08 18:15 ` Manasi Navare
2016-12-08 21:23 ` Jani Nikula
2016-12-08 21:39 ` Manasi Navare
2016-12-08 21:48 ` Manasi Navare
2016-12-13 14:28 ` Jani Nikula [this message]
2016-12-06 0:27 ` [PATCH 3/4] drm/i915: Find fallback link rate/lane count Manasi Navare
2016-12-08 18:19 ` Manasi Navare
2016-12-08 21:46 ` Jani Nikula
2016-12-08 22:05 ` Manasi Navare
2016-12-09 3:05 ` [PATCH v7 " Manasi Navare
2016-12-09 9:54 ` Jani Nikula
2016-12-13 14:36 ` Jani Nikula
2016-12-06 0:27 ` [PATCH 4/4] drm/i915: Implement Link Rate fallback on Link training failure Manasi Navare
2016-12-08 18:23 ` Manasi Navare
2016-12-08 21:51 ` Jani Nikula
2016-12-08 22:09 ` Manasi Navare
2016-12-06 0:53 ` ✗ Fi.CI.BAT: failure for Link Training failure handling by sending Hotplug Uevent Patchwork
2016-12-06 16:53 ` ✗ Fi.CI.BAT: failure for Link Training failure handling by sending Hotplug Uevent (rev2) Patchwork
2016-12-08 19:53 ` ✗ Fi.CI.BAT: failure for Link Training failure handling by sending Hotplug Uevent (rev3) Patchwork
2016-12-08 20:23 ` ✗ Fi.CI.BAT: failure for Link Training failure handling by sending Hotplug Uevent (rev4) Patchwork
2016-12-09 4:23 ` ✗ Fi.CI.BAT: failure for Link Training failure handling by sending Hotplug Uevent (rev5) Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=87y3zjsy2q.fsf@intel.com \
--to=jani.nikula@linux.intel.com \
--cc=daniel.vetter@intel.com \
--cc=dri-devel@lists.freedesktop.org \
--cc=intel-gfx@lists.freedesktop.org \
--cc=manasi.d.navare@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox