From: Mika Kuoppala <mika.kuoppala@linux.intel.com>
To: Chris Wilson <chris@chris-wilson.co.uk>, intel-gfx@lists.freedesktop.org
Subject: Re: [PATCH 2/2] drm/i915: Recover all available ringbuffer space following reset
Date: Fri, 23 Oct 2015 14:07:35 +0300 [thread overview]
Message-ID: <87k2qdyhag.fsf@gaia.fi.intel.com> (raw)
In-Reply-To: <1441281700-17814-2-git-send-email-chris@chris-wilson.co.uk>
Chris Wilson <chris@chris-wilson.co.uk> writes:
> Having flushed all requests from all queues, we know that all
> ringbuffers must now be empty. However, since we do not reclaim
> all space when retiring the request (to prevent HEADs colliding
> with rapid ringbuffer wraparound) the amount of available space
> on each ringbuffer upon reset is less than when we start. Do one
> more pass over all the ringbuffers to reset the available space
>
> Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
> Cc: Arun Siluvery <arun.siluvery@linux.intel.com>
> Cc: Mika Kuoppala <mika.kuoppala@intel.com>
> Cc: Dave Gordon <david.s.gordon@intel.com>
> ---
> drivers/gpu/drm/i915/i915_gem.c | 14 ++++++++++++++
> drivers/gpu/drm/i915/intel_lrc.c | 1 +
> drivers/gpu/drm/i915/intel_ringbuffer.c | 13 ++++++++++---
> drivers/gpu/drm/i915/intel_ringbuffer.h | 2 ++
> 4 files changed, 27 insertions(+), 3 deletions(-)
>
> diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
> index 41263cd4170c..3a42c350fec9 100644
> --- a/drivers/gpu/drm/i915/i915_gem.c
> +++ b/drivers/gpu/drm/i915/i915_gem.c
> @@ -2738,6 +2738,8 @@ static void i915_gem_reset_ring_status(struct drm_i915_private *dev_priv,
> static void i915_gem_reset_ring_cleanup(struct drm_i915_private *dev_priv,
> struct intel_engine_cs *ring)
> {
> + struct intel_ringbuffer *buffer;
> +
> while (!list_empty(&ring->active_list)) {
> struct drm_i915_gem_object *obj;
>
> @@ -2783,6 +2785,18 @@ static void i915_gem_reset_ring_cleanup(struct drm_i915_private *dev_priv,
>
> i915_gem_request_retire(request);
> }
> +
> + /* Having flushed all requests from all queues, we know that all
> + * ringbuffers must now be empty. However, since we do not reclaim
> + * all space when retiring the request (to prevent HEADs colliding
> + * with rapid ringbuffer wraparound) the amount of available space
> + * upon reset is less than when we start. Do one more pass over
> + * all the ringbuffers to reset last_retired_head.
> + */
> + list_for_each_entry(buffer, &ring->buffers, link) {
> + buffer->last_retired_head = buffer->tail;
> + intel_ring_update_space(buffer);
> + }
This is all in vain as the i915_gem_context_reset() ->
intel_lr_context_reset still sets head and tail to zero.
So your last_retired_head will still dangle in a pre-reset
world when the rest of the ringbuf items will be set to post
reset world.
-Mika
> }
>
> void i915_gem_reset(struct drm_device *dev)
> diff --git a/drivers/gpu/drm/i915/intel_lrc.c b/drivers/gpu/drm/i915/intel_lrc.c
> index 28a712e7d2d0..de52ddc108a7 100644
> --- a/drivers/gpu/drm/i915/intel_lrc.c
> +++ b/drivers/gpu/drm/i915/intel_lrc.c
> @@ -1881,6 +1881,7 @@ static int logical_ring_init(struct drm_device *dev, struct intel_engine_cs *rin
> i915_gem_batch_pool_init(dev, &ring->batch_pool);
> init_waitqueue_head(&ring->irq_queue);
>
> + INIT_LIST_HEAD(&ring->buffers);
> INIT_LIST_HEAD(&ring->execlist_queue);
> INIT_LIST_HEAD(&ring->execlist_retired_req_list);
> spin_lock_init(&ring->execlist_lock);
> diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.c b/drivers/gpu/drm/i915/intel_ringbuffer.c
> index 20a75bb516ac..d2e0b3b7efbf 100644
> --- a/drivers/gpu/drm/i915/intel_ringbuffer.c
> +++ b/drivers/gpu/drm/i915/intel_ringbuffer.c
> @@ -2030,10 +2030,14 @@ intel_engine_create_ringbuffer(struct intel_engine_cs *engine, int size)
> int ret;
>
> ring = kzalloc(sizeof(*ring), GFP_KERNEL);
> - if (ring == NULL)
> + if (ring == NULL) {
> + DRM_DEBUG_DRIVER("Failed to allocate ringbuffer %s\n",
> + engine->name);
> return ERR_PTR(-ENOMEM);
> + }
>
> ring->ring = engine;
> + list_add(&ring->link, &engine->buffers);
>
> ring->size = size;
> /* Workaround an erratum on the i830 which causes a hang if
> @@ -2049,8 +2053,9 @@ intel_engine_create_ringbuffer(struct intel_engine_cs *engine, int size)
>
> ret = intel_alloc_ringbuffer_obj(engine->dev, ring);
> if (ret) {
> - DRM_ERROR("Failed to allocate ringbuffer %s: %d\n",
> - engine->name, ret);
> + DRM_DEBUG_DRIVER("Failed to allocate ringbuffer %s: %d\n",
> + engine->name, ret);
> + list_del(&ring->link);
> kfree(ring);
> return ERR_PTR(ret);
> }
> @@ -2062,6 +2067,7 @@ void
> intel_ringbuffer_free(struct intel_ringbuffer *ring)
> {
> intel_destroy_ringbuffer_obj(ring);
> + list_del(&ring->link);
> kfree(ring);
> }
>
> @@ -2077,6 +2083,7 @@ static int intel_init_ring_buffer(struct drm_device *dev,
> INIT_LIST_HEAD(&ring->active_list);
> INIT_LIST_HEAD(&ring->request_list);
> INIT_LIST_HEAD(&ring->execlist_queue);
> + INIT_LIST_HEAD(&ring->buffers);
> i915_gem_batch_pool_init(dev, &ring->batch_pool);
> memset(ring->semaphore.sync_seqno, 0, sizeof(ring->semaphore.sync_seqno));
>
> diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.h b/drivers/gpu/drm/i915/intel_ringbuffer.h
> index 49fa41dc0eb6..58b1976a7d0a 100644
> --- a/drivers/gpu/drm/i915/intel_ringbuffer.h
> +++ b/drivers/gpu/drm/i915/intel_ringbuffer.h
> @@ -100,6 +100,7 @@ struct intel_ringbuffer {
> void __iomem *virtual_start;
>
> struct intel_engine_cs *ring;
> + struct list_head link;
>
> u32 head;
> u32 tail;
> @@ -157,6 +158,7 @@ struct intel_engine_cs {
> u32 mmio_base;
> struct drm_device *dev;
> struct intel_ringbuffer *buffer;
> + struct list_head buffers;
>
> /*
> * A pool of objects to use as shadow copies of client batch buffers
> --
> 2.5.1
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
http://lists.freedesktop.org/mailman/listinfo/intel-gfx
next prev parent reply other threads:[~2015-10-23 11:08 UTC|newest]
Thread overview: 12+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-09-03 12:01 [PATCH 1/2] drm/i915: Refactor common ringbuffer allocation code Chris Wilson
2015-09-03 12:01 ` [PATCH 2/2] drm/i915: Recover all available ringbuffer space following reset Chris Wilson
2015-09-28 15:25 ` Mika Kuoppala
2015-09-28 15:43 ` Chris Wilson
2015-10-23 11:07 ` Mika Kuoppala [this message]
2015-10-23 11:12 ` Chris Wilson
2015-10-23 11:46 ` Mika Kuoppala
2015-10-28 17:18 ` Tvrtko Ursulin
2015-09-03 14:01 ` [PATCH 1/2] drm/i915: Refactor common ringbuffer allocation code Zanoni, Paulo R
2015-09-03 14:47 ` chris
2015-09-03 14:56 ` Mika Kuoppala
2015-09-04 8:17 ` Daniel Vetter
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=87k2qdyhag.fsf@gaia.fi.intel.com \
--to=mika.kuoppala@linux.intel.com \
--cc=chris@chris-wilson.co.uk \
--cc=intel-gfx@lists.freedesktop.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox