public inbox for intel-gfx@lists.freedesktop.org
 help / color / mirror / Atom feed
From: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
To: Chris Wilson <chris@chris-wilson.co.uk>, intel-gfx@lists.freedesktop.org
Subject: Re: [PATCH 51/55] drm/i915: i915_vma_move_to_active prep patch
Date: Tue, 26 Jul 2016 15:53:30 +0300	[thread overview]
Message-ID: <1469537610.4681.90.camel@linux.intel.com> (raw)
In-Reply-To: <1469467954-3920-52-git-send-email-chris@chris-wilson.co.uk>

On ma, 2016-07-25 at 18:32 +0100, Chris Wilson wrote:
> This patch is broken out of the next just to remove the code motion from
> that patch and make it more readable. What we do here is move the
> i915_vma_move_to_active() to i915_gem_execbuffer.c and put the three
> stages (read, write, fenced) together so that future modifications to
> active handling are all located in the same spot. The importance of this
> is so that we can more simply control the order in which the requests
> are place in the retirement list (i.e. control the order at which we
> retire and so control the lifetimes to avoid having to hold onto
> references).
> 

This is much better. So,

Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>

> Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
> ---
>  drivers/gpu/drm/i915/i915_drv.h              |  3 +-
>  drivers/gpu/drm/i915/i915_gem.c              | 18 --------
>  drivers/gpu/drm/i915/i915_gem_context.c      |  9 ++--
>  drivers/gpu/drm/i915/i915_gem_execbuffer.c   | 65 ++++++++++++++++++----------
>  drivers/gpu/drm/i915/i915_gem_render_state.c |  2 +-
>  5 files changed, 51 insertions(+), 46 deletions(-)
> 
> diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
> index 03f12304308a..4876d2a6c2c4 100644
> --- a/drivers/gpu/drm/i915/i915_drv.h
> +++ b/drivers/gpu/drm/i915/i915_drv.h
> @@ -3153,7 +3153,8 @@ int __must_check i915_mutex_lock_interruptible(struct drm_device *dev);
>  int i915_gem_object_sync(struct drm_i915_gem_object *obj,
>  			 struct drm_i915_gem_request *to);
>  void i915_vma_move_to_active(struct i915_vma *vma,
> -			     struct drm_i915_gem_request *req);
> +			     struct drm_i915_gem_request *req,
> +			     unsigned int flags);
>  int i915_gem_dumb_create(struct drm_file *file_priv,
>  			 struct drm_device *dev,
>  			 struct drm_mode_create_dumb *args);
> diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
> index c572c80a6604..2e0b54fa03f9 100644
> --- a/drivers/gpu/drm/i915/i915_gem.c
> +++ b/drivers/gpu/drm/i915/i915_gem.c
> @@ -2330,24 +2330,6 @@ void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj)
>  	return obj->mapping;
>  }
>  
> -void i915_vma_move_to_active(struct i915_vma *vma,
> -			     struct drm_i915_gem_request *req)
> -{
> -	struct drm_i915_gem_object *obj = vma->obj;
> -	struct intel_engine_cs *engine;
> -
> -	engine = i915_gem_request_get_engine(req);
> -
> -	/* Add a reference if we're newly entering the active list. */
> -	if (obj->active == 0)
> -		i915_gem_object_get(obj);
> -	obj->active |= intel_engine_flag(engine);
> -
> -	i915_gem_active_set(&obj->last_read[engine->id], req);
> -
> -	list_move_tail(&vma->vm_link, &vma->vm->active_list);
> -}
> -
>  static void
>  i915_gem_object_retire__fence(struct i915_gem_active *active,
>  			      struct drm_i915_gem_request *req)
> diff --git a/drivers/gpu/drm/i915/i915_gem_context.c b/drivers/gpu/drm/i915/i915_gem_context.c
> index 60861f616f24..29b2547a2b4c 100644
> --- a/drivers/gpu/drm/i915/i915_gem_context.c
> +++ b/drivers/gpu/drm/i915/i915_gem_context.c
> @@ -816,8 +816,8 @@ static int do_rcs_switch(struct drm_i915_gem_request *req)
>  	 * MI_SET_CONTEXT instead of when the next seqno has completed.
>  	 */
>  	if (from != NULL) {
> -		from->engine[RCS].state->base.read_domains = I915_GEM_DOMAIN_INSTRUCTION;
> -		i915_vma_move_to_active(i915_gem_obj_to_ggtt(from->engine[RCS].state), req);
> +		struct drm_i915_gem_object *obj = from->engine[RCS].state;
> +
>  		/* As long as MI_SET_CONTEXT is serializing, ie. it flushes the
>  		 * whole damn pipeline, we don't need to explicitly mark the
>  		 * object dirty. The only exception is that the context must be
> @@ -825,10 +825,11 @@ static int do_rcs_switch(struct drm_i915_gem_request *req)
>  		 * able to defer doing this until we know the object would be
>  		 * swapped, but there is no way to do that yet.
>  		 */
> -		from->engine[RCS].state->dirty = 1;
> +		obj->base.read_domains = I915_GEM_DOMAIN_INSTRUCTION;
> +		i915_vma_move_to_active(i915_gem_obj_to_ggtt(obj), req, 0);
>  
>  		/* obj is kept alive until the next request by its active ref */
> -		i915_gem_object_ggtt_unpin(from->engine[RCS].state);
> +		i915_gem_object_ggtt_unpin(obj);
>  		i915_gem_context_put(from);
>  	}
>  	engine->last_context = i915_gem_context_get(to);
> diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
> index 9778b1bc6336..d0f1da2863e4 100644
> --- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c
> +++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
> @@ -1143,43 +1143,64 @@ i915_gem_validate_context(struct drm_device *dev, struct drm_file *file,
>  	return ctx;
>  }
>  
> +void i915_vma_move_to_active(struct i915_vma *vma,
> +			     struct drm_i915_gem_request *req,
> +			     unsigned int flags)
> +{
> +	struct drm_i915_gem_object *obj = vma->obj;
> +	const unsigned int idx = req->engine->id;
> +
> +	GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
> +
> +	obj->dirty = 1; /* be paranoid  */
> +
> +	/* Add a reference if we're newly entering the active list. */
> +	if (obj->active == 0)
> +		i915_gem_object_get(obj);
> +	obj->active |= 1 << idx;
> +	i915_gem_active_set(&obj->last_read[idx], req);
> +
> +	if (flags & EXEC_OBJECT_WRITE) {
> +		i915_gem_active_set(&obj->last_write, req);
> +
> +		intel_fb_obj_invalidate(obj, ORIGIN_CS);
> +
> +		/* update for the implicit flush after a batch */
> +		obj->base.write_domain &= ~I915_GEM_GPU_DOMAINS;
> +	}
> +
> +	if (flags & EXEC_OBJECT_NEEDS_FENCE) {
> +		i915_gem_active_set(&obj->last_fence, req);
> +		if (flags & __EXEC_OBJECT_HAS_FENCE) {
> +			struct drm_i915_private *dev_priv = req->i915;
> +
> +			list_move_tail(&dev_priv->fence_regs[obj->fence_reg].lru_list,
> +				       &dev_priv->mm.fence_list);
> +		}
> +	}
> +
> +	list_move_tail(&vma->vm_link, &vma->vm->active_list);
> +}
> +
>  static void
>  i915_gem_execbuffer_move_to_active(struct list_head *vmas,
>  				   struct drm_i915_gem_request *req)
>  {
> -	struct intel_engine_cs *engine = i915_gem_request_get_engine(req);
>  	struct i915_vma *vma;
>  
>  	list_for_each_entry(vma, vmas, exec_list) {
> -		struct drm_i915_gem_exec_object2 *entry = vma->exec_entry;
>  		struct drm_i915_gem_object *obj = vma->obj;
>  		u32 old_read = obj->base.read_domains;
>  		u32 old_write = obj->base.write_domain;
>  
> -		obj->dirty = 1; /* be paranoid  */
>  		obj->base.write_domain = obj->base.pending_write_domain;
> -		if (obj->base.write_domain == 0)
> +		if (obj->base.write_domain)
> +			vma->exec_entry->flags |= EXEC_OBJECT_WRITE;
> +		else
>  			obj->base.pending_read_domains |= obj->base.read_domains;
>  		obj->base.read_domains = obj->base.pending_read_domains;
>  
> -		i915_vma_move_to_active(vma, req);
> -		if (obj->base.write_domain) {
> -			i915_gem_active_set(&obj->last_write, req);
> -
> -			intel_fb_obj_invalidate(obj, ORIGIN_CS);
> -
> -			/* update for the implicit flush after a batch */
> -			obj->base.write_domain &= ~I915_GEM_GPU_DOMAINS;
> -		}
> -		if (entry->flags & EXEC_OBJECT_NEEDS_FENCE) {
> -			i915_gem_active_set(&obj->last_fence, req);
> -			if (entry->flags & __EXEC_OBJECT_HAS_FENCE) {
> -				struct drm_i915_private *dev_priv = engine->i915;
> -				list_move_tail(&dev_priv->fence_regs[obj->fence_reg].lru_list,
> -					       &dev_priv->mm.fence_list);
> -			}
> -		}
> -
> +		i915_vma_move_to_active(vma, req, vma->exec_entry->flags);
>  		trace_i915_gem_object_change_domain(obj, old_read, old_write);
>  	}
>  }
> diff --git a/drivers/gpu/drm/i915/i915_gem_render_state.c b/drivers/gpu/drm/i915/i915_gem_render_state.c
> index f85c5505bce2..90236672ac1e 100644
> --- a/drivers/gpu/drm/i915/i915_gem_render_state.c
> +++ b/drivers/gpu/drm/i915/i915_gem_render_state.c
> @@ -217,7 +217,7 @@ int i915_gem_render_state_init(struct drm_i915_gem_request *req)
>  			goto err_unpin;
>  	}
>  
> -	i915_vma_move_to_active(i915_gem_obj_to_ggtt(so.obj), req);
> +	i915_vma_move_to_active(i915_gem_obj_to_ggtt(so.obj), req, 0);
>  err_unpin:
>  	i915_gem_object_ggtt_unpin(so.obj);
>  err_obj:
-- 
Joonas Lahtinen
Open Source Technology Center
Intel Corporation
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

  reply	other threads:[~2016-07-26 12:53 UTC|newest]

Thread overview: 124+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2016-07-25 17:31 Fix the vma leak Chris Wilson
2016-07-25 17:31 ` [PATCH 01/55] drm/i915: Reduce breadcrumb lock coverage for intel_engine_enable_signaling() Chris Wilson
2016-07-26  5:07   ` Joonas Lahtinen
2016-07-25 17:31 ` [PATCH 02/55] drm/i915: Prefer list_first_entry_or_null Chris Wilson
2016-07-25 17:31 ` [PATCH 03/55] drm/i915: Only clear the client pointer when tearing down the file Chris Wilson
2016-07-25 17:31 ` [PATCH 04/55] drm/i915: Only drop the batch-pool's object reference Chris Wilson
2016-07-25 17:31 ` [PATCH 05/55] drm/i915/cmdparser: Remove stray intel_engine_cs *ring Chris Wilson
2016-07-25 17:31 ` [PATCH 06/55] drm/i915: Use engine to refer to the user's BSD intel_engine_cs Chris Wilson
2016-07-25 17:31 ` [PATCH 07/55] drm/i915: Avoid using intel_engine_cs *ring for GPU error capture Chris Wilson
2016-07-26  4:59   ` Joonas Lahtinen
2016-07-26  8:19     ` Chris Wilson
2016-07-27 11:08       ` Joonas Lahtinen
2016-07-26 10:21     ` [PATCH v2] " Chris Wilson
2016-07-26 12:35       ` Joonas Lahtinen
2016-07-25 17:31 ` [PATCH 08/55] drm/i915: Remove stray intel_engine_cs ring identifiers from i915_gem.c Chris Wilson
2016-07-26  5:02   ` Joonas Lahtinen
2016-07-26  8:12     ` Chris Wilson
2016-07-27  6:12       ` Joonas Lahtinen
2016-07-25 17:31 ` [PATCH 09/55] drm/i915: Update a couple of hangcheck comments to talk about engines Chris Wilson
2016-07-25 17:31 ` [PATCH 10/55] drm/i915: Unify intel_logical_ring_emit and intel_ring_emit Chris Wilson
2016-07-25 17:31 ` [PATCH 11/55] drm/i915: Rename request->ringbuf to request->ring Chris Wilson
2016-07-25 17:31 ` [PATCH 12/55] drm/i915: Rename intel_context[engine].ringbuf Chris Wilson
2016-07-25 17:31 ` [PATCH 13/55] drm/i915: Rename struct intel_ringbuffer to struct intel_ring Chris Wilson
2016-07-25 17:31 ` [PATCH 14/55] drm/i915: Rename residual ringbuf parameters Chris Wilson
2016-07-25 17:31 ` [PATCH 15/55] drm/i915: Rename intel_pin_and_map_ring() Chris Wilson
2016-07-25 17:31 ` [PATCH 16/55] drm/i915: Remove obsolete engine->gpu_caches_dirty Chris Wilson
2016-07-26  5:06   ` Joonas Lahtinen
2016-07-25 17:31 ` [PATCH 17/55] drm/i915: Simplify request_alloc by returning the allocated request Chris Wilson
2016-07-26  5:09   ` Joonas Lahtinen
2016-07-25 17:31 ` [PATCH 18/55] drm/i915: Unify legacy/execlists emission of MI_BATCHBUFFER_START Chris Wilson
2016-07-25 17:31 ` [PATCH 19/55] drm/i915: Remove intel_ring_get_tail() Chris Wilson
2016-07-25 17:31 ` [PATCH 20/55] drm/i915: Convert engine->write_tail to operate on a request Chris Wilson
2016-07-25 17:32 ` [PATCH 21/55] drm/i915: Unify request submission Chris Wilson
2016-07-25 17:32 ` [PATCH 22/55] drm/i915/lrc: Update function names to match request flow Chris Wilson
2016-07-25 17:32 ` [PATCH 23/55] drm/i915: Stop passing caller's num_dwords to engine->semaphore.signal() Chris Wilson
2016-07-25 17:32 ` [PATCH 24/55] drm/i915: Reuse legacy breadcrumbs + tail emission Chris Wilson
2016-07-25 17:32 ` [PATCH 25/55] drm/i915/ringbuffer: Specialise SNB+ request emission for semaphores Chris Wilson
2016-07-25 17:32 ` [PATCH 26/55] drm/i915: Remove duplicate golden render state init from execlists Chris Wilson
2016-07-25 17:32 ` [PATCH 27/55] drm/i915: Refactor golden render state emission to unconfuse gcc Chris Wilson
2016-07-25 17:32 ` [PATCH 28/55] drm/i915: Unify legacy/execlists submit_execbuf callbacks Chris Wilson
2016-07-25 17:32 ` [PATCH 29/55] drm/i915: Simplify calling engine->sync_to Chris Wilson
2016-07-25 17:32 ` [PATCH 30/55] drm/i915: Rename engine->semaphore.sync_to, engine->sempahore.signal locals Chris Wilson
2016-07-25 17:32 ` [PATCH 31/55] drm/i915: Amalgamate GGTT/ppGTT vma debug list walkers Chris Wilson
2016-07-26  5:15   ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 32/55] drm/i915: Split early global GTT initialisation Chris Wilson
2016-07-26  7:08   ` Joonas Lahtinen
2016-07-26  7:42     ` Chris Wilson
2016-07-27 10:20       ` Joonas Lahtinen
2016-07-27 10:34         ` Chris Wilson
2016-07-27 11:09           ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 33/55] drm/i915: Store owning file on the i915_address_space Chris Wilson
2016-07-26  7:15   ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 34/55] drm/i915: Count how many VMA are bound for an object Chris Wilson
2016-07-26  7:44   ` Joonas Lahtinen
2016-07-26  8:02     ` Chris Wilson
2016-07-25 17:32 ` [PATCH 35/55] drm/i915: Be more careful when unbinding vma Chris Wilson
2016-07-26  7:59   ` Joonas Lahtinen
2016-07-26  8:08     ` Chris Wilson
2016-07-25 17:32 ` [PATCH 36/55] drm/i915: Kill drop_pages() Chris Wilson
2016-07-25 17:32 ` [PATCH 37/55] drm/i915: Introduce i915_gem_active for request tracking Chris Wilson
2016-07-26  8:23   ` Joonas Lahtinen
2016-07-26  8:28     ` Chris Wilson
2016-07-28  7:21       ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 38/55] drm/i915: Prepare i915_gem_active for annotations Chris Wilson
2016-07-26  8:50   ` Joonas Lahtinen
2016-07-26  9:03     ` Chris Wilson
2016-07-25 17:32 ` [PATCH 39/55] drm/i915: Mark up i915_gem_active for locking annotation Chris Wilson
2016-07-26  8:54   ` Joonas Lahtinen
2016-07-26  9:06     ` Chris Wilson
2016-07-28  7:26       ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 40/55] drm/i915: Refactor blocking waits Chris Wilson
2016-07-27  6:04   ` Joonas Lahtinen
2016-07-27  7:04     ` Chris Wilson
2016-07-27 10:40       ` Joonas Lahtinen
2016-07-27 10:48         ` Chris Wilson
2016-07-27  7:07     ` Chris Wilson
2016-07-27 10:42       ` Joonas Lahtinen
2016-07-27 17:34         ` Chris Wilson
2016-07-28  6:40           ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 41/55] drm/i915: Rename request->list to link for consistency Chris Wilson
2016-07-26  9:26   ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 42/55] drm/i915: Remove obsolete i915_gem_object_flush_active() Chris Wilson
2016-07-26  9:31   ` Joonas Lahtinen
2016-07-26  9:47     ` Chris Wilson
2016-07-25 17:32 ` [PATCH 43/55] drm/i915: Refactor activity tracking for requests Chris Wilson
2016-07-27  7:40   ` Joonas Lahtinen
2016-07-27  7:57     ` Chris Wilson
2016-07-27 10:55       ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 44/55] drm/i915: Track requests inside each intel_ring Chris Wilson
2016-07-26 10:10   ` Joonas Lahtinen
2016-07-26 10:15     ` Chris Wilson
2016-07-25 17:32 ` [PATCH 45/55] drm/i915: Convert intel_overlay to request tracking Chris Wilson
2016-07-27  8:12   ` Joonas Lahtinen
2016-07-27  8:22     ` Chris Wilson
2016-07-27  8:34       ` Chris Wilson
2016-07-27 10:59       ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 46/55] drm/i915: Move the special case wait-request handling to its one caller Chris Wilson
2016-07-26 12:39   ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 47/55] drm/i915: Disable waitboosting for a saturated engine Chris Wilson
2016-07-26 12:40   ` Joonas Lahtinen
2016-07-26 13:11     ` Chris Wilson
2016-07-25 17:32 ` [PATCH 48/55] drm/i915: s/__i915_wait_request/i915_wait_request/ Chris Wilson
2016-07-26 12:42   ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 49/55] drm/i915: Double check activity before relocations Chris Wilson
2016-07-26 12:45   ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 50/55] drm/i915: Move request list retirement to i915_gem_request.c Chris Wilson
2016-07-26 12:48   ` Joonas Lahtinen
2016-07-26 13:39     ` Chris Wilson
2016-07-25 17:32 ` [PATCH 51/55] drm/i915: i915_vma_move_to_active prep patch Chris Wilson
2016-07-26 12:53   ` Joonas Lahtinen [this message]
2016-07-25 17:32 ` [PATCH 52/55] drm/i915: Track active vma requests Chris Wilson
2016-07-27  9:47   ` Joonas Lahtinen
2016-07-27 10:15     ` Chris Wilson
2016-07-25 17:32 ` [PATCH 53/55] drm/i915: Release vma when the handle is closed Chris Wilson
2016-07-27 10:00   ` Joonas Lahtinen
2016-07-27 10:13     ` Chris Wilson
2016-07-28  7:16       ` Joonas Lahtinen
2016-07-25 17:32 ` [PATCH 54/55] drm/i915: Mark the context and address space as closed Chris Wilson
2016-07-27 10:13   ` Joonas Lahtinen
2016-07-27 10:27     ` Chris Wilson
2016-07-25 17:32 ` [PATCH 55/55] Revert "drm/i915: Clean up associated VMAs on context destruction" Chris Wilson
2016-07-27 10:18   ` Joonas Lahtinen
2016-07-26  5:18 ` ✗ Ro.CI.BAT: warning for series starting with [01/55] drm/i915: Reduce breadcrumb lock coverage for intel_engine_enable_signaling() Patchwork
2016-07-26 10:48 ` ✗ Ro.CI.BAT: failure for series starting with [01/55] drm/i915: Reduce breadcrumb lock coverage for intel_engine_enable_signaling() (rev2) Patchwork

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1469537610.4681.90.camel@linux.intel.com \
    --to=joonas.lahtinen@linux.intel.com \
    --cc=chris@chris-wilson.co.uk \
    --cc=intel-gfx@lists.freedesktop.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox