From: Ben Widawsky <ben@bwidawsk.net>
To: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Daniel Vetter <daniel.vetter@ffwll.ch>,
Intel GFX <intel-gfx@lists.freedesktop.org>,
DRI Devel <dri-devel@lists.freedesktop.org>
Subject: Re: [Intel-gfx] [PATCH] [RFC] intel: Non-LLC based non-blocking maps.
Date: Tue, 19 Jun 2012 09:13:20 -0700 [thread overview]
Message-ID: <20120619091320.4a99b8c6@bwidawsk.net> (raw)
In-Reply-To: <1340094148_12273@CP5-2952>
On Tue, 19 Jun 2012 09:22:03 +0100
Chris Wilson <chris@chris-wilson.co.uk> wrote:
> On Mon, 18 Jun 2012 20:38:15 -0700, Ben Widawsky <ben@bwidawsk.net> wrote:
> > The history on this patch goes back quite a way. This time around, the
> > patch builds on top of the map_unsynchronized that Eric pushed. Eric's
> > patch attempted only to solve the problem for LLC machines. Unlike
> > my earlier versions of this patch (with the help from Daniel Vetter), we
> > do not attempt to cpu map objects in a unsynchronized manner.
> >
> > The concept is fairly simple - once a buffer is moved into the GTT
> > domain, we can assume it remains there unless we tell it otherwise (via
> > cpu map). It therefore stands to reason that as long as we can keep the
> > object in the GTT domain, and don't ever count on reading back contents,
> > things might just work. I believe as long as we are doing GTT mappings
> > only, we get to avoid worry about clflushing the dirtied cachelines, but
> > that could use some fact checking.
> >
> > The patch makes some assumptions about how the kernel does buffer
> > tracking, this could be conceived as an ABI dependency, but actually the
> > behavior is pretty confined. It exploits the fact the BOs are only moved
> > into the CPU domain under certain circumstances, and daintily dances
> > around those conditions. The main thing here is we assume MADV_WILLNEED
> > prevents the object from getting evicted.
> >
> > I am not aware of a good way to test it's effectiveness
> > performance-wise; but it introduces no regressions with piglit on my
> > ILK, or SNB.
>
> This is broken wrt to cache invalidation if I want to rewrite part of
> the buffer that already has been read by the GPU.
> -Chris
>
Well if you're talking about what I think you're talking about (ie. not
clflushing, but simply dealing with the GPUs internal caching). It's a
problem that has existed with all of the non-LLC non-blocking map
patches; and sort of the point of non-blocking maps. Play it fast and
loose, submit pipe controls if you get nervous.
Did I catch your meaning, or were you just talking about clflushing
stuff (we also miss chipset flush on really old platforms; I was
thinking of restricting this to ILK only)?
--
Ben Widawsky, Intel Open Source Technology Center
next prev parent reply other threads:[~2012-06-19 16:13 UTC|newest]
Thread overview: 4+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-06-19 3:38 [PATCH] [RFC] intel: Non-LLC based non-blocking maps Ben Widawsky
2012-06-19 8:22 ` Chris Wilson
2012-06-19 16:13 ` Ben Widawsky [this message]
2012-06-19 19:50 ` [Intel-gfx] " Chris Wilson
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20120619091320.4a99b8c6@bwidawsk.net \
--to=ben@bwidawsk.net \
--cc=chris@chris-wilson.co.uk \
--cc=daniel.vetter@ffwll.ch \
--cc=dri-devel@lists.freedesktop.org \
--cc=intel-gfx@lists.freedesktop.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox