qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: Stefan Hajnoczi <stefanha@gmail.com>
To: Alberto Garcia <berto@igalia.com>
Cc: qemu-block@nongnu.org, qemu-devel@nongnu.org,
	Stefan Hajnoczi <stefanha@redhat.com>
Subject: Re: [Qemu-devel] [Qemu-block] [PATCH] qcow2: do lazy allocation of the L2 cache
Date: Fri, 24 Apr 2015 10:26:21 +0100	[thread overview]
Message-ID: <20150424092621.GD13278@stefanha-thinkpad.redhat.com> (raw)
In-Reply-To: <w51oamfjc6z.fsf@maestria.local.igalia.com>

[-- Attachment #1: Type: text/plain, Size: 2823 bytes --]

On Thu, Apr 23, 2015 at 01:50:28PM +0200, Alberto Garcia wrote:
> On Thu 23 Apr 2015 12:15:04 PM CEST, Stefan Hajnoczi wrote:
> 
> >> For a cache size of 128MB, the PSS is actually ~10MB larger without
> >> the patch, which seems to come from posix_memalign().
> >
> > Do you mean RSS or are you using a tool that reports a "PSS" number
> > that I don't know about?
> >
> > We should understand what is going on instead of moving the code
> > around to hide/delay the problem.
> 
> Both RSS and PSS ("proportional set size", also reported by the kernel).
> 
> I'm not an expert in memory allocators, but I measured the overhead like
> this:
> 
> An L2 cache of 128MB implies a refcount cache of 32MB, in total 160MB.
> With a default cluster size of 64k, that's 2560 cache entries.
> 
> So I wrote a test case that allocates 2560 blocks of 64k each using
> posix_memalign and mmap, and here's how their /proc/<pid>/smaps compare:
> 
> -Size:             165184 kB
> -Rss:               10244 kB
> -Pss:               10244 kB
> +Size:             161856 kB
> +Rss:                   0 kB
> +Pss:                   0 kB
>  Shared_Clean:          0 kB
>  Shared_Dirty:          0 kB
>  Private_Clean:         0 kB
> -Private_Dirty:     10244 kB
> -Referenced:        10244 kB
> -Anonymous:         10244 kB
> +Private_Dirty:         0 kB
> +Referenced:            0 kB
> +Anonymous:             0 kB
>  AnonHugePages:         0 kB
>  Swap:                  0 kB
>  KernelPageSize:        4 kB
> 
> Those are the 10MB I saw. For the record I also tried with malloc() and
> the results are similar to those of posix_memalign().

The posix_memalign() call wastes memory.  I compared:

  posix_memalign(&memptr, 65536, 2560 * 65536);
  memset(memptr, 0, 2560 * 65536);

with:

  for (i = 0; i < 2560; i++) {
      posix_memalign(&memptr, 65536, 65536);
      memset(memptr, 0, 65536);
  }

Here are the results:

-Size:             163920 kB
-Rss:              163860 kB
-Pss:              163860 kB
+Size:             337800 kB
+Rss:              183620 kB
+Pss:              183620 kB

Note the memset simulates a fully occupied cache.

The 19 MB RSS difference between the two seems wasteful.  The large
"Size" difference hints that the mmap pattern is very different when
posix_memalign() is called multiple times.

We could avoid the 19 MB overhead by switching to a single allocation.

What's more is that dropping the memset() to simulate no cache entry
usage (like your example) gives us a grand total of 20 kB RSS.  There is
no point in delaying allocations if we do a single big upfront
allocation.

I'd prefer a patch that replaces the small allocations with a single big
one.  That's a win in both empty and full cache cases.

Stefan

[-- Attachment #2: Type: application/pgp-signature, Size: 473 bytes --]

  reply	other threads:[~2015-04-24  9:26 UTC|newest]

Thread overview: 15+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2015-04-21 15:20 [Qemu-devel] [PATCH] qcow2: do lazy allocation of the L2 cache Alberto Garcia
2015-04-22 10:26 ` [Qemu-devel] [Qemu-block] " Stefan Hajnoczi
2015-04-22 14:37   ` Alberto Garcia
2015-04-23 10:15     ` Stefan Hajnoczi
2015-04-23 11:50       ` Alberto Garcia
2015-04-24  9:26         ` Stefan Hajnoczi [this message]
2015-04-24  9:45           ` Kevin Wolf
2015-04-24  9:52           ` Kevin Wolf
2015-04-24 11:10             ` Alberto Garcia
2015-04-24 12:37               ` Stefan Hajnoczi
2015-04-24 12:50                 ` Alberto Garcia
2015-04-24 13:04                   ` Kevin Wolf
2015-05-08  9:00                     ` Alberto Garcia
2015-05-08  9:47                       ` Kevin Wolf
2015-05-08 11:47                         ` Alberto Garcia

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20150424092621.GD13278@stefanha-thinkpad.redhat.com \
    --to=stefanha@gmail.com \
    --cc=berto@igalia.com \
    --cc=qemu-block@nongnu.org \
    --cc=qemu-devel@nongnu.org \
    --cc=stefanha@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).