qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: "Chunguang Li" <lichunguang@hust.edu.cn>
To: quintela@redhat.com
Cc: qemu-devel@nongnu.org, dgilbert@redhat.com, amit.shah@redhat.com,
	pbonzini@redhat.com, stefanha@redhat.com
Subject: Re: [Qemu-devel] Abnormal observation during migration: too many "write-not-dirty" pages
Date: Wed, 15 Nov 2017 22:22:13 +0800 (GMT+08:00)	[thread overview]
Message-ID: <7d022afb.758c.15fc00f11b5.Coremail.lichunguang@hust.edu.cn> (raw)
In-Reply-To: <874lpvdgdz.fsf@secure.laptop>




> -----Original Messages-----
> From: "Juan Quintela" <quintela@redhat.com>
> Sent Time: 2017-11-15 17:45:44 (Wednesday)
> To: "Chunguang Li" <lichunguang@hust.edu.cn>
> Cc: qemu-devel@nongnu.org, dgilbert@redhat.com, amit.shah@redhat.com, pbonzini@redhat.com, stefanha@redhat.com
> Subject: Re: Abnormal observation during migration: too many "write-not-dirty" pages
> 
> "Chunguang Li" <lichunguang@hust.edu.cn> wrote:
> > Hi all! 
> 
> Hi
> 
> Sorry for the delay, I was on vacation an still getting up to speed.

Hi, Juan, thanks for your reply.

> 
> > I got a very abnormal observation for the VM migration. I found that many pages marked as dirty during
> > migration are "not really dirty", which is, their content are the same as the old version. 
> 
> I think your test is quite good, and I am also ashamed that 80% of
> "false" dirty pages is really a lot.
> 
> > I did the migration experiment like this: 
> >
> > During the setup phase of migration, first I suspended the VM. Then I copied all the pages within the guest
> > physical address space to a memory buffer as large as the guest memory size. After that, the dirty tracking
> > began and I resumed the VM. Besides, at the end
> > of each iteration, I also suspended the VM temporarily. During the suspension, I compared the content of all
> > the pages marked as dirty in this iteration byte-by-byte with their former copies inside the buffer. If the
> > content of one page was the same as its former copy, I recorded it as a "write-not-dirty" page (the page is
> > written exactly with the same content as the old version). Otherwise, I replaced this page in the buffer with
> > the new content, for the possible comparison in the future. After the reset of the dirty bitmap, I resumed the
> > VM. Thus, I obtain the proportion of the write-not-dirty pages within all the pages marked as dirty for each
> > pre-copy iteration. 
> 
> 
> vhost and friends could make a small difference here, but in general,
> this approach should be ok.
> 
> > I repeated this experiment with 15 workloads, which are 11 CPU2006 benchmarks, Memcached server,
> > kernel compilation, playing a video, and an idle VM. The CPU2006 benchmarks and Memcached are
> > write-intensive workloads. So almost all of them did not converge to stop-copy. 
> 
> That is the impressive part, 15 workloads.  Thanks for taking the effor.
> 
> BTW, do you have your qemu changes handy, just to be able to test
> locally, and "review" how do you measure things.

Sorry, I do not have my changes handy. But don't worry, I will send them to you tomorrow morning. It's night here.

> 
> 
> > Startlingly, the proportions of the write-not-dirty pages are quite high. Memcached and three CPU2006
> > benchmarks(zeusmp, mcf and bzip2) have the most high proportions. Their proportions of the write-not-dirty
> > pages within all the dirty pages are as high as 45%-80%.
> 
> Or the workload does really stupid things like:
> 
> a = 0;
> a = 1;
> a = 0;
> 
> This makes no sense at all.
> 
> Just in case, could you try to test this with xbzrle?  It should go well
> with this use case (but you need to get a big enough buffer to cache
> enough memory).

In fact, I have tested these workloads (the 45%-80% ones) with xbzrle. And when the buffer is big enough, they really go well. While they do not converge to stop-copy before, now they finish migration quickly.

> 
> 
> > The proportions of the other workloads are about
> > 5%-20%, which are also abnormal. According to my intuition, the proportion of write-not-dirty pages should be
> > far less than these numbers. I think it should be quite a particular case that one page is written with exactly
> > the same content as the former data. 
> 
> I agree with that.
> 
> > Besides, the zero pages are not counted for all the results. Because I think codes like memset() may write
> > large area of pages to zero pages, which are already zero pages before. 
> >
> > I excluded some possible unknown reasons with the machine hardware, because I repeated the experiments
> > with two sets of different machines. Then I guessed it might be related with the huge page feature. However,
> > the result was the same when I turned the huge page feature off in the OS. 
> 
> Huge page could have caused that.  Remember that we have transparent
> huge pages.  I have to look at that code.

In fact, the results are the same no matter I turn on or turn off the transparent huge pages in the OS.

Later, Chunguang.

> 
> > Now there are only two possible reasons in my opinion. 
> >
> > First, there is some bugs in the KVM kernel dirty tracking mechanism. It may mark some pages that do not
> > receive write request as dirty. 
> 
> That is a posibilty.
> 
> > Second, there is some bugs in the OS running inside the VM. It may issue some unnecessary write
> > requests. 
> >
> > What do you think about this abnormal phenomenon? Any advice or possible reasons or even guesses? I
> > appreciate any responses, because it has confused me for a long time. Thank you.
> 
> I would like to reproduce this.
> 
> Thanks for bringing this to our attention.
> 
> Later, Juan.


--
Chunguang Li, Ph.D. Candidate
Wuhan National Laboratory for Optoelectronics (WNLO)
Huazhong University of Science & Technology (HUST)
Wuhan, Hubei Prov., China



  reply	other threads:[~2017-11-15 14:22 UTC|newest]

Thread overview: 8+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2017-11-12  9:26 [Qemu-devel] Abnormal observation during migration: too many "write-not-dirty" pages Chunguang Li
2017-11-15  9:45 ` Juan Quintela
2017-11-15 14:22   ` Chunguang Li [this message]
2017-11-15 10:11 ` Dr. David Alan Gilbert
2017-11-15 13:41   ` Chunguang Li
2017-11-15 14:23     ` Dr. David Alan Gilbert
2017-11-16  3:01       ` Chunguang Li
  -- strict thread matches above, loose matches on Subject: below --
2017-11-15  6:24 Chunguang Li

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=7d022afb.758c.15fc00f11b5.Coremail.lichunguang@hust.edu.cn \
    --to=lichunguang@hust.edu.cn \
    --cc=amit.shah@redhat.com \
    --cc=dgilbert@redhat.com \
    --cc=pbonzini@redhat.com \
    --cc=qemu-devel@nongnu.org \
    --cc=quintela@redhat.com \
    --cc=stefanha@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).