From: Avi Kivity <avi@redhat.com>
To: Dave Hansen <dave@linux.vnet.ibm.com>
Cc: linux-kernel@vger.kernel.org, kvm@vger.kernel.org
Subject: Re: [RFC][PATCH 4/9] create aggregate kvm_total_used_mmu_pages value
Date: Wed, 16 Jun 2010 11:48:00 +0300 [thread overview]
Message-ID: <4C188FC0.3050306@redhat.com> (raw)
In-Reply-To: <20100615135523.25D24A73@kernel.beaverton.ibm.com>
On 06/15/2010 04:55 PM, Dave Hansen wrote:
> Note: this is the real meat of the patch set. It can be applied up
> to this point, and everything will probably be improved, at least
> a bit.
>
> Of slab shrinkers, the VM code says:
>
> * Note that 'shrink' will be passed nr_to_scan == 0 when the VM is
> * querying the cache size, so a fastpath for that case is appropriate.
>
> and it *means* it. Look at how it calls the shrinkers:
>
> nr_before = (*shrinker->shrink)(0, gfp_mask);
> shrink_ret = (*shrinker->shrink)(this_scan, gfp_mask);
>
> So, if you do anything stupid in your shrinker, the VM will doubly
> punish you.
>
Ouch.
> The mmu_shrink() function takes the global kvm_lock, then acquires
> every VM's kvm->mmu_lock in sequence. If we have 100 VMs, then
> we're going to take 101 locks. We do it twice, so each call takes
> 202 locks. If we're under memory pressure, we can have each cpu
> trying to do this. It can get really hairy, and we've seen lock
> spinning in mmu_shrink() be the dominant entry in profiles.
>
> This is guaranteed to optimize at least half of those lock
> aquisitions away. It removes the need to take any of the locks
> when simply trying to count objects.
>
>
> Signed-off-by: Dave Hansen<dave@linux.vnet.ibm.com>
> ---
>
> linux-2.6.git-dave/arch/x86/kvm/mmu.c | 33 +++++++++++++++++++++++----------
> 1 file changed, 23 insertions(+), 10 deletions(-)
>
> diff -puN arch/x86/kvm/mmu.c~make_global_used_value arch/x86/kvm/mmu.c
> --- linux-2.6.git/arch/x86/kvm/mmu.c~make_global_used_value 2010-06-09 15:14:30.000000000 -0700
> +++ linux-2.6.git-dave/arch/x86/kvm/mmu.c 2010-06-09 15:14:30.000000000 -0700
> @@ -891,6 +891,19 @@ static int is_empty_shadow_page(u64 *spt
> }
> #endif
>
> +/*
> + * This value is the sum of all of the kvm instances's
> + * kvm->arch.n_used_mmu_pages values. We need a global,
> + * aggregate version in order to make the slab shrinker
> + * faster
> + */
> +static unsigned int kvm_total_used_mmu_pages;
>
The variable needs to be at the top of the file.
> +static inline void kvm_mod_used_mmu_pages(struct kvm *kvm, int nr)
> +{
> + kvm->arch.n_used_mmu_pages += nr;
> + kvm_total_used_mmu_pages += nr;
>
Needs an atomic operation, since there's no global lock here. To avoid
bouncing this cacheline around, make the variable percpu and make
readers take a sum across all cpus. Side benefit is that you no longer
need an atomic but a local_t, which is considerably cheaper.
> +}
> +
>
--
error compiling committee.c: too many arguments to function
next prev parent reply other threads:[~2010-06-16 8:48 UTC|newest]
Thread overview: 31+ messages / expand[flat|nested] mbox.gz Atom feed top
2010-06-15 13:55 [RFC][PATCH 0/9] rework KVM mmu_shrink() code Dave Hansen
2010-06-15 13:55 ` [RFC][PATCH 1/9] abstract kvm x86 mmu->n_free_mmu_pages Dave Hansen
2010-06-16 8:40 ` Avi Kivity
2010-06-15 13:55 ` [RFC][PATCH 2/9] rename x86 kvm->arch.n_alloc_mmu_pages Dave Hansen
2010-06-15 13:55 ` [RFC][PATCH 3/9] replace x86 kvm n_free_mmu_pages with n_used_mmu_pages Dave Hansen
2010-06-16 14:25 ` Marcelo Tosatti
2010-06-16 15:42 ` Dave Hansen
2010-06-15 13:55 ` [RFC][PATCH 4/9] create aggregate kvm_total_used_mmu_pages value Dave Hansen
2010-06-16 8:48 ` Avi Kivity [this message]
2010-06-16 15:06 ` Dave Hansen
2010-06-17 8:43 ` Avi Kivity
2010-06-16 16:55 ` Dave Hansen
2010-06-17 8:23 ` Avi Kivity
2010-06-15 13:55 ` [RFC][PATCH 5/9] break out some mmu_skrink() code Dave Hansen
2010-06-15 13:55 ` [RFC][PATCH 6/9] remove kvm_freed variable Dave Hansen
2010-06-15 13:55 ` [RFC][PATCH 7/9] make kvm_get_kvm() more robust Dave Hansen
2010-06-15 13:55 ` [RFC][PATCH 8/9] reduce kvm_lock hold times in mmu_skrink() Dave Hansen
2010-06-16 8:54 ` Avi Kivity
2010-06-15 13:55 ` [RFC][PATCH 9/9] make kvm mmu shrinker more aggressive Dave Hansen
2010-06-16 9:24 ` Avi Kivity
2010-06-16 15:25 ` Dave Hansen
2010-06-17 8:37 ` Avi Kivity
2010-06-18 15:49 ` Dave Hansen
2010-06-20 8:11 ` Avi Kivity
2010-06-22 16:32 ` Dave Hansen
2010-07-22 4:36 ` Avi Kivity
2010-07-22 5:36 ` Dave Hansen
2010-07-22 5:42 ` Avi Kivity
2010-06-16 8:38 ` [RFC][PATCH 0/9] rework KVM mmu_shrink() code Avi Kivity
2010-06-16 15:03 ` Dave Hansen
2010-06-17 8:40 ` Avi Kivity
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4C188FC0.3050306@redhat.com \
--to=avi@redhat.com \
--cc=dave@linux.vnet.ibm.com \
--cc=kvm@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).