From: Andre Przywara <andre.przywara@amd.com>
To: "Cui, Dexuan" <dexuan.cui@intel.com>
Cc: "xen-devel@lists.xensource.com" <xen-devel@lists.xensource.com>,
Keir Fraser <keir.fraser@eu.citrix.com>
Subject: Re: [PATCH 0/5] [POST-4.0]: RFC: HVM NUMA guest support
Date: Sun, 4 Apr 2010 23:41:20 +0200 [thread overview]
Message-ID: <4BB90780.9050202@amd.com> (raw)
In-Reply-To: <ED3036A092A28F4C91B0B4360DD128EABE0C06C8@shzsmsx502.ccr.corp.intel.com>
Cui, Dexuan wrote:
> Hi Andre, will you re-post your patches?
Yes, I will do in the next days. I plan to add the missing automatic
assignment patch before posting.
> Now I think for the first implementation, we can make things simple,
> e.g, we should specify how many guest nodes (the "guestnodes" option
> in your patch -- I think "numa_nodes", or "nodes", may be a better
> naming) the hvm guest will see, and we distribute guest memory and
> vcpus uniformly among the guest nodes.
I agree, making things simple in the first step was also my intention.
We have enough time to make it better later if we have more experience
with it.
To be honest, my first try also used "nodes" and later "numa_nodes" to
specify the number, but I learned that it confuses users who don't see
the difference between host and guest NUMA functionality. So I wanted to
make sure that it is clear that this number is from the guest's point of
view.
> And we should add one more option "uniform_nodes" -- this boolean
> option's default value can be True, meaning if we can't construct
> uniform nodes to guest(e.g., on the related host node, no enough
> memory as expected can be allocated to the guest), the guest
> creation should fail. This option is useful to users who want
> predictable guest performance.
I agree that we need to avoid missing user influence, although I'd
prefer to have the word "strict" somewhere in this name. As I wrote in
one my earlier mails, I'd opt for a single option describing the policy,
the "strict" meaning could be integrated in there:
numa_policy="strict|uniform|automatic|none|single|..."
Regards,
Andre.
--
Andre Przywara
AMD-Operating System Research Center (OSRC), Dresden, Germany
Tel: +49 351 488-3567-12
next prev parent reply other threads:[~2010-04-04 21:41 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2010-02-04 21:50 [PATCH 0/5] [POST-4.0]: RFC: HVM NUMA guest support Andre Przywara
2010-02-05 16:35 ` Cui, Dexuan
2010-02-23 9:53 ` Andre Przywara
2010-02-25 13:14 ` Cui, Dexuan
2010-03-31 6:23 ` Cui, Dexuan
2010-04-04 21:41 ` Andre Przywara [this message]
2010-04-07 7:42 ` Cui, Dexuan
2010-04-07 7:52 ` Keir Fraser
2010-04-07 9:03 ` Andre Przywara
2010-04-07 9:28 ` Cui, Dexuan
2010-02-22 10:24 ` Cui, Dexuan
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4BB90780.9050202@amd.com \
--to=andre.przywara@amd.com \
--cc=dexuan.cui@intel.com \
--cc=keir.fraser@eu.citrix.com \
--cc=xen-devel@lists.xensource.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).