From: George Dunlap <george.dunlap@eu.citrix.com>
To: Dario Faggioli <dario.faggioli@citrix.com>
Cc: Marcus Granado <Marcus.Granado@eu.citrix.com>,
Justin Weaver <jtweaver@hawaii.edu>, Matt Wilson <msw@amazon.com>,
Li Yechen <lccycc123@gmail.com>,
Andrew Cooper <Andrew.Cooper3@citrix.com>,
Juergen Gross <juergen.gross@ts.fujitsu.com>,
Ian Jackson <Ian.Jackson@eu.citrix.com>,
Jan Beulich <JBeulich@suse.com>,
xen-devel <xen-devel@lists.xenproject.org>,
Keir Fraser <keir@xen.org>, Elena Ufimtseva <ufimtseva@gmail.com>,
Ian Campbell <Ian.Campbell@citrix.com>
Subject: Re: [PATCH v5 12/17] xen/libxc: sched: DOMCTL_*vcpuaffinity works with hard and soft affinity
Date: Wed, 4 Dec 2013 15:49:36 +0000 [thread overview]
Message-ID: <529F4F10.5040407@eu.citrix.com> (raw)
In-Reply-To: <1386147821.5338.394.camel@Solace>
On 12/04/2013 09:03 AM, Dario Faggioli wrote:
> On mar, 2013-12-03 at 20:06 +0100, Dario Faggioli wrote:
>> On mar, 2013-12-03 at 18:37 +0000, George Dunlap wrote:
>>> It is worth looking at the whole series again to try to see what the
>>> risks are, and if it's still worth taking. I'll probably send something
>>> out tomorrow.
>>>
>> Right. Since you pronounced yourself for the exception fairly early, I
>> never include such analysis in further releases. I think it's on me to
>> provide it, so I will do that (tomorrow too, so feel free to wait for
>> mine, if you want).
>>
> So, risk-vs-benefits analysis.
>
> If this still was only per-vCPU NUMA affinity, as it started, there
> would be no point in having it: no in-tree consumers, unlikely to be
> noticed and used by actual users. However, the way we redesigned and put
> it, makes it general enough to be interesting even independently from
> NUMA. It now is quite an advanced feature which, as far as I know, not
> many other OSes or hypervisors have, and it comes at a very reasonable
> cost, in terms of amount of code and magnitude of infrastructural (in
> the scheduling subsystem) changes. Actually, the latter is a
> simplification wrt what we have now!
>
> Granted that it provides benefits, risks. I think there are two kinds of
> risks: one is related bugs (of course), the other has to do with the
> interface.
>
> Bugs wise, I tend to agree to what George said in his last e-mail. Most
> of the hypervisor work which happens in 'common code' (i.e., will affect
> people not using this feature) is just refactoring and rewiring various
> bits and pieces. Most of the new code is in enabling the feature at Xen,
> libxc and libxl level and bugs there, for one, shouldn't be too hard to
> spot and fix (as it happened right during v5), and could only be
> triggered from dom0 (domU creation or via the new toolstack command
> being introduced).
>
> I think the (potential) interface issues are the more important. In
> fact, the interface not being the optimal one (at the Xen and xc level)
> and not getting much attention (at the libxl level) in the first
> versions of the patch series is what brought us here, this late into
> code freeze. My personal opinion is that we have finally reached a point
> where the interface is consistent and easy to maintain and to extend in
> a compatible way, where that is needed (see, for instance, the
> conversation with Ian Campbell about xl options).
I agree with the general description of the situation. I just went
through all of the patches yesterday and tried to break down the "risk"
by evaluating the complexity of the patch, the potential impact if there
were a bug, and the "reviewer risk" (i.e., how well the reviewers seemed
comfortable with the code / interface). (I'll paste in my notes below
for those interested.)
The patches are fairly straightforward, and the risk for most patches is
fairly small and simple. The core patches I think it likely that the
worst that could happen if there is a bug would be a performance
regression. So from a pure code perspective, I think the risk is on the
low side.
The thing that is more risky is, as you say, the interface. It's still
quite a bit in flux, particularly the command-line part.
And as far as coolness -- while it is definitely a cool feature, I'm not
sure it's so critical that it can't wait for the next release. If we
really want to prioritize a stable, bug-free release, I think we'll
probably have to say 'wait' at this point to the soft affinity feature.
Other views are welcome. :-)
-George
next prev parent reply other threads:[~2013-12-04 15:49 UTC|newest]
Thread overview: 41+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-12-02 18:27 [PATCH v5 00/17] Implement vcpu soft affinity for credit1 Dario Faggioli
2013-12-02 18:27 ` [PATCH v5 01/17] xl: match output of vcpu-list with pinning syntax Dario Faggioli
2013-12-02 18:27 ` [PATCH v5 02/17] libxl: better name for last parameter of libxl_list_vcpu Dario Faggioli
2013-12-04 11:40 ` Ian Jackson
2013-12-06 14:40 ` Dario Faggioli
2013-12-02 18:27 ` [PATCH v5 03/17] libxl: fix memory leak in libxl_list_vcpu Dario Faggioli
2013-12-05 12:07 ` Ian Jackson
2013-12-02 18:27 ` [PATCH v5 04/17] libxc/libxl: sanitize error handling in *_get_max_{cpus, nodes} Dario Faggioli
2013-12-05 12:10 ` Ian Jackson
2013-12-06 10:34 ` Dario Faggioli
2013-12-06 11:52 ` Ian Jackson
2013-12-02 18:27 ` [PATCH v5 05/17] libxc/libxl: allow to retrieve the number of online pCPUs Dario Faggioli
2013-12-02 18:28 ` [PATCH v5 06/17] xl: allow for node-wise specification of vcpu pinning Dario Faggioli
2013-12-02 18:28 ` [PATCH v5 07/17] xl: implement and enable dryrun mode for `xl vcpu-pin' Dario Faggioli
2013-12-02 18:28 ` [PATCH v5 08/17] xl: test script for the cpumap parser (for vCPU pinning) Dario Faggioli
2013-12-02 18:28 ` [PATCH v5 09/17] xen: sched: rename v->cpu_affinity into v->cpu_hard_affinity Dario Faggioli
2013-12-02 18:28 ` [PATCH v5 10/17] xen: sched: introduce soft-affinity and use it instead d->node-affinity Dario Faggioli
2013-12-02 18:28 ` [PATCH v5 11/17] xen: derive NUMA node affinity from hard and soft CPU affinity Dario Faggioli
2013-12-02 18:29 ` [PATCH v5 12/17] xen/libxc: sched: DOMCTL_*vcpuaffinity works with hard and soft affinity Dario Faggioli
2013-12-03 10:02 ` Jan Beulich
2013-12-03 10:06 ` Jan Beulich
2013-12-03 11:08 ` Dario Faggioli
2013-12-03 13:25 ` Dario Faggioli
2013-12-03 18:21 ` George Dunlap
2013-12-03 18:29 ` Dario Faggioli
2013-12-03 18:37 ` George Dunlap
2013-12-03 19:06 ` Dario Faggioli
2013-12-04 9:03 ` Dario Faggioli
2013-12-04 15:49 ` George Dunlap [this message]
2013-12-04 16:03 ` Dario Faggioli
2013-12-04 16:20 ` Jan Beulich
2013-12-11 11:33 ` Jan Beulich
2013-12-03 10:59 ` Dario Faggioli
2013-12-03 11:20 ` Jan Beulich
2013-12-03 11:30 ` Dario Faggioli
2013-12-02 18:29 ` [PATCH v5 13/17] libxc: get and set soft and hard affinity Dario Faggioli
2013-12-02 18:29 ` [PATCH v5 14/17] libxl: get and set soft affinity Dario Faggioli
2013-12-02 18:29 ` [PATCH v5 15/17] xl: enable getting and setting soft Dario Faggioli
2013-12-02 18:29 ` [PATCH v5 16/17] xl: enable for specifying node-affinity in the config file Dario Faggioli
2013-12-02 18:29 ` [PATCH v5 17/17] libxl: automatic NUMA placement affects soft affinity Dario Faggioli
2013-12-03 14:05 ` [PATCH v5 00/17] Implement vcpu soft affinity for credit1 George Dunlap
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=529F4F10.5040407@eu.citrix.com \
--to=george.dunlap@eu.citrix.com \
--cc=Andrew.Cooper3@citrix.com \
--cc=Ian.Campbell@citrix.com \
--cc=Ian.Jackson@eu.citrix.com \
--cc=JBeulich@suse.com \
--cc=Marcus.Granado@eu.citrix.com \
--cc=dario.faggioli@citrix.com \
--cc=jtweaver@hawaii.edu \
--cc=juergen.gross@ts.fujitsu.com \
--cc=keir@xen.org \
--cc=lccycc123@gmail.com \
--cc=msw@amazon.com \
--cc=ufimtseva@gmail.com \
--cc=xen-devel@lists.xenproject.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).