From: Yi Sun <yi.y.sun@linux.intel.com>
To: Jan Beulich <JBeulich@suse.com>
Cc: kevin.tian@intel.com, wei.liu2@citrix.com,
he.chen@linux.intel.com, andrew.cooper3@citrix.com,
dario.faggioli@citrix.com, ian.jackson@eu.citrix.com,
mengxu@cis.upenn.edu, xen-devel@lists.xenproject.org,
chao.p.peng@linux.intel.com, roger.pau@citrix.com
Subject: Re: [PATCH v10 09/25] x86: refactor psr: L3 CAT: set value: implement framework.
Date: Thu, 13 Apr 2017 19:11:54 +0800 [thread overview]
Message-ID: <20170413111153.GH17458@yi.y.sun> (raw)
In-Reply-To: <58EF75DE0200007800150B1E@prv-mh.provo.novell.com>
On 17-04-13 04:58:06, Jan Beulich wrote:
> >>> On 13.04.17 at 12:49, <yi.y.sun@linux.intel.com> wrote:
> > On 17-04-13 03:41:44, Jan Beulich wrote:
> >> >>> On 13.04.17 at 10:11, <yi.y.sun@linux.intel.com> wrote:
> >> > On 17-04-12 06:42:01, Jan Beulich wrote:
> >> >> >>> On 12.04.17 at 14:23, <yi.y.sun@linux.intel.com> wrote:
> >> >> > On 17-04-12 03:09:56, Jan Beulich wrote:
> >> >> >> >>> On 12.04.17 at 07:53, <yi.y.sun@linux.intel.com> wrote:
> >> >> >> > On 17-04-11 09:01:53, Jan Beulich wrote:
> >> >> >> >> >>> On 01.04.17 at 15:53, <yi.y.sun@linux.intel.com> wrote:
> >> >> >> >> Furthermore I'm not at all convinced this is appropriate to do in the
> >> >> >> >> context of a CPU_UP_CANCELED / CPU_DEAD notification: If you
> >> >> >> >> have a few thousand VMs, the loop above may take a while.
> >> >> >> >>
> >> >> >> > Hmm, that may be a potential issue. I have two proposals below. Could you
> >> >> >> > please help to check which one you prefer? Or provide another solution?
> >> >> >> >
> >> >> >> > 1. Start a tasklet in free_socket_resources() to restore
> >> >> > 'psr_cos_ids[socket]'
> >> >> >> > of all domains. The action is protected by 'ref_lock' to avoid
> >> >> > confliction
> >> >> >> > in 'psr_set_val'. We can reduce 'info->cos_ref[cos]' in tasklet or
> > memset
> >> >> >> > the array to 0 in free_socket_resources().
> >> >> >> >
> >> >> >> > 2. Move 'psr_cos_ids[]' from 'domain' to 'psr_socket_info' and change
> > index
> >> >> >> > from 'socket' to 'domain_id'. So we keep all domains' COS IDs per
> > socket
> >> >> >> > and can memset the array to 0 when socket is offline. But here is an
> >> >> > issue
> >> >> >> > that we do not know how many members this array should have. I cannot
> >> >> > find
> >> >> >> > a macro something like 'DOMAIN_MAX_NUMBER'. So, I prefer to use
> >> >> > reallocation
> >> >> >> > in 'psr_alloc_cos' if the newly created domain's id is bigger than
> >> >> > current
> >> >> >> > array number.
> >> >> >>
> >> >> >> The number of domains is limited by the special DOMID_* values.
> >> >> >> However, allocating an array with 32k entries doesn't sound very
> >> >> >> reasonable.
> >> >> >
> >> >> > I think 32K entries should be the extreme case. I can allocate e.g. 100
> > entries
> >> >> > when the first domain is created. If a new domain's id exceeds 100,
> > reallocate
> >> >> > another 100 entries. The total number of entries allocated should be less
> > than
> >> >> > 32K. This is a functional requirement which cannot be avoided. How do you
> >> >> > think?
> >> >>
> >> >> So how many entries would your array have once I start the 32,000th
> >> >> domain (having at any one time at most a single one running, besides
> >> >> Dom0)?
> >> >>
> >> > In such case, we have to keep a 32K array because the domain_id is the
> > index to
> >> > access the array. But this array is per socket so the whole memory used
> > should
> >> > not be too much.
> >>
> >> We carefully avoid any runtime allocations of order > 0, so if you
> >> were to set up such an array, you'd need to use vmalloc()/vzalloc().
> >> But I continue to be unconvinced that we want such a large array
> >> in the first place.
> >>
> >> > After considering this issue more, I think the original codes might not be
> >> > so unacceptable. Per my knowledge, Intel Xeon Phi chip can support at most
> >> > 288 CPUs. So, I think the domains running at same time in reality may not
> > be
> >> > so many (no efficient resources). If this hypothesis is right, a loop to
> > write
> >> > 'psr_cos_ids[socket]' of every domain to 0 may not take much time. If I am
> >> > wrong, please correct me. Thanks!
> >>
> >> What relationship does the number of CPUs have to the number of
> >> domains on a host? There could be thousands with just a few dozen
> >> CPUs, provided none or very few of them have high demands on
> >> CPU resources. Additionally please never forget that system sizes
> >> basically only ever grow. Plus we wouldn't want a latent issue here
> >> in case we ever end up needing to widen domain IDs beyond 16 bits.
> >>
> > How about a per socket array like this:
> > uint32_t domain_switch[1024];
> >
> > Every bit represents a domain id. Then, we can handle this case as below:
> > 1. In 'psr_cpu_init()', clear the array to be 0. I think this place is enough to
> > cover socket offline case. We do not need to clear it in
> > 'free_socket_resources'.
> >
> > 2. In 'psr_ctxt_switch_to()', test_and_set_bit(domain_id, domain_switch) to set
> > the bit to 1 according to domain_id. If the old value is 0 and the
> > 'psr_cos_ids[socket]' is not 0, restore 'psr_cos_ids[socket]' to be 0.
> >
> > 3. In 'psr_set_val()', test_and_set_bit(domain_id, domain_switch) to set the bit
> > to 1 too. Then, update 'psr_cos_ids[socket]' according to find/pick flow.
> >
> > Then, we only use 4KB for one socket.
>
> This looks to come closer to something I'd consider acceptable, but
> I may not understand your intentions in full yet: For one, there's
> nowhere you clear the bit (other than presumably during socket
> cleanup).
Actually, clear the array in 'free_socket_resources' has same effect. I can
move clear action into it.
> And then I don't understand the test_and_ parts of the
> constructs above, i.e. you don't clarify what the return values
> would be used/needed for.
>
Sorry, 0 means this domain has not been scheduled to the socket yet. If
test_and_ returns 0, that is the first time the domain runs on the socket
(the first time the socket is online). So, we need restore 'psr_cos_ids[socket]'
to 0 in 'psr_ctxt_switch_to()'.
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
https://lists.xen.org/xen-devel
next prev parent reply other threads:[~2017-04-13 11:11 UTC|newest]
Thread overview: 114+ messages / expand[flat|nested] mbox.gz Atom feed top
2017-04-01 13:53 [PATCH v10 00/25] Enable L2 Cache Allocation Technology & Refactor psr.c Yi Sun
2017-04-01 13:53 ` [PATCH v10 01/25] docs: create Cache Allocation Technology (CAT) and Code and Data Prioritization (CDP) feature document Yi Sun
2017-04-01 13:53 ` [PATCH v10 02/25] x86: refactor psr: remove L3 CAT/CDP codes Yi Sun
2017-04-01 13:53 ` [PATCH v10 03/25] x86: refactor psr: implement main data structures Yi Sun
2017-04-03 15:50 ` Jan Beulich
2017-04-05 3:12 ` Yi Sun
2017-04-05 8:20 ` Jan Beulich
2017-04-05 8:45 ` Yi Sun
2017-04-01 13:53 ` [PATCH v10 04/25] x86: move cpuid_count_leaf from cpuid.c to processor.h Yi Sun
2017-04-01 13:53 ` [PATCH v10 05/25] x86: refactor psr: L3 CAT: implement CPU init and free flow Yi Sun
2017-04-05 15:10 ` Jan Beulich
2017-04-06 5:49 ` Yi Sun
2017-04-06 8:32 ` Jan Beulich
2017-04-06 9:22 ` Yi Sun
2017-04-06 9:34 ` Jan Beulich
2017-04-06 10:02 ` Yi Sun
2017-04-06 14:02 ` Jan Beulich
2017-04-07 5:17 ` Yi Sun
2017-04-07 8:48 ` Jan Beulich
2017-04-07 9:08 ` Yi Sun
2017-04-07 9:46 ` Jan Beulich
2017-04-10 3:27 ` Yi Sun
2017-04-10 12:43 ` Yi Sun
2017-04-01 13:53 ` [PATCH v10 06/25] x86: refactor psr: L3 CAT: implement Domain init/free and schedule flows Yi Sun
2017-04-05 15:23 ` Jan Beulich
2017-04-06 6:01 ` Yi Sun
2017-04-06 8:34 ` Jan Beulich
2017-04-01 13:53 ` [PATCH v10 07/25] x86: refactor psr: L3 CAT: implement get hw info flow Yi Sun
2017-04-05 15:37 ` Jan Beulich
2017-04-06 6:05 ` Yi Sun
2017-04-06 8:36 ` Jan Beulich
2017-04-06 11:16 ` Yi Sun
2017-04-06 14:04 ` Jan Beulich
2017-04-07 5:39 ` Yi Sun
2017-04-01 13:53 ` [PATCH v10 08/25] x86: refactor psr: L3 CAT: implement get value flow Yi Sun
2017-04-05 15:51 ` Jan Beulich
2017-04-06 6:10 ` Yi Sun
2017-04-06 8:40 ` Jan Beulich
2017-04-06 11:13 ` Yi Sun
2017-04-06 14:08 ` Jan Beulich
2017-04-07 5:40 ` Yi Sun
2017-04-01 13:53 ` [PATCH v10 09/25] x86: refactor psr: L3 CAT: set value: implement framework Yi Sun
2017-04-11 15:01 ` Jan Beulich
2017-04-12 5:53 ` Yi Sun
2017-04-12 9:09 ` Jan Beulich
2017-04-12 12:23 ` Yi Sun
2017-04-12 12:42 ` Jan Beulich
2017-04-13 8:11 ` Yi Sun
2017-04-13 9:41 ` Jan Beulich
2017-04-13 10:49 ` Yi Sun
2017-04-13 10:58 ` Jan Beulich
2017-04-13 11:11 ` Yi Sun [this message]
2017-04-13 11:26 ` Yi Sun
2017-04-13 11:31 ` Jan Beulich
2017-04-13 11:44 ` Yi Sun
2017-04-13 11:50 ` Jan Beulich
2017-04-18 10:55 ` Yi Sun
2017-04-18 11:46 ` Jan Beulich
2017-04-19 8:22 ` Yi Sun
2017-04-19 9:00 ` Jan Beulich
2017-04-20 2:14 ` Yi Sun
2017-04-20 9:43 ` Jan Beulich
2017-04-20 13:02 ` Lars Kurth
2017-04-20 13:21 ` Jan Beulich
2017-04-20 16:52 ` Lars Kurth
2017-04-21 6:11 ` Jan Beulich
2017-04-21 1:13 ` Konrad Rzeszutek Wilk
2017-04-21 6:18 ` Jan Beulich
2017-04-24 6:40 ` Yi Sun
2017-04-24 6:55 ` Jan Beulich
2017-04-25 7:15 ` Yi Sun
2017-04-25 8:24 ` Jan Beulich
2017-04-25 8:40 ` Yi Sun
2017-04-20 5:38 ` [PATCH] dom_ids array implementation Yi Sun
2017-04-26 10:04 ` Jan Beulich
2017-04-27 2:38 ` Yi Sun
2017-04-27 6:48 ` Jan Beulich
2017-04-27 9:30 ` Yi Sun
2017-04-27 9:39 ` Jan Beulich
2017-04-27 12:03 ` Yi Sun
2017-04-01 13:53 ` [PATCH v10 10/25] x86: refactor psr: L3 CAT: set value: assemble features value array Yi Sun
2017-04-11 15:11 ` Jan Beulich
2017-04-12 5:55 ` Yi Sun
2017-04-12 9:13 ` Jan Beulich
2017-04-12 12:26 ` Yi Sun
2017-04-01 13:53 ` [PATCH v10 11/25] x86: refactor psr: L3 CAT: set value: implement cos finding flow Yi Sun
2017-04-11 15:17 ` Jan Beulich
2017-04-01 13:53 ` [PATCH v10 12/25] x86: refactor psr: L3 CAT: set value: implement cos id picking flow Yi Sun
2017-04-11 15:20 ` Jan Beulich
2017-04-01 13:53 ` [PATCH v10 13/25] x86: refactor psr: L3 CAT: set value: implement write msr flow Yi Sun
2017-04-11 15:25 ` Jan Beulich
2017-04-12 6:04 ` Yi Sun
2017-04-01 13:53 ` [PATCH v10 14/25] x86: refactor psr: CDP: implement CPU init and free flow Yi Sun
2017-04-01 13:53 ` [PATCH v10 15/25] x86: refactor psr: CDP: implement get hw info flow Yi Sun
2017-04-01 13:53 ` [PATCH v10 16/25] x86: refactor psr: CDP: implement get value flow Yi Sun
2017-04-11 15:39 ` Jan Beulich
2017-04-12 6:05 ` Yi Sun
2017-04-12 9:14 ` Jan Beulich
2017-04-01 13:53 ` [PATCH v10 17/25] x86: refactor psr: CDP: implement set value callback functions Yi Sun
2017-04-11 16:03 ` Jan Beulich
2017-04-12 6:14 ` Yi Sun
2017-04-01 13:53 ` [PATCH v10 18/25] x86: L2 CAT: implement CPU init and free flow Yi Sun
2017-04-12 15:18 ` Jan Beulich
2017-04-13 8:12 ` Yi Sun
2017-04-13 8:16 ` Jan Beulich
2017-04-01 13:53 ` [PATCH v10 19/25] x86: L2 CAT: implement get hw info flow Yi Sun
2017-04-01 13:53 ` [PATCH v10 20/25] x86: L2 CAT: implement get value flow Yi Sun
2017-04-01 13:53 ` [PATCH v10 21/25] x86: L2 CAT: implement set " Yi Sun
2017-04-12 15:23 ` Jan Beulich
2017-04-01 13:53 ` [PATCH v10 22/25] tools: L2 CAT: support get HW info for L2 CAT Yi Sun
2017-04-12 15:24 ` Jan Beulich
2017-04-01 13:53 ` [PATCH v10 23/25] tools: L2 CAT: support show cbm " Yi Sun
2017-04-01 13:53 ` [PATCH v10 24/25] tools: L2 CAT: support set " Yi Sun
2017-04-01 13:53 ` [PATCH v10 25/25] docs: add L2 CAT description in docs Yi Sun
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20170413111153.GH17458@yi.y.sun \
--to=yi.y.sun@linux.intel.com \
--cc=JBeulich@suse.com \
--cc=andrew.cooper3@citrix.com \
--cc=chao.p.peng@linux.intel.com \
--cc=dario.faggioli@citrix.com \
--cc=he.chen@linux.intel.com \
--cc=ian.jackson@eu.citrix.com \
--cc=kevin.tian@intel.com \
--cc=mengxu@cis.upenn.edu \
--cc=roger.pau@citrix.com \
--cc=wei.liu2@citrix.com \
--cc=xen-devel@lists.xenproject.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).