xen-devel.lists.xenproject.org archive mirror
 help / color / mirror / Atom feed
From: Chong Li <lichong659@gmail.com>
To: Jan Beulich <JBeulich@suse.com>
Cc: Chong Li <chong.li@wustl.edu>, Wei Liu <wei.liu2@citrix.com>,
	Sisu Xi <xisisu@gmail.com>,
	"george.dunlap" <george.dunlap@eu.citrix.com>,
	"dario.faggioli" <dario.faggioli@citrix.com>,
	xen-devel <xen-devel@lists.xen.org>,
	Meng Xu <mengxu@cis.upenn.edu>,
	Dagaen Golomb <dgolomb@seas.upenn.edu>
Subject: Re: [PATCH v1 1/4] xen: enabling XL to set per-VCPU parameters of a domain for RTDS scheduler
Date: Thu, 14 May 2015 17:27:22 -0500	[thread overview]
Message-ID: <CAGHO-iosmqj3odmpTXkZrvp2G8QALa3OxA8aU1Aays4Zazkc-w@mail.gmail.com> (raw)
In-Reply-To: <55506EF90200007800078B4B@mail.emea.novell.com>


[-- Attachment #1.1: Type: text/plain, Size: 5737 bytes --]

On Mon, May 11, 2015 at 1:57 AM, Jan Beulich <JBeulich@suse.com> wrote:

> >>> On 11.05.15 at 00:04, <lichong659@gmail.com> wrote:
> > On Fri, May 8, 2015 at 2:49 AM, Jan Beulich <JBeulich@suse.com> wrote:
> >> >>> On 07.05.15 at 19:05, <lichong659@gmail.com> wrote:
> >> > @@ -1110,6 +1113,67 @@ rt_dom_cntl(
> >> >          }
> >> >          spin_unlock_irqrestore(&prv->lock, flags);
> >> >          break;
> >> > +    case XEN_DOMCTL_SCHEDOP_getvcpuinfo:
> >> > +        op->u.rtds.nr_vcpus = 0;
> >> > +        spin_lock_irqsave(&prv->lock, flags);
> >> > +        list_for_each( iter, &sdom->vcpu )
> >> > +            vcpu_index++;
> >> > +        spin_unlock_irqrestore(&prv->lock, flags);
> >> > +        op->u.rtds.nr_vcpus = vcpu_index;
> >>
> >> Does dropping of the lock here and re-acquiring it below really work
> >> race free?
> >>
> >
> > Here, the lock is used in the same way as the ones in the two cases
> > above (XEN_DOMCTL_SCHEDOP_get/putinfo). So I think if race free
> > is guaranteed in that two cases, the lock in this case works race free
> > as well.
>
> No - the difference is that in the {get,put}info cases it is being
> acquired just once each.
>

I see. I changed it based on Dario's suggestions.

>
> >> > +        vcpu_index = 0;
> >> > +        spin_lock_irqsave(&prv->lock, flags);
> >> > +        list_for_each( iter, &sdom->vcpu )
> >> > +        {
> >> > +            struct rt_vcpu *svc = list_entry(iter, struct rt_vcpu,
> >> sdom_elem);
> >> > +
> >> > +            local_sched[vcpu_index].budget = svc->budget /
> MICROSECS(1);
> >> > +            local_sched[vcpu_index].period = svc->period /
> MICROSECS(1);
> >> > +            local_sched[vcpu_index].index = vcpu_index;
> >>
> >> What use is this index to the caller? I think you rather want to tell it
> >> the vCPU number. That's especially also taking the use case of a
> >> get/set pair into account - unless you tell me that these indexes can
> >> never change, the indexes passed back into the set operation would
> >> risk to have become stale by the time the hypervisor processes the
> >> request.
> >>
> >
> > I don't quite understand what the "stale" means. The array here
> > (local_sched[ ])
> > and the array (in libxc) that local_sched[ ] is copied to are both used
> for
> > this get
> > operation only. When users set per-vcpu parameters, there are also
> > dedicated
> > arrays for that set operation.
>
> Just clarify this for me (and maybe yourself): Is the vCPU number
> <-> vcpu_index mapping invariable for the lifetime of a domain?
> If it isn't, the vCPU for a particular vcpu_index during a "get"
> may be different from that for the same vcpu_index during a
> subsequent "set".
>

Here the vcpu_index means the vcpu_id. I'll use svc->vcpu.vcpu_id instead
of the
vcpu_index in next version.


>
> >> > +        if( local_sched == NULL )
> >> > +        {
> >> > +            return -ENOMEM;
> >> > +        }
> >> > +        copy_from_guest(local_sched, op->u.rtds.vcpus,
> >> op->u.rtds.nr_vcpus);
> >> > +
> >> > +        for( i = 0; i < op->u.rtds.nr_vcpus; i++ )
> >> > +        {
> >> > +            vcpu_index = 0;
> >> > +            spin_lock_irqsave(&prv->lock, flags);
> >> > +            list_for_each( iter, &sdom->vcpu )
> >> > +            {
> >> > +                struct rt_vcpu *svc = list_entry(iter, struct
> rt_vcpu,
> >> sdom_elem);
> >> > +                if ( local_sched[i].index == vcpu_index )
> >> > +                {
> >> > +                    if ( local_sched[i].period <= 0 ||
> >> local_sched[i].budget <= 0 )
> >> > +                         return -EINVAL;
> >> > +
> >> > +                    svc->period = MICROSECS(local_sched[i].period);
> >> > +                    svc->budget = MICROSECS(local_sched[i].budget);
> >> > +                    break;
> >> > +                }
> >> > +                vcpu_index++;
> >> > +            }
> >> > +            spin_unlock_irqrestore(&prv->lock, flags);
> >> > +        }
> >>
> >> Considering a maximum size guest, these two nested loops could
> >> require a couple of million iterations. That's too much without any
> >> preemption checks in the middle.
> >>
> >
> > The section protected by the lock is only the "list_for_each" loop, whose
> > running time is limited by the number of vcpus of a domain (32 at most).
>
> Since when is 32 the limit on the number of vCPU-s in a domain?
>

Based on Dario's suggestion, I'll use vcpu_id to locate the vcpu to set,
which cost much
less time.


>
> > If this does cause problems, I think adding a "hypercall_preempt_check()"
> > at the outside "for" loop may help. Is that right?
>
> Yes.
>
> >> > --- a/xen/common/schedule.c
> >> > +++ b/xen/common/schedule.c
> >> > @@ -1093,7 +1093,9 @@ long sched_adjust(struct domain *d, struct
> >> xen_domctl_scheduler_op *op)
> >> >
> >> >      if ( (op->sched_id != DOM2OP(d)->sched_id) ||
> >> >           ((op->cmd != XEN_DOMCTL_SCHEDOP_putinfo) &&
> >> > -          (op->cmd != XEN_DOMCTL_SCHEDOP_getinfo)) )
> >> > +          (op->cmd != XEN_DOMCTL_SCHEDOP_getinfo) &&
> >> > +          (op->cmd != XEN_DOMCTL_SCHEDOP_putvcpuinfo) &&
> >> > +          (op->cmd != XEN_DOMCTL_SCHEDOP_getvcpuinfo)) )
> >>
> >> Imo this should become a switch now.
> >>
> >
> > Do you mean "switch ( op->cmd )" ? I'm afraid that would make it look
> more
> > complicated.
>
> This may be a matter of taste to a certain degree, but I personally
> don't think a series of four almost identical comparisons reads any
> better than its switch() replacement. But it being a style issue, the
> ultimate decision is with George as the maintainer anyway.
>
> Jan
>



-- 
Chong Li
Department of Computer Science and Engineering
Washington University in St.louis

[-- Attachment #1.2: Type: text/html, Size: 8710 bytes --]

[-- Attachment #2: Type: text/plain, Size: 126 bytes --]

_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
http://lists.xen.org/xen-devel

  reply	other threads:[~2015-05-14 22:27 UTC|newest]

Thread overview: 25+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2015-05-07 17:05 [PATCH v1 0/4] Enabling XL to set per-VCPU parameters of a domain for RTDS scheduler Chong Li
2015-05-07 17:05 ` [PATCH v1 1/4] xen: enabling " Chong Li
2015-05-08  7:49   ` Jan Beulich
2015-05-10 22:04     ` Chong Li
2015-05-11  6:57       ` Jan Beulich
2015-05-14 22:27         ` Chong Li [this message]
2015-05-15 14:42           ` Dario Faggioli
2015-05-11 13:11   ` Dario Faggioli
2015-05-14 22:15     ` Chong Li
2015-05-07 17:05 ` [PATCH v1 2/4] libxc: " Chong Li
2015-05-11 13:27   ` Dario Faggioli
2015-05-07 17:05 ` [PATCH v1 3/4] libxl: " Chong Li
2015-05-11 14:06   ` Dario Faggioli
2015-05-15 15:24     ` Chong Li
2015-05-15 23:09       ` Dario Faggioli
2015-05-12 10:01   ` Dario Faggioli
2015-05-15 16:35     ` Chong Li
2015-05-15 23:02       ` Dario Faggioli
2015-05-22 17:57     ` Chong Li
2015-05-07 17:05 ` [PATCH v1 4/4] xl: " Chong Li
2015-05-14 14:24   ` Meng Xu
2015-05-14 14:39     ` Dario Faggioli
2015-05-14 14:43       ` Meng Xu
2015-05-11  9:56 ` [PATCH v1 0/4] Enabling " Dario Faggioli
2015-05-14 17:08   ` Chong Li

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=CAGHO-iosmqj3odmpTXkZrvp2G8QALa3OxA8aU1Aays4Zazkc-w@mail.gmail.com \
    --to=lichong659@gmail.com \
    --cc=JBeulich@suse.com \
    --cc=chong.li@wustl.edu \
    --cc=dario.faggioli@citrix.com \
    --cc=dgolomb@seas.upenn.edu \
    --cc=george.dunlap@eu.citrix.com \
    --cc=mengxu@cis.upenn.edu \
    --cc=wei.liu2@citrix.com \
    --cc=xen-devel@lists.xen.org \
    --cc=xisisu@gmail.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).