xen-devel.lists.xenproject.org archive mirror
 help / color / mirror / Atom feed
From: anshul makkar <anshul.makkar@citrix.com>
To: Dario Faggioli <dario.faggioli@citrix.com>,
	xen-devel@lists.xenproject.org
Cc: George Dunlap <george.dunlap@citrix.com>,
	David Vrabel <david.vrabel@citrix.com>
Subject: Re: [PATCH 18/19] xen: credit2: implement SMT support independent runq arrangement
Date: Mon, 27 Jun 2016 16:20:01 +0100	[thread overview]
Message-ID: <57714421.3040501@citrix.com> (raw)
In-Reply-To: <146620520979.29766.17431818083809592415.stgit@Solace.fritz.box>

On 18/06/16 00:13, Dario Faggioli wrote:
> In fact, right now, we recommend keepeing runqueues
> arranged per-core, so that it is the inter-runqueue load
> balancing code that automatically spreads the work in an
> SMT friendly way. This means that any other runq
> arrangement one may want to use falls short of SMT
> scheduling optimizations.
>
> This commit implements SMT awareness --similar to the
> one we have in Credit1-- for any possible runq
> arrangement. This turned out to be pretty easy to do,
> as the logic can live entirely in runq_tickle()
> (although, in order to avoid for_each_cpu loops in
> that function, we use a new cpumask which indeed needs
> to be updated in other places).
>
> In addition to disentangling SMT awareness from load
> balancing, this also allows us to support the
> sched_smt_power_savings parametar in Credit2 as well.
>
> Signed-off-by: Dario Faggioli <dario.faggioli@citrix.com>

Reviewed-by: Anshul Makkar <anshul.makkar@citrix.com>
> ---
> Cc: George Dunlap <george.dunlap@citrix.com>
> Cc: Anshul Makkar <anshul.makkar@citrix.com>
> Cc: David Vrabel <david.vrabel@citrix.com>
> ---
>   xen/common/sched_credit2.c |  141 +++++++++++++++++++++++++++++++++++++++-----
>   1 file changed, 126 insertions(+), 15 deletions(-)
>
> diff --git a/xen/common/sched_credit2.c b/xen/common/sched_credit2.c
> index 93943fa..a8b3a85 100644
> --- a/xen/common/sched_credit2.c
> +++ b/xen/common/sched_credit2.c
> @@ -351,7 +351,8 @@ struct csched2_runqueue_data {
>       unsigned int max_weight;
>
>       cpumask_t idle,        /* Currently idle */
> -        tickled;           /* Another cpu in the queue is already targeted for this one */
> +        smt_idle,          /* Fully idle cores (as in all the siblings are idle) */
> +        tickled;           /* Have been asked to go through schedule */
>       int load;              /* Instantaneous load: Length of queue  + num non-idle threads */
>       s_time_t load_last_update;  /* Last time average was updated */
>       s_time_t avgload;           /* Decaying queue load */
> @@ -412,6 +413,73 @@ struct csched2_dom {
>   };
>
>   /*
> + * Hyperthreading (SMT) support.
> + *
> + * We use a special per-runq mask (smt_idle) and update it according to the
> + * following logic:
> + *  - when _all_ the SMT sibling in a core are idle, all their corresponding
> + *    bits are set in the smt_idle mask;
> + *  - when even _just_one_ of the SMT siblings in a core is not idle, all the
> + *    bits correspondings to it and to all its siblings are clear in the
> + *    smt_idle mask.
> + *
> + * Once we have such a mask, it is easy to implement a policy that, either:
> + *  - uses fully idle cores first: it is enough to try to schedule the vcpus
> + *    on pcpus from smt_idle mask first. This is what happens if
> + *    sched_smt_power_savings was not set at boot (default), and it maximizes
> + *    true parallelism, and hence performance;
> + *  - uses already busy cores first: it is enough to try to schedule the vcpus
> + *    on pcpus that are idle, but are not in smt_idle. This is what happens if
> + *    sched_smt_power_savings is set at boot, and it allows as more cores as
> + *    possible to stay in low power states, minimizing power consumption.
> + *
> + * This logic is entirely implemented in runq_tickle(), and that is enough.
> + * In fact, in this scheduler, placement of a vcpu on one of the pcpus of a
> + * runq, _always_ happens by means of tickling:
> + *  - when a vcpu wakes up, it calls csched2_vcpu_wake(), which calls
> + *    runq_tickle();
> + *  - when a migration is initiated in schedule.c, we call csched2_cpu_pick(),
> + *    csched2_vcpu_migrate() (which calls migrate()) and csched2_vcpu_wake().
> + *    csched2_cpu_pick() looks for the least loaded runq and return just any
> + *    of its processors. Then, csched2_vcpu_migrate() just moves the vcpu to
> + *    the chosen runq, and it is again runq_tickle(), called by
> + *    csched2_vcpu_wake() that actually decides what pcpu to use within the
> + *    chosen runq;
> + *  - when a migration is initiated in sched_credit2.c, by calling  migrate()
> + *    directly, that again temporarily use a random pcpu from the new runq,
> + *    and then calls runq_tickle(), by itself.
> + */
> +
> +/*
> + * If all the siblings of cpu (including cpu itself) are in idlers,
> + * set all their bits in mask.
> + *
> + * In order to properly take into account tickling, idlers needs to be
> + * set qeual to something like:
> + *
> + *  rqd->idle & (~rqd->tickled)
> + *
> + * This is because cpus that have been tickled will very likely pick up some
> + * work as soon as the manage to schedule, and hence we should really consider
> + * them as busy.
> + */
> +static inline
> +void smt_idle_mask_set(unsigned int cpu, cpumask_t *idlers, cpumask_t *mask)
> +{
> +    if ( cpumask_subset( per_cpu(cpu_sibling_mask, cpu), idlers) )
> +        cpumask_or(mask, mask, per_cpu(cpu_sibling_mask, cpu));
> +}
> +
> +/*
> + * Clear the bits of all the siblings of cpu from mask.
> + */
> +static inline
> +void smt_idle_mask_clear(unsigned int cpu, cpumask_t *mask)
> +{
> +    cpumask_andnot(mask, mask, per_cpu(cpu_sibling_mask, cpu));
> +}
> +
> +/*
>    * When a hard affinity change occurs, we may not be able to check some
>    * (any!) of the other runqueues, when looking for the best new processor
>    * for svc (as trylock-s in csched2_cpu_pick() can fail). If that happens, we
> @@ -851,9 +919,30 @@ runq_tickle(const struct scheduler *ops, struct csched2_vcpu *new, s_time_t now)
>       }
>
>       /*
> -     * Get a mask of idle, but not tickled, processors that new is
> -     * allowed to run on. If that's not empty, choose someone from there
> -     * (preferrably, the one were new was running on already).
> +     * First of all, consider idle cpus, checking if we can just
> +     * re-use the pcpu where we were running before.
> +     *
> +     * If there are cores where all the siblings are idle, consider
> +     * them first, honoring whatever the spreading-vs-consolidation
> +     * SMT policy wants us to do.
> +     */
> +    if ( unlikely(sched_smt_power_savings) )
> +        cpumask_andnot(&mask, &rqd->idle, &rqd->smt_idle);
> +    else
> +        cpumask_copy(&mask, &rqd->smt_idle);
> +    cpumask_and(&mask, &mask, new->vcpu->cpu_hard_affinity);
> +    i = cpumask_test_or_cycle(cpu, &mask);
> +    if ( i < nr_cpu_ids )
> +    {
> +        SCHED_STAT_CRANK(tickled_idle_cpu);
> +        ipid = i;
> +        goto tickle;
> +    }
> +
> +    /*
> +     * If there are no fully idle cores, check all idlers, after
> +     * having filtered out pcpus that have been tickled but haven't
> +     * gone through the scheduler yet.
>        */
>       cpumask_andnot(&mask, &rqd->idle, &rqd->tickled);
>       cpumask_and(&mask, &mask, new->vcpu->cpu_hard_affinity);
> @@ -945,6 +1034,7 @@ runq_tickle(const struct scheduler *ops, struct csched2_vcpu *new, s_time_t now)
>                       (unsigned char *)&d);
>       }
>       __cpumask_set_cpu(ipid, &rqd->tickled);
> +    //smt_idle_mask_clear(ipid, &rqd->smt_idle); XXX
>       cpu_raise_softirq(ipid, SCHEDULE_SOFTIRQ);
>   }
>
> @@ -1435,13 +1525,15 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
>
>       if ( !read_trylock(&prv->lock) )
>       {
> -        /* We may be here because someon requested us to migrate */
> +        /* We may be here because someone requested us to migrate */
>           __clear_bit(__CSFLAG_runq_migrate_request, &svc->flags);
>           return get_fallback_cpu(svc);
>       }
>
> -    /* First check to see if we're here because someone else suggested a place
> -     * for us to move. */
> +    /*
> +     * First check to see if we're here because someone else suggested a place
> +     * for us to move.
> +     */
>       if ( __test_and_clear_bit(__CSFLAG_runq_migrate_request, &svc->flags) )
>       {
>           if ( unlikely(svc->migrate_rqd->id < 0) )
> @@ -1462,7 +1554,7 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
>
>       min_avgload = MAX_LOAD;
>
> -    /* Find the runqueue with the lowest instantaneous load */
> +    /* Find the runqueue with the lowest average load. */
>       for_each_cpu(i, &prv->active_queues)
>       {
>           struct csched2_runqueue_data *rqd;
> @@ -1505,16 +1597,17 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
>
>       /* We didn't find anyone (most likely because of spinlock contention). */
>       if ( min_rqi == -1 )
> -        new_cpu = get_fallback_cpu(svc);
> -    else
>       {
> -        cpumask_and(cpumask_scratch, vc->cpu_hard_affinity,
> -                    &prv->rqd[min_rqi].active);
> -        new_cpu = cpumask_any(cpumask_scratch);
> -        BUG_ON(new_cpu >= nr_cpu_ids);
> +        new_cpu = get_fallback_cpu(svc);
> +        goto out_up;
>       }
>
> -out_up:
> +    cpumask_and(cpumask_scratch, vc->cpu_hard_affinity,
> +                &prv->rqd[min_rqi].active);
> +    new_cpu = cpumask_any(cpumask_scratch);
> +    BUG_ON(new_cpu >= nr_cpu_ids);
> +
> + out_up:
>       read_unlock(&prv->lock);
>
>       if ( unlikely(tb_init_done) )
> @@ -2166,7 +2259,11 @@ csched2_schedule(
>
>       /* Clear "tickled" bit now that we've been scheduled */
>       if ( cpumask_test_cpu(cpu, &rqd->tickled) )
> +    {
>           __cpumask_clear_cpu(cpu, &rqd->tickled);
> +        cpumask_andnot(cpumask_scratch, &rqd->idle, &rqd->tickled);
> +        smt_idle_mask_set(cpu, cpumask_scratch, &rqd->smt_idle); // XXX
> +    }
>
>       /* Update credits */
>       burn_credits(rqd, scurr, now);
> @@ -2228,7 +2325,10 @@ csched2_schedule(
>
>           /* Clear the idle mask if necessary */
>           if ( cpumask_test_cpu(cpu, &rqd->idle) )
> +        {
>               __cpumask_clear_cpu(cpu, &rqd->idle);
> +            smt_idle_mask_clear(cpu, &rqd->smt_idle);
> +        }
>
>           snext->start_time = now;
>
> @@ -2250,10 +2350,17 @@ csched2_schedule(
>           if ( tasklet_work_scheduled )
>           {
>               if ( cpumask_test_cpu(cpu, &rqd->idle) )
> +            {
>                   __cpumask_clear_cpu(cpu, &rqd->idle);
> +                smt_idle_mask_clear(cpu, &rqd->smt_idle);
> +            }
>           }
>           else if ( !cpumask_test_cpu(cpu, &rqd->idle) )
> +        {
>               __cpumask_set_cpu(cpu, &rqd->idle);
> +            cpumask_andnot(cpumask_scratch, &rqd->idle, &rqd->tickled);
> +            smt_idle_mask_set(cpu, cpumask_scratch, &rqd->smt_idle);
> +        }
>           /* Make sure avgload gets updated periodically even
>            * if there's no activity */
>           update_load(ops, rqd, NULL, 0, now);
> @@ -2383,6 +2490,8 @@ csched2_dump(const struct scheduler *ops)
>           printk("\tidlers: %s\n", cpustr);
>           cpumask_scnprintf(cpustr, sizeof(cpustr), &prv->rqd[i].tickled);
>           printk("\ttickled: %s\n", cpustr);
> +        cpumask_scnprintf(cpustr, sizeof(cpustr), &prv->rqd[i].smt_idle);
> +        printk("\tfully idle cores: %s\n", cpustr);
>       }
>
>       printk("Domain info:\n");
> @@ -2536,6 +2645,7 @@ init_pdata(struct csched2_private *prv, unsigned int cpu)
>       __cpumask_set_cpu(cpu, &rqd->idle);
>       __cpumask_set_cpu(cpu, &rqd->active);
>       __cpumask_set_cpu(cpu, &prv->initialized);
> +    __cpumask_set_cpu(cpu, &rqd->smt_idle);
>
>       return rqi;
>   }
> @@ -2641,6 +2751,7 @@ csched2_deinit_pdata(const struct scheduler *ops, void *pcpu, int cpu)
>       printk(XENLOG_INFO "Removing cpu %d from runqueue %d\n", cpu, rqi);
>
>       __cpumask_clear_cpu(cpu, &rqd->idle);
> +    __cpumask_clear_cpu(cpu, &rqd->smt_idle);
>       __cpumask_clear_cpu(cpu, &rqd->active);
>
>       if ( cpumask_empty(&rqd->active) )
>


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
http://lists.xen.org/xen-devel

  parent reply	other threads:[~2016-06-27 15:21 UTC|newest]

Thread overview: 64+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2016-06-17 23:11 [PATCH 00/19] xen: sched: assorted fixes and improvements to Credit2 Dario Faggioli
2016-06-17 23:11 ` [PATCH 01/19] xen: sched: leave CPUs doing tasklet work alone Dario Faggioli
2016-06-20  7:48   ` Jan Beulich
2016-07-07 10:11     ` Dario Faggioli
2016-06-21 16:17   ` anshul makkar
2016-07-06 15:41   ` George Dunlap
2016-07-07 10:25     ` Dario Faggioli
2016-06-17 23:11 ` [PATCH 02/19] xen: sched: make the 'tickled' perf counter clearer Dario Faggioli
2016-06-18  0:36   ` Meng Xu
2016-07-06 15:52   ` George Dunlap
2016-06-17 23:11 ` [PATCH 03/19] xen: credit2: insert and tickle don't need a cpu parameter Dario Faggioli
2016-06-21 16:41   ` anshul makkar
2016-07-06 15:59   ` George Dunlap
2016-06-17 23:11 ` [PATCH 04/19] xen: credit2: kill useless helper function choose_cpu Dario Faggioli
2016-07-06 16:02   ` George Dunlap
2016-07-07 10:26     ` Dario Faggioli
2016-06-17 23:11 ` [PATCH 05/19] xen: credit2: do not warn if calling burn_credits more than once Dario Faggioli
2016-07-06 16:05   ` George Dunlap
2016-06-17 23:12 ` [PATCH 06/19] xen: credit2: read NOW() with the proper runq lock held Dario Faggioli
2016-06-20  7:56   ` Jan Beulich
2016-07-06 16:10     ` George Dunlap
2016-07-07 10:28       ` Dario Faggioli
2016-06-17 23:12 ` [PATCH 07/19] xen: credit2: prevent load balancing to go mad if time goes backwards Dario Faggioli
2016-06-20  8:02   ` Jan Beulich
2016-07-06 16:21     ` George Dunlap
2016-07-07  7:29       ` Jan Beulich
2016-07-07  9:09         ` George Dunlap
2016-07-07  9:18           ` Jan Beulich
2016-07-07 10:53             ` Dario Faggioli
2016-06-17 23:12 ` [PATCH 08/19] xen: credit2: when tickling, check idle cpus first Dario Faggioli
2016-07-06 16:36   ` George Dunlap
2016-06-17 23:12 ` [PATCH 09/19] xen: credit2: avoid calling __update_svc_load() multiple times on the same vcpu Dario Faggioli
2016-07-06 16:40   ` George Dunlap
2016-06-17 23:12 ` [PATCH 10/19] xen: credit2: rework load tracking logic Dario Faggioli
2016-07-06 17:33   ` George Dunlap
2016-06-17 23:12 ` [PATCH 11/19] tools: tracing: adapt Credit2 load tracking events to new format Dario Faggioli
2016-06-21  9:27   ` Wei Liu
2016-06-17 23:12 ` [PATCH 12/19] xen: credit2: use non-atomic cpumask and bit operations Dario Faggioli
2016-07-07  9:45   ` George Dunlap
2016-06-17 23:12 ` [PATCH 13/19] xen: credit2: make the code less experimental Dario Faggioli
2016-06-20  8:13   ` Jan Beulich
2016-07-07 10:59     ` Dario Faggioli
2016-07-07 15:17   ` George Dunlap
2016-07-07 16:43     ` Dario Faggioli
2016-06-17 23:12 ` [PATCH 14/19] xen: credit2: add yet some more tracing Dario Faggioli
2016-06-20  8:15   ` Jan Beulich
2016-07-07 15:34     ` George Dunlap
2016-07-07 15:34   ` George Dunlap
2016-06-17 23:13 ` [PATCH 15/19] xen: credit2: only marshall trace point arguments if tracing enabled Dario Faggioli
2016-07-07 15:37   ` George Dunlap
2016-06-17 23:13 ` [PATCH 16/19] tools: tracing: deal with new Credit2 events Dario Faggioli
2016-07-07 15:39   ` George Dunlap
2016-06-17 23:13 ` [PATCH 17/19] xen: credit2: the private scheduler lock can be an rwlock Dario Faggioli
2016-07-07 16:00   ` George Dunlap
2016-06-17 23:13 ` [PATCH 18/19] xen: credit2: implement SMT support independent runq arrangement Dario Faggioli
2016-06-20  8:26   ` Jan Beulich
2016-06-20 10:38     ` Dario Faggioli
2016-06-27 15:20   ` anshul makkar [this message]
2016-07-12 13:40   ` George Dunlap
2016-06-17 23:13 ` [PATCH 19/19] xen: credit2: use cpumask_first instead of cpumask_any when choosing cpu Dario Faggioli
2016-06-20  8:30   ` Jan Beulich
2016-06-20 11:28     ` Dario Faggioli
2016-06-21 10:42   ` David Vrabel
2016-07-07 16:55     ` Dario Faggioli

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=57714421.3040501@citrix.com \
    --to=anshul.makkar@citrix.com \
    --cc=dario.faggioli@citrix.com \
    --cc=david.vrabel@citrix.com \
    --cc=george.dunlap@citrix.com \
    --cc=xen-devel@lists.xenproject.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).