From: Yuyang Du <yuyang.du@intel.com>
To: Morten Rasmussen <morten.rasmussen@arm.com>
Cc: Andrey Ryabinin <aryabinin@virtuozzo.com>,
Peter Zijlstra <peterz@infradead.org>,
mingo@redhat.com, linux-kernel@vger.kernel.org,
Paul Turner <pjt@google.com>, Ben Segall <bsegall@google.com>
Subject: Re: [PATCH] sched/fair: fix mul overflow on 32-bit systems
Date: Mon, 14 Dec 2015 06:42:24 +0800 [thread overview]
Message-ID: <20151213224224.GC28098@intel.com> (raw)
In-Reply-To: <20151211175751.GA27552@e105550-lin.cambridge.arm.com>
On Fri, Dec 11, 2015 at 05:57:51PM +0000, Morten Rasmussen wrote:
> > >>> if (atomic_long_read(&cfs_rq->removed_load_avg)) {
> > >>> - long r = atomic_long_xchg(&cfs_rq->removed_load_avg, 0);
> > >>> + s64 r = atomic_long_xchg(&cfs_rq->removed_load_avg, 0);
> > >>> sa->load_avg = max_t(long, sa->load_avg - r, 0);
> > >>> sa->load_sum = max_t(s64, sa->load_sum - r * LOAD_AVG_MAX, 0);
> > >>
> > >> This makes sense, because sched_avg::load_sum is u64.
>
> A single removed nice=-20 task should be sufficient to cause the
> overflow.
Oh yes, it was a wreck, sorry.
> > >>> if (atomic_long_read(&cfs_rq->removed_util_avg)) {
> > >>> - long r = atomic_long_xchg(&cfs_rq->removed_util_avg, 0);
> > >>> + s64 r = atomic_long_xchg(&cfs_rq->removed_util_avg, 0);
> > >>> sa->util_avg = max_t(long, sa->util_avg - r, 0);
> > >>> sa->util_sum = max_t(s32, sa->util_sum - r * LOAD_AVG_MAX, 0);
> > >>> }
> > >>
> > >> However sched_avg::util_sum is u32, so this is still wrecked.
> > >
> > > I seems to have wrecked that in:
> > >
> > > 006cdf025a33 ("sched/fair: Optimize per entity utilization tracking")
> > >
> > > maybe just make util_load u64 too?
>
> It isn't as bad, but the optimization does increase the normal range
> (not guaranteed) for util_sum from 47742 to
> scale_down(SCHED_LOAD_SCALE)*47742 (= 1024*47742, unless you mess with
> the scaling).
>
> > Is there any guarantee that the final result of expression 'util_sum - r * LOAD_AVG_MAX' always can be represented by s32?
> >
> > If yes, than we could just do this:
> > max_t(s32, (u64)sa->util_sum - r * LOAD_AVG_MAX, 0)
>
> In most cases 'r' shouldn't exceed 1024 and util_sum not significantly
> exceed 1024*47742, but in extreme cases like spawning lots of new tasks
> it may potentially overflow 32 bit. Newly created tasks contribute
> 1024*47742 each to the rq util_sum, which means that more than ~87 new
> tasks on a single rq will get us in trouble I think.
>
> Without Peter's optimization referenced above, that number should
> increase to ~87k tasks as each task only contributed 47742 before, but
> 'r' could still cause 32-bit overflow if we remove more than ~87 newly
> created tasks in one go. But I'm not sure if that is a situation we
> should worry about?
>
> I think we have to either make util_sum u64 too or look at the
> optimizations again.
Both can workaround the issue with additional overhead. But I suspectthey
will end up going in the wrong direction for util_avg. The question is a
big util_sum (much bigger than 1024) may not be in a right range for it
to be used in load balancing.
The problem is that it is not so good to initiate a new task's util_avg
to 1024. At least, it makes much less sense than a new task's load_avg
being initiated to its full weight. Because the top util_avg should be
well bounded by 1024 - the CPU's full utilization.
So, maybe give the initial util_sum to an average of its cfs_rq, like:
cfs_rq->avg.util_sum / cfs_rq->load.weight * task->load.weight
And make sure that initial value's is bounded on various conditions.
Thanks,
Yuyang
next prev parent reply other threads:[~2015-12-14 6:29 UTC|newest]
Thread overview: 20+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-12-11 12:55 [PATCH] sched/fair: fix mul overflow on 32-bit systems Andrey Ryabinin
2015-12-11 13:25 ` Peter Zijlstra
2015-12-11 13:36 ` Peter Zijlstra
2015-12-11 14:00 ` Andrey Ryabinin
2015-12-11 17:57 ` Morten Rasmussen
2015-12-11 18:32 ` Dietmar Eggemann
2015-12-11 19:18 ` bsegall
2015-12-13 21:02 ` Yuyang Du
2015-12-14 12:32 ` Morten Rasmussen
2015-12-14 17:51 ` bsegall
2015-12-13 22:42 ` Yuyang Du [this message]
2015-12-14 11:54 ` Peter Zijlstra
2015-12-14 13:07 ` Morten Rasmussen
2015-12-14 14:20 ` Peter Zijlstra
2015-12-14 14:46 ` Morten Rasmussen
2015-12-15 2:22 ` Yuyang Du
2015-12-15 21:56 ` Steve Muckle
2015-12-18 2:33 ` Yuyang Du
2016-01-03 23:14 ` Yuyang Du
2015-12-11 17:58 ` bsegall
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20151213224224.GC28098@intel.com \
--to=yuyang.du@intel.com \
--cc=aryabinin@virtuozzo.com \
--cc=bsegall@google.com \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@redhat.com \
--cc=morten.rasmussen@arm.com \
--cc=peterz@infradead.org \
--cc=pjt@google.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox