From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-13.7 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH,MAILING_LIST_MULTI, MENTIONS_GIT_HOSTING,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 0BB7BC433DB for ; Tue, 23 Feb 2021 12:08:32 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id D1B1A64E21 for ; Tue, 23 Feb 2021 12:08:31 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232835AbhBWMHw (ORCPT ); Tue, 23 Feb 2021 07:07:52 -0500 Received: from foss.arm.com ([217.140.110.172]:45940 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232503AbhBWME7 (ORCPT ); Tue, 23 Feb 2021 07:04:59 -0500 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id E990B31B; Tue, 23 Feb 2021 04:03:56 -0800 (PST) Received: from e113632-lin (e113632-lin.cambridge.arm.com [10.1.194.46]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 5E06C3F70D; Tue, 23 Feb 2021 04:03:55 -0800 (PST) From: Valentin Schneider To: syzbot , akpm@linux-foundation.org, bp@alien8.de, hpa@zytor.com, linux-kernel@vger.kernel.org, luto@kernel.org, mingo@redhat.com, peterz@infradead.org, syzkaller-bugs@googlegroups.com, tglx@linutronix.de, x86@kernel.org, Vincent Guittot Subject: Re: UBSAN: shift-out-of-bounds in load_balance In-Reply-To: <000000000000a43f1f05bbefe703@google.com> References: <000000000000a43f1f05bbefe703@google.com> User-Agent: Notmuch/0.21 (http://notmuchmail.org) Emacs/26.3 (x86_64-pc-linux-gnu) Date: Tue, 23 Feb 2021 12:03:49 +0000 Message-ID: MIME-Version: 1.0 Content-Type: text/plain Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org +Vincent On 22/02/21 09:12, syzbot wrote: > syzbot has found a reproducer for the following issue on: > > HEAD commit: 31caf8b2 Merge branch 'linus' of git://git.kernel.org/pub/.. > git tree: upstream > console output: https://syzkaller.appspot.com/x/log.txt?x=16ab2682d00000 > kernel config: https://syzkaller.appspot.com/x/.config?x=b81388f0b32761d4 > dashboard link: https://syzkaller.appspot.com/bug?extid=d7581744d5fd27c9fbe1 > syz repro: https://syzkaller.appspot.com/x/repro.syz?x=1277457f500000 > > IMPORTANT: if you fix the issue, please add the following tag to the commit: > Reported-by: syzbot+d7581744d5fd27c9fbe1@syzkaller.appspotmail.com > > ================================================================================ > UBSAN: shift-out-of-bounds in kernel/sched/fair.c:7712:14 > shift exponent 149 is too large for 64-bit type 'long unsigned int' That 149 is surprising. sd->cache_nice_tries is \in {1, 2}, and sd->nr_balanced_failed should be in the same ballpark. A successful load_balance() resets it to 0; a failed one increments it. Once it gets to sd->cache_nice_tries + 3, this should trigger an active balance, which will either set it to sd->cache_nice_tries+1 or reset it to 0. There is this one condition that could let it creep up uncontrollably: /* * Don't kick the active_load_balance_cpu_stop, * if the curr task on busiest CPU can't be * moved to this_cpu: */ if (!cpumask_test_cpu(this_cpu, busiest->curr->cpus_ptr)) { raw_spin_unlock_irqrestore(&busiest->lock, flags); goto out_one_pinned; } So despite the resulting sd->balance_interval increase, repeatedly hitting this might yield the above. Would we then want something like this? --- diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 8a8bd7b13634..b65c24b5ae91 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -7422,6 +7422,11 @@ struct lb_env { struct list_head tasks; }; +static inline unsigned int sd_balance_failed_cap(struct sched_domain *sd) +{ + return sd->cache_nice_tries + 3; +} + /* * Is this task likely cache-hot: */ @@ -9493,7 +9498,7 @@ imbalanced_active_balance(struct lb_env *env) * threads on a system with spare capacity */ if ((env->migration_type == migrate_task) && - (sd->nr_balance_failed > sd->cache_nice_tries+2)) + (sd->nr_balance_failed >= sd_balance_failed_cap(sd))) return 1; return 0; @@ -9737,8 +9742,10 @@ static int load_balance(int this_cpu, struct rq *this_rq, * frequent, pollute the failure counter causing * excessive cache_hot migrations and active balances. */ - if (idle != CPU_NEWLY_IDLE) - sd->nr_balance_failed++; + if (idle != CPU_NEWLY_IDLE) { + sd->nr_balance_failed = min(sd->nr_balance_failed + 1, + sd_balance_failed_cap(sd)); + } if (need_active_balance(&env)) { unsigned long flags;