From: "Srivatsa S. Bhat" <srivatsa.bhat@linux.vnet.ibm.com> To: tglx@linutronix.de, tj@kernel.org, oleg@redhat.com, paulmck@linux.vnet.ibm.com, rusty@rustcorp.com.au, mingo@kernel.org, akpm@linux-foundation.org, namhyung@kernel.org, walken@google.com, vincent.guittot@linaro.org, laijs@cn.fujitsu.com, David.Laight@aculab.com Cc: rostedt@goodmis.org, wangyun@linux.vnet.ibm.com, xiaoguangrong@linux.vnet.ibm.com, sbw@mit.edu, fweisbec@gmail.com, zhong@linux.vnet.ibm.com, nikunj@linux.vnet.ibm.com, srivatsa.bhat@linux.vnet.ibm.com, linux-pm@vger.kernel.org, linux-arch@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, netdev@vger.kernel.org, linux-kernel@vger.kernel.org, Ingo Molnar <mingo@redhat.com>, Peter Zijlstra <peterz@infradead.org>"Srivatsa S. Bhat" <srivatsa.bhat@linux.vnet.ibm.com> Subject: [PATCH v3 13/45] sched/fair: Use get/put_online_cpus_atomic() to prevent CPU offline Date: Fri, 28 Jun 2013 01:24:39 +0530 [thread overview] Message-ID: <20130627195439.29830.96122.stgit@srivatsabhat.in.ibm.com> (raw) In-Reply-To: <20130627195136.29830.10445.stgit@srivatsabhat.in.ibm.com> Once stop_machine() is gone from the CPU offline path, we won't be able to depend on disabling preemption to prevent CPUs from going offline from under us. Use the get/put_online_cpus_atomic() APIs to prevent CPUs from going offline, while invoking from atomic context. Cc: Ingo Molnar <mingo@redhat.com> Cc: Peter Zijlstra <peterz@infradead.org> Signed-off-by: Srivatsa S. Bhat <srivatsa.bhat@linux.vnet.ibm.com> --- kernel/sched/fair.c | 14 +++++++++++++- 1 file changed, 13 insertions(+), 1 deletion(-) diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index f77f9c5..62d98dd 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -3369,7 +3369,8 @@ done: * * Returns the target CPU number, or the same CPU if no balancing is needed. * - * preempt must be disabled. + * Must be called within get/put_online_cpus_atomic(), to prevent CPUs + * from going offline from under us. */ static int select_task_rq_fair(struct task_struct *p, int sd_flag, int wake_flags) @@ -5289,6 +5290,8 @@ void idle_balance(int this_cpu, struct rq *this_rq) raw_spin_unlock(&this_rq->lock); update_blocked_averages(this_cpu); + + get_online_cpus_atomic(); rcu_read_lock(); for_each_domain(this_cpu, sd) { unsigned long interval; @@ -5312,6 +5315,7 @@ void idle_balance(int this_cpu, struct rq *this_rq) } } rcu_read_unlock(); + put_online_cpus_atomic(); raw_spin_lock(&this_rq->lock); @@ -5338,6 +5342,7 @@ static int active_load_balance_cpu_stop(void *data) struct rq *target_rq = cpu_rq(target_cpu); struct sched_domain *sd; + get_online_cpus_atomic(); raw_spin_lock_irq(&busiest_rq->lock); /* make sure the requested cpu hasn't gone down in the meantime */ @@ -5389,6 +5394,7 @@ static int active_load_balance_cpu_stop(void *data) out_unlock: busiest_rq->active_balance = 0; raw_spin_unlock_irq(&busiest_rq->lock); + put_online_cpus_atomic(); return 0; } @@ -5549,6 +5555,7 @@ static void rebalance_domains(int cpu, enum cpu_idle_type idle) update_blocked_averages(cpu); + get_online_cpus_atomic(); rcu_read_lock(); for_each_domain(cpu, sd) { if (!(sd->flags & SD_LOAD_BALANCE)) @@ -5597,6 +5604,7 @@ out: break; } rcu_read_unlock(); + put_online_cpus_atomic(); /* * next_balance will be updated only when there is a need. @@ -5728,6 +5736,7 @@ static void run_rebalance_domains(struct softirq_action *h) enum cpu_idle_type idle = this_rq->idle_balance ? CPU_IDLE : CPU_NOT_IDLE; + get_online_cpus_atomic(); rebalance_domains(this_cpu, idle); /* @@ -5736,6 +5745,7 @@ static void run_rebalance_domains(struct softirq_action *h) * stopped. */ nohz_idle_balance(this_cpu, idle); + put_online_cpus_atomic(); } static inline int on_null_domain(int cpu) @@ -5753,8 +5763,10 @@ void trigger_load_balance(struct rq *rq, int cpu) likely(!on_null_domain(cpu))) raise_softirq(SCHED_SOFTIRQ); #ifdef CONFIG_NO_HZ_COMMON + get_online_cpus_atomic(); if (nohz_kick_needed(rq, cpu) && likely(!on_null_domain(cpu))) nohz_balancer_kick(cpu); + put_online_cpus_atomic(); #endif }
WARNING: multiple messages have this Message-ID (diff)
From: "Srivatsa S. Bhat" <srivatsa.bhat@linux.vnet.ibm.com> To: tglx@linutronix.de, peterz@infradead.org, tj@kernel.org, oleg@redhat.com, paulmck@linux.vnet.ibm.com, rusty@rustcorp.com.au, mingo@kernel.org, akpm@linux-foundation.org, namhyung@kernel.org, walken@google.com, vincent.guittot@linaro.org, laijs@cn.fujitsu.com, David.Laight@aculab.com Cc: rostedt@goodmis.org, wangyun@linux.vnet.ibm.com, xiaoguangrong@linux.vnet.ibm.com, sbw@mit.edu, fweisbec@gmail.com, zhong@linux.vnet.ibm.com, nikunj@linux.vnet.ibm.com, srivatsa.bhat@linux.vnet.ibm.com, linux-pm@vger.kernel.org, linux-arch@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, netdev@vger.kernel.org, linux-kernel@vger.kernel.org, Ingo Molnar <mingo@redhat.com>"Srivatsa S. Bhat" <srivatsa.bhat@linux.vnet.ibm.com> Subject: [PATCH v3 13/45] sched/fair: Use get/put_online_cpus_atomic() to prevent CPU offline Date: Fri, 28 Jun 2013 01:24:39 +0530 [thread overview] Message-ID: <20130627195439.29830.96122.stgit@srivatsabhat.in.ibm.com> (raw) Message-ID: <20130627195439.dlD9h07mmqOQl4Nkqi963LId-tOrSQuItqg7pF0UnTU@z> (raw) In-Reply-To: <20130627195136.29830.10445.stgit@srivatsabhat.in.ibm.com> Once stop_machine() is gone from the CPU offline path, we won't be able to depend on disabling preemption to prevent CPUs from going offline from under us. Use the get/put_online_cpus_atomic() APIs to prevent CPUs from going offline, while invoking from atomic context. Cc: Ingo Molnar <mingo@redhat.com> Cc: Peter Zijlstra <peterz@infradead.org> Signed-off-by: Srivatsa S. Bhat <srivatsa.bhat@linux.vnet.ibm.com> --- kernel/sched/fair.c | 14 +++++++++++++- 1 file changed, 13 insertions(+), 1 deletion(-) diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index f77f9c5..62d98dd 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -3369,7 +3369,8 @@ done: * * Returns the target CPU number, or the same CPU if no balancing is needed. * - * preempt must be disabled. + * Must be called within get/put_online_cpus_atomic(), to prevent CPUs + * from going offline from under us. */ static int select_task_rq_fair(struct task_struct *p, int sd_flag, int wake_flags) @@ -5289,6 +5290,8 @@ void idle_balance(int this_cpu, struct rq *this_rq) raw_spin_unlock(&this_rq->lock); update_blocked_averages(this_cpu); + + get_online_cpus_atomic(); rcu_read_lock(); for_each_domain(this_cpu, sd) { unsigned long interval; @@ -5312,6 +5315,7 @@ void idle_balance(int this_cpu, struct rq *this_rq) } } rcu_read_unlock(); + put_online_cpus_atomic(); raw_spin_lock(&this_rq->lock); @@ -5338,6 +5342,7 @@ static int active_load_balance_cpu_stop(void *data) struct rq *target_rq = cpu_rq(target_cpu); struct sched_domain *sd; + get_online_cpus_atomic(); raw_spin_lock_irq(&busiest_rq->lock); /* make sure the requested cpu hasn't gone down in the meantime */ @@ -5389,6 +5394,7 @@ static int active_load_balance_cpu_stop(void *data) out_unlock: busiest_rq->active_balance = 0; raw_spin_unlock_irq(&busiest_rq->lock); + put_online_cpus_atomic(); return 0; } @@ -5549,6 +5555,7 @@ static void rebalance_domains(int cpu, enum cpu_idle_type idle) update_blocked_averages(cpu); + get_online_cpus_atomic(); rcu_read_lock(); for_each_domain(cpu, sd) { if (!(sd->flags & SD_LOAD_BALANCE)) @@ -5597,6 +5604,7 @@ out: break; } rcu_read_unlock(); + put_online_cpus_atomic(); /* * next_balance will be updated only when there is a need. @@ -5728,6 +5736,7 @@ static void run_rebalance_domains(struct softirq_action *h) enum cpu_idle_type idle = this_rq->idle_balance ? CPU_IDLE : CPU_NOT_IDLE; + get_online_cpus_atomic(); rebalance_domains(this_cpu, idle); /* @@ -5736,6 +5745,7 @@ static void run_rebalance_domains(struct softirq_action *h) * stopped. */ nohz_idle_balance(this_cpu, idle); + put_online_cpus_atomic(); } static inline int on_null_domain(int cpu) @@ -5753,8 +5763,10 @@ void trigger_load_balance(struct rq *rq, int cpu) likely(!on_null_domain(cpu))) raise_softirq(SCHED_SOFTIRQ); #ifdef CONFIG_NO_HZ_COMMON + get_online_cpus_atomic(); if (nohz_kick_needed(rq, cpu) && likely(!on_null_domain(cpu))) nohz_balancer_kick(cpu); + put_online_cpus_atomic(); #endif }
next prev parent reply other threads:[~2013-06-27 19:54 UTC|newest] Thread overview: 101+ messages / expand[flat|nested] mbox.gz Atom feed top 2013-06-27 19:52 [PATCH v3 00/45] CPU hotplug: stop_machine()-free CPU hotplug, part 1 Srivatsa S. Bhat 2013-06-27 19:52 ` [PATCH v3 01/45] CPU hotplug: Provide APIs to prevent CPU offline from atomic context Srivatsa S. Bhat 2013-06-27 19:52 ` Srivatsa S. Bhat 2013-06-27 19:52 ` [PATCH v3 02/45] CPU hotplug: Clarify the usage of different synchronization APIs Srivatsa S. Bhat 2013-06-27 19:52 ` Srivatsa S. Bhat 2013-06-27 19:52 ` [PATCH v3 03/45] Documentation, CPU hotplug: Recommend usage of get/put_online_cpus_atomic() Srivatsa S. Bhat 2013-06-27 19:52 ` Srivatsa S. Bhat 2013-06-27 19:53 ` [PATCH v3 04/45] CPU hotplug: Add infrastructure to check lacking hotplug synchronization Srivatsa S. Bhat 2013-06-27 19:53 ` Srivatsa S. Bhat 2013-06-27 19:53 ` [PATCH v3 05/45] CPU hotplug: Protect set_cpu_online() to avoid false-positives Srivatsa S. Bhat 2013-06-27 19:53 ` Srivatsa S. Bhat 2013-06-27 19:53 ` [PATCH v3 06/45] CPU hotplug: Sprinkle debugging checks to catch locking bugs Srivatsa S. Bhat 2013-06-27 19:53 ` Srivatsa S. Bhat 2013-06-27 19:53 ` [PATCH v3 07/45] CPU hotplug: Add _nocheck() variants of accessor functions Srivatsa S. Bhat 2013-06-27 19:53 ` Srivatsa S. Bhat 2013-06-27 19:53 ` [PATCH v3 08/45] CPU hotplug: Expose the new debug config option Srivatsa S. Bhat 2013-06-27 19:53 ` Srivatsa S. Bhat 2013-06-27 19:54 ` [PATCH v3 09/45] CPU hotplug: Convert preprocessor macros to static inline functions Srivatsa S. Bhat 2013-06-27 19:54 ` Srivatsa S. Bhat 2013-06-27 19:54 ` [PATCH v3 10/45] smp: Use get/put_online_cpus_atomic() to prevent CPU offline Srivatsa S. Bhat 2013-06-27 19:54 ` Srivatsa S. Bhat 2013-07-02 5:32 ` Michael Wang 2013-07-02 5:32 ` Michael Wang 2013-07-02 8:25 ` Srivatsa S. Bhat 2013-07-02 8:25 ` Srivatsa S. Bhat 2013-07-02 8:47 ` Michael Wang 2013-07-02 9:51 ` Srivatsa S. Bhat 2013-07-02 9:51 ` Srivatsa S. Bhat 2013-07-02 10:08 ` Michael Wang 2013-07-02 10:08 ` Michael Wang 2013-06-27 19:54 ` [PATCH v3 11/45] sched/core: " Srivatsa S. Bhat 2013-06-27 19:54 ` Srivatsa S. Bhat 2013-06-27 19:54 ` [PATCH v3 12/45] migration: Use raw_spin_lock/unlock since interrupts are already disabled Srivatsa S. Bhat 2013-06-27 19:54 ` Srivatsa S. Bhat 2013-06-27 19:54 ` Srivatsa S. Bhat [this message] 2013-06-27 19:54 ` [PATCH v3 13/45] sched/fair: Use get/put_online_cpus_atomic() to prevent CPU offline Srivatsa S. Bhat 2013-06-27 19:54 ` [PATCH v3 14/45] timer: " Srivatsa S. Bhat 2013-06-27 19:54 ` Srivatsa S. Bhat 2013-06-27 19:54 ` [PATCH v3 15/45] sched/rt: " Srivatsa S. Bhat 2013-06-27 19:54 ` Srivatsa S. Bhat 2013-06-27 19:55 ` [PATCH v3 16/45] rcu: Use cpu_is_offline_nocheck() to avoid false-positive warnings Srivatsa S. Bhat 2013-06-27 19:55 ` Srivatsa S. Bhat 2013-06-27 20:12 ` Paul E. McKenney 2013-06-27 20:12 ` Paul E. McKenney 2013-06-27 19:55 ` [PATCH v3 17/45] tick-broadcast: Use get/put_online_cpus_atomic() to prevent CPU offline Srivatsa S. Bhat 2013-06-27 19:55 ` Srivatsa S. Bhat 2013-06-27 19:55 ` [PATCH v3 18/45] time/clocksource: " Srivatsa S. Bhat 2013-06-27 19:55 ` Srivatsa S. Bhat 2013-06-27 19:56 ` [PATCH v3 19/45] softirq: " Srivatsa S. Bhat 2013-06-27 19:56 ` Srivatsa S. Bhat 2013-06-27 19:56 ` [PATCH v3 20/45] irq: " Srivatsa S. Bhat 2013-06-27 19:56 ` Srivatsa S. Bhat 2013-06-27 19:56 ` [PATCH v3 21/45] net: " Srivatsa S. Bhat 2013-06-27 19:56 ` Srivatsa S. Bhat 2013-06-27 19:56 ` [PATCH v3 22/45] block: " Srivatsa S. Bhat 2013-06-27 19:56 ` Srivatsa S. Bhat 2013-06-27 19:56 ` [PATCH v3 23/45] percpu_counter: Use _nocheck version of for_each_online_cpu() Srivatsa S. Bhat 2013-06-27 19:56 ` Srivatsa S. Bhat 2013-06-27 19:56 ` [PATCH v3 24/45] infiniband: ehca: Use get/put_online_cpus_atomic() to prevent CPU offline Srivatsa S. Bhat 2013-06-27 19:56 ` Srivatsa S. Bhat 2013-06-27 19:56 ` [PATCH v3 25/45] [SCSI] fcoe: " Srivatsa S. Bhat 2013-06-27 19:56 ` Srivatsa S. Bhat 2013-06-27 19:57 ` [PATCH v3 26/45] staging/octeon: " Srivatsa S. Bhat 2013-06-27 19:57 ` Srivatsa S. Bhat 2013-06-27 19:57 ` [PATCH v3 27/45] x86: " Srivatsa S. Bhat 2013-06-27 19:57 ` Srivatsa S. Bhat 2013-06-27 19:57 ` [PATCH v3 28/45] perf/x86: " Srivatsa S. Bhat 2013-06-27 19:57 ` Srivatsa S. Bhat 2013-06-27 19:57 ` [PATCH v3 29/45] KVM: " Srivatsa S. Bhat 2013-06-27 19:57 ` Srivatsa S. Bhat 2013-06-27 19:57 ` [PATCH v3 30/45] x86/xen: " Srivatsa S. Bhat 2013-06-27 19:57 ` Srivatsa S. Bhat 2013-06-27 19:57 ` [PATCH v3 31/45] alpha/smp: " Srivatsa S. Bhat 2013-06-27 19:57 ` Srivatsa S. Bhat 2013-06-27 19:58 ` [PATCH v3 32/45] blackfin/smp: " Srivatsa S. Bhat 2013-06-27 19:58 ` Srivatsa S. Bhat 2013-06-27 19:58 ` [PATCH v3 33/45] cris/smp: " Srivatsa S. Bhat 2013-06-27 19:58 ` Srivatsa S. Bhat 2013-06-27 19:58 ` [PATCH v3 34/45] hexagon/smp: " Srivatsa S. Bhat 2013-06-27 19:58 ` Srivatsa S. Bhat 2013-06-27 19:58 ` [PATCH v3 35/45] ia64: irq, perfmon: " Srivatsa S. Bhat 2013-06-27 19:58 ` Srivatsa S. Bhat 2013-06-27 19:59 ` [PATCH v3 36/45] ia64: smp, tlb: " Srivatsa S. Bhat 2013-06-27 19:59 ` Srivatsa S. Bhat 2013-06-27 19:59 ` [PATCH v3 37/45] m32r: " Srivatsa S. Bhat 2013-06-27 19:59 ` [PATCH v3 38/45] MIPS: " Srivatsa S. Bhat 2013-06-27 19:59 ` Srivatsa S. Bhat 2013-06-27 19:59 ` [PATCH v3 39/45] mn10300: " Srivatsa S. Bhat 2013-06-27 19:59 ` Srivatsa S. Bhat 2013-06-27 19:59 ` [PATCH v3 40/45] powerpc, irq: Use GFP_ATOMIC allocations in atomic context Srivatsa S. Bhat 2013-06-27 19:59 ` Srivatsa S. Bhat 2013-06-27 20:00 ` [PATCH v3 41/45] powerpc: Use get/put_online_cpus_atomic() to prevent CPU offline Srivatsa S. Bhat 2013-06-27 20:00 ` Srivatsa S. Bhat 2013-06-27 20:00 ` [PATCH v3 42/45] powerpc: Use get/put_online_cpus_atomic() to avoid false-positive warning Srivatsa S. Bhat 2013-06-27 20:00 ` Srivatsa S. Bhat 2013-06-27 20:00 ` [PATCH v3 43/45] sh: Use get/put_online_cpus_atomic() to prevent CPU offline Srivatsa S. Bhat 2013-06-27 20:00 ` Srivatsa S. Bhat 2013-06-27 20:00 ` [PATCH v3 44/45] sparc: " Srivatsa S. Bhat 2013-06-27 20:00 ` Srivatsa S. Bhat 2013-06-27 20:00 ` [PATCH v3 45/45] tile: " Srivatsa S. Bhat 2013-06-27 20:00 ` Srivatsa S. Bhat
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20130627195439.29830.96122.stgit@srivatsabhat.in.ibm.com \ --to=srivatsa.bhat@linux.vnet.ibm.com \ --cc=David.Laight@aculab.com \ --cc=akpm@linux-foundation.org \ --cc=fweisbec@gmail.com \ --cc=laijs@cn.fujitsu.com \ --cc=linux-arch@vger.kernel.org \ --cc=linux-kernel@vger.kernel.org \ --cc=linux-pm@vger.kernel.org \ --cc=linuxppc-dev@lists.ozlabs.org \ --cc=mingo@kernel.org \ --cc=mingo@redhat.com \ --cc=namhyung@kernel.org \ --cc=netdev@vger.kernel.org \ --cc=nikunj@linux.vnet.ibm.com \ --cc=oleg@redhat.com \ --cc=paulmck@linux.vnet.ibm.com \ --cc=peterz@infradead.org \ --cc=rostedt@goodmis.org \ --cc=rusty@rustcorp.com.au \ --cc=sbw@mit.edu \ --cc=tglx@linutronix.de \ --cc=tj@kernel.org \ --cc=vincent.guittot@linaro.org \ --cc=walken@google.com \ --cc=wangyun@linux.vnet.ibm.com \ --cc=xiaoguangrong@linux.vnet.ibm.com \ --cc=zhong@linux.vnet.ibm.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).