public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
* [PATCH v3] sched: fix tsk->pi_lock isn't held when do_set_cpus_allowed()
@ 2015-08-28  6:55 Wanpeng Li
  2015-08-28 13:29 ` Peter Zijlstra
  2015-09-13 10:56 ` [tip:sched/core] sched: 'Annotate' migrate_tasks() tip-bot for Wanpeng Li
  0 siblings, 2 replies; 7+ messages in thread
From: Wanpeng Li @ 2015-08-28  6:55 UTC (permalink / raw)
  To: Ingo Molnar, Peter Zijlstra
  Cc: Sasha Levin, kernel test robot, Boqun Feng, linux-kernel,
	Wanpeng Li

------------[ cut here ]------------
WARNING: CPU: 0 PID: 13 at kernel/sched/core.c:1156 do_set_cpus_allowed+0x7e/0x80()
Modules linked in:
CPU: 0 PID: 13 Comm: migration/0 Not tainted 4.2.0-rc1-00049-g25834c7 #2
Call Trace:
	dump_stack+0x4b/0x75
	warn_slowpath_common+0x8b/0xc0
	? do_set_cpus_allowed+0x7e/0x80
	? do_set_cpus_allowed+0x7e/0x80
	warn_slowpath_null+0x22/0x30
	do_set_cpus_allowed+0x7e/0x80
	cpuset_cpus_allowed_fallback+0x7c/0x170
	? cpuset_cpus_allowed+0x180/0x180
	select_fallback_rq+0x221/0x280
	migration_call+0xe3/0x250
	notifier_call_chain+0x53/0x70
	__raw_notifier_call_chain+0x1e/0x30
	cpu_notify+0x28/0x50
	take_cpu_down+0x22/0x40
	multi_cpu_stop+0xd5/0x140
	? __stop_cpus+0x80/0x80
	cpu_stopper_thread+0xbc/0x170
	? preempt_count_sub+0x9/0x50
	? _raw_spin_unlock_irq+0x37/0x50
	? _raw_spin_unlock_irqrestore+0x55/0x70
	? trace_hardirqs_on_caller+0x144/0x1e0
	? cpu_stop_should_run+0x35/0x40
	? preempt_count_sub+0x9/0x50
	? _raw_spin_unlock_irqrestore+0x41/0x70
	smpboot_thread_fn+0x174/0x2f0
	? sort_range+0x30/0x30
	kthread+0xc4/0xe0
	ret_from_kernel_thread+0x21/0x30
	? kthread_create_on_node+0x180/0x180
---[ end trace 15f4c86d404693b0 ]---

As Peterz pointed out:

| So the normal rules for changing task_struct::cpus_allowed are holding
| both pi_lock and rq->lock, such that holding either stabilizes the mask.
| 
| This is so that wakeup can happen without rq->lock and load-balance
| without pi_lock.
|
| From this we already get the relaxation that we can omit acquiring
| rq->lock if the task is not on the rq, because in that case
| load-balancing will not apply to it.
|
| ** these are the rules currently tested in do_set_cpus_allowed() **
|
| Now, since __set_cpus_allowed_ptr() uses task_rq_lock() which
| unconditionally acquires both locks, we could get away with holding just
| rq->lock when on_rq for modification because that'd still exclude
| __set_cpus_allowed_ptr(), it would also work against
| __kthread_bind_mask() because that assumes !on_rq.
|
| That said, this is all somewhat fragile.
|
| Now, I don't think dropping rq->lock is quite as disastrous as it
| usually is because !cpu_active at this point, which means load-balance
| will not interfere, but that too is somewhat fragile.
| 
| So we end up with a choice of two fragile..

This patch fix it by following the rules for changing task_struct::cpus_allowed
w/ both pi_lock and rq->lock are held. 

Reported-by: kernel test robot <ying.huang@intel.com>
Reported-by: Sasha Levin <sasha.levin@oracle.com>
Signed-off-by: Wanpeng Li <wanpeng.li@hotmail.com>
---
 v2 -> v3:
  * drop the unnecessary lockdep_unpin and unlock rq->lock
 v1 -> v2:
  * fix the silly double lock stuff
  * follow the rules for changing task_struct::cpus_allowed

 kernel/sched/core.c |   20 ++++++++++++++++++++
 1 files changed, 20 insertions(+), 0 deletions(-)

diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index b3386c6..56d19cc 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -5186,6 +5186,25 @@ static void migrate_tasks(struct rq *dead_rq)
 		BUG_ON(!next);
 		next->sched_class->put_prev_task(rq, next);
 
+		/*
+		 * Rules for changing task_struct::cpus_allowed are holding
+		 * both pi_lock and rq->lock, such that holding either
+		 * stabilizes the mask.
+		 *
+		 * Drop rq->lock is not quite as disastrous as it usually is
+		 * because !cpu_active at this point, which means load-balance
+		 * will not interfere.
+		 */
+		lockdep_unpin_lock(&rq->lock);
+		raw_spin_unlock(&rq->lock);
+		raw_spin_lock(&next->pi_lock);
+		raw_spin_lock(&rq->lock);
+		lockdep_pin_lock(&rq->lock);
+		if (!(task_rq(next) == rq && task_on_rq_queued(next))) {
+			raw_spin_unlock(&next->pi_lock);
+			continue;
+		}
+
 		/* Find suitable destination for @next, with force if needed. */
 		dest_cpu = select_fallback_rq(dead_rq->cpu, next);
 
@@ -5196,6 +5215,7 @@ static void migrate_tasks(struct rq *dead_rq)
 			rq = dead_rq;
 			raw_spin_lock(&rq->lock);
 		}
+		raw_spin_unlock(&next->pi_lock);
 	}
 
 	rq->stop = stop;
-- 
1.7.1


^ permalink raw reply related	[flat|nested] 7+ messages in thread

* Re: [PATCH v3] sched: fix tsk->pi_lock isn't held when do_set_cpus_allowed()
  2015-08-28  6:55 [PATCH v3] sched: fix tsk->pi_lock isn't held when do_set_cpus_allowed() Wanpeng Li
@ 2015-08-28 13:29 ` Peter Zijlstra
  2015-08-28 13:58   ` Wanpeng Li
  2015-09-01  6:11   ` Wanpeng Li
  2015-09-13 10:56 ` [tip:sched/core] sched: 'Annotate' migrate_tasks() tip-bot for Wanpeng Li
  1 sibling, 2 replies; 7+ messages in thread
From: Peter Zijlstra @ 2015-08-28 13:29 UTC (permalink / raw)
  To: Wanpeng Li
  Cc: Ingo Molnar, Sasha Levin, kernel test robot, Boqun Feng,
	linux-kernel

On Fri, Aug 28, 2015 at 02:55:56PM +0800, Wanpeng Li wrote:
> This patch fix it by following the rules for changing task_struct::cpus_allowed
> w/ both pi_lock and rq->lock are held. 

Thanks, I made that the below. There was a pin leak and I turned the
safety check into a WARN_ON because it really should not happen.

I also munged some of the comments a bit and did some slight edits to
the Changelog.

---
Subject: sched: 'Annotate' migrate_tasks()
From: Wanpeng Li <wanpeng.li@hotmail.com>
Date: Fri, 28 Aug 2015 14:55:56 +0800

| WARNING: CPU: 0 PID: 13 at kernel/sched/core.c:1156 do_set_cpus_allowed+0x7e/0x80()
| Modules linked in:
| CPU: 0 PID: 13 Comm: migration/0 Not tainted 4.2.0-rc1-00049-g25834c7 #2
| Call Trace:
|   dump_stack+0x4b/0x75
|   warn_slowpath_common+0x8b/0xc0
|   warn_slowpath_null+0x22/0x30
|   do_set_cpus_allowed+0x7e/0x80
|   cpuset_cpus_allowed_fallback+0x7c/0x170
|   select_fallback_rq+0x221/0x280
|   migration_call+0xe3/0x250
|   notifier_call_chain+0x53/0x70
|   __raw_notifier_call_chain+0x1e/0x30
|   cpu_notify+0x28/0x50
|   take_cpu_down+0x22/0x40
|   multi_cpu_stop+0xd5/0x140
|   cpu_stopper_thread+0xbc/0x170
|   smpboot_thread_fn+0x174/0x2f0
|   kthread+0xc4/0xe0
|   ret_from_kernel_thread+0x21/0x30

As Peterz pointed out:

| So the normal rules for changing task_struct::cpus_allowed are holding
| both pi_lock and rq->lock, such that holding either stabilizes the mask.
|
| This is so that wakeup can happen without rq->lock and load-balance
| without pi_lock.
|
| From this we already get the relaxation that we can omit acquiring
| rq->lock if the task is not on the rq, because in that case
| load-balancing will not apply to it.
|
| ** these are the rules currently tested in do_set_cpus_allowed() **
|
| Now, since __set_cpus_allowed_ptr() uses task_rq_lock() which
| unconditionally acquires both locks, we could get away with holding just
| rq->lock when on_rq for modification because that'd still exclude
| __set_cpus_allowed_ptr(), it would also work against
| __kthread_bind_mask() because that assumes !on_rq.
|
| That said, this is all somewhat fragile.
|
| Now, I don't think dropping rq->lock is quite as disastrous as it
| usually is because !cpu_active at this point, which means load-balance
| will not interfere, but that too is somewhat fragile.
|
| So we end up with a choice of two fragile..

This patch fixes it by following the rules for changing
task_struct::cpus_allowed with both pi_lock and rq->lock held.

Cc: Ingo Molnar <mingo@kernel.org>
Reported-by: kernel test robot <ying.huang@intel.com>
Reported-by: Sasha Levin <sasha.levin@oracle.com>
Signed-off-by: Wanpeng Li <wanpeng.li@hotmail.com>
[Modified changelog and patch]
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Link: http://lkml.kernel.org/r/BLU436-SMTP1660820490DE202E3934ED3806E0@phx.gbl
---

 kernel/sched/core.c |   29 ++++++++++++++++++++++++++---
 1 file changed, 26 insertions(+), 3 deletions(-)

--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -5178,24 +5178,47 @@ static void migrate_tasks(struct rq *dea
 			break;
 
 		/*
-		 * Ensure rq->lock covers the entire task selection
-		 * until the migration.
+		 * pick_next_task assumes pinned rq->lock.
 		 */
 		lockdep_pin_lock(&rq->lock);
 		next = pick_next_task(rq, &fake_task);
 		BUG_ON(!next);
 		next->sched_class->put_prev_task(rq, next);
 
+		/*
+		 * Rules for changing task_struct::cpus_allowed are holding
+		 * both pi_lock and rq->lock, such that holding either
+		 * stabilizes the mask.
+		 *
+		 * Drop rq->lock is not quite as disastrous as it usually is
+		 * because !cpu_active at this point, which means load-balance
+		 * will not interfere. Also, stop-machine.
+		 */
+		lockdep_unpin_lock(&rq->lock);
+		raw_spin_unlock(&rq->lock);
+		raw_spin_lock(&next->pi_lock);
+		raw_spin_lock(&rq->lock);
+
+		/*
+		 * Since we're inside stop-machine, _nothing_ should have
+		 * changed the task, WARN if weird stuff happened, because in
+		 * that case the above rq->lock drop is a fail too.
+		 */
+		if (WARN_ON(task_rq(next) != rq || !task_on_rq_queued(next))) {
+			raw_spin_unlock(&next->pi_lock);
+			continue;
+		}
+
 		/* Find suitable destination for @next, with force if needed. */
 		dest_cpu = select_fallback_rq(dead_rq->cpu, next);
 
-		lockdep_unpin_lock(&rq->lock);
 		rq = __migrate_task(rq, next, dest_cpu);
 		if (rq != dead_rq) {
 			raw_spin_unlock(&rq->lock);
 			rq = dead_rq;
 			raw_spin_lock(&rq->lock);
 		}
+		raw_spin_unlock(&next->pi_lock);
 	}
 
 	rq->stop = stop;

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH v3] sched: fix tsk->pi_lock isn't held when do_set_cpus_allowed()
  2015-08-28 13:29 ` Peter Zijlstra
@ 2015-08-28 13:58   ` Wanpeng Li
  2015-09-01  6:11   ` Wanpeng Li
  1 sibling, 0 replies; 7+ messages in thread
From: Wanpeng Li @ 2015-08-28 13:58 UTC (permalink / raw)
  To: Peter Zijlstra
  Cc: Ingo Molnar, Sasha Levin, kernel test robot, Boqun Feng,
	linux-kernel

On 8/28/15 9:29 PM, Peter Zijlstra wrote:
> On Fri, Aug 28, 2015 at 02:55:56PM +0800, Wanpeng Li wrote:
>> This patch fix it by following the rules for changing task_struct::cpus_allowed
>> w/ both pi_lock and rq->lock are held.
> Thanks, I made that the below. There was a pin leak and I turned the
> safety check into a WARN_ON because it really should not happen.
>
> I also munged some of the comments a bit and did some slight edits to
> the Changelog.

Cool, thanks for the help. :-)

Regards,
Wanpeng Li

>
> ---
> Subject: sched: 'Annotate' migrate_tasks()
> From: Wanpeng Li <wanpeng.li@hotmail.com>
> Date: Fri, 28 Aug 2015 14:55:56 +0800
>
> | WARNING: CPU: 0 PID: 13 at kernel/sched/core.c:1156 do_set_cpus_allowed+0x7e/0x80()
> | Modules linked in:
> | CPU: 0 PID: 13 Comm: migration/0 Not tainted 4.2.0-rc1-00049-g25834c7 #2
> | Call Trace:
> |   dump_stack+0x4b/0x75
> |   warn_slowpath_common+0x8b/0xc0
> |   warn_slowpath_null+0x22/0x30
> |   do_set_cpus_allowed+0x7e/0x80
> |   cpuset_cpus_allowed_fallback+0x7c/0x170
> |   select_fallback_rq+0x221/0x280
> |   migration_call+0xe3/0x250
> |   notifier_call_chain+0x53/0x70
> |   __raw_notifier_call_chain+0x1e/0x30
> |   cpu_notify+0x28/0x50
> |   take_cpu_down+0x22/0x40
> |   multi_cpu_stop+0xd5/0x140
> |   cpu_stopper_thread+0xbc/0x170
> |   smpboot_thread_fn+0x174/0x2f0
> |   kthread+0xc4/0xe0
> |   ret_from_kernel_thread+0x21/0x30
>
> As Peterz pointed out:
>
> | So the normal rules for changing task_struct::cpus_allowed are holding
> | both pi_lock and rq->lock, such that holding either stabilizes the mask.
> |
> | This is so that wakeup can happen without rq->lock and load-balance
> | without pi_lock.
> |
> | From this we already get the relaxation that we can omit acquiring
> | rq->lock if the task is not on the rq, because in that case
> | load-balancing will not apply to it.
> |
> | ** these are the rules currently tested in do_set_cpus_allowed() **
> |
> | Now, since __set_cpus_allowed_ptr() uses task_rq_lock() which
> | unconditionally acquires both locks, we could get away with holding just
> | rq->lock when on_rq for modification because that'd still exclude
> | __set_cpus_allowed_ptr(), it would also work against
> | __kthread_bind_mask() because that assumes !on_rq.
> |
> | That said, this is all somewhat fragile.
> |
> | Now, I don't think dropping rq->lock is quite as disastrous as it
> | usually is because !cpu_active at this point, which means load-balance
> | will not interfere, but that too is somewhat fragile.
> |
> | So we end up with a choice of two fragile..
>
> This patch fixes it by following the rules for changing
> task_struct::cpus_allowed with both pi_lock and rq->lock held.
>
> Cc: Ingo Molnar <mingo@kernel.org>
> Reported-by: kernel test robot <ying.huang@intel.com>
> Reported-by: Sasha Levin <sasha.levin@oracle.com>
> Signed-off-by: Wanpeng Li <wanpeng.li@hotmail.com>
> [Modified changelog and patch]
> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
> Link: http://lkml.kernel.org/r/BLU436-SMTP1660820490DE202E3934ED3806E0@phx.gbl
> ---
>
>   kernel/sched/core.c |   29 ++++++++++++++++++++++++++---
>   1 file changed, 26 insertions(+), 3 deletions(-)
>
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -5178,24 +5178,47 @@ static void migrate_tasks(struct rq *dea
>   			break;
>   
>   		/*
> -		 * Ensure rq->lock covers the entire task selection
> -		 * until the migration.
> +		 * pick_next_task assumes pinned rq->lock.
>   		 */
>   		lockdep_pin_lock(&rq->lock);
>   		next = pick_next_task(rq, &fake_task);
>   		BUG_ON(!next);
>   		next->sched_class->put_prev_task(rq, next);
>   
> +		/*
> +		 * Rules for changing task_struct::cpus_allowed are holding
> +		 * both pi_lock and rq->lock, such that holding either
> +		 * stabilizes the mask.
> +		 *
> +		 * Drop rq->lock is not quite as disastrous as it usually is
> +		 * because !cpu_active at this point, which means load-balance
> +		 * will not interfere. Also, stop-machine.
> +		 */
> +		lockdep_unpin_lock(&rq->lock);
> +		raw_spin_unlock(&rq->lock);
> +		raw_spin_lock(&next->pi_lock);
> +		raw_spin_lock(&rq->lock);
> +
> +		/*
> +		 * Since we're inside stop-machine, _nothing_ should have
> +		 * changed the task, WARN if weird stuff happened, because in
> +		 * that case the above rq->lock drop is a fail too.
> +		 */
> +		if (WARN_ON(task_rq(next) != rq || !task_on_rq_queued(next))) {
> +			raw_spin_unlock(&next->pi_lock);
> +			continue;
> +		}
> +
>   		/* Find suitable destination for @next, with force if needed. */
>   		dest_cpu = select_fallback_rq(dead_rq->cpu, next);
>   
> -		lockdep_unpin_lock(&rq->lock);
>   		rq = __migrate_task(rq, next, dest_cpu);
>   		if (rq != dead_rq) {
>   			raw_spin_unlock(&rq->lock);
>   			rq = dead_rq;
>   			raw_spin_lock(&rq->lock);
>   		}
> +		raw_spin_unlock(&next->pi_lock);
>   	}
>   
>   	rq->stop = stop;


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH v3] sched: fix tsk->pi_lock isn't held when do_set_cpus_allowed()
  2015-08-28 13:29 ` Peter Zijlstra
  2015-08-28 13:58   ` Wanpeng Li
@ 2015-09-01  6:11   ` Wanpeng Li
  1 sibling, 0 replies; 7+ messages in thread
From: Wanpeng Li @ 2015-09-01  6:11 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: Peter Zijlstra, Sasha Levin, kernel test robot, Boqun Feng,
	linux-kernel

Ping Ingo, ;-)
On 8/28/15 9:29 PM, Peter Zijlstra wrote:
> On Fri, Aug 28, 2015 at 02:55:56PM +0800, Wanpeng Li wrote:
>> This patch fix it by following the rules for changing task_struct::cpus_allowed
>> w/ both pi_lock and rq->lock are held.
> Thanks, I made that the below. There was a pin leak and I turned the
> safety check into a WARN_ON because it really should not happen.
>
> I also munged some of the comments a bit and did some slight edits to
> the Changelog.
>
> ---
> Subject: sched: 'Annotate' migrate_tasks()
> From: Wanpeng Li <wanpeng.li@hotmail.com>
> Date: Fri, 28 Aug 2015 14:55:56 +0800
>
> | WARNING: CPU: 0 PID: 13 at kernel/sched/core.c:1156 do_set_cpus_allowed+0x7e/0x80()
> | Modules linked in:
> | CPU: 0 PID: 13 Comm: migration/0 Not tainted 4.2.0-rc1-00049-g25834c7 #2
> | Call Trace:
> |   dump_stack+0x4b/0x75
> |   warn_slowpath_common+0x8b/0xc0
> |   warn_slowpath_null+0x22/0x30
> |   do_set_cpus_allowed+0x7e/0x80
> |   cpuset_cpus_allowed_fallback+0x7c/0x170
> |   select_fallback_rq+0x221/0x280
> |   migration_call+0xe3/0x250
> |   notifier_call_chain+0x53/0x70
> |   __raw_notifier_call_chain+0x1e/0x30
> |   cpu_notify+0x28/0x50
> |   take_cpu_down+0x22/0x40
> |   multi_cpu_stop+0xd5/0x140
> |   cpu_stopper_thread+0xbc/0x170
> |   smpboot_thread_fn+0x174/0x2f0
> |   kthread+0xc4/0xe0
> |   ret_from_kernel_thread+0x21/0x30
>
> As Peterz pointed out:
>
> | So the normal rules for changing task_struct::cpus_allowed are holding
> | both pi_lock and rq->lock, such that holding either stabilizes the mask.
> |
> | This is so that wakeup can happen without rq->lock and load-balance
> | without pi_lock.
> |
> | From this we already get the relaxation that we can omit acquiring
> | rq->lock if the task is not on the rq, because in that case
> | load-balancing will not apply to it.
> |
> | ** these are the rules currently tested in do_set_cpus_allowed() **
> |
> | Now, since __set_cpus_allowed_ptr() uses task_rq_lock() which
> | unconditionally acquires both locks, we could get away with holding just
> | rq->lock when on_rq for modification because that'd still exclude
> | __set_cpus_allowed_ptr(), it would also work against
> | __kthread_bind_mask() because that assumes !on_rq.
> |
> | That said, this is all somewhat fragile.
> |
> | Now, I don't think dropping rq->lock is quite as disastrous as it
> | usually is because !cpu_active at this point, which means load-balance
> | will not interfere, but that too is somewhat fragile.
> |
> | So we end up with a choice of two fragile..
>
> This patch fixes it by following the rules for changing
> task_struct::cpus_allowed with both pi_lock and rq->lock held.
>
> Cc: Ingo Molnar <mingo@kernel.org>
> Reported-by: kernel test robot <ying.huang@intel.com>
> Reported-by: Sasha Levin <sasha.levin@oracle.com>
> Signed-off-by: Wanpeng Li <wanpeng.li@hotmail.com>
> [Modified changelog and patch]
> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
> Link: http://lkml.kernel.org/r/BLU436-SMTP1660820490DE202E3934ED3806E0@phx.gbl
> ---
>
>   kernel/sched/core.c |   29 ++++++++++++++++++++++++++---
>   1 file changed, 26 insertions(+), 3 deletions(-)
>
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -5178,24 +5178,47 @@ static void migrate_tasks(struct rq *dea
>   			break;
>   
>   		/*
> -		 * Ensure rq->lock covers the entire task selection
> -		 * until the migration.
> +		 * pick_next_task assumes pinned rq->lock.
>   		 */
>   		lockdep_pin_lock(&rq->lock);
>   		next = pick_next_task(rq, &fake_task);
>   		BUG_ON(!next);
>   		next->sched_class->put_prev_task(rq, next);
>   
> +		/*
> +		 * Rules for changing task_struct::cpus_allowed are holding
> +		 * both pi_lock and rq->lock, such that holding either
> +		 * stabilizes the mask.
> +		 *
> +		 * Drop rq->lock is not quite as disastrous as it usually is
> +		 * because !cpu_active at this point, which means load-balance
> +		 * will not interfere. Also, stop-machine.
> +		 */
> +		lockdep_unpin_lock(&rq->lock);
> +		raw_spin_unlock(&rq->lock);
> +		raw_spin_lock(&next->pi_lock);
> +		raw_spin_lock(&rq->lock);
> +
> +		/*
> +		 * Since we're inside stop-machine, _nothing_ should have
> +		 * changed the task, WARN if weird stuff happened, because in
> +		 * that case the above rq->lock drop is a fail too.
> +		 */
> +		if (WARN_ON(task_rq(next) != rq || !task_on_rq_queued(next))) {
> +			raw_spin_unlock(&next->pi_lock);
> +			continue;
> +		}
> +
>   		/* Find suitable destination for @next, with force if needed. */
>   		dest_cpu = select_fallback_rq(dead_rq->cpu, next);
>   
> -		lockdep_unpin_lock(&rq->lock);
>   		rq = __migrate_task(rq, next, dest_cpu);
>   		if (rq != dead_rq) {
>   			raw_spin_unlock(&rq->lock);
>   			rq = dead_rq;
>   			raw_spin_lock(&rq->lock);
>   		}
> +		raw_spin_unlock(&next->pi_lock);
>   	}
>   
>   	rq->stop = stop;


^ permalink raw reply	[flat|nested] 7+ messages in thread

* [tip:sched/core] sched: 'Annotate' migrate_tasks()
  2015-08-28  6:55 [PATCH v3] sched: fix tsk->pi_lock isn't held when do_set_cpus_allowed() Wanpeng Li
  2015-08-28 13:29 ` Peter Zijlstra
@ 2015-09-13 10:56 ` tip-bot for Wanpeng Li
  2015-09-14 23:58   ` Linus Torvalds
  1 sibling, 1 reply; 7+ messages in thread
From: tip-bot for Wanpeng Li @ 2015-09-13 10:56 UTC (permalink / raw)
  To: linux-tip-commits
  Cc: torvalds, wanpeng.li, ying.huang, hpa, linux-kernel, tglx, mingo,
	peterz, sasha.levin

Commit-ID:  5473e0cc37c03c576adbda7591a6cc8e37c1bb7f
Gitweb:     http://git.kernel.org/tip/5473e0cc37c03c576adbda7591a6cc8e37c1bb7f
Author:     Wanpeng Li <wanpeng.li@hotmail.com>
AuthorDate: Fri, 28 Aug 2015 14:55:56 +0800
Committer:  Ingo Molnar <mingo@kernel.org>
CommitDate: Fri, 11 Sep 2015 07:57:50 +0200

sched: 'Annotate' migrate_tasks()

Kernel testing triggered this warning:

| WARNING: CPU: 0 PID: 13 at kernel/sched/core.c:1156 do_set_cpus_allowed+0x7e/0x80()
| Modules linked in:
| CPU: 0 PID: 13 Comm: migration/0 Not tainted 4.2.0-rc1-00049-g25834c7 #2
| Call Trace:
|   dump_stack+0x4b/0x75
|   warn_slowpath_common+0x8b/0xc0
|   warn_slowpath_null+0x22/0x30
|   do_set_cpus_allowed+0x7e/0x80
|   cpuset_cpus_allowed_fallback+0x7c/0x170
|   select_fallback_rq+0x221/0x280
|   migration_call+0xe3/0x250
|   notifier_call_chain+0x53/0x70
|   __raw_notifier_call_chain+0x1e/0x30
|   cpu_notify+0x28/0x50
|   take_cpu_down+0x22/0x40
|   multi_cpu_stop+0xd5/0x140
|   cpu_stopper_thread+0xbc/0x170
|   smpboot_thread_fn+0x174/0x2f0
|   kthread+0xc4/0xe0
|   ret_from_kernel_thread+0x21/0x30

As Peterz pointed out:

| So the normal rules for changing task_struct::cpus_allowed are holding
| both pi_lock and rq->lock, such that holding either stabilizes the mask.
|
| This is so that wakeup can happen without rq->lock and load-balance
| without pi_lock.
|
| From this we already get the relaxation that we can omit acquiring
| rq->lock if the task is not on the rq, because in that case
| load-balancing will not apply to it.
|
| ** these are the rules currently tested in do_set_cpus_allowed() **
|
| Now, since __set_cpus_allowed_ptr() uses task_rq_lock() which
| unconditionally acquires both locks, we could get away with holding just
| rq->lock when on_rq for modification because that'd still exclude
| __set_cpus_allowed_ptr(), it would also work against
| __kthread_bind_mask() because that assumes !on_rq.
|
| That said, this is all somewhat fragile.
|
| Now, I don't think dropping rq->lock is quite as disastrous as it
| usually is because !cpu_active at this point, which means load-balance
| will not interfere, but that too is somewhat fragile.
|
| So we end up with a choice of two fragile..

This patch fixes it by following the rules for changing
task_struct::cpus_allowed with both pi_lock and rq->lock held.

Reported-by: kernel test robot <ying.huang@intel.com>
Reported-by: Sasha Levin <sasha.levin@oracle.com>
Signed-off-by: Wanpeng Li <wanpeng.li@hotmail.com>
[ Modified changelog and patch. ]
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/BLU436-SMTP1660820490DE202E3934ED3806E0@phx.gbl
Signed-off-by: Ingo Molnar <mingo@kernel.org>
---
 kernel/sched/core.c | 29 ++++++++++++++++++++++++++---
 1 file changed, 26 insertions(+), 3 deletions(-)

diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 0902e4d..9b78670 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -5183,24 +5183,47 @@ static void migrate_tasks(struct rq *dead_rq)
 			break;
 
 		/*
-		 * Ensure rq->lock covers the entire task selection
-		 * until the migration.
+		 * pick_next_task assumes pinned rq->lock.
 		 */
 		lockdep_pin_lock(&rq->lock);
 		next = pick_next_task(rq, &fake_task);
 		BUG_ON(!next);
 		next->sched_class->put_prev_task(rq, next);
 
+		/*
+		 * Rules for changing task_struct::cpus_allowed are holding
+		 * both pi_lock and rq->lock, such that holding either
+		 * stabilizes the mask.
+		 *
+		 * Drop rq->lock is not quite as disastrous as it usually is
+		 * because !cpu_active at this point, which means load-balance
+		 * will not interfere. Also, stop-machine.
+		 */
+		lockdep_unpin_lock(&rq->lock);
+		raw_spin_unlock(&rq->lock);
+		raw_spin_lock(&next->pi_lock);
+		raw_spin_lock(&rq->lock);
+
+		/*
+		 * Since we're inside stop-machine, _nothing_ should have
+		 * changed the task, WARN if weird stuff happened, because in
+		 * that case the above rq->lock drop is a fail too.
+		 */
+		if (WARN_ON(task_rq(next) != rq || !task_on_rq_queued(next))) {
+			raw_spin_unlock(&next->pi_lock);
+			continue;
+		}
+
 		/* Find suitable destination for @next, with force if needed. */
 		dest_cpu = select_fallback_rq(dead_rq->cpu, next);
 
-		lockdep_unpin_lock(&rq->lock);
 		rq = __migrate_task(rq, next, dest_cpu);
 		if (rq != dead_rq) {
 			raw_spin_unlock(&rq->lock);
 			rq = dead_rq;
 			raw_spin_lock(&rq->lock);
 		}
+		raw_spin_unlock(&next->pi_lock);
 	}
 
 	rq->stop = stop;

^ permalink raw reply related	[flat|nested] 7+ messages in thread

* Re: [tip:sched/core] sched: 'Annotate' migrate_tasks()
  2015-09-13 10:56 ` [tip:sched/core] sched: 'Annotate' migrate_tasks() tip-bot for Wanpeng Li
@ 2015-09-14 23:58   ` Linus Torvalds
  2015-09-15  7:57     ` Peter Zijlstra
  0 siblings, 1 reply; 7+ messages in thread
From: Linus Torvalds @ 2015-09-14 23:58 UTC (permalink / raw)
  To: Ingo Molnar, Peter Zijlstra, Sasha Levin, wanpeng.li, Peter Anvin,
	Linus Torvalds, Huang Ying, Linux Kernel Mailing List,
	Thomas Gleixner
  Cc: linux-tip-commits@vger.kernel.org

On Sun, Sep 13, 2015 at 3:56 AM, tip-bot for Wanpeng Li
<tipbot@zytor.com> wrote:
> +               lockdep_unpin_lock(&rq->lock);
> +               raw_spin_unlock(&rq->lock);
> +               raw_spin_lock(&next->pi_lock);
> +               raw_spin_lock(&rq->lock);

So _if_ this is ever likely to be a performance-critical piece of
code, it might be worth doing

    if (!raw_spin_trylock(&next->pi_lock)) {
        .. do the above unlock and relock in the right order sequence ..
    }

to avoid doing that whole "unlock just to relock in the right order" thing.

That's particularly true in cases where dropping one of the locks can
necessitate re-doing some checks.

              Linus

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [tip:sched/core] sched: 'Annotate' migrate_tasks()
  2015-09-14 23:58   ` Linus Torvalds
@ 2015-09-15  7:57     ` Peter Zijlstra
  0 siblings, 0 replies; 7+ messages in thread
From: Peter Zijlstra @ 2015-09-15  7:57 UTC (permalink / raw)
  To: Linus Torvalds
  Cc: Ingo Molnar, Sasha Levin, wanpeng.li, Peter Anvin, Huang Ying,
	Linux Kernel Mailing List, Thomas Gleixner,
	linux-tip-commits@vger.kernel.org

On Mon, Sep 14, 2015 at 04:58:21PM -0700, Linus Torvalds wrote:
> On Sun, Sep 13, 2015 at 3:56 AM, tip-bot for Wanpeng Li
> <tipbot@zytor.com> wrote:
> > +               lockdep_unpin_lock(&rq->lock);
> > +               raw_spin_unlock(&rq->lock);
> > +               raw_spin_lock(&next->pi_lock);
> > +               raw_spin_lock(&rq->lock);
> 
> So _if_ this is ever likely to be a performance-critical piece of
> code, it might be worth doing
> 
>     if (!raw_spin_trylock(&next->pi_lock)) {
>         .. do the above unlock and relock in the right order sequence ..
>     }
> 
> to avoid doing that whole "unlock just to relock in the right order" thing.
> 
> That's particularly true in cases where dropping one of the locks can
> necessitate re-doing some checks.

Correct, but this is the hot-unplug, migrate the few tasks that are now
on a dead CPU away path, so meh ;-)

^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2015-09-15  7:57 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2015-08-28  6:55 [PATCH v3] sched: fix tsk->pi_lock isn't held when do_set_cpus_allowed() Wanpeng Li
2015-08-28 13:29 ` Peter Zijlstra
2015-08-28 13:58   ` Wanpeng Li
2015-09-01  6:11   ` Wanpeng Li
2015-09-13 10:56 ` [tip:sched/core] sched: 'Annotate' migrate_tasks() tip-bot for Wanpeng Li
2015-09-14 23:58   ` Linus Torvalds
2015-09-15  7:57     ` Peter Zijlstra

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox