From: Peter Zijlstra <peterz@infradead.org>
To: Dave Jones <davej@redhat.com>
Cc: Linux Kernel <linux-kernel@vger.kernel.org>,
Andrew Morton <akpm@linux-foundation.org>,
tglx <tglx@linutronix.de>, Oleg Nesterov <oleg@redhat.com>,
"Paul E. McKenney" <paulmck@linux.vnet.ibm.com>
Subject: Re: rcu_read_lock/unlock protect find_task_by_vpid call in posix_cpu_timer_create
Date: Thu, 25 Nov 2010 09:40:36 +0100 [thread overview]
Message-ID: <1290674436.2072.562.camel@laptop> (raw)
In-Reply-To: <20101125010948.GA1371@redhat.com>
On Wed, 2010-11-24 at 20:09 -0500, Dave Jones wrote:
> ===================================================
> [ INFO: suspicious rcu_dereference_check() usage. ]
> ---------------------------------------------------
> kernel/pid.c:419 invoked rcu_dereference_check() without protection!
>
> other info that might help us debug this:
>
>
> rcu_scheduler_active = 1, debug_locks = 1
> 1 lock held by scrashme/20820:
> #0: (tasklist_lock){.?.?..}, at: [<ffffffff8106e30f>] posix_cpu_timer_create+0x50/0xee
>
> stack backtrace:
> Pid: 20820, comm: scrashme Not tainted 2.6.37-rc3+ #7
> Call Trace:
> [<ffffffff8107cfd5>] lockdep_rcu_dereference+0x9d/0xa5
> [<ffffffff81069d08>] find_task_by_pid_ns+0x44/0x5d
> [<ffffffff81069d43>] find_task_by_vpid+0x22/0x24
> [<ffffffff8106e32d>] posix_cpu_timer_create+0x6e/0xee
> [<ffffffff8106eb88>] do_cpu_nanosleep+0x83/0x1ad
> [<ffffffff8106f50a>] posix_cpu_nsleep+0x6d/0xf6
> [<ffffffff810f9a64>] ? might_fault+0xa5/0xac
> [<ffffffff810f9a1b>] ? might_fault+0x5c/0xac
> [<ffffffff8106bf57>] sys_clock_nanosleep+0x7c/0xcb
> [<ffffffff81009cb2>] system_call_fastpath+0x16/0x1b
>
> Signed-off-by: Dave Jones <davej@redhat.com>
>
> diff --git a/kernel/posix-cpu-timers.c b/kernel/posix-cpu-timers.c
> index 6842eeb..2658955 100644
> --- a/kernel/posix-cpu-timers.c
> +++ b/kernel/posix-cpu-timers.c
> @@ -391,6 +391,7 @@ int posix_cpu_timer_create(struct k_itimer *new_timer)
> INIT_LIST_HEAD(&new_timer->it.cpu.entry);
>
> read_lock(&tasklist_lock);
> + rcu_read_lock();
> if (CPUCLOCK_PERTHREAD(new_timer->it_clock)) {
> if (pid == 0) {
> p = current;
> @@ -414,6 +415,7 @@ int posix_cpu_timer_create(struct k_itimer *new_timer)
> } else {
> ret = -EINVAL;
> }
> + rcu_read_unlock();
> read_unlock(&tasklist_lock);
>
> return ret;
Do we still need the tasklist_lock in this case?
Also, why is that think complaining, surely the tasklist_lock pins any
and all PID objects?
next prev parent reply other threads:[~2010-11-25 8:40 UTC|newest]
Thread overview: 7+ messages / expand[flat|nested] mbox.gz Atom feed top
2010-11-25 1:09 rcu_read_lock/unlock protect find_task_by_vpid call in posix_cpu_timer_create Dave Jones
2010-11-25 1:35 ` rcu_read_lock/unlock protect find_task_by_vpid call in check_clock Dave Jones
2010-11-25 8:40 ` Peter Zijlstra
2010-11-25 8:40 ` Peter Zijlstra [this message]
2010-11-25 11:00 ` rcu_read_lock/unlock protect find_task_by_vpid call in posix_cpu_timer_create Oleg Nesterov
2010-11-25 11:02 ` Oleg Nesterov
2010-11-25 11:28 ` Sergey Senozhatsky
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1290674436.2072.562.camel@laptop \
--to=peterz@infradead.org \
--cc=akpm@linux-foundation.org \
--cc=davej@redhat.com \
--cc=linux-kernel@vger.kernel.org \
--cc=oleg@redhat.com \
--cc=paulmck@linux.vnet.ibm.com \
--cc=tglx@linutronix.de \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox