From: Nicholas Mc Guire <der.herr@hofr.at>
To: Steven Rostedt <rostedt@goodmis.org>
Cc: linux-rt-users@vger.kernel.org,
Sebastian Andrzej Siewior <bigeasy@linutronix.de>,
Andreas Platschek <platschek@ict.tuwien.ac.at>,
Peter Zijlstra <peterz@infradead.org>,
LKML <linux-kernel@vger.kernel.org>
Subject: Re: [PATCH 2/3] read_lock migrate_disable pushdown to rt_read_lock
Date: Fri, 6 Dec 2013 03:33:34 +0100 [thread overview]
Message-ID: <20131206023334.GA26623@opentech.at> (raw)
In-Reply-To: <20131205211951.5c2eaa0f@gandalf.local.home>
On Thu, 05 Dec 2013, Steven Rostedt wrote:
> On Fri, 6 Dec 2013 00:44:49 +0100
> Nicholas Mc Guire <der.herr@hofr.at> wrote:
>
> >
> >
> > pushdown of migrate_disable/enable from read_*lock* to the rt_read_*lock*
> > api level
> >
> > general mapping to mutexes:
> >
> > read_*lock*
> > `-> rt_read_*lock*
> > `-> __spin_lock (the sleeping spin locks)
> > `-> rt_mutex
> >
> > The real read_lock* mapping:
> >
> > read_lock_irqsave -.
> > read_lock_irq `-> rt_read_lock_irqsave()
> > `->read_lock ---------. \
> > read_lock_bh ------+ \
> > `--> rt_read_lock()
> > if (rt_mutex_owner(lock) != current){
> > `-> __rt_spin_lock()
> > rt_spin_lock_fastlock()
> > `->rt_mutex_cmpxchg()
> > migrate_disable()
> > }
> > rwlock->read_depth++;
> > read_trylock mapping:
> >
> > read_trylock
> > `-> rt_read_trylock
> > if (rt_mutex_owner(lock) != current){
> > `-> rt_mutex_trylock()
> > rt_mutex_fasttrylock()
> > rt_mutex_cmpxchg()
> > migrate_disable()
> > }
> > rwlock->read_depth++;
> >
> > read_unlock* mapping:
> >
> > read_unlock_bh --------+
> > read_unlock_irq -------+
> > read_unlock_irqrestore +
> > read_unlock -----------+
> > `-> rt_read_unlock()
> > if(--rwlock->read_depth==0){
> > `-> __rt_spin_unlock()
> > rt_spin_lock_fastunlock()
> > `-> rt_mutex_cmpxchg()
> > migrate_disable()
> > }
> >
> > So calls to migrate_disable/enable() are better placed at the rt_read_*
> > level of lock/trylock/unlock as all of the read_*lock* API has this as a
> > common path. In the rt_read* API of lock/trylock/unlock the nesting level
> > is already being recorded in rwlock->read_depth, so we can push down the
> > migrate disable/enable to that level and condition it on the read_depth
> > going from 0 to 1 -> migrate_disable and 1 to 0 -> migrate_enable. This
> > eliminates the recursive calls that were needed when migrate_disable/enable
> > was done at the read_*lock* level.
> >
> > The approach to read_*_bh also eliminates the concerns raised with the
> > regards to api inbalances (read_lock_bh -> read_unlock+local_bh_enable)
> >
> > this is on top of 3.12.1-rt4 with the first batch of migration_cleanup
> > patches applied.
> >
> > No change of functional behavior
> >
> > Signed-off-by: Nicholas Mc Guire <der.herr@hofr.at>
> > ---
> > include/linux/rwlock_rt.h | 6 ------
> > kernel/rt.c | 9 +++++----
> > 2 files changed, 5 insertions(+), 10 deletions(-)
> >
> > diff --git a/include/linux/rwlock_rt.h b/include/linux/rwlock_rt.h
> > index 853ee36..6d86c33 100644
> > --- a/include/linux/rwlock_rt.h
> > +++ b/include/linux/rwlock_rt.h
> > @@ -33,7 +33,6 @@ extern void __rt_rwlock_init(rwlock_t *rwlock, char *name, struct lock_class_key
> > #define read_lock_irqsave(lock, flags) \
> > do { \
> > typecheck(unsigned long, flags); \
> > - migrate_disable(); \
> > flags = rt_read_lock_irqsave(lock); \
> > } while (0)
> >
> > @@ -46,14 +45,12 @@ extern void __rt_rwlock_init(rwlock_t *rwlock, char *name, struct lock_class_key
> >
> > #define read_lock(lock) \
> > do { \
> > - migrate_disable(); \
> > rt_read_lock(lock); \
> > } while (0)
> >
> > #define read_lock_bh(lock) \
> > do { \
> > local_bh_disable(); \
> > - migrate_disable(); \
> > rt_read_lock(lock); \
> > } while (0)
> >
> > @@ -77,13 +74,11 @@ extern void __rt_rwlock_init(rwlock_t *rwlock, char *name, struct lock_class_key
> > #define read_unlock(lock) \
> > do { \
> > rt_read_unlock(lock); \
> > - migrate_enable(); \
> > } while (0)
> >
> > #define read_unlock_bh(lock) \
> > do { \
> > rt_read_unlock(lock); \
> > - migrate_enable(); \
> > local_bh_enable(); \
> > } while (0)
> >
> > @@ -109,7 +104,6 @@ extern void __rt_rwlock_init(rwlock_t *rwlock, char *name, struct lock_class_key
> > typecheck(unsigned long, flags); \
> > (void) flags; \
> > rt_read_unlock(lock); \
> > - migrate_enable(); \
> > } while (0)
> >
> > #define write_unlock_irqrestore(lock, flags) \
> > diff --git a/kernel/rt.c b/kernel/rt.c
> > index 29771e2..3403000 100644
> > --- a/kernel/rt.c
> > +++ b/kernel/rt.c
> > @@ -213,19 +213,18 @@ int __lockfunc rt_read_trylock(rwlock_t *rwlock)
> > * but not when read_depth == 0 which means that the lock is
> > * write locked.
> > */
> > - migrate_disable();
> > if (rt_mutex_owner(lock) != current) {
> > ret = rt_mutex_trylock(lock);
> > - if (ret)
> > + if (ret) {
> > + migrate_disable();
> > rwlock_acquire(&rwlock->dep_map, 0, 1, _RET_IP_);
> > + }
>
> I like this patch in general, but I'm nervous about this part.
>
> What was the original reason to disable migration before checking the
> mutex owner? Seems like we should have only disabled it on success. I'm
> assuming there was some subtle reason not to.
>
I think that it was simply being treated semantically as a lock - but it
is actually not.
hold a
migrate_disable/lock per_cpu unlock/migrate_enable
lock/migrate_disable per_cpu migrate_enable/unlcok
lock/migrate_disable per_cpu unlock/migrate_enable
migrate_disable/lock per_cpu migrate_enable/unlcok
reference
are all equivalent - the only thing you need to ensure that the per cpu
object will not be accessed before both lock and migration_disable have
been sucessful. So from my understanding this is safe.
if we get migrated after a succesful trylock what would go wrong ?
the protected object will not be accessed until after the spin_trylock
returned so migration is disabled
> If there was some strange reason, I'm not convinced that your change
> makes that reason go away.
>
IF there is a reason then this patch is broken - my conclusion up to now
is that there is no such reason.
thx!
hofrat
next prev parent reply other threads:[~2013-12-06 2:33 UTC|newest]
Thread overview: 7+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-12-05 23:44 [PATCH 2/3] read_lock migrate_disable pushdown to rt_read_lock Nicholas Mc Guire
2013-12-06 2:19 ` Steven Rostedt
2013-12-06 2:33 ` Nicholas Mc Guire [this message]
2013-12-06 15:25 ` Steven Rostedt
2013-12-15 15:15 ` Sebastian Andrzej Siewior
2014-01-24 12:42 ` Sebastian Andrzej Siewior
2014-01-25 8:29 ` Nicholas Mc Guire
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20131206023334.GA26623@opentech.at \
--to=der.herr@hofr.at \
--cc=bigeasy@linutronix.de \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-rt-users@vger.kernel.org \
--cc=peterz@infradead.org \
--cc=platschek@ict.tuwien.ac.at \
--cc=rostedt@goodmis.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).