public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
* [PATCH v3 6/9] rcu: use __this_cpu_read helper instead of per_cpu_ptr(p, raw_smp_processor_id())
@ 2012-11-09  2:21 Shan Wei
  2012-11-09 20:14 ` Christoph Lameter
  0 siblings, 1 reply; 7+ messages in thread
From: Shan Wei @ 2012-11-09  2:21 UTC (permalink / raw)
  To: dipankar, paulmck, Kernel-Maillist, cl, Shan Wei

From: Shan Wei <davidshan@tencent.com>

Signed-off-by: Shan Wei <davidshan@tencent.com>
---
no changes vs v2.
---
 kernel/rcutree.c |    2 +-
 1 files changed, 1 insertions(+), 1 deletions(-)

diff --git a/kernel/rcutree.c b/kernel/rcutree.c
index 74df86b..441b945 100644
--- a/kernel/rcutree.c
+++ b/kernel/rcutree.c
@@ -1960,7 +1960,7 @@ static void force_quiescent_state(struct rcu_state *rsp)
 	struct rcu_node *rnp_old = NULL;
 
 	/* Funnel through hierarchy to reduce memory contention. */
-	rnp = per_cpu_ptr(rsp->rda, raw_smp_processor_id())->mynode;
+	rnp = __this_cpu_read(rsp->rda->mynode);
 	for (; rnp != NULL; rnp = rnp->parent) {
 		ret = (ACCESS_ONCE(rsp->gp_flags) & RCU_GP_FLAG_FQS) ||
 		      !raw_spin_trylock(&rnp->fqslock);
-- 
1.7.1

^ permalink raw reply related	[flat|nested] 7+ messages in thread

* Re: [PATCH v3 6/9] rcu: use __this_cpu_read helper instead of per_cpu_ptr(p, raw_smp_processor_id())
  2012-11-09  2:21 Shan Wei
@ 2012-11-09 20:14 ` Christoph Lameter
  0 siblings, 0 replies; 7+ messages in thread
From: Christoph Lameter @ 2012-11-09 20:14 UTC (permalink / raw)
  To: Shan Wei; +Cc: dipankar, paulmck, Kernel-Maillist

On Fri, 9 Nov 2012, Shan Wei wrote:

>  	/* Funnel through hierarchy to reduce memory contention. */
> -	rnp = per_cpu_ptr(rsp->rda, raw_smp_processor_id())->mynode;
> +	rnp = __this_cpu_read(rsp->rda->mynode);
>  	for (; rnp != NULL; rnp = rnp->parent) {

Acked-by: Christoph Lameter <cl@linux.com>


^ permalink raw reply	[flat|nested] 7+ messages in thread

* [PATCH v3 6/9] rcu: use __this_cpu_read helper instead of per_cpu_ptr(p, raw_smp_processor_id())
@ 2012-11-13  1:52 Shan Wei
  2012-11-16  8:32 ` Shan Wei
  0 siblings, 1 reply; 7+ messages in thread
From: Shan Wei @ 2012-11-13  1:52 UTC (permalink / raw)
  To: dipankar, paulmck, Kernel-Maillist, cl, Shan Wei

From: Shan Wei <davidshan@tencent.com>

Signed-off-by: Shan Wei <davidshan@tencent.com>
Acked-by: Christoph Lameter <cl@linux.com>
---
no changes vs v3,v2.
---
 kernel/rcutree.c |    2 +-
 1 files changed, 1 insertions(+), 1 deletions(-)

diff --git a/kernel/rcutree.c b/kernel/rcutree.c
index 74df86b..441b945 100644
--- a/kernel/rcutree.c
+++ b/kernel/rcutree.c
@@ -1960,7 +1960,7 @@ static void force_quiescent_state(struct rcu_state *rsp)
 	struct rcu_node *rnp_old = NULL;
 
 	/* Funnel through hierarchy to reduce memory contention. */
-	rnp = per_cpu_ptr(rsp->rda, raw_smp_processor_id())->mynode;
+	rnp = __this_cpu_read(rsp->rda->mynode);
 	for (; rnp != NULL; rnp = rnp->parent) {
 		ret = (ACCESS_ONCE(rsp->gp_flags) & RCU_GP_FLAG_FQS) ||
 		      !raw_spin_trylock(&rnp->fqslock);
-- 
1.7.1


^ permalink raw reply related	[flat|nested] 7+ messages in thread

* Re: [PATCH v3 6/9] rcu: use __this_cpu_read helper instead of per_cpu_ptr(p, raw_smp_processor_id())
  2012-11-13  1:52 [PATCH v3 6/9] rcu: use __this_cpu_read helper instead of per_cpu_ptr(p, raw_smp_processor_id()) Shan Wei
@ 2012-11-16  8:32 ` Shan Wei
  2012-11-16 16:43   ` Paul E. McKenney
  0 siblings, 1 reply; 7+ messages in thread
From: Shan Wei @ 2012-11-16  8:32 UTC (permalink / raw)
  To: paulmck; +Cc: Shan Wei, dipankar, Kernel-Maillist, cl, Tejun Heo

Shan Wei said, at 2012/11/13 9:52:
> From: Shan Wei <davidshan@tencent.com>
> 
> Signed-off-by: Shan Wei <davidshan@tencent.com>
> Acked-by: Christoph Lameter <cl@linux.com>

Paul, would you like to pick it up to your tree?

> ---
> no changes vs v3,v2.
> ---
>  kernel/rcutree.c |    2 +-
>  1 files changed, 1 insertions(+), 1 deletions(-)
> 
> diff --git a/kernel/rcutree.c b/kernel/rcutree.c
> index 74df86b..441b945 100644
> --- a/kernel/rcutree.c
> +++ b/kernel/rcutree.c
> @@ -1960,7 +1960,7 @@ static void force_quiescent_state(struct rcu_state *rsp)
>  	struct rcu_node *rnp_old = NULL;
>  
>  	/* Funnel through hierarchy to reduce memory contention. */
> -	rnp = per_cpu_ptr(rsp->rda, raw_smp_processor_id())->mynode;
> +	rnp = __this_cpu_read(rsp->rda->mynode);
>  	for (; rnp != NULL; rnp = rnp->parent) {
>  		ret = (ACCESS_ONCE(rsp->gp_flags) & RCU_GP_FLAG_FQS) ||
>  		      !raw_spin_trylock(&rnp->fqslock);
> 


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH v3 6/9] rcu: use __this_cpu_read helper instead of per_cpu_ptr(p, raw_smp_processor_id())
  2012-11-16  8:32 ` Shan Wei
@ 2012-11-16 16:43   ` Paul E. McKenney
  2012-11-16 17:42     ` Christoph Lameter
  0 siblings, 1 reply; 7+ messages in thread
From: Paul E. McKenney @ 2012-11-16 16:43 UTC (permalink / raw)
  To: Shan Wei; +Cc: dipankar, Kernel-Maillist, cl, Tejun Heo

On Fri, Nov 16, 2012 at 04:32:50PM +0800, Shan Wei wrote:
> Shan Wei said, at 2012/11/13 9:52:
> > From: Shan Wei <davidshan@tencent.com>
> > 
> > Signed-off-by: Shan Wei <davidshan@tencent.com>
> > Acked-by: Christoph Lameter <cl@linux.com>
> 
> Paul, would you like to pick it up to your tree?

Hello, Shan Wei,

If you either show me some significant performance benefits or get me
an independent Tested-by, in both cases on a range of hardware (e.g.,
x86 on the one hand and ARM or Power on the other), then I will queue it.

I wasn't prioritizing this one very high because it does not appear
to be on any sort of fastpath.  If I am wrong about that, then you
have a good performance-benefit case, right?  ;-)

							Thanx, Paul

> > ---
> > no changes vs v3,v2.
> > ---
> >  kernel/rcutree.c |    2 +-
> >  1 files changed, 1 insertions(+), 1 deletions(-)
> > 
> > diff --git a/kernel/rcutree.c b/kernel/rcutree.c
> > index 74df86b..441b945 100644
> > --- a/kernel/rcutree.c
> > +++ b/kernel/rcutree.c
> > @@ -1960,7 +1960,7 @@ static void force_quiescent_state(struct rcu_state *rsp)
> >  	struct rcu_node *rnp_old = NULL;
> >  
> >  	/* Funnel through hierarchy to reduce memory contention. */
> > -	rnp = per_cpu_ptr(rsp->rda, raw_smp_processor_id())->mynode;
> > +	rnp = __this_cpu_read(rsp->rda->mynode);
> >  	for (; rnp != NULL; rnp = rnp->parent) {
> >  		ret = (ACCESS_ONCE(rsp->gp_flags) & RCU_GP_FLAG_FQS) ||
> >  		      !raw_spin_trylock(&rnp->fqslock);
> > 
> 


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH v3 6/9] rcu: use __this_cpu_read helper instead of per_cpu_ptr(p, raw_smp_processor_id())
  2012-11-16 16:43   ` Paul E. McKenney
@ 2012-11-16 17:42     ` Christoph Lameter
  2012-11-16 18:09       ` Paul E. McKenney
  0 siblings, 1 reply; 7+ messages in thread
From: Christoph Lameter @ 2012-11-16 17:42 UTC (permalink / raw)
  To: Paul E. McKenney; +Cc: Shan Wei, dipankar, Kernel-Maillist, Tejun Heo


On Fri, 16 Nov 2012, Paul E. McKenney wrote:

> If you either show me some significant performance benefits or get me
> an independent Tested-by, in both cases on a range of hardware (e.g.,
> x86 on the one hand and ARM or Power on the other), then I will queue it.

Just putting the code generated for x86 before and after side
by side would be enough to convince you I think.

> I wasn't prioritizing this one very high because it does not appear
> to be on any sort of fastpath.  If I am wrong about that, then you
> have a good performance-benefit case, right?  ;-)

I do not think this needs to be a priority item. Just stick it in the tree
somewhere to merge for the next merge period.


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH v3 6/9] rcu: use __this_cpu_read helper instead of per_cpu_ptr(p, raw_smp_processor_id())
  2012-11-16 17:42     ` Christoph Lameter
@ 2012-11-16 18:09       ` Paul E. McKenney
  0 siblings, 0 replies; 7+ messages in thread
From: Paul E. McKenney @ 2012-11-16 18:09 UTC (permalink / raw)
  To: Christoph Lameter; +Cc: Shan Wei, dipankar, Kernel-Maillist, Tejun Heo

On Fri, Nov 16, 2012 at 05:42:14PM +0000, Christoph Lameter wrote:
> 
> On Fri, 16 Nov 2012, Paul E. McKenney wrote:
> 
> > If you either show me some significant performance benefits or get me
> > an independent Tested-by, in both cases on a range of hardware (e.g.,
> > x86 on the one hand and ARM or Power on the other), then I will queue it.
> 
> Just putting the code generated for x86 before and after side
> by side would be enough to convince you I think.

If accompanied by similar before/after code for ARM or Power, sure.

							Thanx, Paul

> > I wasn't prioritizing this one very high because it does not appear
> > to be on any sort of fastpath.  If I am wrong about that, then you
> > have a good performance-benefit case, right?  ;-)
> 
> I do not think this needs to be a priority item. Just stick it in the tree
> somewhere to merge for the next merge period.
> 


^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2012-11-16 18:10 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2012-11-13  1:52 [PATCH v3 6/9] rcu: use __this_cpu_read helper instead of per_cpu_ptr(p, raw_smp_processor_id()) Shan Wei
2012-11-16  8:32 ` Shan Wei
2012-11-16 16:43   ` Paul E. McKenney
2012-11-16 17:42     ` Christoph Lameter
2012-11-16 18:09       ` Paul E. McKenney
  -- strict thread matches above, loose matches on Subject: below --
2012-11-09  2:21 Shan Wei
2012-11-09 20:14 ` Christoph Lameter

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox