public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: "BALBIR SINGH" <balbir.singh@wipro.com>
To: <dipankar@in.ibm.com>, "'Mala Anand'" <manand@us.ibm.com>
Cc: <linux-kernel@vger.kernel.org>,
	<lse-tech-admin@lists.sourceforge.net>,
	"'Paul McKenney'" <Paul.McKenney@us.ibm.com>,
	"'Rusty Russell'" <rusty@rustcorp.com.au>
Subject: RE: [Lse-tech] Re: [RFC] Dynamic percpu data allocator
Date: Fri, 31 May 2002 13:27:44 +0530	[thread overview]
Message-ID: <00eb01c20878$d952b890$290806c0@wipro.com> (raw)
In-Reply-To: <20020530232513.C3575@in.ibm.com>

[-- Attachment #1: Type: text/plain, Size: 2920 bytes --]

|Actually I don't know for sure what plans are afoot to fix the 
|slab allocator
|for per-cpu. One plan I heard about was allocating from per-cpu pools
|rather than per-cpu copies. My requirements are similar to
|the hot list skbs. I want to do this -
|
|	int *ctrp1, *ctrp2;
|	
|	ctrp1 = kmalloc_percpu(sizeof(*ctrp1), GFP_ATOMIC);
|	if (ctrp1 == NULL) {
|		/* recover */
|	}
|	ctrp2 = kmalloc_percpu(sizeof(*ctrp2), GFP_ATOMIC);
|	if (ctrp2 == NULL) {
|		/* recover */
|	}
|
|	*per_cpu_ptr(ctrp1, smp_processor_id())++;
|	this_cpu_ptr(ctrp2)++;
|
|Now I can allocate by making ctrp1/ctrp2 point to an array
|of NR_CPUS and kmalloc() memory for each CPU's copy of the
|int. This is simple and will work. 
|
|	void **ptrs = kmalloc(sizeof(*ptrs) * NR_CPUS, flags);
|
|	if (!ptrs) return NULL;
|	for (i = 0; i < NR_CPUS; i++) {
|	      ptrs[i] = kmalloc(size, flags);
|	      if (!ptrs[i])
|		      goto unwind_oom;
|	}
|
|
|However I would like to use kmalloc_percpu() for allocating very 
|small objects - typlically integer counters or small structures
|to be used as per-cpu counters for things like dst entries and 
|dentries.
|kmalloc will waste the rest of the cache line for such small objects.
|The alternative is to use a layer of code to interleave small objects
|and save on space.
|
|
|   CPU #0          CPU#1
|
| ---------       ---------         Start of cache line
|   *ctrp1         *ctrp1
|   *ctrp2         *ctrp2
|
|   .               .
|   .               .
|   .               .
|   .               .
|   .               .
|
| ---------       ----------        End of cache line


Won't this result in a lot of false sharing, if any of the CPUs
tried to access any of the counters, the entire cache line would be
moved from the current CPU to that CPU. Isn't this a very bad thing or
am I missing something? Do all your counters fit into one cache line.

For sometime now, I have been thinking of implementing/supporting
PME's (Peformance Monitoring Events and Counters), so that we can
get real values (atleast on x86) as compared to our guesses about
cacheline bouncing, etc. Do you know if somebody is already doing
this?

Regards,
Balbir

|
|I have an allocator that interleaves objects like this if they 
|can be fitted
|into size that is a factor of SMP_CACHE_BYTES. 
|
|I hope someone can tell me that I don't even have to do this. Otherwise
|I will go ahead and do my thing.
|
|Thanks
|-- 
|Dipankar Sarma  <dipankar@in.ibm.com> http://lse.sourceforge.net
|Linux Technology Center, IBM Software Lab, Bangalore, India.
|
|_______________________________________________________________
|
|Don't miss the 2002 Sprint PCS Application Developer's Conference
|August 25-28 in Las Vegas -- http://devcon.sprintpcs.com/adp/index.cfm
|
|_______________________________________________
|Lse-tech mailing list
|Lse-tech@lists.sourceforge.net
|https://lists.sourceforge.net/lists/listinfo/lse-tech
|


[-- Attachment #2: Wipro_Disclaimer.txt --]
[-- Type: text/plain, Size: 490 bytes --]

**************************Disclaimer************************************

Information contained in this E-MAIL being proprietary to Wipro Limited is 
'privileged' and 'confidential' and intended for use only by the individual
 or entity to which it is addressed. You are notified that any use, copying 
or dissemination of the information contained in the E-MAIL in any manner 
whatsoever is strictly prohibited.

***************************************************************************

  reply	other threads:[~2002-05-31  7:54 UTC|newest]

Thread overview: 12+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2002-05-30 13:56 [Lse-tech] Re: [RFC] Dynamic percpu data allocator Mala Anand
2002-05-30 17:55 ` Dipankar Sarma
2002-05-31  7:57   ` BALBIR SINGH [this message]
2002-05-31  8:40     ` Dipankar Sarma
  -- strict thread matches above, loose matches on Subject: below --
2002-06-04 21:11 Paul McKenney
2002-06-04 12:05 Mala Anand
2002-06-03 19:12 Mala Anand
2002-06-03 19:48 ` Dipankar Sarma
2002-05-24  6:13 Dipankar Sarma
2002-05-24  8:38 ` [Lse-tech] " BALBIR SINGH
2002-05-24  9:13   ` Dipankar Sarma
2002-05-24 11:59     ` BALBIR SINGH
2002-05-24 14:38   ` Martin J. Bligh

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to='00eb01c20878$d952b890$290806c0@wipro.com' \
    --to=balbir.singh@wipro.com \
    --cc=Paul.McKenney@us.ibm.com \
    --cc=dipankar@in.ibm.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=lse-tech-admin@lists.sourceforge.net \
    --cc=manand@us.ibm.com \
    --cc=rusty@rustcorp.com.au \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox