* [PATCH 5/7] perf: Optimise topology iteration @ 2010-12-27 15:38 Lin Ming 2011-01-03 11:02 ` Peter Zijlstra 2011-01-04 12:06 ` Peter Zijlstra 0 siblings, 2 replies; 9+ messages in thread From: Lin Ming @ 2010-12-27 15:38 UTC (permalink / raw) To: Peter Zijlstra, Ingo Molnar, Andi Kleen, Stephane Eranian, robert.richter Cc: lkml Currently we iterate the full machine looking for a matching core_id/nb for the percore and the amd northbridge stuff , using a smaller topology mask makes sense. Signed-off-by: Lin Ming <ming.m.lin@intel.com> --- arch/x86/kernel/cpu/perf_event_amd.c | 2 +- arch/x86/kernel/cpu/perf_event_intel.c | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/arch/x86/kernel/cpu/perf_event_amd.c b/arch/x86/kernel/cpu/perf_event_amd.c index 67e2202..5a3b7b8 100644 --- a/arch/x86/kernel/cpu/perf_event_amd.c +++ b/arch/x86/kernel/cpu/perf_event_amd.c @@ -323,7 +323,7 @@ static void amd_pmu_cpu_starting(int cpu) nb_id = amd_get_nb_id(cpu); WARN_ON_ONCE(nb_id == BAD_APICID); - for_each_online_cpu(i) { + for_each_cpu(i, topology_core_cpumask(cpu)) { nb = per_cpu(cpu_hw_events, i).amd_nb; if (WARN_ON_ONCE(!nb)) continue; diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c index 354d1de..ad70c2c 100644 --- a/arch/x86/kernel/cpu/perf_event_intel.c +++ b/arch/x86/kernel/cpu/perf_event_intel.c @@ -1111,7 +1111,7 @@ static void intel_pmu_cpu_starting(int cpu) if (!ht_enabled(cpu)) return; - for_each_online_cpu(i) { + for_each_cpu(i, topology_thread_cpumask(cpu)) { struct intel_percore *pc = per_cpu(cpu_hw_events, i).per_core; if (pc && pc->core_id == core_id) { -- 1.7.3 ^ permalink raw reply related [flat|nested] 9+ messages in thread
* Re: [PATCH 5/7] perf: Optimise topology iteration 2010-12-27 15:38 [PATCH 5/7] perf: Optimise topology iteration Lin Ming @ 2011-01-03 11:02 ` Peter Zijlstra 2011-01-03 15:20 ` Andi Kleen 2011-01-04 6:18 ` Lin Ming 2011-01-04 12:06 ` Peter Zijlstra 1 sibling, 2 replies; 9+ messages in thread From: Peter Zijlstra @ 2011-01-03 11:02 UTC (permalink / raw) To: Lin Ming; +Cc: Ingo Molnar, Andi Kleen, Stephane Eranian, robert.richter, lkml On Mon, 2010-12-27 at 23:38 +0800, Lin Ming wrote: > Currently we iterate the full machine looking for a matching core_id/nb > for the percore and the amd northbridge stuff , using a smaller topology > mask makes sense. Does topology_thread_cpumask() include offline cpus? I tried looking at it, but I cannot find any code clearing bits in that mask on offline. > Signed-off-by: Lin Ming <ming.m.lin@intel.com> > --- > arch/x86/kernel/cpu/perf_event_amd.c | 2 +- > arch/x86/kernel/cpu/perf_event_intel.c | 2 +- > 2 files changed, 2 insertions(+), 2 deletions(-) > > diff --git a/arch/x86/kernel/cpu/perf_event_amd.c b/arch/x86/kernel/cpu/perf_event_amd.c > index 67e2202..5a3b7b8 100644 > --- a/arch/x86/kernel/cpu/perf_event_amd.c > +++ b/arch/x86/kernel/cpu/perf_event_amd.c > @@ -323,7 +323,7 @@ static void amd_pmu_cpu_starting(int cpu) > nb_id = amd_get_nb_id(cpu); > WARN_ON_ONCE(nb_id == BAD_APICID); > > - for_each_online_cpu(i) { > + for_each_cpu(i, topology_core_cpumask(cpu)) { > nb = per_cpu(cpu_hw_events, i).amd_nb; > if (WARN_ON_ONCE(!nb)) > continue; > diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c > index 354d1de..ad70c2c 100644 > --- a/arch/x86/kernel/cpu/perf_event_intel.c > +++ b/arch/x86/kernel/cpu/perf_event_intel.c > @@ -1111,7 +1111,7 @@ static void intel_pmu_cpu_starting(int cpu) > if (!ht_enabled(cpu)) > return; > > - for_each_online_cpu(i) { > + for_each_cpu(i, topology_thread_cpumask(cpu)) { > struct intel_percore *pc = per_cpu(cpu_hw_events, i).per_core; > > if (pc && pc->core_id == core_id) { ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 5/7] perf: Optimise topology iteration 2011-01-03 11:02 ` Peter Zijlstra @ 2011-01-03 15:20 ` Andi Kleen 2011-01-04 7:13 ` Lin Ming 2011-01-04 6:18 ` Lin Ming 1 sibling, 1 reply; 9+ messages in thread From: Andi Kleen @ 2011-01-03 15:20 UTC (permalink / raw) To: Peter Zijlstra Cc: Lin Ming, Ingo Molnar, Andi Kleen, Stephane Eranian, robert.richter, lkml On Mon, Jan 03, 2011 at 12:02:10PM +0100, Peter Zijlstra wrote: > On Mon, 2010-12-27 at 23:38 +0800, Lin Ming wrote: > > Currently we iterate the full machine looking for a matching core_id/nb > > for the percore and the amd northbridge stuff , using a smaller topology > > mask makes sense. > > Does topology_thread_cpumask() include offline cpus? I tried looking at > it, but I cannot find any code clearing bits in that mask on offline. The problem is not only at offline, but also at online between CPUs going online. I don't think the patch is a good idea and it doesn't even have any advantages either since this is a initialization only slow path. -Andi ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 5/7] perf: Optimise topology iteration 2011-01-03 15:20 ` Andi Kleen @ 2011-01-04 7:13 ` Lin Ming 0 siblings, 0 replies; 9+ messages in thread From: Lin Ming @ 2011-01-04 7:13 UTC (permalink / raw) To: Andi Kleen Cc: Peter Zijlstra, Ingo Molnar, Stephane Eranian, robert.richter@amd.com, lkml On Mon, 2011-01-03 at 23:20 +0800, Andi Kleen wrote: > On Mon, Jan 03, 2011 at 12:02:10PM +0100, Peter Zijlstra wrote: > > On Mon, 2010-12-27 at 23:38 +0800, Lin Ming wrote: > > > Currently we iterate the full machine looking for a matching core_id/nb > > > for the percore and the amd northbridge stuff , using a smaller topology > > > mask makes sense. > > > > Does topology_thread_cpumask() include offline cpus? I tried looking at > > it, but I cannot find any code clearing bits in that mask on offline. > > The problem is not only at offline, but also at online between CPUs > going online. I don't think the patch is a good idea and it doesn't I didn't see the problem. Assume logical cpu 3, 7 are 2 threads in a core, and they are plug/unpluged as below sequence, CPU 3 offline, CPU7 offline, CPU3 online, CPU7 online 1. After cpu3 offline topology_thread_cpumask(3) returns empty topology_thread_cpumask(7) returns 7 2. After CPU7 offline topology_thread_cpumask(3) returns empty topology_thread_cpumask(7) returns empty 3. When CPU3 online, calling intel_pmu_cpu_starting topology_thread_cpumask(3) returns 3 topology_thread_cpumask(7) returns empty for_each_cpu(i, topology_thread_cpumask(cpu)) { struct intel_percore *pc = per_cpu(cpu_hw_events, i).per_core; if (pc && pc->core_id == core_id) { kfree(cpuc->per_core); cpuc->per_core = pc; break; } } Above "if" statement will not be executed, because pc->core_id was initialized to -1 in intel_pmu_cpu_prepare. 4. When CPU7 online, calling intel_pmu_cpu_starting topology_thread_cpumask(3) returns 3, 7 topology_thread_cpumask(7) returns 3, 7 for_each_cpu(i, topology_thread_cpumask(cpu)) { struct intel_percore *pc = per_cpu(cpu_hw_events, i).per_core; if (pc && pc->core_id == core_id) { kfree(cpuc->per_core); cpuc->per_core = pc; break; } } cpuc->per_core->core_id = core_id; cpuc->per_core->refcnt++; Above "if" statement will be executed and the per_core data allocated for cpu7 will be freed. All above is right, or could you explain more about the problem at CPUs offline and online? Thanks, Lin Ming > even have any advantages either since this is a initialization only > slow path. > > -Andi ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 5/7] perf: Optimise topology iteration 2011-01-03 11:02 ` Peter Zijlstra 2011-01-03 15:20 ` Andi Kleen @ 2011-01-04 6:18 ` Lin Ming 1 sibling, 0 replies; 9+ messages in thread From: Lin Ming @ 2011-01-04 6:18 UTC (permalink / raw) To: Peter Zijlstra Cc: Ingo Molnar, Andi Kleen, Stephane Eranian, robert.richter@amd.com, lkml On Mon, 2011-01-03 at 19:02 +0800, Peter Zijlstra wrote: > On Mon, 2010-12-27 at 23:38 +0800, Lin Ming wrote: > > Currently we iterate the full machine looking for a matching core_id/nb > > for the percore and the amd northbridge stuff , using a smaller topology > > mask makes sense. > > Does topology_thread_cpumask() include offline cpus? I tried looking at > it, but I cannot find any code clearing bits in that mask on offline. No, it does not include offline cpus. For x86 code, remove_siblinginfo() clears the bits. take_cpu_down -> __cpu_disable -> native_cpu_disable -> cpu_disable_common -> remove_siblinginfo static void remove_siblinginfo(int cpu) { int sibling; struct cpuinfo_x86 *c = &cpu_data(cpu); for_each_cpu(sibling, cpu_core_mask(cpu)) { cpumask_clear_cpu(cpu, cpu_core_mask(sibling)); /*/ * last thread sibling in this cpu core going down */ if (cpumask_weight(cpu_sibling_mask(cpu)) == 1) cpu_data(sibling).booted_cores--; } for_each_cpu(sibling, cpu_sibling_mask(cpu)) cpumask_clear_cpu(cpu, cpu_sibling_mask(sibling)); cpumask_clear(cpu_sibling_mask(cpu)); cpumask_clear(cpu_core_mask(cpu)); c->phys_proc_id = 0; c->cpu_core_id = 0; cpumask_clear_cpu(cpu, cpu_sibling_setup_mask); } Lin Ming > > > Signed-off-by: Lin Ming <ming.m.lin@intel.com> > > --- > > arch/x86/kernel/cpu/perf_event_amd.c | 2 +- > > arch/x86/kernel/cpu/perf_event_intel.c | 2 +- > > 2 files changed, 2 insertions(+), 2 deletions(-) > > > > diff --git a/arch/x86/kernel/cpu/perf_event_amd.c b/arch/x86/kernel/cpu/perf_event_amd.c > > index 67e2202..5a3b7b8 100644 > > --- a/arch/x86/kernel/cpu/perf_event_amd.c > > +++ b/arch/x86/kernel/cpu/perf_event_amd.c > > @@ -323,7 +323,7 @@ static void amd_pmu_cpu_starting(int cpu) > > nb_id = amd_get_nb_id(cpu); > > WARN_ON_ONCE(nb_id == BAD_APICID); > > > > - for_each_online_cpu(i) { > > + for_each_cpu(i, topology_core_cpumask(cpu)) { > > nb = per_cpu(cpu_hw_events, i).amd_nb; > > if (WARN_ON_ONCE(!nb)) > > continue; > > diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c > > index 354d1de..ad70c2c 100644 > > --- a/arch/x86/kernel/cpu/perf_event_intel.c > > +++ b/arch/x86/kernel/cpu/perf_event_intel.c > > @@ -1111,7 +1111,7 @@ static void intel_pmu_cpu_starting(int cpu) > > if (!ht_enabled(cpu)) > > return; > > > > - for_each_online_cpu(i) { > > + for_each_cpu(i, topology_thread_cpumask(cpu)) { > > struct intel_percore *pc = per_cpu(cpu_hw_events, i).per_core; > > > > if (pc && pc->core_id == core_id) { > > ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 5/7] perf: Optimise topology iteration 2010-12-27 15:38 [PATCH 5/7] perf: Optimise topology iteration Lin Ming 2011-01-03 11:02 ` Peter Zijlstra @ 2011-01-04 12:06 ` Peter Zijlstra 2011-01-04 14:22 ` Borislav Petkov 1 sibling, 1 reply; 9+ messages in thread From: Peter Zijlstra @ 2011-01-04 12:06 UTC (permalink / raw) To: Lin Ming Cc: Ingo Molnar, Andi Kleen, Stephane Eranian, robert.richter, lkml, Borislav Petkov On Mon, 2010-12-27 at 23:38 +0800, Lin Ming wrote: > Currently we iterate the full machine looking for a matching core_id/nb > for the percore and the amd northbridge stuff , using a smaller topology > mask makes sense. > > Signed-off-by: Lin Ming <ming.m.lin@intel.com> > --- > arch/x86/kernel/cpu/perf_event_amd.c | 2 +- > arch/x86/kernel/cpu/perf_event_intel.c | 2 +- > 2 files changed, 2 insertions(+), 2 deletions(-) > > diff --git a/arch/x86/kernel/cpu/perf_event_amd.c b/arch/x86/kernel/cpu/perf_event_amd.c > index 67e2202..5a3b7b8 100644 > --- a/arch/x86/kernel/cpu/perf_event_amd.c > +++ b/arch/x86/kernel/cpu/perf_event_amd.c > @@ -323,7 +323,7 @@ static void amd_pmu_cpu_starting(int cpu) > nb_id = amd_get_nb_id(cpu); > WARN_ON_ONCE(nb_id == BAD_APICID); > > - for_each_online_cpu(i) { > + for_each_cpu(i, topology_core_cpumask(cpu)) { > nb = per_cpu(cpu_hw_events, i).amd_nb; > if (WARN_ON_ONCE(!nb)) > continue; Borislav, is topology_core_cpumask() the right mask for northbridge_id span? I could imagine Magny-Cours would have all 12 cores in the core_cpumask() and have the node_mask() be half that. ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 5/7] perf: Optimise topology iteration 2011-01-04 12:06 ` Peter Zijlstra @ 2011-01-04 14:22 ` Borislav Petkov 2011-01-05 5:24 ` Lin Ming 0 siblings, 1 reply; 9+ messages in thread From: Borislav Petkov @ 2011-01-04 14:22 UTC (permalink / raw) To: Peter Zijlstra Cc: Lin Ming, Ingo Molnar, Andi Kleen, Stephane Eranian, Richter, Robert, lkml, Andreas Herrmann Adding Andreas since this is his code. On Tue, Jan 04, 2011 at 07:06:16AM -0500, Peter Zijlstra wrote: > On Mon, 2010-12-27 at 23:38 +0800, Lin Ming wrote: > > Currently we iterate the full machine looking for a matching core_id/nb > > for the percore and the amd northbridge stuff , using a smaller topology > > mask makes sense. > > > > Signed-off-by: Lin Ming <ming.m.lin@intel.com> > > --- > > arch/x86/kernel/cpu/perf_event_amd.c | 2 +- > > arch/x86/kernel/cpu/perf_event_intel.c | 2 +- > > 2 files changed, 2 insertions(+), 2 deletions(-) > > > > diff --git a/arch/x86/kernel/cpu/perf_event_amd.c b/arch/x86/kernel/cpu/perf_event_amd.c > > index 67e2202..5a3b7b8 100644 > > --- a/arch/x86/kernel/cpu/perf_event_amd.c > > +++ b/arch/x86/kernel/cpu/perf_event_amd.c > > @@ -323,7 +323,7 @@ static void amd_pmu_cpu_starting(int cpu) > > nb_id = amd_get_nb_id(cpu); > > WARN_ON_ONCE(nb_id == BAD_APICID); > > > > - for_each_online_cpu(i) { > > + for_each_cpu(i, topology_core_cpumask(cpu)) { > > nb = per_cpu(cpu_hw_events, i).amd_nb; > > if (WARN_ON_ONCE(!nb)) > > continue; > > Borislav, is topology_core_cpumask() the right mask for northbridge_id > span? I could imagine Magny-Cours would have all 12 cores in the > core_cpumask() and have the node_mask() be half that. So, topology_core_cpumask() or cpu_core_mask() both are cpu_core_map which represents the socket mask. I.e., on a multisocket cpu you'll have in it all the cores on one socket. A 12-cores Magny-Cours contains two internal northbridges and this mask will have 12 bits set. AFAICT, you want to iterate over the cores on a single node here (an internal node in the Magny-Cours case) so for this we have the llc_shared_map. See near the top of cache_shared_cpu_map_setup() in <arch/x86/kernel/cpu/intel_cacheinfo.c> for an example. node_mask() is roughly the same but contains correct info only with CONFIG_NUMA on and correct SRAT table on the system. HTH. -- Regards/Gruss, Boris. Advanced Micro Devices GmbH Einsteinring 24, 85609 Dornach General Managers: Alberto Bozzo, Andrew Bowd Registration: Dornach, Gemeinde Aschheim, Landkreis Muenchen Registergericht Muenchen, HRB Nr. 43632 ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 5/7] perf: Optimise topology iteration 2011-01-04 14:22 ` Borislav Petkov @ 2011-01-05 5:24 ` Lin Ming 2011-01-05 9:51 ` Peter Zijlstra 0 siblings, 1 reply; 9+ messages in thread From: Lin Ming @ 2011-01-05 5:24 UTC (permalink / raw) To: Borislav Petkov Cc: Peter Zijlstra, Ingo Molnar, Andi Kleen, Stephane Eranian, Richter, Robert, lkml, Andreas Herrmann On Tue, 2011-01-04 at 22:22 +0800, Borislav Petkov wrote: > Adding Andreas since this is his code. > > On Tue, Jan 04, 2011 at 07:06:16AM -0500, Peter Zijlstra wrote: > > On Mon, 2010-12-27 at 23:38 +0800, Lin Ming wrote: > > > Currently we iterate the full machine looking for a matching core_id/nb > > > for the percore and the amd northbridge stuff , using a smaller topology > > > mask makes sense. > > > > > > Signed-off-by: Lin Ming <ming.m.lin@intel.com> > > > --- > > > arch/x86/kernel/cpu/perf_event_amd.c | 2 +- > > > arch/x86/kernel/cpu/perf_event_intel.c | 2 +- > > > 2 files changed, 2 insertions(+), 2 deletions(-) > > > > > > diff --git a/arch/x86/kernel/cpu/perf_event_amd.c b/arch/x86/kernel/cpu/perf_event_amd.c > > > index 67e2202..5a3b7b8 100644 > > > --- a/arch/x86/kernel/cpu/perf_event_amd.c > > > +++ b/arch/x86/kernel/cpu/perf_event_amd.c > > > @@ -323,7 +323,7 @@ static void amd_pmu_cpu_starting(int cpu) > > > nb_id = amd_get_nb_id(cpu); > > > WARN_ON_ONCE(nb_id == BAD_APICID); > > > > > > - for_each_online_cpu(i) { > > > + for_each_cpu(i, topology_core_cpumask(cpu)) { > > > nb = per_cpu(cpu_hw_events, i).amd_nb; > > > if (WARN_ON_ONCE(!nb)) > > > continue; > > > > Borislav, is topology_core_cpumask() the right mask for northbridge_id > > span? I could imagine Magny-Cours would have all 12 cores in the > > core_cpumask() and have the node_mask() be half that. > > So, topology_core_cpumask() or cpu_core_mask() both are cpu_core_map > which represents the socket mask. I.e., on a multisocket cpu you'll have > in it all the cores on one socket. A 12-cores Magny-Cours contains two > internal northbridges and this mask will have 12 bits set. > > AFAICT, you want to iterate over the cores on a single node here > (an internal node in the Magny-Cours case) so for this we have the > llc_shared_map. See near the top of cache_shared_cpu_map_setup() in > <arch/x86/kernel/cpu/intel_cacheinfo.c> for an example. cpu_coregroup_mask() seems the right mask for northbridge_id span. arch/x86/kernel/smpboot.c: /* maps the cpu to the sched domain representing multi-core */ const struct cpumask *cpu_coregroup_mask(int cpu) { struct cpuinfo_x86 *c = &cpu_data(cpu); /* * For perf, we return last level cache shared map. * And for power savings, we return cpu_core_map */ if ((sched_mc_power_savings || sched_smt_power_savings) && !(cpu_has(c, X86_FEATURE_AMD_DCM))) return cpu_core_mask(cpu); else return c->llc_shared_map; } > > node_mask() is roughly the same but contains correct info only with > CONFIG_NUMA on and correct SRAT table on the system. > > HTH. > ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 5/7] perf: Optimise topology iteration 2011-01-05 5:24 ` Lin Ming @ 2011-01-05 9:51 ` Peter Zijlstra 0 siblings, 0 replies; 9+ messages in thread From: Peter Zijlstra @ 2011-01-05 9:51 UTC (permalink / raw) To: Lin Ming Cc: Borislav Petkov, Ingo Molnar, Andi Kleen, Stephane Eranian, Richter, Robert, lkml, Andreas Herrmann On Wed, 2011-01-05 at 13:24 +0800, Lin Ming wrote: > > > > + for_each_cpu(i, topology_core_cpumask(cpu)) { > > > > nb = per_cpu(cpu_hw_events, i).amd_nb; > > > > if (WARN_ON_ONCE(!nb)) > > > > continue; > > > > > > Borislav, is topology_core_cpumask() the right mask for northbridge_id > > > span? I could imagine Magny-Cours would have all 12 cores in the > > > core_cpumask() and have the node_mask() be half that. > > > > So, topology_core_cpumask() or cpu_core_mask() both are cpu_core_map > > which represents the socket mask. I.e., on a multisocket cpu you'll have > > in it all the cores on one socket. A 12-cores Magny-Cours contains two > > internal northbridges and this mask will have 12 bits set. > > > > AFAICT, you want to iterate over the cores on a single node here > > (an internal node in the Magny-Cours case) so for this we have the > > llc_shared_map. See near the top of cache_shared_cpu_map_setup() in > > <arch/x86/kernel/cpu/intel_cacheinfo.c> for an example. > > cpu_coregroup_mask() seems the right mask for northbridge_id span. > > arch/x86/kernel/smpboot.c: > > /* maps the cpu to the sched domain representing multi-core */ > const struct cpumask *cpu_coregroup_mask(int cpu) > { > struct cpuinfo_x86 *c = &cpu_data(cpu); > /* > * For perf, we return last level cache shared map. > * And for power savings, we return cpu_core_map > */ > if ((sched_mc_power_savings || sched_smt_power_savings) && > !(cpu_has(c, X86_FEATURE_AMD_DCM))) > return cpu_core_mask(cpu); > else > return c->llc_shared_map; > } Argh, that function really must die, its the most horrible brain damage around. Andreas promised he'd clean that up after making it worse for Magny-Cours. But yes, assuming all Magny-Cours have this AMD_DCM thing set, it seems to return the right map. ^ permalink raw reply [flat|nested] 9+ messages in thread
end of thread, other threads:[~2011-01-05 9:50 UTC | newest] Thread overview: 9+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2010-12-27 15:38 [PATCH 5/7] perf: Optimise topology iteration Lin Ming 2011-01-03 11:02 ` Peter Zijlstra 2011-01-03 15:20 ` Andi Kleen 2011-01-04 7:13 ` Lin Ming 2011-01-04 6:18 ` Lin Ming 2011-01-04 12:06 ` Peter Zijlstra 2011-01-04 14:22 ` Borislav Petkov 2011-01-05 5:24 ` Lin Ming 2011-01-05 9:51 ` Peter Zijlstra
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox