From: Yinghai Lu <yinghai@kernel.org>
To: Thomas Gleixner <tglx@linutronix.de>, Ingo Molnar <mingo@elte.hu>,
"H. Peter Anvin" <hpa@zytor.com>,
Andrew Morton <akpm@linux-foundation.org>,
Tejun Heo <tj@kernel.org>, Thomas Renninger <trenn@suse.de>,
Tang Chen <tangchen@cn.fujitsu.com>
Cc: linux-kernel@vger.kernel.org, Yinghai Lu <yinghai@kernel.org>
Subject: [PATCH v2 09/20] x86, mm, numa: Move two functions calling on successful path later
Date: Sat, 9 Mar 2013 22:44:36 -0800 [thread overview]
Message-ID: <1362897887-30808-10-git-send-email-yinghai@kernel.org> (raw)
In-Reply-To: <1362897887-30808-1-git-send-email-yinghai@kernel.org>
We need to have numa info ready before init_mem_mapping, so we
can call init_mem_mapping per nodes also can trim node mem range to
big alignment.
Current numa parsing need to allocate some buffer and need to be
called after init_mem_mapping.
So try to split parsing numa info to two stages, and early one will be
before init_mem_mapping, and it should not need allocate buffers.
At last we will have early_initmem_init() and initmem_init().
This one is first one for separation.
setup_node_data() and numa_init_array() are only called for successful
path, so we can move calling to x86_numa_init(). That will also make
numa_init() small and readable.
-v2: remove online_node_map clear in numa_init(), as it is only
set in setup_node_data() at last in successful path.
Signed-off-by: Yinghai Lu <yinghai@kernel.org>
---
arch/x86/mm/numa.c | 69 +++++++++++++++++++++++++++++-----------------------
1 file changed, 39 insertions(+), 30 deletions(-)
diff --git a/arch/x86/mm/numa.c b/arch/x86/mm/numa.c
index 72fe01e..d545638 100644
--- a/arch/x86/mm/numa.c
+++ b/arch/x86/mm/numa.c
@@ -480,7 +480,7 @@ static bool __init numa_meminfo_cover_memory(const struct numa_meminfo *mi)
static int __init numa_register_memblks(struct numa_meminfo *mi)
{
unsigned long uninitialized_var(pfn_align);
- int i, nid;
+ int i;
/* Account for nodes with cpus and no memory */
node_possible_map = numa_nodes_parsed;
@@ -509,24 +509,6 @@ static int __init numa_register_memblks(struct numa_meminfo *mi)
if (!numa_meminfo_cover_memory(mi))
return -EINVAL;
- /* Finally register nodes. */
- for_each_node_mask(nid, node_possible_map) {
- u64 start = PFN_PHYS(max_pfn);
- u64 end = 0;
-
- for (i = 0; i < mi->nr_blks; i++) {
- if (nid != mi->blk[i].nid)
- continue;
- start = min(mi->blk[i].start, start);
- end = max(mi->blk[i].end, end);
- }
-
- if (start < end)
- setup_node_data(nid, start, end);
- }
-
- /* Dump memblock with node info and return. */
- memblock_dump_all();
return 0;
}
@@ -562,7 +544,6 @@ static int __init numa_init(int (*init_func)(void))
nodes_clear(numa_nodes_parsed);
nodes_clear(node_possible_map);
- nodes_clear(node_online_map);
memset(&numa_meminfo, 0, sizeof(numa_meminfo));
WARN_ON(memblock_set_node(0, ULLONG_MAX, MAX_NUMNODES));
numa_reset_distance();
@@ -580,15 +561,6 @@ static int __init numa_init(int (*init_func)(void))
if (ret < 0)
return ret;
- for (i = 0; i < nr_cpu_ids; i++) {
- int nid = early_cpu_to_node(i);
-
- if (nid == NUMA_NO_NODE)
- continue;
- if (!node_online(nid))
- numa_clear_node(i);
- }
- numa_init_array();
return 0;
}
@@ -621,7 +593,7 @@ static int __init dummy_numa_init(void)
* last fallback is dummy single node config encomapssing whole memory and
* never fails.
*/
-void __init x86_numa_init(void)
+static void __init early_x86_numa_init(void)
{
if (!numa_off) {
#ifdef CONFIG_X86_NUMAQ
@@ -641,6 +613,43 @@ void __init x86_numa_init(void)
numa_init(dummy_numa_init);
}
+void __init x86_numa_init(void)
+{
+ int i, nid;
+ struct numa_meminfo *mi = &numa_meminfo;
+
+ early_x86_numa_init();
+
+ /* Finally register nodes. */
+ for_each_node_mask(nid, node_possible_map) {
+ u64 start = PFN_PHYS(max_pfn);
+ u64 end = 0;
+
+ for (i = 0; i < mi->nr_blks; i++) {
+ if (nid != mi->blk[i].nid)
+ continue;
+ start = min(mi->blk[i].start, start);
+ end = max(mi->blk[i].end, end);
+ }
+
+ if (start < end)
+ setup_node_data(nid, start, end); /* online is set */
+ }
+
+ /* Dump memblock with node info */
+ memblock_dump_all();
+
+ for (i = 0; i < nr_cpu_ids; i++) {
+ int nid = early_cpu_to_node(i);
+
+ if (nid == NUMA_NO_NODE)
+ continue;
+ if (!node_online(nid))
+ numa_clear_node(i);
+ }
+ numa_init_array();
+}
+
static __init int find_near_online_node(int node)
{
int n, val;
--
1.7.10.4
next prev parent reply other threads:[~2013-03-10 6:46 UTC|newest]
Thread overview: 43+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-03-10 6:44 [PATCH v2 00/20] x86, ACPI, numa: Parse numa info early Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 01/20] x86: Change get_ramdisk_image() to global Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 02/20] x86, microcode: Use common get_ramdisk_image() Yinghai Lu
2013-04-04 17:48 ` Tejun Heo
2013-04-04 17:59 ` Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 03/20] x86, ACPI, mm: Kill max_low_pfn_mapped Yinghai Lu
2013-04-04 17:36 ` Tejun Heo
2013-04-04 18:20 ` Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 04/20] x86, ACPI: Increase override tables number limit Yinghai Lu
2013-04-04 17:50 ` Tejun Heo
2013-04-04 18:03 ` Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 05/20] x86, ACPI: Split acpi_initrd_override to find/copy two functions Yinghai Lu
2013-04-04 18:07 ` Tejun Heo
2013-04-04 19:29 ` Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 06/20] x86, ACPI: Store override acpi tables phys addr in cpio files info array Yinghai Lu
2013-04-04 18:27 ` Tejun Heo
2013-04-04 18:30 ` Tejun Heo
2013-04-04 19:40 ` Yinghai Lu
2013-04-04 20:03 ` Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 07/20] x86, ACPI: Make acpi_initrd_override_find work with 32bit flat mode Yinghai Lu
2013-04-04 18:35 ` Tejun Heo
2013-04-04 20:22 ` Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 08/20] x86, ACPI: Find acpi tables in initrd early from head_32.S/head64.c Yinghai Lu
2013-03-10 10:25 ` Pekka Enberg
2013-03-10 16:47 ` Yinghai Lu
2013-03-10 17:42 ` H. Peter Anvin
2013-04-04 20:25 ` H. Peter Anvin
2013-03-10 6:44 ` Yinghai Lu [this message]
2013-03-10 6:44 ` [PATCH v2 10/20] x86, mm, numa: Call numa_meminfo_cover_memory() checking early Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 11/20] x86, mm, numa: Move node_map_pfn alignment() to x86 Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 12/20] x86, mm, numa: Use numa_meminfo to check node_map_pfn alignment Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 13/20] x86, mm, numa: Set memblock nid later Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 14/20] x86, mm, numa: Move node_possible_map setting later Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 15/20] x86, mm, numa: Move emulation handling down Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 16/20] x86, ACPI, numa, ia64: split SLIT handling out Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 17/20] x86, mm, numa: Add early_initmem_init() stub Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 18/20] x86, mm: Parse numa info early Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 19/20] x86, mm: Make init_mem_mapping be able to be called several times Yinghai Lu
2013-03-11 13:16 ` Konrad Rzeszutek Wilk
2013-03-11 20:28 ` Yinghai Lu
2013-03-10 6:44 ` [PATCH v2 20/20] x86, mm, numa: Put pagetable on local node ram for 64bit Yinghai Lu
2013-03-11 5:49 ` Tang Chen
2013-03-11 6:29 ` Yinghai Lu
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1362897887-30808-10-git-send-email-yinghai@kernel.org \
--to=yinghai@kernel.org \
--cc=akpm@linux-foundation.org \
--cc=hpa@zytor.com \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@elte.hu \
--cc=tangchen@cn.fujitsu.com \
--cc=tglx@linutronix.de \
--cc=tj@kernel.org \
--cc=trenn@suse.de \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox