linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes
       [not found] <cover.1660897864.git.remckee0@gmail.com>
@ 2022-08-19  9:05 ` Rebecca Mckeever
  2022-08-30 11:17   ` David Hildenbrand
  2022-08-31 15:15   ` Mike Rapoport
       [not found] ` <957966f06474e3885796247ad1beaa6b3841ebd1.1660897864.git.remckee0@gmail.com>
  1 sibling, 2 replies; 11+ messages in thread
From: Rebecca Mckeever @ 2022-08-19  9:05 UTC (permalink / raw)
  To: Mike Rapoport, linux-mm, linux-kernel; +Cc: David Hildenbrand, Rebecca Mckeever

Add functions setup_numa_memblock_generic() and setup_numa_memblock()
for setting up a memory layout with multiple NUMA nodes in a previously
allocated dummy physical memory. These functions can be used in place of
setup_memblock() in tests that need to simulate a NUMA system.

setup_numa_memblock_generic():
- allows for setting up a custom memory layout by specifying the amount
  of memory in each node, the number of nodes, and a factor that will be
  used to scale the memory in each node

setup_numa_memblock():
- allows for setting up a default memory layout

Introduce constant MEM_FACTOR, which is used to scale the default memory
layout based on MEM_SIZE.

Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
16 NUMA nodes.

Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
---
 .../testing/memblock/scripts/Makefile.include |  2 +-
 tools/testing/memblock/tests/common.c         | 38 +++++++++++++++++++
 tools/testing/memblock/tests/common.h         |  9 ++++-
 3 files changed, 47 insertions(+), 2 deletions(-)

diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
index aa6d82d56a23..998281723590 100644
--- a/tools/testing/memblock/scripts/Makefile.include
+++ b/tools/testing/memblock/scripts/Makefile.include
@@ -3,7 +3,7 @@
 
 # Simulate CONFIG_NUMA=y
 ifeq ($(NUMA), 1)
-	CFLAGS += -D CONFIG_NUMA
+	CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
 endif
 
 # Use 32 bit physical addresses.
diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
index eec6901081af..15d8767dc70c 100644
--- a/tools/testing/memblock/tests/common.c
+++ b/tools/testing/memblock/tests/common.c
@@ -34,6 +34,10 @@ static const char * const help_opts[] = {
 
 static int verbose;
 
+static const phys_addr_t node_sizes[] = {
+	SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
+};
+
 /* sets global variable returned by movable_node_is_enabled() stub */
 bool movable_node_enabled;
 
@@ -72,6 +76,40 @@ void setup_memblock(void)
 	fill_memblock();
 }
 
+/**
+ * setup_numa_memblock_generic:
+ * Set up a memory layout with multiple NUMA nodes in a previously allocated
+ * dummy physical memory.
+ * @nodes: an array containing the amount of memory in each node
+ * @node_cnt: the size of @nodes
+ * @factor: a factor that will be used to scale the memory in each node
+ *
+ * The nids will be set to 0 through node_cnt - 1.
+ */
+void setup_numa_memblock_generic(const phys_addr_t nodes[],
+				 int node_cnt, int factor)
+{
+	phys_addr_t base;
+	int flags;
+
+	reset_memblock_regions();
+	base = (phys_addr_t)memory_block.base;
+	flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
+
+	for (int i = 0; i < node_cnt; i++) {
+		phys_addr_t size = factor * nodes[i];
+
+		memblock_add_node(base, size, i, flags);
+		base += size;
+	}
+	fill_memblock();
+}
+
+void setup_numa_memblock(void)
+{
+	setup_numa_memblock_generic(node_sizes, NUMA_NODES, MEM_FACTOR);
+}
+
 void dummy_physical_memory_init(void)
 {
 	memory_block.base = malloc(MEM_SIZE);
diff --git a/tools/testing/memblock/tests/common.h b/tools/testing/memblock/tests/common.h
index 4fd3534ff955..e5117d959d6c 100644
--- a/tools/testing/memblock/tests/common.h
+++ b/tools/testing/memblock/tests/common.h
@@ -10,7 +10,11 @@
 #include <linux/printk.h>
 #include <../selftests/kselftest.h>
 
-#define MEM_SIZE SZ_16K
+#define MEM_SIZE		SZ_16K
+#define NUMA_NODES		8
+
+/* used to resize values that need to scale with MEM_SIZE */
+#define MEM_FACTOR		(MEM_SIZE / SZ_16K)
 
 enum test_flags {
 	TEST_ZEROED = 0x0,
@@ -100,6 +104,9 @@ struct region {
 void reset_memblock_regions(void);
 void reset_memblock_attributes(void);
 void setup_memblock(void);
+void setup_numa_memblock_generic(const phys_addr_t nodes[],
+				 int node_cnt, int factor);
+void setup_numa_memblock(void);
 void dummy_physical_memory_init(void);
 void dummy_physical_memory_cleanup(void);
 void parse_args(int argc, char **argv);
-- 
2.25.1



^ permalink raw reply related	[flat|nested] 11+ messages in thread

* Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes
  2022-08-19  9:05 ` [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes Rebecca Mckeever
@ 2022-08-30 11:17   ` David Hildenbrand
  2022-08-31  3:49     ` Rebecca Mckeever
  2022-08-31 15:15   ` Mike Rapoport
  1 sibling, 1 reply; 11+ messages in thread
From: David Hildenbrand @ 2022-08-30 11:17 UTC (permalink / raw)
  To: Rebecca Mckeever, Mike Rapoport, linux-mm, linux-kernel

On 19.08.22 11:05, Rebecca Mckeever wrote:
> Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> for setting up a memory layout with multiple NUMA nodes in a previously
> allocated dummy physical memory. These functions can be used in place of
> setup_memblock() in tests that need to simulate a NUMA system.
> 
> setup_numa_memblock_generic():
> - allows for setting up a custom memory layout by specifying the amount
>   of memory in each node, the number of nodes, and a factor that will be
>   used to scale the memory in each node
> 
> setup_numa_memblock():
> - allows for setting up a default memory layout
> 
> Introduce constant MEM_FACTOR, which is used to scale the default memory
> layout based on MEM_SIZE.
> 
> Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> 16 NUMA nodes.
> 
> Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
> ---
>  .../testing/memblock/scripts/Makefile.include |  2 +-
>  tools/testing/memblock/tests/common.c         | 38 +++++++++++++++++++
>  tools/testing/memblock/tests/common.h         |  9 ++++-
>  3 files changed, 47 insertions(+), 2 deletions(-)
> 
> diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
> index aa6d82d56a23..998281723590 100644
> --- a/tools/testing/memblock/scripts/Makefile.include
> +++ b/tools/testing/memblock/scripts/Makefile.include
> @@ -3,7 +3,7 @@
>  
>  # Simulate CONFIG_NUMA=y
>  ifeq ($(NUMA), 1)
> -	CFLAGS += -D CONFIG_NUMA
> +	CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
>  endif
>  
>  # Use 32 bit physical addresses.
> diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
> index eec6901081af..15d8767dc70c 100644
> --- a/tools/testing/memblock/tests/common.c
> +++ b/tools/testing/memblock/tests/common.c
> @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
>  
>  static int verbose;
>  
> +static const phys_addr_t node_sizes[] = {
> +	SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
> +};
> +
>  /* sets global variable returned by movable_node_is_enabled() stub */
>  bool movable_node_enabled;
>  
> @@ -72,6 +76,40 @@ void setup_memblock(void)
>  	fill_memblock();
>  }
>  
> +/**
> + * setup_numa_memblock_generic:
> + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> + * dummy physical memory.
> + * @nodes: an array containing the amount of memory in each node
> + * @node_cnt: the size of @nodes
> + * @factor: a factor that will be used to scale the memory in each node
> + *
> + * The nids will be set to 0 through node_cnt - 1.
> + */
> +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> +				 int node_cnt, int factor)
> +{
> +	phys_addr_t base;
> +	int flags;
> +
> +	reset_memblock_regions();
> +	base = (phys_addr_t)memory_block.base;
> +	flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> +
> +	for (int i = 0; i < node_cnt; i++) {
> +		phys_addr_t size = factor * nodes[i];

I'm a bit lost why we need the factor if we already provide sizes in the
array.

Can you enlighten me? :)

Why can't we just stick to the sizes in the array?

-- 
Thanks,

David / dhildenb



^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: [PATCH v2 2/4] memblock tests: add top-down NUMA tests for memblock_alloc_try_nid*
       [not found] ` <957966f06474e3885796247ad1beaa6b3841ebd1.1660897864.git.remckee0@gmail.com>
@ 2022-08-30 11:56   ` David Hildenbrand
  2022-09-02  0:37     ` Rebecca Mckeever
  0 siblings, 1 reply; 11+ messages in thread
From: David Hildenbrand @ 2022-08-30 11:56 UTC (permalink / raw)
  To: Rebecca Mckeever, Mike Rapoport, linux-mm, linux-kernel

On 19.08.22 11:05, Rebecca Mckeever wrote:
> Add tests for memblock_alloc_try_nid() and memblock_alloc_try_nid_raw()
> where the simulated physical memory is set up with multiple NUMA nodes.
> Additionally, all of these tests set nid != NUMA_NO_NODE. These tests are
> run with a top-down allocation direction.
> 
> The tested scenarios are:
> 
> Range unrestricted:
> - region can be allocated in the specific node requested:
>       + there are no previously reserved regions
>       + the requested node is partially reserved but has enough space
> - the specific node requested cannot accommodate the request, but the
>   region can be allocated in a different node:
>       + there are no previously reserved regions, but node is too small
>       + the requested node is fully reserved
>       + the requested node is partially reserved and does not have
>         enough space
> 
> Range restricted:
> - region can be allocated in the specific node requested after dropping
>   min_addr:
>       + range partially overlaps with two different nodes, where the first
>         node is the requested node
>       + range partially overlaps with two different nodes, where the
>         requested node ends before min_addr
> - region cannot be allocated in the specific node requested, but it can be
>   allocated in the requested range:
>       + range overlaps with multiple nodes along node boundaries, and the
>         requested node ends before min_addr
>       + range overlaps with multiple nodes along node boundaries, and the
>         requested node starts after max_addr
> - region cannot be allocated in the specific node requested, but it can be
>   allocated after dropping min_addr:
>       + range partially overlaps with two different nodes, where the
>         second node is the requested node
> 
> Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
> ---
>  tools/testing/memblock/tests/alloc_nid_api.c | 702 ++++++++++++++++++-
>  tools/testing/memblock/tests/alloc_nid_api.h |  16 +
>  tools/testing/memblock/tests/common.h        |  18 +
>  3 files changed, 725 insertions(+), 11 deletions(-)
> 
> diff --git a/tools/testing/memblock/tests/alloc_nid_api.c b/tools/testing/memblock/tests/alloc_nid_api.c
> index 2c1d5035e057..a410f1318402 100644
> --- a/tools/testing/memblock/tests/alloc_nid_api.c
> +++ b/tools/testing/memblock/tests/alloc_nid_api.c
> @@ -1102,7 +1102,7 @@ static int alloc_try_nid_bottom_up_cap_min_check(void)
>  	return 0;
>  }
>  
> -/* Test case wrappers */
> +/* Test case wrappers for range tests */
>  static int alloc_try_nid_simple_check(void)
>  {
>  	test_print("\tRunning %s...\n", __func__);
> @@ -1234,17 +1234,10 @@ static int alloc_try_nid_low_max_check(void)
>  	return 0;
>  }
>  
> -static int memblock_alloc_nid_checks_internal(int flags)
> +static int memblock_alloc_nid_range_checks(void)
>  {
> -	const char *func = get_func_testing(flags);
> -
> -	alloc_nid_test_flags = flags;
> -	prefix_reset();
> -	prefix_push(func);
> -	test_print("Running %s tests...\n", func);
> -
> -	reset_memblock_attributes();
> -	dummy_physical_memory_init();
> +	test_print("Running %s range tests...\n",
> +		   get_func_testing(alloc_nid_test_flags));
>  
>  	alloc_try_nid_simple_check();
>  	alloc_try_nid_misaligned_check();
> @@ -1261,6 +1254,693 @@ static int memblock_alloc_nid_checks_internal(int flags)
>  	alloc_try_nid_reserved_all_check();
>  	alloc_try_nid_low_max_check();
>  
> +	return 0;
> +}
> +
> +/*
> + * A test that tries to allocate a memory region in a specific NUMA node that
> + * has enough memory to allocate a region of the requested size.
> + * Expect to allocate an aligned region at the end of the requested node.
> + */
> +static int alloc_try_nid_top_down_numa_simple_check(void)
> +{
> +	int nid_req = 3;
> +	struct memblock_region *new_rgn = &memblock.reserved.regions[0];
> +	struct memblock_region *req_node = &memblock.memory.regions[nid_req];
> +	void *allocated_ptr = NULL;
> +
> +	PREFIX_PUSH();
> +
> +	phys_addr_t size;
> +	phys_addr_t min_addr;
> +	phys_addr_t max_addr;

Usually we define variables in a single block. So, before the
PREFIX_PUSH(). Same applies to the other functions.

> +
> +	setup_numa_memblock();
> +
> +	ASSERT_LE(SZ_4, req_node->size);
> +	size = req_node->size / SZ_4;
> +	min_addr = memblock_start_of_DRAM();
> +	max_addr = memblock_end_of_DRAM();
> +
> +	allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
> +						   min_addr, max_addr, nid_req);
> +
> +	ASSERT_NE(allocated_ptr, NULL);
> +	verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
> +
> +	ASSERT_EQ(new_rgn->size, size);
> +	ASSERT_EQ(new_rgn->base, region_end(req_node) - size);
> +	ASSERT_LE(req_node->base, new_rgn->base);
> +
> +	ASSERT_EQ(memblock.reserved.cnt, 1);
> +	ASSERT_EQ(memblock.reserved.total_size, size);
> +
> +	test_pass_pop();
> +
> +	return 0;
> +}
> +

[...]

> +
> +/*
> + * A test that tries to allocate a memory region that spans over the min_addr
> + * and max_addr range and overlaps with two different nodes, where the first
> + * node is the requested node:
> + *
> + *                                min_addr
> + *                                |           max_addr
> + *                                |           |
> + *                                v           v
> + *  |           +-----------------------+-----------+              |
> + *  |           |       requested       |   node3   |              |
> + *  +-----------+-----------------------+-----------+--------------+
> + *                                +           +
> + *  |                       +-----------+                          |
> + *  |                       |    rgn    |                          |
> + *  +-----------------------+-----------+--------------------------+
> + *
> + * Expect to drop the lower limit and allocate a cleared memory region that
> + * ends at the end of the requested node.

Interesting, allocating out-of-range is expected behavior? At least to
me that wasn't immediately clear :)

[...]

> +
> +/*
> + * A test that tries to allocate a memory region that spans over the min_addr
> + * and max_addr range and overlaps with two different nodes, where the second
> + * node is the requested node:
> + *
> + *                               min_addr
> + *                               |         max_addr
> + *                               |         |
> + *                               v         v
> + *  |      +--------------------------+---------+                |
> + *  |      |         expected         |requested|                |
> + *  +------+--------------------------+---------+----------------+
> + *                               +         +
> + *  |                       +---------+                          |
> + *  |                       |   rgn   |                          |
> + *  +-----------------------+---------+--------------------------+
> + *
> + * Expect to drop the lower limit and allocate a cleared memory region that

Does the "cleared memory region" part still apply? Or would we also end
up calling the raw variant from run_memblock_alloc_try_nid() ?

> + * ends at the end of the first node that overlaps with the range.
> + */
> +static int alloc_try_nid_top_down_numa_split_range_high_check(void)
> +{
> +	int nid_req = 3;
> +	int nid_exp = nid_req - 1;
> +	struct memblock_region *new_rgn = &memblock.reserved.regions[0];
> +	struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
> +	void *allocated_ptr = NULL;
> +
> +	PREFIX_PUSH();
> +
> +	phys_addr_t size = SZ_512;
> +	phys_addr_t min_addr;
> +	phys_addr_t max_addr;
> +	phys_addr_t exp_node_end;
> +
> +	setup_numa_memblock();
> +
> +	exp_node_end = region_end(exp_node);
> +	min_addr = exp_node_end - SZ_256;
> +	max_addr = min_addr + size;
> +
> +	allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
> +						   min_addr, max_addr, nid_req);
> +
> +	ASSERT_NE(allocated_ptr, NULL);
> +	verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
> +
> +	ASSERT_EQ(new_rgn->size, size);
> +	ASSERT_EQ(new_rgn->base, exp_node_end - size);
> +	ASSERT_LE(exp_node->base, new_rgn->base);
> +
> +	ASSERT_EQ(memblock.reserved.cnt, 1);
> +	ASSERT_EQ(memblock.reserved.total_size, size);
> +
> +	test_pass_pop();
> +
> +	return 0;
> +}


[...]

> +int __memblock_alloc_nid_numa_checks(void)
> +{
> +	test_print("Running %s NUMA tests...\n",
> +		   get_func_testing(alloc_nid_test_flags));
> +
> +	alloc_try_nid_numa_simple_check();
> +	alloc_try_nid_numa_small_node_check();
> +	alloc_try_nid_numa_node_reserved_check();
> +	alloc_try_nid_numa_part_reserved_check();
> +	alloc_try_nid_numa_part_reserved_fallback_check();
> +	alloc_try_nid_numa_split_range_low_check();
> +	alloc_try_nid_numa_split_range_high_check();
> +
> +	alloc_try_nid_numa_no_overlap_split_check();
> +	alloc_try_nid_numa_no_overlap_low_check();
> +	alloc_try_nid_numa_no_overlap_high_check();
> +
> +	return 0;
> +}
> +
> +static int memblock_alloc_nid_checks_internal(int flags)
> +{
> +	alloc_nid_test_flags = flags;

Empty line missing

> +	prefix_reset();
> +	prefix_push(get_func_testing(flags));
> +
> +	reset_memblock_attributes();
> +	dummy_physical_memory_init();
> +
> +	memblock_alloc_nid_range_checks();
> +	memblock_alloc_nid_numa_checks();
> +
>  	dummy_physical_memory_cleanup();
>  
>  	prefix_pop();


-- 
Thanks,

David / dhildenb



^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes
  2022-08-30 11:17   ` David Hildenbrand
@ 2022-08-31  3:49     ` Rebecca Mckeever
  2022-08-31 15:12       ` Mike Rapoport
  2022-09-01  8:06       ` David Hildenbrand
  0 siblings, 2 replies; 11+ messages in thread
From: Rebecca Mckeever @ 2022-08-31  3:49 UTC (permalink / raw)
  To: David Hildenbrand; +Cc: Mike Rapoport, linux-mm, linux-kernel

On Tue, Aug 30, 2022 at 01:17:56PM +0200, David Hildenbrand wrote:
> On 19.08.22 11:05, Rebecca Mckeever wrote:
> > Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> > for setting up a memory layout with multiple NUMA nodes in a previously
> > allocated dummy physical memory. These functions can be used in place of
> > setup_memblock() in tests that need to simulate a NUMA system.
> > 
> > setup_numa_memblock_generic():
> > - allows for setting up a custom memory layout by specifying the amount
> >   of memory in each node, the number of nodes, and a factor that will be
> >   used to scale the memory in each node
> > 
> > setup_numa_memblock():
> > - allows for setting up a default memory layout
> > 
> > Introduce constant MEM_FACTOR, which is used to scale the default memory
> > layout based on MEM_SIZE.
> > 
> > Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> > 16 NUMA nodes.
> > 
> > Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
> > ---
> >  .../testing/memblock/scripts/Makefile.include |  2 +-
> >  tools/testing/memblock/tests/common.c         | 38 +++++++++++++++++++
> >  tools/testing/memblock/tests/common.h         |  9 ++++-
> >  3 files changed, 47 insertions(+), 2 deletions(-)
> > 
> > diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
> > index aa6d82d56a23..998281723590 100644
> > --- a/tools/testing/memblock/scripts/Makefile.include
> > +++ b/tools/testing/memblock/scripts/Makefile.include
> > @@ -3,7 +3,7 @@
> >  
> >  # Simulate CONFIG_NUMA=y
> >  ifeq ($(NUMA), 1)
> > -	CFLAGS += -D CONFIG_NUMA
> > +	CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
> >  endif
> >  
> >  # Use 32 bit physical addresses.
> > diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
> > index eec6901081af..15d8767dc70c 100644
> > --- a/tools/testing/memblock/tests/common.c
> > +++ b/tools/testing/memblock/tests/common.c
> > @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
> >  
> >  static int verbose;
> >  
> > +static const phys_addr_t node_sizes[] = {
> > +	SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
> > +};
> > +
> >  /* sets global variable returned by movable_node_is_enabled() stub */
> >  bool movable_node_enabled;
> >  
> > @@ -72,6 +76,40 @@ void setup_memblock(void)
> >  	fill_memblock();
> >  }
> >  
> > +/**
> > + * setup_numa_memblock_generic:
> > + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> > + * dummy physical memory.
> > + * @nodes: an array containing the amount of memory in each node
> > + * @node_cnt: the size of @nodes
> > + * @factor: a factor that will be used to scale the memory in each node
> > + *
> > + * The nids will be set to 0 through node_cnt - 1.
> > + */
> > +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> > +				 int node_cnt, int factor)
> > +{
> > +	phys_addr_t base;
> > +	int flags;
> > +
> > +	reset_memblock_regions();
> > +	base = (phys_addr_t)memory_block.base;
> > +	flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> > +
> > +	for (int i = 0; i < node_cnt; i++) {
> > +		phys_addr_t size = factor * nodes[i];
> 
> I'm a bit lost why we need the factor if we already provide sizes in the
> array.
> 
> Can you enlighten me? :)
> 
> Why can't we just stick to the sizes in the array?
> 
Without the factor, some of the tests will break if we increase MEM_SIZE
in the future (which we may need to do). I could rewrite them so that the
factor is not needed, but I thought the code would be over-complicated if
I did.

> -- 
> Thanks,
> 
> David / dhildenb
> 
Thanks,
Rebecca


^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes
  2022-08-31  3:49     ` Rebecca Mckeever
@ 2022-08-31 15:12       ` Mike Rapoport
  2022-09-01 22:53         ` Rebecca Mckeever
  2022-09-01  8:06       ` David Hildenbrand
  1 sibling, 1 reply; 11+ messages in thread
From: Mike Rapoport @ 2022-08-31 15:12 UTC (permalink / raw)
  To: Rebecca Mckeever; +Cc: David Hildenbrand, linux-mm, linux-kernel

On Tue, Aug 30, 2022 at 10:49:09PM -0500, Rebecca Mckeever wrote:
> On Tue, Aug 30, 2022 at 01:17:56PM +0200, David Hildenbrand wrote:
> > On 19.08.22 11:05, Rebecca Mckeever wrote:
> > > Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> > > for setting up a memory layout with multiple NUMA nodes in a previously
> > > allocated dummy physical memory. These functions can be used in place of
> > > setup_memblock() in tests that need to simulate a NUMA system.
> > > 
> > > setup_numa_memblock_generic():
> > > - allows for setting up a custom memory layout by specifying the amount
> > >   of memory in each node, the number of nodes, and a factor that will be
> > >   used to scale the memory in each node
> > > 
> > > setup_numa_memblock():
> > > - allows for setting up a default memory layout
> > > 
> > > Introduce constant MEM_FACTOR, which is used to scale the default memory
> > > layout based on MEM_SIZE.
> > > 
> > > Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> > > 16 NUMA nodes.
> > > 
> > > Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
> > > ---
> > >  .../testing/memblock/scripts/Makefile.include |  2 +-
> > >  tools/testing/memblock/tests/common.c         | 38 +++++++++++++++++++
> > >  tools/testing/memblock/tests/common.h         |  9 ++++-
> > >  3 files changed, 47 insertions(+), 2 deletions(-)
> > > 
> > > diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
> > > index aa6d82d56a23..998281723590 100644
> > > --- a/tools/testing/memblock/scripts/Makefile.include
> > > +++ b/tools/testing/memblock/scripts/Makefile.include
> > > @@ -3,7 +3,7 @@
> > >  
> > >  # Simulate CONFIG_NUMA=y
> > >  ifeq ($(NUMA), 1)
> > > -	CFLAGS += -D CONFIG_NUMA
> > > +	CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
> > >  endif
> > >  
> > >  # Use 32 bit physical addresses.
> > > diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
> > > index eec6901081af..15d8767dc70c 100644
> > > --- a/tools/testing/memblock/tests/common.c
> > > +++ b/tools/testing/memblock/tests/common.c
> > > @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
> > >  
> > >  static int verbose;
> > >  
> > > +static const phys_addr_t node_sizes[] = {
> > > +	SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
> > > +};
> > > +
> > >  /* sets global variable returned by movable_node_is_enabled() stub */
> > >  bool movable_node_enabled;
> > >  
> > > @@ -72,6 +76,40 @@ void setup_memblock(void)
> > >  	fill_memblock();
> > >  }
> > >  
> > > +/**
> > > + * setup_numa_memblock_generic:
> > > + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> > > + * dummy physical memory.
> > > + * @nodes: an array containing the amount of memory in each node
> > > + * @node_cnt: the size of @nodes
> > > + * @factor: a factor that will be used to scale the memory in each node
> > > + *
> > > + * The nids will be set to 0 through node_cnt - 1.
> > > + */
> > > +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> > > +				 int node_cnt, int factor)
> > > +{
> > > +	phys_addr_t base;
> > > +	int flags;
> > > +
> > > +	reset_memblock_regions();
> > > +	base = (phys_addr_t)memory_block.base;
> > > +	flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> > > +
> > > +	for (int i = 0; i < node_cnt; i++) {
> > > +		phys_addr_t size = factor * nodes[i];
> > 
> > I'm a bit lost why we need the factor if we already provide sizes in the
> > array.
> > 
> > Can you enlighten me? :)
> > 
> > Why can't we just stick to the sizes in the array?
> > 
> Without the factor, some of the tests will break if we increase MEM_SIZE
> in the future (which we may need to do). I could rewrite them so that the
> factor is not needed, but I thought the code would be over-complicated if
> I did.

What if we make nodes[] to represent the fraction of the memory rather than
a node size? Then the factor won't be required.

> Thanks,
> Rebecca

-- 
Sincerely yours,
Mike.


^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes
  2022-08-19  9:05 ` [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes Rebecca Mckeever
  2022-08-30 11:17   ` David Hildenbrand
@ 2022-08-31 15:15   ` Mike Rapoport
  2022-09-02  0:14     ` Rebecca Mckeever
  1 sibling, 1 reply; 11+ messages in thread
From: Mike Rapoport @ 2022-08-31 15:15 UTC (permalink / raw)
  To: Rebecca Mckeever; +Cc: linux-mm, linux-kernel, David Hildenbrand

On Fri, Aug 19, 2022 at 02:05:31AM -0700, Rebecca Mckeever wrote:
> Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> for setting up a memory layout with multiple NUMA nodes in a previously
> allocated dummy physical memory. These functions can be used in place of
> setup_memblock() in tests that need to simulate a NUMA system.
> 
> setup_numa_memblock_generic():
> - allows for setting up a custom memory layout by specifying the amount
>   of memory in each node, the number of nodes, and a factor that will be
>   used to scale the memory in each node
> 
> setup_numa_memblock():
> - allows for setting up a default memory layout
> 
> Introduce constant MEM_FACTOR, which is used to scale the default memory
> layout based on MEM_SIZE.
> 
> Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> 16 NUMA nodes.
> 
> Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
> ---
>  .../testing/memblock/scripts/Makefile.include |  2 +-
>  tools/testing/memblock/tests/common.c         | 38 +++++++++++++++++++
>  tools/testing/memblock/tests/common.h         |  9 ++++-
>  3 files changed, 47 insertions(+), 2 deletions(-)

...
  
> +/**
> + * setup_numa_memblock_generic:
> + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> + * dummy physical memory.
> + * @nodes: an array containing the amount of memory in each node
> + * @node_cnt: the size of @nodes
> + * @factor: a factor that will be used to scale the memory in each node
> + *
> + * The nids will be set to 0 through node_cnt - 1.
> + */
> +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> +				 int node_cnt, int factor)

I only had time for a quick look and it seems this function is never used
on its own.
Let's fold it into setup_numa_memblock() for now.

> +{
> +	phys_addr_t base;
> +	int flags;
> +
> +	reset_memblock_regions();
> +	base = (phys_addr_t)memory_block.base;
> +	flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> +
> +	for (int i = 0; i < node_cnt; i++) {
> +		phys_addr_t size = factor * nodes[i];
> +
> +		memblock_add_node(base, size, i, flags);
> +		base += size;
> +	}
> +	fill_memblock();
> +}
> +
> +void setup_numa_memblock(void)
> +{
> +	setup_numa_memblock_generic(node_sizes, NUMA_NODES, MEM_FACTOR);
> +}
> +

-- 
Sincerely yours,
Mike.


^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes
  2022-08-31  3:49     ` Rebecca Mckeever
  2022-08-31 15:12       ` Mike Rapoport
@ 2022-09-01  8:06       ` David Hildenbrand
  2022-09-02  0:08         ` Rebecca Mckeever
  1 sibling, 1 reply; 11+ messages in thread
From: David Hildenbrand @ 2022-09-01  8:06 UTC (permalink / raw)
  To: Rebecca Mckeever; +Cc: Mike Rapoport, linux-mm, linux-kernel

On 31.08.22 05:49, Rebecca Mckeever wrote:
> On Tue, Aug 30, 2022 at 01:17:56PM +0200, David Hildenbrand wrote:
>> On 19.08.22 11:05, Rebecca Mckeever wrote:
>>> Add functions setup_numa_memblock_generic() and setup_numa_memblock()
>>> for setting up a memory layout with multiple NUMA nodes in a previously
>>> allocated dummy physical memory. These functions can be used in place of
>>> setup_memblock() in tests that need to simulate a NUMA system.
>>>
>>> setup_numa_memblock_generic():
>>> - allows for setting up a custom memory layout by specifying the amount
>>>   of memory in each node, the number of nodes, and a factor that will be
>>>   used to scale the memory in each node
>>>
>>> setup_numa_memblock():
>>> - allows for setting up a default memory layout
>>>
>>> Introduce constant MEM_FACTOR, which is used to scale the default memory
>>> layout based on MEM_SIZE.
>>>
>>> Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
>>> 16 NUMA nodes.
>>>
>>> Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
>>> ---
>>>  .../testing/memblock/scripts/Makefile.include |  2 +-
>>>  tools/testing/memblock/tests/common.c         | 38 +++++++++++++++++++
>>>  tools/testing/memblock/tests/common.h         |  9 ++++-
>>>  3 files changed, 47 insertions(+), 2 deletions(-)
>>>
>>> diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
>>> index aa6d82d56a23..998281723590 100644
>>> --- a/tools/testing/memblock/scripts/Makefile.include
>>> +++ b/tools/testing/memblock/scripts/Makefile.include
>>> @@ -3,7 +3,7 @@
>>>  
>>>  # Simulate CONFIG_NUMA=y
>>>  ifeq ($(NUMA), 1)
>>> -	CFLAGS += -D CONFIG_NUMA
>>> +	CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
>>>  endif
>>>  
>>>  # Use 32 bit physical addresses.
>>> diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
>>> index eec6901081af..15d8767dc70c 100644
>>> --- a/tools/testing/memblock/tests/common.c
>>> +++ b/tools/testing/memblock/tests/common.c
>>> @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
>>>  
>>>  static int verbose;
>>>  
>>> +static const phys_addr_t node_sizes[] = {
>>> +	SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
>>> +};
>>> +
>>>  /* sets global variable returned by movable_node_is_enabled() stub */
>>>  bool movable_node_enabled;
>>>  
>>> @@ -72,6 +76,40 @@ void setup_memblock(void)
>>>  	fill_memblock();
>>>  }
>>>  
>>> +/**
>>> + * setup_numa_memblock_generic:
>>> + * Set up a memory layout with multiple NUMA nodes in a previously allocated
>>> + * dummy physical memory.
>>> + * @nodes: an array containing the amount of memory in each node
>>> + * @node_cnt: the size of @nodes
>>> + * @factor: a factor that will be used to scale the memory in each node
>>> + *
>>> + * The nids will be set to 0 through node_cnt - 1.
>>> + */
>>> +void setup_numa_memblock_generic(const phys_addr_t nodes[],
>>> +				 int node_cnt, int factor)
>>> +{
>>> +	phys_addr_t base;
>>> +	int flags;
>>> +
>>> +	reset_memblock_regions();
>>> +	base = (phys_addr_t)memory_block.base;
>>> +	flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
>>> +
>>> +	for (int i = 0; i < node_cnt; i++) {
>>> +		phys_addr_t size = factor * nodes[i];
>>
>> I'm a bit lost why we need the factor if we already provide sizes in the
>> array.
>>
>> Can you enlighten me? :)
>>
>> Why can't we just stick to the sizes in the array?
>>
> Without the factor, some of the tests will break if we increase MEM_SIZE
> in the future (which we may need to do). I could rewrite them so that the
> factor is not needed, but I thought the code would be over-complicated if
> I did.

Independent of the suggestion from Mike, I wonder if we should really
care about (eventual) MEM_SIZE changes for now if not caring simplifies
the current code.

-- 
Thanks,

David / dhildenb



^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes
  2022-08-31 15:12       ` Mike Rapoport
@ 2022-09-01 22:53         ` Rebecca Mckeever
  0 siblings, 0 replies; 11+ messages in thread
From: Rebecca Mckeever @ 2022-09-01 22:53 UTC (permalink / raw)
  To: Mike Rapoport; +Cc: David Hildenbrand, linux-mm, linux-kernel

On Wed, Aug 31, 2022 at 06:12:10PM +0300, Mike Rapoport wrote:
> On Tue, Aug 30, 2022 at 10:49:09PM -0500, Rebecca Mckeever wrote:
> > On Tue, Aug 30, 2022 at 01:17:56PM +0200, David Hildenbrand wrote:
> > > On 19.08.22 11:05, Rebecca Mckeever wrote:
> > > > Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> > > > for setting up a memory layout with multiple NUMA nodes in a previously
> > > > allocated dummy physical memory. These functions can be used in place of
> > > > setup_memblock() in tests that need to simulate a NUMA system.
> > > > 
> > > > setup_numa_memblock_generic():
> > > > - allows for setting up a custom memory layout by specifying the amount
> > > >   of memory in each node, the number of nodes, and a factor that will be
> > > >   used to scale the memory in each node
> > > > 
> > > > setup_numa_memblock():
> > > > - allows for setting up a default memory layout
> > > > 
> > > > Introduce constant MEM_FACTOR, which is used to scale the default memory
> > > > layout based on MEM_SIZE.
> > > > 
> > > > Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> > > > 16 NUMA nodes.
> > > > 
> > > > Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
> > > > ---
> > > >  .../testing/memblock/scripts/Makefile.include |  2 +-
> > > >  tools/testing/memblock/tests/common.c         | 38 +++++++++++++++++++
> > > >  tools/testing/memblock/tests/common.h         |  9 ++++-
> > > >  3 files changed, 47 insertions(+), 2 deletions(-)
> > > > 
> > > > diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
> > > > index aa6d82d56a23..998281723590 100644
> > > > --- a/tools/testing/memblock/scripts/Makefile.include
> > > > +++ b/tools/testing/memblock/scripts/Makefile.include
> > > > @@ -3,7 +3,7 @@
> > > >  
> > > >  # Simulate CONFIG_NUMA=y
> > > >  ifeq ($(NUMA), 1)
> > > > -	CFLAGS += -D CONFIG_NUMA
> > > > +	CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
> > > >  endif
> > > >  
> > > >  # Use 32 bit physical addresses.
> > > > diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
> > > > index eec6901081af..15d8767dc70c 100644
> > > > --- a/tools/testing/memblock/tests/common.c
> > > > +++ b/tools/testing/memblock/tests/common.c
> > > > @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
> > > >  
> > > >  static int verbose;
> > > >  
> > > > +static const phys_addr_t node_sizes[] = {
> > > > +	SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
> > > > +};
> > > > +
> > > >  /* sets global variable returned by movable_node_is_enabled() stub */
> > > >  bool movable_node_enabled;
> > > >  
> > > > @@ -72,6 +76,40 @@ void setup_memblock(void)
> > > >  	fill_memblock();
> > > >  }
> > > >  
> > > > +/**
> > > > + * setup_numa_memblock_generic:
> > > > + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> > > > + * dummy physical memory.
> > > > + * @nodes: an array containing the amount of memory in each node
> > > > + * @node_cnt: the size of @nodes
> > > > + * @factor: a factor that will be used to scale the memory in each node
> > > > + *
> > > > + * The nids will be set to 0 through node_cnt - 1.
> > > > + */
> > > > +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> > > > +				 int node_cnt, int factor)
> > > > +{
> > > > +	phys_addr_t base;
> > > > +	int flags;
> > > > +
> > > > +	reset_memblock_regions();
> > > > +	base = (phys_addr_t)memory_block.base;
> > > > +	flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> > > > +
> > > > +	for (int i = 0; i < node_cnt; i++) {
> > > > +		phys_addr_t size = factor * nodes[i];
> > > 
> > > I'm a bit lost why we need the factor if we already provide sizes in the
> > > array.
> > > 
> > > Can you enlighten me? :)
> > > 
> > > Why can't we just stick to the sizes in the array?
> > > 
> > Without the factor, some of the tests will break if we increase MEM_SIZE
> > in the future (which we may need to do). I could rewrite them so that the
> > factor is not needed, but I thought the code would be over-complicated if
> > I did.
> 
> What if we make nodes[] to represent the fraction of the memory rather than
> a node size? Then the factor won't be required.
> 
I think that will work. I'll try it.

> > Thanks,
> > Rebecca
> 
> -- 
> Sincerely yours,
> Mike.

Thanks,
Rebecca


^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes
  2022-09-01  8:06       ` David Hildenbrand
@ 2022-09-02  0:08         ` Rebecca Mckeever
  0 siblings, 0 replies; 11+ messages in thread
From: Rebecca Mckeever @ 2022-09-02  0:08 UTC (permalink / raw)
  To: David Hildenbrand; +Cc: Mike Rapoport, linux-mm, linux-kernel

On Thu, Sep 01, 2022 at 10:06:48AM +0200, David Hildenbrand wrote:
> On 31.08.22 05:49, Rebecca Mckeever wrote:
> > On Tue, Aug 30, 2022 at 01:17:56PM +0200, David Hildenbrand wrote:
> >> On 19.08.22 11:05, Rebecca Mckeever wrote:
> >>> Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> >>> for setting up a memory layout with multiple NUMA nodes in a previously
> >>> allocated dummy physical memory. These functions can be used in place of
> >>> setup_memblock() in tests that need to simulate a NUMA system.
> >>>
> >>> setup_numa_memblock_generic():
> >>> - allows for setting up a custom memory layout by specifying the amount
> >>>   of memory in each node, the number of nodes, and a factor that will be
> >>>   used to scale the memory in each node
> >>>
> >>> setup_numa_memblock():
> >>> - allows for setting up a default memory layout
> >>>
> >>> Introduce constant MEM_FACTOR, which is used to scale the default memory
> >>> layout based on MEM_SIZE.
> >>>
> >>> Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> >>> 16 NUMA nodes.
> >>>
> >>> Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
> >>> ---
> >>>  .../testing/memblock/scripts/Makefile.include |  2 +-
> >>>  tools/testing/memblock/tests/common.c         | 38 +++++++++++++++++++
> >>>  tools/testing/memblock/tests/common.h         |  9 ++++-
> >>>  3 files changed, 47 insertions(+), 2 deletions(-)
> >>>
> >>> diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
> >>> index aa6d82d56a23..998281723590 100644
> >>> --- a/tools/testing/memblock/scripts/Makefile.include
> >>> +++ b/tools/testing/memblock/scripts/Makefile.include
> >>> @@ -3,7 +3,7 @@
> >>>  
> >>>  # Simulate CONFIG_NUMA=y
> >>>  ifeq ($(NUMA), 1)
> >>> -	CFLAGS += -D CONFIG_NUMA
> >>> +	CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
> >>>  endif
> >>>  
> >>>  # Use 32 bit physical addresses.
> >>> diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
> >>> index eec6901081af..15d8767dc70c 100644
> >>> --- a/tools/testing/memblock/tests/common.c
> >>> +++ b/tools/testing/memblock/tests/common.c
> >>> @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
> >>>  
> >>>  static int verbose;
> >>>  
> >>> +static const phys_addr_t node_sizes[] = {
> >>> +	SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
> >>> +};
> >>> +
> >>>  /* sets global variable returned by movable_node_is_enabled() stub */
> >>>  bool movable_node_enabled;
> >>>  
> >>> @@ -72,6 +76,40 @@ void setup_memblock(void)
> >>>  	fill_memblock();
> >>>  }
> >>>  
> >>> +/**
> >>> + * setup_numa_memblock_generic:
> >>> + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> >>> + * dummy physical memory.
> >>> + * @nodes: an array containing the amount of memory in each node
> >>> + * @node_cnt: the size of @nodes
> >>> + * @factor: a factor that will be used to scale the memory in each node
> >>> + *
> >>> + * The nids will be set to 0 through node_cnt - 1.
> >>> + */
> >>> +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> >>> +				 int node_cnt, int factor)
> >>> +{
> >>> +	phys_addr_t base;
> >>> +	int flags;
> >>> +
> >>> +	reset_memblock_regions();
> >>> +	base = (phys_addr_t)memory_block.base;
> >>> +	flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> >>> +
> >>> +	for (int i = 0; i < node_cnt; i++) {
> >>> +		phys_addr_t size = factor * nodes[i];
> >>
> >> I'm a bit lost why we need the factor if we already provide sizes in the
> >> array.
> >>
> >> Can you enlighten me? :)
> >>
> >> Why can't we just stick to the sizes in the array?
> >>
> > Without the factor, some of the tests will break if we increase MEM_SIZE
> > in the future (which we may need to do). I could rewrite them so that the
> > factor is not needed, but I thought the code would be over-complicated if
> > I did.
> 
> Independent of the suggestion from Mike, I wonder if we should really
> care about (eventual) MEM_SIZE changes for now if not caring simplifies
> the current code.
> 
Maybe not. I'm going to try Mike's suggestion, but I will keep this in
mind if the code seems too complicated.

> -- 
> Thanks,
> 
> David / dhildenb
> 
Thanks,
Rebecca


^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes
  2022-08-31 15:15   ` Mike Rapoport
@ 2022-09-02  0:14     ` Rebecca Mckeever
  0 siblings, 0 replies; 11+ messages in thread
From: Rebecca Mckeever @ 2022-09-02  0:14 UTC (permalink / raw)
  To: Mike Rapoport; +Cc: linux-mm, linux-kernel, David Hildenbrand

On Wed, Aug 31, 2022 at 06:15:41PM +0300, Mike Rapoport wrote:
> On Fri, Aug 19, 2022 at 02:05:31AM -0700, Rebecca Mckeever wrote:
> > Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> > for setting up a memory layout with multiple NUMA nodes in a previously
> > allocated dummy physical memory. These functions can be used in place of
> > setup_memblock() in tests that need to simulate a NUMA system.
> > 
> > setup_numa_memblock_generic():
> > - allows for setting up a custom memory layout by specifying the amount
> >   of memory in each node, the number of nodes, and a factor that will be
> >   used to scale the memory in each node
> > 
> > setup_numa_memblock():
> > - allows for setting up a default memory layout
> > 
> > Introduce constant MEM_FACTOR, which is used to scale the default memory
> > layout based on MEM_SIZE.
> > 
> > Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> > 16 NUMA nodes.
> > 
> > Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
> > ---
> >  .../testing/memblock/scripts/Makefile.include |  2 +-
> >  tools/testing/memblock/tests/common.c         | 38 +++++++++++++++++++
> >  tools/testing/memblock/tests/common.h         |  9 ++++-
> >  3 files changed, 47 insertions(+), 2 deletions(-)
> 
> ...
>   
> > +/**
> > + * setup_numa_memblock_generic:
> > + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> > + * dummy physical memory.
> > + * @nodes: an array containing the amount of memory in each node
> > + * @node_cnt: the size of @nodes
> > + * @factor: a factor that will be used to scale the memory in each node
> > + *
> > + * The nids will be set to 0 through node_cnt - 1.
> > + */
> > +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> > +				 int node_cnt, int factor)
> 
> I only had time for a quick look and it seems this function is never used
> on its own.
> Let's fold it into setup_numa_memblock() for now.
> 
Okay, will do.

> > +{
> > +	phys_addr_t base;
> > +	int flags;
> > +
> > +	reset_memblock_regions();
> > +	base = (phys_addr_t)memory_block.base;
> > +	flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> > +
> > +	for (int i = 0; i < node_cnt; i++) {
> > +		phys_addr_t size = factor * nodes[i];
> > +
> > +		memblock_add_node(base, size, i, flags);
> > +		base += size;
> > +	}
> > +	fill_memblock();
> > +}
> > +
> > +void setup_numa_memblock(void)
> > +{
> > +	setup_numa_memblock_generic(node_sizes, NUMA_NODES, MEM_FACTOR);
> > +}
> > +
> 
> -- 
> Sincerely yours,
> Mike.

Thanks,
Rebecca


^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: [PATCH v2 2/4] memblock tests: add top-down NUMA tests for memblock_alloc_try_nid*
  2022-08-30 11:56   ` [PATCH v2 2/4] memblock tests: add top-down NUMA tests for memblock_alloc_try_nid* David Hildenbrand
@ 2022-09-02  0:37     ` Rebecca Mckeever
  0 siblings, 0 replies; 11+ messages in thread
From: Rebecca Mckeever @ 2022-09-02  0:37 UTC (permalink / raw)
  To: David Hildenbrand; +Cc: Mike Rapoport, linux-mm, linux-kernel

On Tue, Aug 30, 2022 at 01:56:00PM +0200, David Hildenbrand wrote:
> On 19.08.22 11:05, Rebecca Mckeever wrote:
> > Add tests for memblock_alloc_try_nid() and memblock_alloc_try_nid_raw()
> > where the simulated physical memory is set up with multiple NUMA nodes.
> > Additionally, all of these tests set nid != NUMA_NO_NODE. These tests are
> > run with a top-down allocation direction.
> > 
> > The tested scenarios are:
> > 
> > Range unrestricted:
> > - region can be allocated in the specific node requested:
> >       + there are no previously reserved regions
> >       + the requested node is partially reserved but has enough space
> > - the specific node requested cannot accommodate the request, but the
> >   region can be allocated in a different node:
> >       + there are no previously reserved regions, but node is too small
> >       + the requested node is fully reserved
> >       + the requested node is partially reserved and does not have
> >         enough space
> > 
> > Range restricted:
> > - region can be allocated in the specific node requested after dropping
> >   min_addr:
> >       + range partially overlaps with two different nodes, where the first
> >         node is the requested node
> >       + range partially overlaps with two different nodes, where the
> >         requested node ends before min_addr
> > - region cannot be allocated in the specific node requested, but it can be
> >   allocated in the requested range:
> >       + range overlaps with multiple nodes along node boundaries, and the
> >         requested node ends before min_addr
> >       + range overlaps with multiple nodes along node boundaries, and the
> >         requested node starts after max_addr
> > - region cannot be allocated in the specific node requested, but it can be
> >   allocated after dropping min_addr:
> >       + range partially overlaps with two different nodes, where the
> >         second node is the requested node
> > 
> > Signed-off-by: Rebecca Mckeever <remckee0@gmail.com>
> > ---
> >  tools/testing/memblock/tests/alloc_nid_api.c | 702 ++++++++++++++++++-
> >  tools/testing/memblock/tests/alloc_nid_api.h |  16 +
> >  tools/testing/memblock/tests/common.h        |  18 +
> >  3 files changed, 725 insertions(+), 11 deletions(-)
> > 
> > diff --git a/tools/testing/memblock/tests/alloc_nid_api.c b/tools/testing/memblock/tests/alloc_nid_api.c
> > index 2c1d5035e057..a410f1318402 100644
> > --- a/tools/testing/memblock/tests/alloc_nid_api.c
> > +++ b/tools/testing/memblock/tests/alloc_nid_api.c
> > @@ -1102,7 +1102,7 @@ static int alloc_try_nid_bottom_up_cap_min_check(void)
> >  	return 0;
> >  }
> >  
> > -/* Test case wrappers */
> > +/* Test case wrappers for range tests */
> >  static int alloc_try_nid_simple_check(void)
> >  {
> >  	test_print("\tRunning %s...\n", __func__);
> > @@ -1234,17 +1234,10 @@ static int alloc_try_nid_low_max_check(void)
> >  	return 0;
> >  }
> >  
> > -static int memblock_alloc_nid_checks_internal(int flags)
> > +static int memblock_alloc_nid_range_checks(void)
> >  {
> > -	const char *func = get_func_testing(flags);
> > -
> > -	alloc_nid_test_flags = flags;
> > -	prefix_reset();
> > -	prefix_push(func);
> > -	test_print("Running %s tests...\n", func);
> > -
> > -	reset_memblock_attributes();
> > -	dummy_physical_memory_init();
> > +	test_print("Running %s range tests...\n",
> > +		   get_func_testing(alloc_nid_test_flags));
> >  
> >  	alloc_try_nid_simple_check();
> >  	alloc_try_nid_misaligned_check();
> > @@ -1261,6 +1254,693 @@ static int memblock_alloc_nid_checks_internal(int flags)
> >  	alloc_try_nid_reserved_all_check();
> >  	alloc_try_nid_low_max_check();
> >  
> > +	return 0;
> > +}
> > +
> > +/*
> > + * A test that tries to allocate a memory region in a specific NUMA node that
> > + * has enough memory to allocate a region of the requested size.
> > + * Expect to allocate an aligned region at the end of the requested node.
> > + */
> > +static int alloc_try_nid_top_down_numa_simple_check(void)
> > +{
> > +	int nid_req = 3;
> > +	struct memblock_region *new_rgn = &memblock.reserved.regions[0];
> > +	struct memblock_region *req_node = &memblock.memory.regions[nid_req];
> > +	void *allocated_ptr = NULL;
> > +
> > +	PREFIX_PUSH();
> > +
> > +	phys_addr_t size;
> > +	phys_addr_t min_addr;
> > +	phys_addr_t max_addr;
> 
> Usually we define variables in a single block. So, before the
> PREFIX_PUSH(). Same applies to the other functions.
> 
Got it.

> > +
> > +	setup_numa_memblock();
> > +
> > +	ASSERT_LE(SZ_4, req_node->size);
> > +	size = req_node->size / SZ_4;
> > +	min_addr = memblock_start_of_DRAM();
> > +	max_addr = memblock_end_of_DRAM();
> > +
> > +	allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
> > +						   min_addr, max_addr, nid_req);
> > +
> > +	ASSERT_NE(allocated_ptr, NULL);
> > +	verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
> > +
> > +	ASSERT_EQ(new_rgn->size, size);
> > +	ASSERT_EQ(new_rgn->base, region_end(req_node) - size);
> > +	ASSERT_LE(req_node->base, new_rgn->base);
> > +
> > +	ASSERT_EQ(memblock.reserved.cnt, 1);
> > +	ASSERT_EQ(memblock.reserved.total_size, size);
> > +
> > +	test_pass_pop();
> > +
> > +	return 0;
> > +}
> > +
> 
> [...]
> 
> > +
> > +/*
> > + * A test that tries to allocate a memory region that spans over the min_addr
> > + * and max_addr range and overlaps with two different nodes, where the first
> > + * node is the requested node:
> > + *
> > + *                                min_addr
> > + *                                |           max_addr
> > + *                                |           |
> > + *                                v           v
> > + *  |           +-----------------------+-----------+              |
> > + *  |           |       requested       |   node3   |              |
> > + *  +-----------+-----------------------+-----------+--------------+
> > + *                                +           +
> > + *  |                       +-----------+                          |
> > + *  |                       |    rgn    |                          |
> > + *  +-----------------------+-----------+--------------------------+
> > + *
> > + * Expect to drop the lower limit and allocate a cleared memory region that
> > + * ends at the end of the requested node.
> 
> Interesting, allocating out-of-range is expected behavior? At least to
> me that wasn't immediately clear :)
> 
Yeah, it seems that memblock avoids allocations that would overlap with
more than one node. Do you think I should explain that in the comment?

> [...]
> 
> > +
> > +/*
> > + * A test that tries to allocate a memory region that spans over the min_addr
> > + * and max_addr range and overlaps with two different nodes, where the second
> > + * node is the requested node:
> > + *
> > + *                               min_addr
> > + *                               |         max_addr
> > + *                               |         |
> > + *                               v         v
> > + *  |      +--------------------------+---------+                |
> > + *  |      |         expected         |requested|                |
> > + *  +------+--------------------------+---------+----------------+
> > + *                               +         +
> > + *  |                       +---------+                          |
> > + *  |                       |   rgn   |                          |
> > + *  +-----------------------+---------+--------------------------+
> > + *
> > + * Expect to drop the lower limit and allocate a cleared memory region that
> 
> Does the "cleared memory region" part still apply? Or would we also end
> up calling the raw variant from run_memblock_alloc_try_nid() ?
> 
No, it doesn't apply. Thanks for catching this. I should probably add
another patch to update the wording in the pre-existing tests too.

> > + * ends at the end of the first node that overlaps with the range.
> > + */
> > +static int alloc_try_nid_top_down_numa_split_range_high_check(void)
> > +{
> > +	int nid_req = 3;
> > +	int nid_exp = nid_req - 1;
> > +	struct memblock_region *new_rgn = &memblock.reserved.regions[0];
> > +	struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
> > +	void *allocated_ptr = NULL;
> > +
> > +	PREFIX_PUSH();
> > +
> > +	phys_addr_t size = SZ_512;
> > +	phys_addr_t min_addr;
> > +	phys_addr_t max_addr;
> > +	phys_addr_t exp_node_end;
> > +
> > +	setup_numa_memblock();
> > +
> > +	exp_node_end = region_end(exp_node);
> > +	min_addr = exp_node_end - SZ_256;
> > +	max_addr = min_addr + size;
> > +
> > +	allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
> > +						   min_addr, max_addr, nid_req);
> > +
> > +	ASSERT_NE(allocated_ptr, NULL);
> > +	verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
> > +
> > +	ASSERT_EQ(new_rgn->size, size);
> > +	ASSERT_EQ(new_rgn->base, exp_node_end - size);
> > +	ASSERT_LE(exp_node->base, new_rgn->base);
> > +
> > +	ASSERT_EQ(memblock.reserved.cnt, 1);
> > +	ASSERT_EQ(memblock.reserved.total_size, size);
> > +
> > +	test_pass_pop();
> > +
> > +	return 0;
> > +}
> 
> 
> [...]
> 
> > +int __memblock_alloc_nid_numa_checks(void)
> > +{
> > +	test_print("Running %s NUMA tests...\n",
> > +		   get_func_testing(alloc_nid_test_flags));
> > +
> > +	alloc_try_nid_numa_simple_check();
> > +	alloc_try_nid_numa_small_node_check();
> > +	alloc_try_nid_numa_node_reserved_check();
> > +	alloc_try_nid_numa_part_reserved_check();
> > +	alloc_try_nid_numa_part_reserved_fallback_check();
> > +	alloc_try_nid_numa_split_range_low_check();
> > +	alloc_try_nid_numa_split_range_high_check();
> > +
> > +	alloc_try_nid_numa_no_overlap_split_check();
> > +	alloc_try_nid_numa_no_overlap_low_check();
> > +	alloc_try_nid_numa_no_overlap_high_check();
> > +
> > +	return 0;
> > +}
> > +
> > +static int memblock_alloc_nid_checks_internal(int flags)
> > +{
> > +	alloc_nid_test_flags = flags;
> 
> Empty line missing
> 
Got it.

> > +	prefix_reset();
> > +	prefix_push(get_func_testing(flags));
> > +
> > +	reset_memblock_attributes();
> > +	dummy_physical_memory_init();
> > +
> > +	memblock_alloc_nid_range_checks();
> > +	memblock_alloc_nid_numa_checks();
> > +
> >  	dummy_physical_memory_cleanup();
> >  
> >  	prefix_pop();
> 
> 
> -- 
> Thanks,
> 
> David / dhildenb
> 
>
Thanks,
Rebecca


^ permalink raw reply	[flat|nested] 11+ messages in thread

end of thread, other threads:[~2022-09-02  0:37 UTC | newest]

Thread overview: 11+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
     [not found] <cover.1660897864.git.remckee0@gmail.com>
2022-08-19  9:05 ` [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes Rebecca Mckeever
2022-08-30 11:17   ` David Hildenbrand
2022-08-31  3:49     ` Rebecca Mckeever
2022-08-31 15:12       ` Mike Rapoport
2022-09-01 22:53         ` Rebecca Mckeever
2022-09-01  8:06       ` David Hildenbrand
2022-09-02  0:08         ` Rebecca Mckeever
2022-08-31 15:15   ` Mike Rapoport
2022-09-02  0:14     ` Rebecca Mckeever
     [not found] ` <957966f06474e3885796247ad1beaa6b3841ebd1.1660897864.git.remckee0@gmail.com>
2022-08-30 11:56   ` [PATCH v2 2/4] memblock tests: add top-down NUMA tests for memblock_alloc_try_nid* David Hildenbrand
2022-09-02  0:37     ` Rebecca Mckeever

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).