2022-08-19 09:30:01

by Rebecca Mckeever

[permalink] [raw]
Subject: [PATCH v2 0/4] memblock tests: add NUMA tests for memblock_alloc_try_nid*

These patches add additional tests for memblock_alloc_try_nid() and
memblock_alloc_try_nid_raw() that use a simulated physical
memory that is set up with multiple NUMA nodes. Additionally, most of
these tests set nid != NUMA_NO_NODE.

To set up a simulated physical memory with multiple NUMA nodes, patch 1
introduces setup_numa_memblock_generic() and setup_numa_memblock().
These functions use a previously allocated dummy physical memory. They
can be used in place of setup_memblock() in tests that need to simulate
a NUMA system.

These tests are run twice, once for memblock_alloc_try_nid() and once
for memblock_alloc_try_nid_raw(), so that both functions are tested with
the same set of tests. When the tests run memblock_alloc_try_nid(), they
test that the entire memory region is zero. When the tests run
memblock_alloc_try_nid_raw(), they test that the entire memory region is
nonzero.

Note:
This patch set depends on the following patch set:
https://lore.kernel.org/all/[email protected]

---
Changelog

v1 -> v2
Updates based on changes to the dependent patch set noted above:
PATCH 2, PATCH 3, PATCH 4:
- tests/alloc_nid_api.c:
- Update calls to verify_mem_content() to include the flags parameter
required by the new, common verify_mem_content() function.
---

Rebecca Mckeever (4):
memblock tests: add simulation of physical memory with multiple NUMA
nodes
memblock tests: add top-down NUMA tests for memblock_alloc_try_nid*
memblock tests: add bottom-up NUMA tests for memblock_alloc_try_nid*
memblock tests: add generic NUMA tests for memblock_alloc_try_nid*

.../testing/memblock/scripts/Makefile.include | 2 +-
tools/testing/memblock/tests/alloc_nid_api.c | 1489 ++++++++++++++++-
tools/testing/memblock/tests/alloc_nid_api.h | 16 +
tools/testing/memblock/tests/common.c | 38 +
tools/testing/memblock/tests/common.h | 27 +-
5 files changed, 1559 insertions(+), 13 deletions(-)

--
2.25.1


2022-08-19 09:51:01

by Rebecca Mckeever

[permalink] [raw]
Subject: [PATCH v2 3/4] memblock tests: add bottom-up NUMA tests for memblock_alloc_try_nid*

Add tests for memblock_alloc_try_nid() and memblock_alloc_try_nid_raw()
where the simulated physical memory is set up with multiple NUMA nodes.
Additionally, all of these tests set nid != NUMA_NO_NODE. These tests are
run with a bottom-up allocation direction.

The tested scenarios are:

Range unrestricted:
- region can be allocated in the specific node requested:
+ there are no previously reserved regions
+ the requested node is partially reserved but has enough space
- the specific node requested cannot accommodate the request, but the
region can be allocated in a different node:
+ there are no previously reserved regions, but node is too small
+ the requested node is fully reserved
+ the requested node is partially reserved and does not have
enough space

Range restricted:
- region can be allocated in the specific node requested after dropping
min_addr:
+ range partially overlaps with two different nodes, where the first
node is the requested node
+ range partially overlaps with two different nodes, where the
requested node ends before min_addr
- region cannot be allocated in the specific node requested, but it can be
allocated in the requested range:
+ range overlaps with multiple nodes along node boundaries, and the
requested node ends before min_addr
+ range overlaps with multiple nodes along node boundaries, and the
requested node starts after max_addr
- region cannot be allocated in the specific node requested, but it can be
allocated after dropping min_addr:
+ range partially overlaps with two different nodes, where the
second node is the requested node

Signed-off-by: Rebecca Mckeever <[email protected]>
---
tools/testing/memblock/tests/alloc_nid_api.c | 584 +++++++++++++++++++
1 file changed, 584 insertions(+)

diff --git a/tools/testing/memblock/tests/alloc_nid_api.c b/tools/testing/memblock/tests/alloc_nid_api.c
index a410f1318402..0a7a7494a157 100644
--- a/tools/testing/memblock/tests/alloc_nid_api.c
+++ b/tools/testing/memblock/tests/alloc_nid_api.c
@@ -1818,12 +1818,578 @@ static int alloc_try_nid_numa_top_down_no_overlap_high_check(void)
return 0;
}

+/*
+ * A test that tries to allocate a memory region in a specific NUMA node that
+ * has enough memory to allocate a region of the requested size.
+ * Expect to allocate an aligned region at the beginning of the requested node.
+ */
+static int alloc_try_nid_bottom_up_numa_simple_check(void)
+{
+ int nid_req = 3;
+ struct memblock_region *new_rgn = &memblock.reserved.regions[0];
+ struct memblock_region *req_node = &memblock.memory.regions[nid_req];
+ void *allocated_ptr = NULL;
+
+ PREFIX_PUSH();
+
+ phys_addr_t size;
+ phys_addr_t min_addr;
+ phys_addr_t max_addr;
+
+ setup_numa_memblock();
+
+ ASSERT_LE(SZ_4, req_node->size);
+ size = req_node->size / SZ_4;
+ min_addr = memblock_start_of_DRAM();
+ max_addr = memblock_end_of_DRAM();
+
+ allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
+ min_addr, max_addr, nid_req);
+
+ ASSERT_NE(allocated_ptr, NULL);
+ verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
+
+ ASSERT_EQ(new_rgn->size, size);
+ ASSERT_EQ(new_rgn->base, req_node->base);
+ ASSERT_LE(region_end(new_rgn), region_end(req_node));
+
+ ASSERT_EQ(memblock.reserved.cnt, 1);
+ ASSERT_EQ(memblock.reserved.total_size, size);
+
+ test_pass_pop();
+
+ return 0;
+}
+
+/*
+ * A test that tries to allocate a memory region in a specific NUMA node that
+ * does not have enough memory to allocate a region of the requested size:
+ *
+ * |----------------------+-----+ |
+ * | expected | req | |
+ * +----------------------+-----+----------------+
+ *
+ * |---------+ |
+ * | rgn | |
+ * +---------+-----------------------------------+
+ *
+ * Expect to allocate an aligned region at the beginning of the first node that
+ * has enough memory (in this case, nid = 0) after falling back to NUMA_NO_NODE.
+ */
+static int alloc_try_nid_bottom_up_numa_small_node_check(void)
+{
+ int nid_req = 1;
+ int nid_exp = 0;
+ struct memblock_region *new_rgn = &memblock.reserved.regions[0];
+ struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
+ void *allocated_ptr = NULL;
+
+ PREFIX_PUSH();
+
+ phys_addr_t size = SZ_2K * MEM_FACTOR;
+ phys_addr_t min_addr;
+ phys_addr_t max_addr;
+
+ setup_numa_memblock();
+
+ min_addr = memblock_start_of_DRAM();
+ max_addr = memblock_end_of_DRAM();
+
+ allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
+ min_addr, max_addr, nid_req);
+
+ ASSERT_NE(allocated_ptr, NULL);
+ verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
+
+ ASSERT_EQ(new_rgn->size, size);
+ ASSERT_EQ(new_rgn->base, exp_node->base);
+ ASSERT_LE(region_end(new_rgn), region_end(exp_node));
+
+ ASSERT_EQ(memblock.reserved.cnt, 1);
+ ASSERT_EQ(memblock.reserved.total_size, size);
+
+ test_pass_pop();
+
+ return 0;
+}
+
+/*
+ * A test that tries to allocate a memory region in a specific NUMA node that
+ * is fully reserved:
+ *
+ * |----------------------+ +-----------+ |
+ * | expected | | requested | |
+ * +----------------------+-----+-----------+--------------------+
+ *
+ * |-----------+ +-----------+ |
+ * | new | | reserved | |
+ * +-----------+----------------+-----------+--------------------+
+ *
+ * Expect to allocate an aligned region at the beginning of the first node that
+ * is large enough and has enough unreserved memory (in this case, nid = 0)
+ * after falling back to NUMA_NO_NODE. The region count and total size get
+ * updated.
+ */
+static int alloc_try_nid_bottom_up_numa_node_reserved_check(void)
+{
+ int nid_req = 2;
+ int nid_exp = 0;
+ struct memblock_region *new_rgn = &memblock.reserved.regions[0];
+ struct memblock_region *req_node = &memblock.memory.regions[nid_req];
+ struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
+ void *allocated_ptr = NULL;
+
+ PREFIX_PUSH();
+
+ phys_addr_t size = SZ_2K * MEM_FACTOR;
+ phys_addr_t min_addr;
+ phys_addr_t max_addr;
+
+ setup_numa_memblock();
+
+ min_addr = memblock_start_of_DRAM();
+ max_addr = memblock_end_of_DRAM();
+
+ memblock_reserve(req_node->base, req_node->size);
+ allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
+ min_addr, max_addr, nid_req);
+
+ ASSERT_NE(allocated_ptr, NULL);
+ verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
+
+ ASSERT_EQ(new_rgn->size, size);
+ ASSERT_EQ(new_rgn->base, exp_node->base);
+ ASSERT_LE(region_end(new_rgn), region_end(exp_node));
+
+ ASSERT_EQ(memblock.reserved.cnt, 2);
+ ASSERT_EQ(memblock.reserved.total_size, size + req_node->size);
+
+ test_pass_pop();
+
+ return 0;
+}
+
+/*
+ * A test that tries to allocate a memory region in a specific NUMA node that
+ * is partially reserved but has enough memory for the allocated region:
+ *
+ * | +---------------------------------------+ |
+ * | | requested | |
+ * +-----------+---------------------------------------+---------+
+ *
+ * | +------------------+-----+ |
+ * | | reserved | new | |
+ * +-----------+------------------+-----+------------------------+
+ *
+ * Expect to allocate an aligned region in the requested node that merges with
+ * the existing reserved region. The total size gets updated.
+ */
+static int alloc_try_nid_bottom_up_numa_part_reserved_check(void)
+{
+ int nid_req = 4;
+ struct memblock_region *new_rgn = &memblock.reserved.regions[0];
+ struct memblock_region *req_node = &memblock.memory.regions[nid_req];
+ void *allocated_ptr = NULL;
+ struct region r1;
+
+ PREFIX_PUSH();
+
+ phys_addr_t size;
+ phys_addr_t min_addr;
+ phys_addr_t max_addr;
+ phys_addr_t total_size;
+
+ setup_numa_memblock();
+
+ r1.base = req_node->base;
+ r1.size = SZ_512 * MEM_FACTOR;
+ size = SZ_128 * MEM_FACTOR;
+
+ min_addr = memblock_start_of_DRAM();
+ max_addr = memblock_end_of_DRAM();
+ total_size = size + r1.size;
+
+ memblock_reserve(r1.base, r1.size);
+ allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
+ min_addr, max_addr, nid_req);
+
+ ASSERT_NE(allocated_ptr, NULL);
+ verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
+
+ ASSERT_EQ(new_rgn->size, total_size);
+ ASSERT_EQ(new_rgn->base, req_node->base);
+ ASSERT_LE(region_end(new_rgn), region_end(req_node));
+
+ ASSERT_EQ(memblock.reserved.cnt, 1);
+ ASSERT_EQ(memblock.reserved.total_size, total_size);
+
+ test_pass_pop();
+
+ return 0;
+}
+
+/*
+ * A test that tries to allocate a memory region in a specific NUMA node that
+ * is partially reserved and does not have enough contiguous memory for the
+ * allocated region:
+ *
+ * |----------------------+ +-----------------------+ |
+ * | expected | | requested | |
+ * +----------------------+-------+-----------------------+---------+
+ *
+ * |-----------+ +----------+ |
+ * | new | | reserved | |
+ * +-----------+------------------------+----------+----------------+
+ *
+ * Expect to allocate an aligned region at the beginning of the first
+ * node that is large enough and has enough unreserved memory (in this case,
+ * nid = 0) after falling back to NUMA_NO_NODE. The region count and total size
+ * get updated.
+ */
+static int alloc_try_nid_bottom_up_numa_part_reserved_fallback_check(void)
+{
+ int nid_req = 4;
+ int nid_exp = 0;
+ struct memblock_region *new_rgn = &memblock.reserved.regions[0];
+ struct memblock_region *req_node = &memblock.memory.regions[nid_req];
+ struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
+ void *allocated_ptr = NULL;
+ struct region r1;
+
+ PREFIX_PUSH();
+
+ phys_addr_t size;
+ phys_addr_t min_addr;
+ phys_addr_t max_addr;
+
+ setup_numa_memblock();
+
+ size = SZ_512 * MEM_FACTOR;
+ r1.base = req_node->base + SZ_256 * MEM_FACTOR;
+ r1.size = size;
+
+ min_addr = memblock_start_of_DRAM();
+ max_addr = memblock_end_of_DRAM();
+
+ memblock_reserve(r1.base, r1.size);
+ allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
+ min_addr, max_addr, nid_req);
+
+ ASSERT_NE(allocated_ptr, NULL);
+ verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
+
+ ASSERT_EQ(new_rgn->size, size);
+ ASSERT_EQ(new_rgn->base, exp_node->base);
+ ASSERT_LE(region_end(new_rgn), region_end(exp_node));
+
+ ASSERT_EQ(memblock.reserved.cnt, 2);
+ ASSERT_EQ(memblock.reserved.total_size, size + r1.size);
+
+ test_pass_pop();
+
+ return 0;
+}
+
+/*
+ * A test that tries to allocate a memory region that spans over the min_addr
+ * and max_addr range and overlaps with two different nodes, where the first
+ * node is the requested node:
+ *
+ * min_addr
+ * | max_addr
+ * | |
+ * v v
+ * | +-----------------------+-----------+ |
+ * | | requested | node3 | |
+ * +-----------+-----------------------+-----------+--------------+
+ * + +
+ * | +-----------+ |
+ * | | rgn | |
+ * +-----------+-----------+--------------------------------------+
+ *
+ * Expect to drop the lower limit and allocate a cleared memory region at the
+ * beginning of the requested node.
+ */
+static int alloc_try_nid_bottom_up_numa_split_range_low_check(void)
+{
+ int nid_req = 2;
+ struct memblock_region *new_rgn = &memblock.reserved.regions[0];
+ struct memblock_region *req_node = &memblock.memory.regions[nid_req];
+ void *allocated_ptr = NULL;
+
+ PREFIX_PUSH();
+
+ phys_addr_t size = SZ_512;
+ phys_addr_t min_addr;
+ phys_addr_t max_addr;
+ phys_addr_t req_node_end;
+
+ setup_numa_memblock();
+
+ req_node_end = region_end(req_node);
+ min_addr = req_node_end - SZ_256;
+ max_addr = min_addr + size;
+
+ allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
+ min_addr, max_addr, nid_req);
+
+ ASSERT_NE(allocated_ptr, NULL);
+ verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
+
+ ASSERT_EQ(new_rgn->size, size);
+ ASSERT_EQ(new_rgn->base, req_node->base);
+ ASSERT_LE(region_end(new_rgn), req_node_end);
+
+ ASSERT_EQ(memblock.reserved.cnt, 1);
+ ASSERT_EQ(memblock.reserved.total_size, size);
+
+ test_pass_pop();
+
+ return 0;
+}
+
+/*
+ * A test that tries to allocate a memory region that spans over the min_addr
+ * and max_addr range and overlaps with two different nodes, where the second
+ * node is the requested node:
+ *
+ * min_addr
+ * | max_addr
+ * | |
+ * v v
+ * |------------------+ +----------------------+---------+ |
+ * | expected | | previous |requested| |
+ * +------------------+--------+----------------------+---------+------+
+ * + +
+ * |---------+ |
+ * | rgn | |
+ * +---------+---------------------------------------------------------+
+ *
+ * Expect to drop the lower limit and allocate a cleared memory region at the
+ * beginning of the first node that has enough memory.
+ */
+static int alloc_try_nid_bottom_up_numa_split_range_high_check(void)
+{
+ int nid_req = 3;
+ int nid_exp = 0;
+ struct memblock_region *new_rgn = &memblock.reserved.regions[0];
+ struct memblock_region *req_node = &memblock.memory.regions[nid_req];
+ struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
+ void *allocated_ptr = NULL;
+
+ PREFIX_PUSH();
+
+ phys_addr_t size = SZ_512;
+ phys_addr_t min_addr;
+ phys_addr_t max_addr;
+ phys_addr_t exp_node_end;
+
+ setup_numa_memblock();
+
+ exp_node_end = region_end(req_node);
+ min_addr = req_node->base - SZ_256;
+ max_addr = min_addr + size;
+
+ allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
+ min_addr, max_addr, nid_req);
+
+ ASSERT_NE(allocated_ptr, NULL);
+ verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
+
+ ASSERT_EQ(new_rgn->size, size);
+ ASSERT_EQ(new_rgn->base, exp_node->base);
+ ASSERT_LE(region_end(new_rgn), exp_node_end);
+
+ ASSERT_EQ(memblock.reserved.cnt, 1);
+ ASSERT_EQ(memblock.reserved.total_size, size);
+
+ test_pass_pop();
+
+ return 0;
+}
+
+/*
+ * A test that tries to allocate a memory region that spans over the min_addr
+ * and max_addr range and overlaps with two different nodes, where the requested
+ * node ends before min_addr:
+ *
+ * min_addr
+ * | max_addr
+ * | |
+ * v v
+ * | +---------------+ +-------------+---------+ |
+ * | | requested | | node1 | node2 | |
+ * +----+---------------+--------+-------------+---------+---------+
+ * + +
+ * | +---------+ |
+ * | | rgn | |
+ * +----+---------+------------------------------------------------+
+ *
+ * Expect to drop the lower limit and allocate a cleared memory region that
+ * starts at the beginning of the requested node.
+ */
+static int alloc_try_nid_bottom_up_numa_no_overlap_split_check(void)
+{
+ int nid_req = 2;
+ struct memblock_region *new_rgn = &memblock.reserved.regions[0];
+ struct memblock_region *req_node = &memblock.memory.regions[nid_req];
+ struct memblock_region *node2 = &memblock.memory.regions[6];
+ void *allocated_ptr = NULL;
+
+ PREFIX_PUSH();
+
+ phys_addr_t size;
+ phys_addr_t min_addr;
+ phys_addr_t max_addr;
+
+ setup_numa_memblock();
+
+ size = SZ_512;
+ min_addr = node2->base - SZ_256;
+ max_addr = min_addr + size;
+
+ allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
+ min_addr, max_addr, nid_req);
+
+ ASSERT_NE(allocated_ptr, NULL);
+ verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
+
+ ASSERT_EQ(new_rgn->size, size);
+ ASSERT_EQ(new_rgn->base, req_node->base);
+ ASSERT_LE(region_end(new_rgn), region_end(req_node));
+
+ ASSERT_EQ(memblock.reserved.cnt, 1);
+ ASSERT_EQ(memblock.reserved.total_size, size);
+
+ test_pass_pop();
+
+ return 0;
+}
+
+/*
+ * A test that tries to allocate memory within min_addr and max_add range when
+ * the requested node and the range do not overlap, and requested node ends
+ * before min_addr. The range overlaps with multiple nodes along node
+ * boundaries:
+ *
+ * min_addr
+ * | max_addr
+ * | |
+ * v v
+ * |-----------+ +----------+----...----+----------+ |
+ * | requested | | min node | ... | max node | |
+ * +-----------+-----------+----------+----...----+----------+------+
+ * + +
+ * | +-----+ |
+ * | | rgn | |
+ * +-----------------------+-----+----------------------------------+
+ *
+ * Expect to allocate a cleared memory region at the beginning of the first node
+ * in the range after falling back to NUMA_NO_NODE.
+ */
+static int alloc_try_nid_numa_bottom_up_no_overlap_low_check(void)
+{
+ int nid_req = 0;
+ struct memblock_region *new_rgn = &memblock.reserved.regions[0];
+ struct memblock_region *min_node = &memblock.memory.regions[2];
+ struct memblock_region *max_node = &memblock.memory.regions[5];
+ void *allocated_ptr = NULL;
+
+ PREFIX_PUSH();
+
+ phys_addr_t size = SZ_64;
+ phys_addr_t max_addr;
+ phys_addr_t min_addr;
+
+ setup_numa_memblock();
+
+ min_addr = min_node->base;
+ max_addr = region_end(max_node);
+
+ allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
+ min_addr, max_addr, nid_req);
+
+ ASSERT_NE(allocated_ptr, NULL);
+ verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
+
+ ASSERT_EQ(new_rgn->size, size);
+ ASSERT_EQ(new_rgn->base, min_addr);
+ ASSERT_LE(region_end(new_rgn), region_end(min_node));
+
+ ASSERT_EQ(memblock.reserved.cnt, 1);
+ ASSERT_EQ(memblock.reserved.total_size, size);
+
+ test_pass_pop();
+
+ return 0;
+}
+
+/*
+ * A test that tries to allocate memory within min_addr and max_add range when
+ * the requested node and the range do not overlap, and requested node starts
+ * after max_addr. The range overlaps with multiple nodes along node
+ * boundaries:
+ *
+ * min_addr
+ * | max_addr
+ * | |
+ * v v
+ * | +----------+----...----+----------+ +---------+ |
+ * | | min node | ... | max node | |requested| |
+ * +-----+----------+----...----+----------+---------+---------+---+
+ * + +
+ * | +-----+ |
+ * | | rgn | |
+ * +-----+-----+---------------------------------------------------+
+ *
+ * Expect to allocate a cleared memory region at the beginning of the first node
+ * in the range after falling back to NUMA_NO_NODE.
+ */
+static int alloc_try_nid_numa_bottom_up_no_overlap_high_check(void)
+{
+ int nid_req = 7;
+ struct memblock_region *new_rgn = &memblock.reserved.regions[0];
+ struct memblock_region *min_node = &memblock.memory.regions[2];
+ struct memblock_region *max_node = &memblock.memory.regions[5];
+ void *allocated_ptr = NULL;
+
+ PREFIX_PUSH();
+
+ phys_addr_t size = SZ_64;
+ phys_addr_t max_addr;
+ phys_addr_t min_addr;
+
+ setup_numa_memblock();
+
+ min_addr = min_node->base;
+ max_addr = region_end(max_node);
+
+ allocated_ptr = run_memblock_alloc_try_nid(size, SMP_CACHE_BYTES,
+ min_addr, max_addr, nid_req);
+
+ ASSERT_NE(allocated_ptr, NULL);
+ verify_mem_content(allocated_ptr, size, alloc_nid_test_flags);
+
+ ASSERT_EQ(new_rgn->size, size);
+ ASSERT_EQ(new_rgn->base, min_addr);
+ ASSERT_LE(region_end(new_rgn), region_end(min_node));
+
+ ASSERT_EQ(memblock.reserved.cnt, 1);
+ ASSERT_EQ(memblock.reserved.total_size, size);
+
+ test_pass_pop();
+
+ return 0;
+}
+
/* Test case wrappers for NUMA tests */
static int alloc_try_nid_numa_simple_check(void)
{
test_print("\tRunning %s...\n", __func__);
memblock_set_bottom_up(false);
alloc_try_nid_top_down_numa_simple_check();
+ memblock_set_bottom_up(true);
+ alloc_try_nid_bottom_up_numa_simple_check();

return 0;
}
@@ -1833,6 +2399,8 @@ static int alloc_try_nid_numa_small_node_check(void)
test_print("\tRunning %s...\n", __func__);
memblock_set_bottom_up(false);
alloc_try_nid_top_down_numa_small_node_check();
+ memblock_set_bottom_up(true);
+ alloc_try_nid_bottom_up_numa_small_node_check();

return 0;
}
@@ -1842,6 +2410,8 @@ static int alloc_try_nid_numa_node_reserved_check(void)
test_print("\tRunning %s...\n", __func__);
memblock_set_bottom_up(false);
alloc_try_nid_top_down_numa_node_reserved_check();
+ memblock_set_bottom_up(true);
+ alloc_try_nid_bottom_up_numa_node_reserved_check();

return 0;
}
@@ -1851,6 +2421,8 @@ static int alloc_try_nid_numa_part_reserved_check(void)
test_print("\tRunning %s...\n", __func__);
memblock_set_bottom_up(false);
alloc_try_nid_top_down_numa_part_reserved_check();
+ memblock_set_bottom_up(true);
+ alloc_try_nid_bottom_up_numa_part_reserved_check();

return 0;
}
@@ -1860,6 +2432,8 @@ static int alloc_try_nid_numa_part_reserved_fallback_check(void)
test_print("\tRunning %s...\n", __func__);
memblock_set_bottom_up(false);
alloc_try_nid_top_down_numa_part_reserved_fallback_check();
+ memblock_set_bottom_up(true);
+ alloc_try_nid_bottom_up_numa_part_reserved_fallback_check();

return 0;
}
@@ -1869,6 +2443,8 @@ static int alloc_try_nid_numa_split_range_low_check(void)
test_print("\tRunning %s...\n", __func__);
memblock_set_bottom_up(false);
alloc_try_nid_top_down_numa_split_range_low_check();
+ memblock_set_bottom_up(true);
+ alloc_try_nid_bottom_up_numa_split_range_low_check();

return 0;
}
@@ -1878,6 +2454,8 @@ static int alloc_try_nid_numa_split_range_high_check(void)
test_print("\tRunning %s...\n", __func__);
memblock_set_bottom_up(false);
alloc_try_nid_top_down_numa_split_range_high_check();
+ memblock_set_bottom_up(true);
+ alloc_try_nid_bottom_up_numa_split_range_high_check();

return 0;
}
@@ -1887,6 +2465,8 @@ static int alloc_try_nid_numa_no_overlap_split_check(void)
test_print("\tRunning %s...\n", __func__);
memblock_set_bottom_up(false);
alloc_try_nid_top_down_numa_no_overlap_split_check();
+ memblock_set_bottom_up(true);
+ alloc_try_nid_bottom_up_numa_no_overlap_split_check();

return 0;
}
@@ -1896,6 +2476,8 @@ static int alloc_try_nid_numa_no_overlap_low_check(void)
test_print("\tRunning %s...\n", __func__);
memblock_set_bottom_up(false);
alloc_try_nid_numa_top_down_no_overlap_low_check();
+ memblock_set_bottom_up(true);
+ alloc_try_nid_numa_bottom_up_no_overlap_low_check();

return 0;
}
@@ -1905,6 +2487,8 @@ static int alloc_try_nid_numa_no_overlap_high_check(void)
test_print("\tRunning %s...\n", __func__);
memblock_set_bottom_up(false);
alloc_try_nid_numa_top_down_no_overlap_high_check();
+ memblock_set_bottom_up(true);
+ alloc_try_nid_numa_bottom_up_no_overlap_high_check();

return 0;
}
--
2.25.1

2022-08-19 09:54:44

by Rebecca Mckeever

[permalink] [raw]
Subject: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes

Add functions setup_numa_memblock_generic() and setup_numa_memblock()
for setting up a memory layout with multiple NUMA nodes in a previously
allocated dummy physical memory. These functions can be used in place of
setup_memblock() in tests that need to simulate a NUMA system.

setup_numa_memblock_generic():
- allows for setting up a custom memory layout by specifying the amount
of memory in each node, the number of nodes, and a factor that will be
used to scale the memory in each node

setup_numa_memblock():
- allows for setting up a default memory layout

Introduce constant MEM_FACTOR, which is used to scale the default memory
layout based on MEM_SIZE.

Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
16 NUMA nodes.

Signed-off-by: Rebecca Mckeever <[email protected]>
---
.../testing/memblock/scripts/Makefile.include | 2 +-
tools/testing/memblock/tests/common.c | 38 +++++++++++++++++++
tools/testing/memblock/tests/common.h | 9 ++++-
3 files changed, 47 insertions(+), 2 deletions(-)

diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
index aa6d82d56a23..998281723590 100644
--- a/tools/testing/memblock/scripts/Makefile.include
+++ b/tools/testing/memblock/scripts/Makefile.include
@@ -3,7 +3,7 @@

# Simulate CONFIG_NUMA=y
ifeq ($(NUMA), 1)
- CFLAGS += -D CONFIG_NUMA
+ CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
endif

# Use 32 bit physical addresses.
diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
index eec6901081af..15d8767dc70c 100644
--- a/tools/testing/memblock/tests/common.c
+++ b/tools/testing/memblock/tests/common.c
@@ -34,6 +34,10 @@ static const char * const help_opts[] = {

static int verbose;

+static const phys_addr_t node_sizes[] = {
+ SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
+};
+
/* sets global variable returned by movable_node_is_enabled() stub */
bool movable_node_enabled;

@@ -72,6 +76,40 @@ void setup_memblock(void)
fill_memblock();
}

+/**
+ * setup_numa_memblock_generic:
+ * Set up a memory layout with multiple NUMA nodes in a previously allocated
+ * dummy physical memory.
+ * @nodes: an array containing the amount of memory in each node
+ * @node_cnt: the size of @nodes
+ * @factor: a factor that will be used to scale the memory in each node
+ *
+ * The nids will be set to 0 through node_cnt - 1.
+ */
+void setup_numa_memblock_generic(const phys_addr_t nodes[],
+ int node_cnt, int factor)
+{
+ phys_addr_t base;
+ int flags;
+
+ reset_memblock_regions();
+ base = (phys_addr_t)memory_block.base;
+ flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
+
+ for (int i = 0; i < node_cnt; i++) {
+ phys_addr_t size = factor * nodes[i];
+
+ memblock_add_node(base, size, i, flags);
+ base += size;
+ }
+ fill_memblock();
+}
+
+void setup_numa_memblock(void)
+{
+ setup_numa_memblock_generic(node_sizes, NUMA_NODES, MEM_FACTOR);
+}
+
void dummy_physical_memory_init(void)
{
memory_block.base = malloc(MEM_SIZE);
diff --git a/tools/testing/memblock/tests/common.h b/tools/testing/memblock/tests/common.h
index 4fd3534ff955..e5117d959d6c 100644
--- a/tools/testing/memblock/tests/common.h
+++ b/tools/testing/memblock/tests/common.h
@@ -10,7 +10,11 @@
#include <linux/printk.h>
#include <../selftests/kselftest.h>

-#define MEM_SIZE SZ_16K
+#define MEM_SIZE SZ_16K
+#define NUMA_NODES 8
+
+/* used to resize values that need to scale with MEM_SIZE */
+#define MEM_FACTOR (MEM_SIZE / SZ_16K)

enum test_flags {
TEST_ZEROED = 0x0,
@@ -100,6 +104,9 @@ struct region {
void reset_memblock_regions(void);
void reset_memblock_attributes(void);
void setup_memblock(void);
+void setup_numa_memblock_generic(const phys_addr_t nodes[],
+ int node_cnt, int factor);
+void setup_numa_memblock(void);
void dummy_physical_memory_init(void);
void dummy_physical_memory_cleanup(void);
void parse_args(int argc, char **argv);
--
2.25.1

2022-08-30 11:59:04

by David Hildenbrand

[permalink] [raw]
Subject: Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes

On 19.08.22 11:05, Rebecca Mckeever wrote:
> Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> for setting up a memory layout with multiple NUMA nodes in a previously
> allocated dummy physical memory. These functions can be used in place of
> setup_memblock() in tests that need to simulate a NUMA system.
>
> setup_numa_memblock_generic():
> - allows for setting up a custom memory layout by specifying the amount
> of memory in each node, the number of nodes, and a factor that will be
> used to scale the memory in each node
>
> setup_numa_memblock():
> - allows for setting up a default memory layout
>
> Introduce constant MEM_FACTOR, which is used to scale the default memory
> layout based on MEM_SIZE.
>
> Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> 16 NUMA nodes.
>
> Signed-off-by: Rebecca Mckeever <[email protected]>
> ---
> .../testing/memblock/scripts/Makefile.include | 2 +-
> tools/testing/memblock/tests/common.c | 38 +++++++++++++++++++
> tools/testing/memblock/tests/common.h | 9 ++++-
> 3 files changed, 47 insertions(+), 2 deletions(-)
>
> diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
> index aa6d82d56a23..998281723590 100644
> --- a/tools/testing/memblock/scripts/Makefile.include
> +++ b/tools/testing/memblock/scripts/Makefile.include
> @@ -3,7 +3,7 @@
>
> # Simulate CONFIG_NUMA=y
> ifeq ($(NUMA), 1)
> - CFLAGS += -D CONFIG_NUMA
> + CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
> endif
>
> # Use 32 bit physical addresses.
> diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
> index eec6901081af..15d8767dc70c 100644
> --- a/tools/testing/memblock/tests/common.c
> +++ b/tools/testing/memblock/tests/common.c
> @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
>
> static int verbose;
>
> +static const phys_addr_t node_sizes[] = {
> + SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
> +};
> +
> /* sets global variable returned by movable_node_is_enabled() stub */
> bool movable_node_enabled;
>
> @@ -72,6 +76,40 @@ void setup_memblock(void)
> fill_memblock();
> }
>
> +/**
> + * setup_numa_memblock_generic:
> + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> + * dummy physical memory.
> + * @nodes: an array containing the amount of memory in each node
> + * @node_cnt: the size of @nodes
> + * @factor: a factor that will be used to scale the memory in each node
> + *
> + * The nids will be set to 0 through node_cnt - 1.
> + */
> +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> + int node_cnt, int factor)
> +{
> + phys_addr_t base;
> + int flags;
> +
> + reset_memblock_regions();
> + base = (phys_addr_t)memory_block.base;
> + flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> +
> + for (int i = 0; i < node_cnt; i++) {
> + phys_addr_t size = factor * nodes[i];

I'm a bit lost why we need the factor if we already provide sizes in the
array.

Can you enlighten me? :)

Why can't we just stick to the sizes in the array?

--
Thanks,

David / dhildenb

2022-08-31 04:20:06

by Rebecca Mckeever

[permalink] [raw]
Subject: Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes

On Tue, Aug 30, 2022 at 01:17:56PM +0200, David Hildenbrand wrote:
> On 19.08.22 11:05, Rebecca Mckeever wrote:
> > Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> > for setting up a memory layout with multiple NUMA nodes in a previously
> > allocated dummy physical memory. These functions can be used in place of
> > setup_memblock() in tests that need to simulate a NUMA system.
> >
> > setup_numa_memblock_generic():
> > - allows for setting up a custom memory layout by specifying the amount
> > of memory in each node, the number of nodes, and a factor that will be
> > used to scale the memory in each node
> >
> > setup_numa_memblock():
> > - allows for setting up a default memory layout
> >
> > Introduce constant MEM_FACTOR, which is used to scale the default memory
> > layout based on MEM_SIZE.
> >
> > Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> > 16 NUMA nodes.
> >
> > Signed-off-by: Rebecca Mckeever <[email protected]>
> > ---
> > .../testing/memblock/scripts/Makefile.include | 2 +-
> > tools/testing/memblock/tests/common.c | 38 +++++++++++++++++++
> > tools/testing/memblock/tests/common.h | 9 ++++-
> > 3 files changed, 47 insertions(+), 2 deletions(-)
> >
> > diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
> > index aa6d82d56a23..998281723590 100644
> > --- a/tools/testing/memblock/scripts/Makefile.include
> > +++ b/tools/testing/memblock/scripts/Makefile.include
> > @@ -3,7 +3,7 @@
> >
> > # Simulate CONFIG_NUMA=y
> > ifeq ($(NUMA), 1)
> > - CFLAGS += -D CONFIG_NUMA
> > + CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
> > endif
> >
> > # Use 32 bit physical addresses.
> > diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
> > index eec6901081af..15d8767dc70c 100644
> > --- a/tools/testing/memblock/tests/common.c
> > +++ b/tools/testing/memblock/tests/common.c
> > @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
> >
> > static int verbose;
> >
> > +static const phys_addr_t node_sizes[] = {
> > + SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
> > +};
> > +
> > /* sets global variable returned by movable_node_is_enabled() stub */
> > bool movable_node_enabled;
> >
> > @@ -72,6 +76,40 @@ void setup_memblock(void)
> > fill_memblock();
> > }
> >
> > +/**
> > + * setup_numa_memblock_generic:
> > + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> > + * dummy physical memory.
> > + * @nodes: an array containing the amount of memory in each node
> > + * @node_cnt: the size of @nodes
> > + * @factor: a factor that will be used to scale the memory in each node
> > + *
> > + * The nids will be set to 0 through node_cnt - 1.
> > + */
> > +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> > + int node_cnt, int factor)
> > +{
> > + phys_addr_t base;
> > + int flags;
> > +
> > + reset_memblock_regions();
> > + base = (phys_addr_t)memory_block.base;
> > + flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> > +
> > + for (int i = 0; i < node_cnt; i++) {
> > + phys_addr_t size = factor * nodes[i];
>
> I'm a bit lost why we need the factor if we already provide sizes in the
> array.
>
> Can you enlighten me? :)
>
> Why can't we just stick to the sizes in the array?
>
Without the factor, some of the tests will break if we increase MEM_SIZE
in the future (which we may need to do). I could rewrite them so that the
factor is not needed, but I thought the code would be over-complicated if
I did.

> --
> Thanks,
>
> David / dhildenb
>
Thanks,
Rebecca

2022-08-31 15:40:32

by Mike Rapoport

[permalink] [raw]
Subject: Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes

On Tue, Aug 30, 2022 at 10:49:09PM -0500, Rebecca Mckeever wrote:
> On Tue, Aug 30, 2022 at 01:17:56PM +0200, David Hildenbrand wrote:
> > On 19.08.22 11:05, Rebecca Mckeever wrote:
> > > Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> > > for setting up a memory layout with multiple NUMA nodes in a previously
> > > allocated dummy physical memory. These functions can be used in place of
> > > setup_memblock() in tests that need to simulate a NUMA system.
> > >
> > > setup_numa_memblock_generic():
> > > - allows for setting up a custom memory layout by specifying the amount
> > > of memory in each node, the number of nodes, and a factor that will be
> > > used to scale the memory in each node
> > >
> > > setup_numa_memblock():
> > > - allows for setting up a default memory layout
> > >
> > > Introduce constant MEM_FACTOR, which is used to scale the default memory
> > > layout based on MEM_SIZE.
> > >
> > > Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> > > 16 NUMA nodes.
> > >
> > > Signed-off-by: Rebecca Mckeever <[email protected]>
> > > ---
> > > .../testing/memblock/scripts/Makefile.include | 2 +-
> > > tools/testing/memblock/tests/common.c | 38 +++++++++++++++++++
> > > tools/testing/memblock/tests/common.h | 9 ++++-
> > > 3 files changed, 47 insertions(+), 2 deletions(-)
> > >
> > > diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
> > > index aa6d82d56a23..998281723590 100644
> > > --- a/tools/testing/memblock/scripts/Makefile.include
> > > +++ b/tools/testing/memblock/scripts/Makefile.include
> > > @@ -3,7 +3,7 @@
> > >
> > > # Simulate CONFIG_NUMA=y
> > > ifeq ($(NUMA), 1)
> > > - CFLAGS += -D CONFIG_NUMA
> > > + CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
> > > endif
> > >
> > > # Use 32 bit physical addresses.
> > > diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
> > > index eec6901081af..15d8767dc70c 100644
> > > --- a/tools/testing/memblock/tests/common.c
> > > +++ b/tools/testing/memblock/tests/common.c
> > > @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
> > >
> > > static int verbose;
> > >
> > > +static const phys_addr_t node_sizes[] = {
> > > + SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
> > > +};
> > > +
> > > /* sets global variable returned by movable_node_is_enabled() stub */
> > > bool movable_node_enabled;
> > >
> > > @@ -72,6 +76,40 @@ void setup_memblock(void)
> > > fill_memblock();
> > > }
> > >
> > > +/**
> > > + * setup_numa_memblock_generic:
> > > + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> > > + * dummy physical memory.
> > > + * @nodes: an array containing the amount of memory in each node
> > > + * @node_cnt: the size of @nodes
> > > + * @factor: a factor that will be used to scale the memory in each node
> > > + *
> > > + * The nids will be set to 0 through node_cnt - 1.
> > > + */
> > > +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> > > + int node_cnt, int factor)
> > > +{
> > > + phys_addr_t base;
> > > + int flags;
> > > +
> > > + reset_memblock_regions();
> > > + base = (phys_addr_t)memory_block.base;
> > > + flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> > > +
> > > + for (int i = 0; i < node_cnt; i++) {
> > > + phys_addr_t size = factor * nodes[i];
> >
> > I'm a bit lost why we need the factor if we already provide sizes in the
> > array.
> >
> > Can you enlighten me? :)
> >
> > Why can't we just stick to the sizes in the array?
> >
> Without the factor, some of the tests will break if we increase MEM_SIZE
> in the future (which we may need to do). I could rewrite them so that the
> factor is not needed, but I thought the code would be over-complicated if
> I did.

What if we make nodes[] to represent the fraction of the memory rather than
a node size? Then the factor won't be required.

> Thanks,
> Rebecca

--
Sincerely yours,
Mike.

2022-08-31 15:41:00

by Mike Rapoport

[permalink] [raw]
Subject: Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes

On Fri, Aug 19, 2022 at 02:05:31AM -0700, Rebecca Mckeever wrote:
> Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> for setting up a memory layout with multiple NUMA nodes in a previously
> allocated dummy physical memory. These functions can be used in place of
> setup_memblock() in tests that need to simulate a NUMA system.
>
> setup_numa_memblock_generic():
> - allows for setting up a custom memory layout by specifying the amount
> of memory in each node, the number of nodes, and a factor that will be
> used to scale the memory in each node
>
> setup_numa_memblock():
> - allows for setting up a default memory layout
>
> Introduce constant MEM_FACTOR, which is used to scale the default memory
> layout based on MEM_SIZE.
>
> Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> 16 NUMA nodes.
>
> Signed-off-by: Rebecca Mckeever <[email protected]>
> ---
> .../testing/memblock/scripts/Makefile.include | 2 +-
> tools/testing/memblock/tests/common.c | 38 +++++++++++++++++++
> tools/testing/memblock/tests/common.h | 9 ++++-
> 3 files changed, 47 insertions(+), 2 deletions(-)

...

> +/**
> + * setup_numa_memblock_generic:
> + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> + * dummy physical memory.
> + * @nodes: an array containing the amount of memory in each node
> + * @node_cnt: the size of @nodes
> + * @factor: a factor that will be used to scale the memory in each node
> + *
> + * The nids will be set to 0 through node_cnt - 1.
> + */
> +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> + int node_cnt, int factor)

I only had time for a quick look and it seems this function is never used
on its own.
Let's fold it into setup_numa_memblock() for now.

> +{
> + phys_addr_t base;
> + int flags;
> +
> + reset_memblock_regions();
> + base = (phys_addr_t)memory_block.base;
> + flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> +
> + for (int i = 0; i < node_cnt; i++) {
> + phys_addr_t size = factor * nodes[i];
> +
> + memblock_add_node(base, size, i, flags);
> + base += size;
> + }
> + fill_memblock();
> +}
> +
> +void setup_numa_memblock(void)
> +{
> + setup_numa_memblock_generic(node_sizes, NUMA_NODES, MEM_FACTOR);
> +}
> +

--
Sincerely yours,
Mike.

2022-09-01 08:15:02

by David Hildenbrand

[permalink] [raw]
Subject: Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes

On 31.08.22 05:49, Rebecca Mckeever wrote:
> On Tue, Aug 30, 2022 at 01:17:56PM +0200, David Hildenbrand wrote:
>> On 19.08.22 11:05, Rebecca Mckeever wrote:
>>> Add functions setup_numa_memblock_generic() and setup_numa_memblock()
>>> for setting up a memory layout with multiple NUMA nodes in a previously
>>> allocated dummy physical memory. These functions can be used in place of
>>> setup_memblock() in tests that need to simulate a NUMA system.
>>>
>>> setup_numa_memblock_generic():
>>> - allows for setting up a custom memory layout by specifying the amount
>>> of memory in each node, the number of nodes, and a factor that will be
>>> used to scale the memory in each node
>>>
>>> setup_numa_memblock():
>>> - allows for setting up a default memory layout
>>>
>>> Introduce constant MEM_FACTOR, which is used to scale the default memory
>>> layout based on MEM_SIZE.
>>>
>>> Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
>>> 16 NUMA nodes.
>>>
>>> Signed-off-by: Rebecca Mckeever <[email protected]>
>>> ---
>>> .../testing/memblock/scripts/Makefile.include | 2 +-
>>> tools/testing/memblock/tests/common.c | 38 +++++++++++++++++++
>>> tools/testing/memblock/tests/common.h | 9 ++++-
>>> 3 files changed, 47 insertions(+), 2 deletions(-)
>>>
>>> diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
>>> index aa6d82d56a23..998281723590 100644
>>> --- a/tools/testing/memblock/scripts/Makefile.include
>>> +++ b/tools/testing/memblock/scripts/Makefile.include
>>> @@ -3,7 +3,7 @@
>>>
>>> # Simulate CONFIG_NUMA=y
>>> ifeq ($(NUMA), 1)
>>> - CFLAGS += -D CONFIG_NUMA
>>> + CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
>>> endif
>>>
>>> # Use 32 bit physical addresses.
>>> diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
>>> index eec6901081af..15d8767dc70c 100644
>>> --- a/tools/testing/memblock/tests/common.c
>>> +++ b/tools/testing/memblock/tests/common.c
>>> @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
>>>
>>> static int verbose;
>>>
>>> +static const phys_addr_t node_sizes[] = {
>>> + SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
>>> +};
>>> +
>>> /* sets global variable returned by movable_node_is_enabled() stub */
>>> bool movable_node_enabled;
>>>
>>> @@ -72,6 +76,40 @@ void setup_memblock(void)
>>> fill_memblock();
>>> }
>>>
>>> +/**
>>> + * setup_numa_memblock_generic:
>>> + * Set up a memory layout with multiple NUMA nodes in a previously allocated
>>> + * dummy physical memory.
>>> + * @nodes: an array containing the amount of memory in each node
>>> + * @node_cnt: the size of @nodes
>>> + * @factor: a factor that will be used to scale the memory in each node
>>> + *
>>> + * The nids will be set to 0 through node_cnt - 1.
>>> + */
>>> +void setup_numa_memblock_generic(const phys_addr_t nodes[],
>>> + int node_cnt, int factor)
>>> +{
>>> + phys_addr_t base;
>>> + int flags;
>>> +
>>> + reset_memblock_regions();
>>> + base = (phys_addr_t)memory_block.base;
>>> + flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
>>> +
>>> + for (int i = 0; i < node_cnt; i++) {
>>> + phys_addr_t size = factor * nodes[i];
>>
>> I'm a bit lost why we need the factor if we already provide sizes in the
>> array.
>>
>> Can you enlighten me? :)
>>
>> Why can't we just stick to the sizes in the array?
>>
> Without the factor, some of the tests will break if we increase MEM_SIZE
> in the future (which we may need to do). I could rewrite them so that the
> factor is not needed, but I thought the code would be over-complicated if
> I did.

Independent of the suggestion from Mike, I wonder if we should really
care about (eventual) MEM_SIZE changes for now if not caring simplifies
the current code.

--
Thanks,

David / dhildenb

2022-09-01 23:00:15

by Rebecca Mckeever

[permalink] [raw]
Subject: Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes

On Wed, Aug 31, 2022 at 06:12:10PM +0300, Mike Rapoport wrote:
> On Tue, Aug 30, 2022 at 10:49:09PM -0500, Rebecca Mckeever wrote:
> > On Tue, Aug 30, 2022 at 01:17:56PM +0200, David Hildenbrand wrote:
> > > On 19.08.22 11:05, Rebecca Mckeever wrote:
> > > > Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> > > > for setting up a memory layout with multiple NUMA nodes in a previously
> > > > allocated dummy physical memory. These functions can be used in place of
> > > > setup_memblock() in tests that need to simulate a NUMA system.
> > > >
> > > > setup_numa_memblock_generic():
> > > > - allows for setting up a custom memory layout by specifying the amount
> > > > of memory in each node, the number of nodes, and a factor that will be
> > > > used to scale the memory in each node
> > > >
> > > > setup_numa_memblock():
> > > > - allows for setting up a default memory layout
> > > >
> > > > Introduce constant MEM_FACTOR, which is used to scale the default memory
> > > > layout based on MEM_SIZE.
> > > >
> > > > Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> > > > 16 NUMA nodes.
> > > >
> > > > Signed-off-by: Rebecca Mckeever <[email protected]>
> > > > ---
> > > > .../testing/memblock/scripts/Makefile.include | 2 +-
> > > > tools/testing/memblock/tests/common.c | 38 +++++++++++++++++++
> > > > tools/testing/memblock/tests/common.h | 9 ++++-
> > > > 3 files changed, 47 insertions(+), 2 deletions(-)
> > > >
> > > > diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
> > > > index aa6d82d56a23..998281723590 100644
> > > > --- a/tools/testing/memblock/scripts/Makefile.include
> > > > +++ b/tools/testing/memblock/scripts/Makefile.include
> > > > @@ -3,7 +3,7 @@
> > > >
> > > > # Simulate CONFIG_NUMA=y
> > > > ifeq ($(NUMA), 1)
> > > > - CFLAGS += -D CONFIG_NUMA
> > > > + CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
> > > > endif
> > > >
> > > > # Use 32 bit physical addresses.
> > > > diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
> > > > index eec6901081af..15d8767dc70c 100644
> > > > --- a/tools/testing/memblock/tests/common.c
> > > > +++ b/tools/testing/memblock/tests/common.c
> > > > @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
> > > >
> > > > static int verbose;
> > > >
> > > > +static const phys_addr_t node_sizes[] = {
> > > > + SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
> > > > +};
> > > > +
> > > > /* sets global variable returned by movable_node_is_enabled() stub */
> > > > bool movable_node_enabled;
> > > >
> > > > @@ -72,6 +76,40 @@ void setup_memblock(void)
> > > > fill_memblock();
> > > > }
> > > >
> > > > +/**
> > > > + * setup_numa_memblock_generic:
> > > > + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> > > > + * dummy physical memory.
> > > > + * @nodes: an array containing the amount of memory in each node
> > > > + * @node_cnt: the size of @nodes
> > > > + * @factor: a factor that will be used to scale the memory in each node
> > > > + *
> > > > + * The nids will be set to 0 through node_cnt - 1.
> > > > + */
> > > > +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> > > > + int node_cnt, int factor)
> > > > +{
> > > > + phys_addr_t base;
> > > > + int flags;
> > > > +
> > > > + reset_memblock_regions();
> > > > + base = (phys_addr_t)memory_block.base;
> > > > + flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> > > > +
> > > > + for (int i = 0; i < node_cnt; i++) {
> > > > + phys_addr_t size = factor * nodes[i];
> > >
> > > I'm a bit lost why we need the factor if we already provide sizes in the
> > > array.
> > >
> > > Can you enlighten me? :)
> > >
> > > Why can't we just stick to the sizes in the array?
> > >
> > Without the factor, some of the tests will break if we increase MEM_SIZE
> > in the future (which we may need to do). I could rewrite them so that the
> > factor is not needed, but I thought the code would be over-complicated if
> > I did.
>
> What if we make nodes[] to represent the fraction of the memory rather than
> a node size? Then the factor won't be required.
>
I think that will work. I'll try it.

> > Thanks,
> > Rebecca
>
> --
> Sincerely yours,
> Mike.

Thanks,
Rebecca

2022-09-02 00:43:03

by Rebecca Mckeever

[permalink] [raw]
Subject: Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes

On Thu, Sep 01, 2022 at 10:06:48AM +0200, David Hildenbrand wrote:
> On 31.08.22 05:49, Rebecca Mckeever wrote:
> > On Tue, Aug 30, 2022 at 01:17:56PM +0200, David Hildenbrand wrote:
> >> On 19.08.22 11:05, Rebecca Mckeever wrote:
> >>> Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> >>> for setting up a memory layout with multiple NUMA nodes in a previously
> >>> allocated dummy physical memory. These functions can be used in place of
> >>> setup_memblock() in tests that need to simulate a NUMA system.
> >>>
> >>> setup_numa_memblock_generic():
> >>> - allows for setting up a custom memory layout by specifying the amount
> >>> of memory in each node, the number of nodes, and a factor that will be
> >>> used to scale the memory in each node
> >>>
> >>> setup_numa_memblock():
> >>> - allows for setting up a default memory layout
> >>>
> >>> Introduce constant MEM_FACTOR, which is used to scale the default memory
> >>> layout based on MEM_SIZE.
> >>>
> >>> Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> >>> 16 NUMA nodes.
> >>>
> >>> Signed-off-by: Rebecca Mckeever <[email protected]>
> >>> ---
> >>> .../testing/memblock/scripts/Makefile.include | 2 +-
> >>> tools/testing/memblock/tests/common.c | 38 +++++++++++++++++++
> >>> tools/testing/memblock/tests/common.h | 9 ++++-
> >>> 3 files changed, 47 insertions(+), 2 deletions(-)
> >>>
> >>> diff --git a/tools/testing/memblock/scripts/Makefile.include b/tools/testing/memblock/scripts/Makefile.include
> >>> index aa6d82d56a23..998281723590 100644
> >>> --- a/tools/testing/memblock/scripts/Makefile.include
> >>> +++ b/tools/testing/memblock/scripts/Makefile.include
> >>> @@ -3,7 +3,7 @@
> >>>
> >>> # Simulate CONFIG_NUMA=y
> >>> ifeq ($(NUMA), 1)
> >>> - CFLAGS += -D CONFIG_NUMA
> >>> + CFLAGS += -D CONFIG_NUMA -D CONFIG_NODES_SHIFT=4
> >>> endif
> >>>
> >>> # Use 32 bit physical addresses.
> >>> diff --git a/tools/testing/memblock/tests/common.c b/tools/testing/memblock/tests/common.c
> >>> index eec6901081af..15d8767dc70c 100644
> >>> --- a/tools/testing/memblock/tests/common.c
> >>> +++ b/tools/testing/memblock/tests/common.c
> >>> @@ -34,6 +34,10 @@ static const char * const help_opts[] = {
> >>>
> >>> static int verbose;
> >>>
> >>> +static const phys_addr_t node_sizes[] = {
> >>> + SZ_4K, SZ_1K, SZ_2K, SZ_2K, SZ_1K, SZ_1K, SZ_4K, SZ_1K
> >>> +};
> >>> +
> >>> /* sets global variable returned by movable_node_is_enabled() stub */
> >>> bool movable_node_enabled;
> >>>
> >>> @@ -72,6 +76,40 @@ void setup_memblock(void)
> >>> fill_memblock();
> >>> }
> >>>
> >>> +/**
> >>> + * setup_numa_memblock_generic:
> >>> + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> >>> + * dummy physical memory.
> >>> + * @nodes: an array containing the amount of memory in each node
> >>> + * @node_cnt: the size of @nodes
> >>> + * @factor: a factor that will be used to scale the memory in each node
> >>> + *
> >>> + * The nids will be set to 0 through node_cnt - 1.
> >>> + */
> >>> +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> >>> + int node_cnt, int factor)
> >>> +{
> >>> + phys_addr_t base;
> >>> + int flags;
> >>> +
> >>> + reset_memblock_regions();
> >>> + base = (phys_addr_t)memory_block.base;
> >>> + flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> >>> +
> >>> + for (int i = 0; i < node_cnt; i++) {
> >>> + phys_addr_t size = factor * nodes[i];
> >>
> >> I'm a bit lost why we need the factor if we already provide sizes in the
> >> array.
> >>
> >> Can you enlighten me? :)
> >>
> >> Why can't we just stick to the sizes in the array?
> >>
> > Without the factor, some of the tests will break if we increase MEM_SIZE
> > in the future (which we may need to do). I could rewrite them so that the
> > factor is not needed, but I thought the code would be over-complicated if
> > I did.
>
> Independent of the suggestion from Mike, I wonder if we should really
> care about (eventual) MEM_SIZE changes for now if not caring simplifies
> the current code.
>
Maybe not. I'm going to try Mike's suggestion, but I will keep this in
mind if the code seems too complicated.

> --
> Thanks,
>
> David / dhildenb
>
Thanks,
Rebecca

2022-09-02 00:48:00

by Rebecca Mckeever

[permalink] [raw]
Subject: Re: [PATCH v2 1/4] memblock tests: add simulation of physical memory with multiple NUMA nodes

On Wed, Aug 31, 2022 at 06:15:41PM +0300, Mike Rapoport wrote:
> On Fri, Aug 19, 2022 at 02:05:31AM -0700, Rebecca Mckeever wrote:
> > Add functions setup_numa_memblock_generic() and setup_numa_memblock()
> > for setting up a memory layout with multiple NUMA nodes in a previously
> > allocated dummy physical memory. These functions can be used in place of
> > setup_memblock() in tests that need to simulate a NUMA system.
> >
> > setup_numa_memblock_generic():
> > - allows for setting up a custom memory layout by specifying the amount
> > of memory in each node, the number of nodes, and a factor that will be
> > used to scale the memory in each node
> >
> > setup_numa_memblock():
> > - allows for setting up a default memory layout
> >
> > Introduce constant MEM_FACTOR, which is used to scale the default memory
> > layout based on MEM_SIZE.
> >
> > Set CONFIG_NODES_SHIFT to 4 when building with NUMA=1 to allow for up to
> > 16 NUMA nodes.
> >
> > Signed-off-by: Rebecca Mckeever <[email protected]>
> > ---
> > .../testing/memblock/scripts/Makefile.include | 2 +-
> > tools/testing/memblock/tests/common.c | 38 +++++++++++++++++++
> > tools/testing/memblock/tests/common.h | 9 ++++-
> > 3 files changed, 47 insertions(+), 2 deletions(-)
>
> ...
>
> > +/**
> > + * setup_numa_memblock_generic:
> > + * Set up a memory layout with multiple NUMA nodes in a previously allocated
> > + * dummy physical memory.
> > + * @nodes: an array containing the amount of memory in each node
> > + * @node_cnt: the size of @nodes
> > + * @factor: a factor that will be used to scale the memory in each node
> > + *
> > + * The nids will be set to 0 through node_cnt - 1.
> > + */
> > +void setup_numa_memblock_generic(const phys_addr_t nodes[],
> > + int node_cnt, int factor)
>
> I only had time for a quick look and it seems this function is never used
> on its own.
> Let's fold it into setup_numa_memblock() for now.
>
Okay, will do.

> > +{
> > + phys_addr_t base;
> > + int flags;
> > +
> > + reset_memblock_regions();
> > + base = (phys_addr_t)memory_block.base;
> > + flags = (movable_node_is_enabled()) ? MEMBLOCK_NONE : MEMBLOCK_HOTPLUG;
> > +
> > + for (int i = 0; i < node_cnt; i++) {
> > + phys_addr_t size = factor * nodes[i];
> > +
> > + memblock_add_node(base, size, i, flags);
> > + base += size;
> > + }
> > + fill_memblock();
> > +}
> > +
> > +void setup_numa_memblock(void)
> > +{
> > + setup_numa_memblock_generic(node_sizes, NUMA_NODES, MEM_FACTOR);
> > +}
> > +
>
> --
> Sincerely yours,
> Mike.

Thanks,
Rebecca