Kdamond is implemented as a periodical split-merge pattern, which will
create and destroy regions possibly on high frequency(hundreds or even
thousands of per sec), depending on number of regions and aggregation
period. In that case, kmalloc and kfree could bring considerable overhead
to system, which can be improved by private kmem cache.
Signed-off-by: Dawei Li <[email protected]>
---
mm/damon/core.c | 20 ++++++++++++++++++--
1 file changed, 18 insertions(+), 2 deletions(-)
diff --git a/mm/damon/core.c b/mm/damon/core.c
index 029ae384e6ff..0b1eb945c68a 100644
--- a/mm/damon/core.c
+++ b/mm/damon/core.c
@@ -29,6 +29,8 @@ static bool running_exclusive_ctxs;
static DEFINE_MUTEX(damon_ops_lock);
static struct damon_operations damon_registered_ops[NR_DAMON_OPS];
+static struct kmem_cache *damon_region_cache __ro_after_init;
+
/* Should be called under damon_ops_lock with id smaller than NR_DAMON_OPS */
static bool __damon_is_registered_ops(enum damon_ops_id id)
{
@@ -119,7 +121,7 @@ struct damon_region *damon_new_region(unsigned long start, unsigned long end)
{
struct damon_region *region;
- region = kmalloc(sizeof(*region), GFP_KERNEL);
+ region = kmem_cache_alloc(damon_region_cache, GFP_KERNEL);
if (!region)
return NULL;
@@ -148,7 +150,7 @@ static void damon_del_region(struct damon_region *r, struct damon_target *t)
static void damon_free_region(struct damon_region *r)
{
- kfree(r);
+ kmem_cache_free(damon_region_cache, r);
}
void damon_destroy_region(struct damon_region *r, struct damon_target *t)
@@ -1279,4 +1281,18 @@ bool damon_find_biggest_system_ram(unsigned long *start, unsigned long *end)
return true;
}
+static int __init damon_init(void)
+{
+ damon_region_cache = kmem_cache_create("damon_region_cache", sizeof(struct damon_region),
+ 0, 0, NULL);
+ if (unlikely(!damon_region_cache)) {
+ pr_err("creating damon_region_cache fails\n");
+ return -ENOMEM;
+ }
+
+ return 0;
+}
+
+subsys_initcall(damon_init);
+
#include "core-test.h"
--
2.25.1
On Mon, 12 Sep 2022 22:39:03 +0800 Dawei Li <[email protected]> wrote:
> Kdamond is implemented as a periodical split-merge pattern, which will
> create and destroy regions possibly on high frequency(hundreds or even
> thousands of per sec), depending on number of regions and aggregation
> period. In that case, kmalloc and kfree could bring considerable overhead
> to system, which can be improved by private kmem cache.
>
A dedicated slab cache should be faster and will also consume less
memory, due to better packing into the underlying pages. So I redid
the changelog thusly:
: Kdamond is implemented as a periodical split-merge pattern, which will
: create and destroy regions possibly at high frequency (hundreds or even
: thousands of per sec), depending on the number of regions and aggregation
: period. In that case, kmalloc and kfree could bring speed and space
: overheads, which can be improved by using a private kmem cache.
> +static int __init damon_init(void)
> +{
> + damon_region_cache = kmem_cache_create("damon_region_cache", sizeof(struct damon_region),
Should be able to use just
damon_region_cache = KMEM_CACHE(damon_region, 0);
here. Please test that and send along a fixup patch?