From: Ma Wupeng <[email protected]>
System may have little normal zone memory and huge movable memory in the
following situations:
- for system with kernelcore=nn% or kernelcore=mirror, movable zone will
be added and movable zone is bigger than normal zone in most cases.
- system with movable nodes, they will have multiple numa nodes with
only movable zone and movable zone will have plenty of memory.
Since kernel/driver can only use memory from non-movable zone in most
cases, normal zone need to increase its watermark to reserve more memory.
However, current watermark_scale_factor is used to control all zones
at once and can't be set separately. To reserve memory in non-movable
zones, the watermark is increased in movable zones as well. Which will
lead to inefficient kswapd.
To solve this problem, per zone watermark is introduced to tune each zone's
watermark separately. This can bring the following advantages:
- each zone can set its own watermark which bring flexibility
- lead to more efficient kswapd if this watermark is set fine
Here is real watermark data in my qemu machine(with THP disabled).
With watermark_scale_factor = 10, there is only 1440(772-68+807-71)
pages(5.76M) reserved for a system with 96G of memory. However if the
watermark is set to 100, the movable zone's watermark increased to
231908(93M), which is too much.
This situation is even worse with 32G of normal zone memory and 1T of
movable zone memory.
Modified | Vanilla wm_factor = 10 | Vanilla wm_factor = 30
Node 0, zone DMA | Node 0, zone DMA | Node 0, zone DMA
min 68 | min 68 | min 68
low 7113 | low 772 | low 7113
high **14158** | high **1476** | high **14158**
Node 0, zone Normal | Node 0, zone Normal | Node 0, zone Normal
min 71 | min 71 | min 71
low 7438 | low 807 | low 7438
high 14805 | high 1543 | high 14805
Node 0, zone Movable | Node 0, zone Movable | Node 0, zone Movable
min 1455 | min 1455 | min 1455
low 16388 | low 16386 | low 150787
high **31321** | high **31317** | high **300119**
Node 1, zone Movable | Node 1, zone Movable | Node 1, zone Movable
min 804 | min 804 | min 804
low 9061 | low 9061 | low 83379
high **17318** | high **17318** | high **165954**
With the modified per zone watermark_scale_factor, only dma/normal zone
will increase its watermark via the following command which the huge
movable zone stay the same.
% echo 100 100 100 10 > /proc/sys/vm/watermark_scale_factor
The reason to disable THP is khugepaged_min_free_kbytes_update() will
update min watermark.
Signed-off-by: Ma Wupeng <[email protected]>
Reported-by: kernel test robot <[email protected]>
---
Documentation/admin-guide/sysctl/vm.rst | 6 ++++
include/linux/mm.h | 2 +-
kernel/sysctl.c | 2 --
mm/page_alloc.c | 37 ++++++++++++++++++++-----
4 files changed, 37 insertions(+), 10 deletions(-)
diff --git a/Documentation/admin-guide/sysctl/vm.rst b/Documentation/admin-guide/sysctl/vm.rst
index 9b833e439f09..ec240aa45322 100644
--- a/Documentation/admin-guide/sysctl/vm.rst
+++ b/Documentation/admin-guide/sysctl/vm.rst
@@ -1002,6 +1002,12 @@ that the number of free pages kswapd maintains for latency reasons is
too small for the allocation bursts occurring in the system. This knob
can then be used to tune kswapd aggressiveness accordingly.
+The watermark_scale_factor is an array. You can set each zone's watermark
+separately and can be seen by reading this file::
+
+ % cat /proc/sys/vm/watermark_scale_factor
+ 10 10 10 10
+
zone_reclaim_mode
=================
diff --git a/include/linux/mm.h b/include/linux/mm.h
index 21f8b27bd9fd..b291c795f9db 100644
--- a/include/linux/mm.h
+++ b/include/linux/mm.h
@@ -2532,7 +2532,7 @@ extern void setup_per_cpu_pageset(void);
/* page_alloc.c */
extern int min_free_kbytes;
extern int watermark_boost_factor;
-extern int watermark_scale_factor;
+extern int watermark_scale_factor[MAX_NR_ZONES];
extern bool arch_has_descending_max_zone_pfns(void);
/* nommu.c */
diff --git a/kernel/sysctl.c b/kernel/sysctl.c
index 205d605cacc5..d16d06c71e5a 100644
--- a/kernel/sysctl.c
+++ b/kernel/sysctl.c
@@ -2251,8 +2251,6 @@ static struct ctl_table vm_table[] = {
.maxlen = sizeof(watermark_scale_factor),
.mode = 0644,
.proc_handler = watermark_scale_factor_sysctl_handler,
- .extra1 = SYSCTL_ONE,
- .extra2 = SYSCTL_THREE_THOUSAND,
},
{
.procname = "percpu_pagelist_high_fraction",
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index f1e4474879f1..7a6ac3b4ebb6 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -421,7 +421,6 @@ compound_page_dtor * const compound_page_dtors[NR_COMPOUND_DTORS] = {
int min_free_kbytes = 1024;
int user_min_free_kbytes = -1;
int watermark_boost_factor __read_mostly = 15000;
-int watermark_scale_factor = 10;
static unsigned long nr_kernel_pages __initdata;
static unsigned long nr_all_pages __initdata;
@@ -449,6 +448,20 @@ EXPORT_SYMBOL(nr_online_nodes);
int page_group_by_mobility_disabled __read_mostly;
+int watermark_scale_factor[MAX_NR_ZONES] = {
+#ifdef CONFIG_ZONE_DMA
+ [ZONE_DMA] = 10,
+#endif
+#ifdef CONFIG_ZONE_DMA32
+ [ZONE_DMA32] = 10,
+#endif
+ [ZONE_NORMAL] = 10,
+#ifdef CONFIG_HIGHMEM
+ [ZONE_HIGHMEM] = 10,
+#endif
+ [ZONE_MOVABLE] = 10,
+};
+
#ifdef CONFIG_DEFERRED_STRUCT_PAGE_INIT
/*
* During boot we initialize deferred pages on-demand, as needed, but once
@@ -8643,6 +8656,7 @@ static void __setup_per_zone_wmarks(void)
}
for_each_zone(zone) {
+ int zone_wm_factor;
u64 tmp;
spin_lock_irqsave(&zone->lock, flags);
@@ -8676,9 +8690,10 @@ static void __setup_per_zone_wmarks(void)
* scale factor in proportion to available memory, but
* ensure a minimum size on small systems.
*/
+ zone_wm_factor = watermark_scale_factor[zone_idx(zone)];
tmp = max_t(u64, tmp >> 2,
- mult_frac(zone_managed_pages(zone),
- watermark_scale_factor, 10000));
+ mult_frac(zone_managed_pages(zone), zone_wm_factor,
+ 10000));
zone->watermark_boost = 0;
zone->_watermark[WMARK_LOW] = min_wmark_pages(zone) + tmp;
@@ -8798,11 +8813,19 @@ int min_free_kbytes_sysctl_handler(struct ctl_table *table, int write,
int watermark_scale_factor_sysctl_handler(struct ctl_table *table, int write,
void *buffer, size_t *length, loff_t *ppos)
{
- int rc;
+ int i;
- rc = proc_dointvec_minmax(table, write, buffer, length, ppos);
- if (rc)
- return rc;
+ proc_dointvec_minmax(table, write, buffer, length, ppos);
+
+ /*
+ * The unit is in fractions of 10,000. The default value of 10
+ * means the distances between watermarks are 0.1% of the available
+ * memory in the node/system. The maximum value is 3000, or 30% of
+ * memory.
+ */
+ for (i = 0; i < MAX_NR_ZONES; i++)
+ watermark_scale_factor[i] =
+ clamp(watermark_scale_factor[i], 1, 3000);
if (write)
setup_per_zone_wmarks();
--
2.25.1
On Mon, Sep 05, 2022 at 11:28:58AM +0800, Wupeng Ma wrote:
> diff --git a/kernel/sysctl.c b/kernel/sysctl.c
> index 205d605cacc5..d16d06c71e5a 100644
> --- a/kernel/sysctl.c
> +++ b/kernel/sysctl.c
> @@ -2251,8 +2251,6 @@ static struct ctl_table vm_table[] = {
> .maxlen = sizeof(watermark_scale_factor),
> .mode = 0644,
> .proc_handler = watermark_scale_factor_sysctl_handler,
> - .extra1 = SYSCTL_ONE,
> - .extra2 = SYSCTL_THREE_THOUSAND,
> },
> {
> .procname = "percpu_pagelist_high_fraction",
Please move the sysctl from kernel/sysctl.c to mm/page_alloc.c while
at it, you can git log the kernel/sysctl.c for prior moves and for
the motivations. No need to keep expanding on the existing table.
Luis
On 2022/9/7 2:23, Luis Chamberlain wrote:
> On Mon, Sep 05, 2022 at 11:28:58AM +0800, Wupeng Ma wrote:
>> diff --git a/kernel/sysctl.c b/kernel/sysctl.c
>> index 205d605cacc5..d16d06c71e5a 100644
>> --- a/kernel/sysctl.c
>> +++ b/kernel/sysctl.c
>> @@ -2251,8 +2251,6 @@ static struct ctl_table vm_table[] = {
>> .maxlen = sizeof(watermark_scale_factor),
>> .mode = 0644,
>> .proc_handler = watermark_scale_factor_sysctl_handler,
>> - .extra1 = SYSCTL_ONE,
>> - .extra2 = SYSCTL_THREE_THOUSAND,
>> },
>> {
>> .procname = "percpu_pagelist_high_fraction",
>
> Please move the sysctl from kernel/sysctl.c to mm/page_alloc.c while
> at it, you can git log the kernel/sysctl.c for prior moves and for
> the motivations. No need to keep expanding on the existing table.
>
> Luis
Ok.
I will move this sysctl to mm/page_alloc.c next version.
On Mon, 2022-09-05 at 11:28 +0800, Wupeng Ma wrote:
> From: Ma Wupeng <[email protected]>
>
> System may have little normal zone memory and huge movable memory in
> the
> following situations:
> - for system with kernelcore=nn% or kernelcore=mirror, movable zone
> will
> be added and movable zone is bigger than normal zone in most cases.
> - system with movable nodes, they will have multiple numa nodes
> with
> only movable zone and movable zone will have plenty of memory.
>
> Since kernel/driver can only use memory from non-movable zone in most
> cases, normal zone need to increase its watermark to reserve more
> memory.
>
> However, current watermark_scale_factor is used to control all zones
> at once and can't be set separately. To reserve memory in non-movable
> zones, the watermark is increased in movable zones as well. Which
> will
> lead to inefficient kswapd.
Similar issues happen on systems with large amount of memory (1 TB or
more). Most of the memory ends up in Normal zone while DMA and DMA32
zones have very little memory. For the watermark scale factor that
results in a reasonable low watermark for Normal zone, low watermark
can be too low for DMA and DMA32 zones. Being able to tune those
watermarks independently can be helpful. The trouble with this approach
is it introduces another level of complexity to tuning knobs with no
clear guidelines for system admins on how to tune these. I already see
multiple customers struggling with setting simple min_free_kbytes or
watermark_scale_factor. Once we add the complexity of per zone
watermark scale factor, it only gets to be a more daunting task for
system admins. NUMA systems with multiple nodes with hot-pluggable
memory can have sizeable number of zones.
I see the usefulness of per-zone watermark but what guidance would you
give to a sysadmin on how to set these values for their systems?
Thanks,
Khalid
>
> To solve this problem, per zone watermark is introduced to tune each
> zone's
> watermark separately. This can bring the following advantages:
> - each zone can set its own watermark which bring flexibility
> - lead to more efficient kswapd if this watermark is set fine
>
> Here is real watermark data in my qemu machine(with THP disabled).
>
> With watermark_scale_factor = 10, there is only 1440(772-68+807-71)
> pages(5.76M) reserved for a system with 96G of memory. However if the
> watermark is set to 100, the movable zone's watermark increased to
> 231908(93M), which is too much.
> This situation is even worse with 32G of normal zone memory and 1T of
> movable zone memory.
>
> Modified | Vanilla wm_factor = 10 | Vanilla wm_factor =
> 30
> Node 0, zone DMA | Node 0, zone DMA | Node 0, zone
> DMA
> min 68 | min 68 | min 68
> low 7113 | low 772 | low
> 7113
> high **14158** | high **1476** | high
> **14158**
> Node 0, zone Normal | Node 0, zone Normal | Node 0, zone
> Normal
> min 71 | min 71 | min 71
> low 7438 | low 807 | low
> 7438
> high 14805 | high 1543 | high
> 14805
> Node 0, zone Movable | Node 0, zone Movable | Node 0, zone
> Movable
> min 1455 | min 1455 | min
> 1455
> low 16388 | low 16386 | low
> 150787
> high **31321** | high **31317** | high
> **300119**
> Node 1, zone Movable | Node 1, zone Movable | Node 1, zone
> Movable
> min 804 | min 804 | min
> 804
> low 9061 | low 9061 | low
> 83379
> high **17318** | high **17318** | high
> **165954**
>
> With the modified per zone watermark_scale_factor, only dma/normal
> zone
> will increase its watermark via the following command which the huge
> movable zone stay the same.
>
> % echo 100 100 100 10 > /proc/sys/vm/watermark_scale_factor
>
> The reason to disable THP is khugepaged_min_free_kbytes_update() will
> update min watermark.
>
> Signed-off-by: Ma Wupeng <[email protected]>
> Reported-by: kernel test robot <[email protected]>
> ---
> Documentation/admin-guide/sysctl/vm.rst | 6 ++++
> include/linux/mm.h | 2 +-
> kernel/sysctl.c | 2 --
> mm/page_alloc.c | 37 ++++++++++++++++++++---
> --
> 4 files changed, 37 insertions(+), 10 deletions(-)
>
> diff --git a/Documentation/admin-guide/sysctl/vm.rst
> b/Documentation/admin-guide/sysctl/vm.rst
> index 9b833e439f09..ec240aa45322 100644
> --- a/Documentation/admin-guide/sysctl/vm.rst
> +++ b/Documentation/admin-guide/sysctl/vm.rst
> @@ -1002,6 +1002,12 @@ that the number of free pages kswapd maintains
> for latency reasons is
> too small for the allocation bursts occurring in the system. This
> knob
> can then be used to tune kswapd aggressiveness accordingly.
>
> +The watermark_scale_factor is an array. You can set each zone's
> watermark
> +separately and can be seen by reading this file::
> +
> + % cat /proc/sys/vm/watermark_scale_factor
> + 10 10 10 10
> +
>
> zone_reclaim_mode
> =================
> diff --git a/include/linux/mm.h b/include/linux/mm.h
> index 21f8b27bd9fd..b291c795f9db 100644
> --- a/include/linux/mm.h
> +++ b/include/linux/mm.h
> @@ -2532,7 +2532,7 @@ extern void setup_per_cpu_pageset(void);
> /* page_alloc.c */
> extern int min_free_kbytes;
> extern int watermark_boost_factor;
> -extern int watermark_scale_factor;
> +extern int watermark_scale_factor[MAX_NR_ZONES];
> extern bool arch_has_descending_max_zone_pfns(void);
>
> /* nommu.c */
> diff --git a/kernel/sysctl.c b/kernel/sysctl.c
> index 205d605cacc5..d16d06c71e5a 100644
> --- a/kernel/sysctl.c
> +++ b/kernel/sysctl.c
> @@ -2251,8 +2251,6 @@ static struct ctl_table vm_table[] = {
> .maxlen = sizeof(watermark_scale_factor),
> .mode = 0644,
> .proc_handler =
> watermark_scale_factor_sysctl_handler,
> - .extra1 = SYSCTL_ONE,
> - .extra2 = SYSCTL_THREE_THOUSAND,
> },
> {
> .procname = "percpu_pagelist_high_fraction",
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index f1e4474879f1..7a6ac3b4ebb6 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -421,7 +421,6 @@ compound_page_dtor * const
> compound_page_dtors[NR_COMPOUND_DTORS] = {
> int min_free_kbytes = 1024;
> int user_min_free_kbytes = -1;
> int watermark_boost_factor __read_mostly = 15000;
> -int watermark_scale_factor = 10;
>
> static unsigned long nr_kernel_pages __initdata;
> static unsigned long nr_all_pages __initdata;
> @@ -449,6 +448,20 @@ EXPORT_SYMBOL(nr_online_nodes);
>
> int page_group_by_mobility_disabled __read_mostly;
>
> +int watermark_scale_factor[MAX_NR_ZONES] = {
> +#ifdef CONFIG_ZONE_DMA
> + [ZONE_DMA] = 10,
> +#endif
> +#ifdef CONFIG_ZONE_DMA32
> + [ZONE_DMA32] = 10,
> +#endif
> + [ZONE_NORMAL] = 10,
> +#ifdef CONFIG_HIGHMEM
> + [ZONE_HIGHMEM] = 10,
> +#endif
> + [ZONE_MOVABLE] = 10,
> +};
> +
> #ifdef CONFIG_DEFERRED_STRUCT_PAGE_INIT
> /*
> * During boot we initialize deferred pages on-demand, as needed,
> but once
> @@ -8643,6 +8656,7 @@ static void __setup_per_zone_wmarks(void)
> }
>
> for_each_zone(zone) {
> + int zone_wm_factor;
> u64 tmp;
>
> spin_lock_irqsave(&zone->lock, flags);
> @@ -8676,9 +8690,10 @@ static void __setup_per_zone_wmarks(void)
> * scale factor in proportion to available memory,
> but
> * ensure a minimum size on small systems.
> */
> + zone_wm_factor =
> watermark_scale_factor[zone_idx(zone)];
> tmp = max_t(u64, tmp >> 2,
> - mult_frac(zone_managed_pages(zone),
> - watermark_scale_factor,
> 10000));
> + mult_frac(zone_managed_pages(zone),
> zone_wm_factor,
> + 10000));
>
> zone->watermark_boost = 0;
> zone->_watermark[WMARK_LOW] = min_wmark_pages(zone)
> + tmp;
> @@ -8798,11 +8813,19 @@ int min_free_kbytes_sysctl_handler(struct
> ctl_table *table, int write,
> int watermark_scale_factor_sysctl_handler(struct ctl_table *table,
> int write,
> void *buffer, size_t *length, loff_t *ppos)
> {
> - int rc;
> + int i;
>
> - rc = proc_dointvec_minmax(table, write, buffer, length,
> ppos);
> - if (rc)
> - return rc;
> + proc_dointvec_minmax(table, write, buffer, length, ppos);
> +
> + /*
> + * The unit is in fractions of 10,000. The default value of
> 10
> + * means the distances between watermarks are 0.1% of the
> available
> + * memory in the node/system. The maximum value is 3000, or
> 30% of
> + * memory.
> + */
> + for (i = 0; i < MAX_NR_ZONES; i++)
> + watermark_scale_factor[i] =
> + clamp(watermark_scale_factor[i], 1, 3000);
>
> if (write)
> setup_per_zone_wmarks();
On Tue, 2022-09-13 at 10:09 +0800, mawupeng wrote:
>
>
> On 2022/9/10 5:41, Khalid Aziz wrote:
> > On Mon, 2022-09-05 at 11:28 +0800, Wupeng Ma wrote:
> > > From: Ma Wupeng <[email protected]>
> > >
> > > System may have little normal zone memory and huge movable memory
> > > in
> > > the
> > > following situations:
> > > - for system with kernelcore=nn% or kernelcore=mirror, movable
> > > zone
> > > will
> > > be added and movable zone is bigger than normal zone in most
> > > cases.
> > > - system with movable nodes, they will have multiple numa nodes
> > > with
> > > only movable zone and movable zone will have plenty of memory.
> > >
> > > Since kernel/driver can only use memory from non-movable zone in
> > > most
> > > cases, normal zone need to increase its watermark to reserve more
> > > memory.
> > >
> > > However, current watermark_scale_factor is used to control all
> > > zones
> > > at once and can't be set separately. To reserve memory in non-
> > > movable
> > > zones, the watermark is increased in movable zones as well. Which
> > > will
> > > lead to inefficient kswapd.
> >
> > Similar issues happen on systems with large amount of memory (1 TB
> > or
> > more). Most of the memory ends up in Normal zone while DMA and
> > DMA32
> > zones have very little memory. For the watermark scale factor that
> > results in a reasonable low watermark for Normal zone, low
> > watermark
> > can be too low for DMA and DMA32 zones. Being able to tune those
> > watermarks independently can be helpful. The trouble with this
> > approach
> > is it introduces another level of complexity to tuning knobs with
> > no
> > clear guidelines for system admins on how to tune these. I already
> > see
> > multiple customers struggling with setting simple min_free_kbytes
> > or
> > watermark_scale_factor. Once we add the complexity of per zone
> > watermark scale factor, it only gets to be a more daunting task for
> > system admins. NUMA systems with multiple nodes with hot-pluggable
> > memory can have sizeable number of zones.
> >
> > I see the usefulness of per-zone watermark but what guidance would
> > you
> > give to a sysadmin on how to set these values for their systems?
> >
> > Thanks,
> > Khalid
>
> Thanks for your reply.
>
> Like the vanilla watermark_scale_factor, this per-zone one just
> introduce
> the ability to tune this watermark separately. The overall usage of
> the
> per-zone one is similar to to vanilla one.
>
> Memory allocation below the low watermark will awake the kswapd and
> drop
> cache or swap out until the memory's watermark reach the high
> watermark.
> Since memory allocation can get their page in fast path(watermark
> above
> low), the following memory allocation can benefit from this
> background
> kswapd.
>
> Since memory below min is used for emergency and below min means
> system
> is low on memory, And higher zones(movable) allocations indirectly
> pin
> pages in lower zones(page table..), boost the watermark for the lower
> zones can reserve enough memory for kernel/driver to use.
>
> So bigger watermark for lower zones if the higher zones is huge may
> be
> the solution.
I understand how it works. My question is how would a system admin look
at the workload on their system and know what would be appropriate
values to set for each zone. "bigger watermark for lower zones if the
higher zones is huge" is one way to look at this but is vague and how
would a system admin arrive at those numbers, how would they know if
they got it right? Documentation for watermark_scale_factor says "A
high rate of threads entering direct reclaim (allocstall) or kswapd
going to sleep prematurely (kswapd_low_wmark_hit_quickly) can indicate
that the number of free pages kswapd maintains for latency reasons is
too small for the allocation bursts occurring in the system. This knob
can then be used to tune kswapd aggressiveness accordingly." That
provides some guidance on where to observe problems that can possibly
be mitigated with watermark_scale_factor. It is still vague about how
much to tune the knob and in which direction but at least one can
observe the effects of tuning the knob. With the number of knobs going
up, how would you know which zone(s) to tune.
Thanks,
Khalid
>
> >
> >
> > >
> > > To solve this problem, per zone watermark is introduced to tune
> > > each
> > > zone's
> > > watermark separately. This can bring the following advantages:
> > > - each zone can set its own watermark which bring flexibility
> > > - lead to more efficient kswapd if this watermark is set fine
> > >
> > > Here is real watermark data in my qemu machine(with THP
> > > disabled).
> > >
> > > With watermark_scale_factor = 10, there is only 1440(772-68+807-
> > > 71)
> > > pages(5.76M) reserved for a system with 96G of memory. However if
> > > the
> > > watermark is set to 100, the movable zone's watermark increased
> > > to
> > > 231908(93M), which is too much.
> > > This situation is even worse with 32G of normal zone memory and
> > > 1T of
> > > movable zone memory.
> > >
> > > Modified | Vanilla wm_factor = 10 | Vanilla
> > > wm_factor =
> > > 30
> > > Node 0, zone DMA | Node 0, zone DMA | Node 0,
> > > zone
> > > DMA
> > > min 68 | min 68 |
> > > min 68
> > > low 7113 | low 772 |
> > > low
> > > 7113
> > > high **14158** | high **1476** | high
> > > **14158**
> > > Node 0, zone Normal | Node 0, zone Normal | Node 0, zone
> > > Normal
> > > min 71 | min 71 |
> > > min 71
> > > low 7438 | low 807 |
> > > low
> > > 7438
> > > high 14805 | high 1543 |
> > > high
> > > 14805
> > > Node 0, zone Movable | Node 0, zone Movable | Node 0, zone
> > > Movable
> > > min 1455 | min 1455 |
> > > min
> > > 1455
> > > low 16388 | low 16386 |
> > > low
> > > 150787
> > > high **31321** | high **31317** | high
> > > **300119**
> > > Node 1, zone Movable | Node 1, zone Movable | Node 1, zone
> > > Movable
> > > min 804 | min 804 |
> > > min
> > > 804
> > > low 9061 | low 9061 |
> > > low
> > > 83379
> > > high **17318** | high **17318** | high
> > > **165954**
> > >
> > > With the modified per zone watermark_scale_factor, only
> > > dma/normal
> > > zone
> > > will increase its watermark via the following command which the
> > > huge
> > > movable zone stay the same.
> > >
> > > % echo 100 100 100 10 > /proc/sys/vm/watermark_scale_factor
> > >
> > > The reason to disable THP is khugepaged_min_free_kbytes_update()
> > > will
> > > update min watermark.
> > >
> > > Signed-off-by: Ma Wupeng <[email protected]>
> > > Reported-by: kernel test robot <[email protected]>
> > > ---
> > > Documentation/admin-guide/sysctl/vm.rst | 6 ++++
> > > include/linux/mm.h | 2 +-
> > > kernel/sysctl.c | 2 --
> > > mm/page_alloc.c | 37
> > > ++++++++++++++++++++---
> > > --
> > > 4 files changed, 37 insertions(+), 10 deletions(-)
> > >
> > > diff --git a/Documentation/admin-guide/sysctl/vm.rst
> > > b/Documentation/admin-guide/sysctl/vm.rst
> > > index 9b833e439f09..ec240aa45322 100644
> > > --- a/Documentation/admin-guide/sysctl/vm.rst
> > > +++ b/Documentation/admin-guide/sysctl/vm.rst
> > > @@ -1002,6 +1002,12 @@ that the number of free pages kswapd
> > > maintains
> > > for latency reasons is
> > > too small for the allocation bursts occurring in the system.
> > > This
> > > knob
> > > can then be used to tune kswapd aggressiveness accordingly.
> > >
> > > +The watermark_scale_factor is an array. You can set each zone's
> > > watermark
> > > +separately and can be seen by reading this file::
> > > +
> > > + % cat /proc/sys/vm/watermark_scale_factor
> > > + 10 10 10 10
> > > +
> > >
> > > zone_reclaim_mode
> > > =================
> > > diff --git a/include/linux/mm.h b/include/linux/mm.h
> > > index 21f8b27bd9fd..b291c795f9db 100644
> > > --- a/include/linux/mm.h
> > > +++ b/include/linux/mm.h
> > > @@ -2532,7 +2532,7 @@ extern void setup_per_cpu_pageset(void);
> > > /* page_alloc.c */
> > > extern int min_free_kbytes;
> > > extern int watermark_boost_factor;
> > > -extern int watermark_scale_factor;
> > > +extern int watermark_scale_factor[MAX_NR_ZONES];
> > > extern bool arch_has_descending_max_zone_pfns(void);
> > >
> > > /* nommu.c */
> > > diff --git a/kernel/sysctl.c b/kernel/sysctl.c
> > > index 205d605cacc5..d16d06c71e5a 100644
> > > --- a/kernel/sysctl.c
> > > +++ b/kernel/sysctl.c
> > > @@ -2251,8 +2251,6 @@ static struct ctl_table vm_table[] = {
> > > .maxlen = sizeof(watermark_scale_factor),
> > > .mode = 0644,
> > > .proc_handler =
> > > watermark_scale_factor_sysctl_handler,
> > > - .extra1 = SYSCTL_ONE,
> > > - .extra2 = SYSCTL_THREE_THOUSAND,
> > > },
> > > {
> > > .procname =
> > > "percpu_pagelist_high_fraction",
> > > diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> > > index f1e4474879f1..7a6ac3b4ebb6 100644
> > > --- a/mm/page_alloc.c
> > > +++ b/mm/page_alloc.c
> > > @@ -421,7 +421,6 @@ compound_page_dtor * const
> > > compound_page_dtors[NR_COMPOUND_DTORS] = {
> > > int min_free_kbytes = 1024;
> > > int user_min_free_kbytes = -1;
> > > int watermark_boost_factor __read_mostly = 15000;
> > > -int watermark_scale_factor = 10;
> > >
> > > static unsigned long nr_kernel_pages __initdata;
> > > static unsigned long nr_all_pages __initdata;
> > > @@ -449,6 +448,20 @@ EXPORT_SYMBOL(nr_online_nodes);
> > >
> > > int page_group_by_mobility_disabled __read_mostly;
> > >
> > > +int watermark_scale_factor[MAX_NR_ZONES] = {
> > > +#ifdef CONFIG_ZONE_DMA
> > > + [ZONE_DMA] = 10,
> > > +#endif
> > > +#ifdef CONFIG_ZONE_DMA32
> > > + [ZONE_DMA32] = 10,
> > > +#endif
> > > + [ZONE_NORMAL] = 10,
> > > +#ifdef CONFIG_HIGHMEM
> > > + [ZONE_HIGHMEM] = 10,
> > > +#endif
> > > + [ZONE_MOVABLE] = 10,
> > > +};
> > > +
> > > #ifdef CONFIG_DEFERRED_STRUCT_PAGE_INIT
> > > /*
> > > * During boot we initialize deferred pages on-demand, as
> > > needed,
> > > but once
> > > @@ -8643,6 +8656,7 @@ static void __setup_per_zone_wmarks(void)
> > > }
> > >
> > > for_each_zone(zone) {
> > > + int zone_wm_factor;
> > > u64 tmp;
> > >
> > > spin_lock_irqsave(&zone->lock, flags);
> > > @@ -8676,9 +8690,10 @@ static void __setup_per_zone_wmarks(void)
> > > * scale factor in proportion to available
> > > memory,
> > > but
> > > * ensure a minimum size on small systems.
> > > */
> > > + zone_wm_factor =
> > > watermark_scale_factor[zone_idx(zone)];
> > > tmp = max_t(u64, tmp >> 2,
> > > - mult_frac(zone_managed_pages(zone),
> > > - watermark_scale_factor,
> > > 10000));
> > > + mult_frac(zone_managed_pages(zone),
> > > zone_wm_factor,
> > > + 10000));
> > >
> > > zone->watermark_boost = 0;
> > > zone->_watermark[WMARK_LOW] =
> > > min_wmark_pages(zone)
> > > + tmp;
> > > @@ -8798,11 +8813,19 @@ int min_free_kbytes_sysctl_handler(struct
> > > ctl_table *table, int write,
> > > int watermark_scale_factor_sysctl_handler(struct ctl_table
> > > *table,
> > > int write,
> > > void *buffer, size_t *length, loff_t *ppos)
> > > {
> > > - int rc;
> > > + int i;
> > >
> > > - rc = proc_dointvec_minmax(table, write, buffer, length,
> > > ppos);
> > > - if (rc)
> > > - return rc;
> > > + proc_dointvec_minmax(table, write, buffer, length, ppos);
> > > +
> > > + /*
> > > + * The unit is in fractions of 10,000. The default value
> > > of
> > > 10
> > > + * means the distances between watermarks are 0.1% of the
> > > available
> > > + * memory in the node/system. The maximum value is 3000,
> > > or
> > > 30% of
> > > + * memory.
> > > + */
> > > + for (i = 0; i < MAX_NR_ZONES; i++)
> > > + watermark_scale_factor[i] =
> > > + clamp(watermark_scale_factor[i], 1,
> > > 3000);
> > >
> > > if (write)
> > > setup_per_zone_wmarks();
> >