2018-06-27 21:46:57

by Cannon Matthews

[permalink] [raw]
Subject: [PATCH] mm: hugetlb: yield when prepping struct pages

When booting with very large numbers of gigantic (i.e. 1G) pages, the
operations in the loop of gather_bootmem_prealloc, and specifically
prep_compound_gigantic_page, takes a very long time, and can cause a
softlockup if enough pages are requested at boot.

For example booting with 3844 1G pages requires prepping
(set_compound_head, init the count) over 1 billion 4K tail pages, which
takes considerable time. This should also apply to reserving the same
amount of memory as 2M pages, as the same number of struct pages
are affected in either case.

Add a cond_resched() to the outer loop in gather_bootmem_prealloc() to
prevent this lockup.

Tested: Booted with softlockup_panic=1 hugepagesz=1G hugepages=3844 and
no softlockup is reported, and the hugepages are reported as
successfully setup.

Signed-off-by: Cannon Matthews <[email protected]>
---
mm/hugetlb.c | 1 +
1 file changed, 1 insertion(+)

diff --git a/mm/hugetlb.c b/mm/hugetlb.c
index a963f2034dfc..d38273c32d3b 100644
--- a/mm/hugetlb.c
+++ b/mm/hugetlb.c
@@ -2169,6 +2169,7 @@ static void __init gather_bootmem_prealloc(void)
*/
if (hstate_is_gigantic(h))
adjust_managed_page_count(page, 1 << h->order);
+ cond_resched();
}
}

--
2.18.0.rc2.346.g013aa6912e-goog



2018-06-27 23:54:50

by Mike Kravetz

[permalink] [raw]
Subject: Re: [PATCH] mm: hugetlb: yield when prepping struct pages

On 06/27/2018 02:44 PM, Cannon Matthews wrote:
> When booting with very large numbers of gigantic (i.e. 1G) pages, the
> operations in the loop of gather_bootmem_prealloc, and specifically
> prep_compound_gigantic_page, takes a very long time, and can cause a
> softlockup if enough pages are requested at boot.
>
> For example booting with 3844 1G pages requires prepping

Wow! I wish I had a system with that much memory to test. :)

> (set_compound_head, init the count) over 1 billion 4K tail pages, which
> takes considerable time. This should also apply to reserving the same
> amount of memory as 2M pages, as the same number of struct pages
> are affected in either case.

Actually, this change would not apply to 2M (on x86) pages. The hugetlbfs
initialization code is a bit confusing, but alloc_bootmem_huge_page and
gather_bootmem_prealloc are only exercised in the case where huge page
order >= MAX_ORDER.

Allocation and initialization of 2M pages happens after the normal memory
allocators are setup via the routine hugetlb_hstate_alloc_pages. And,
there is already a cond_resched in that loop today.

Note that 'else if' in the for loop of hugetlb_hstate_alloc_pages. This
allows the same routine to be called for early gigantic page allocations
using the bootmem allocator, and later normal (2M) allocations using the
normal memory allocators. To me, this is a source of confusion and is
something I plan to clean up in the future.

> Add a cond_resched() to the outer loop in gather_bootmem_prealloc() to
> prevent this lockup.
>
> Tested: Booted with softlockup_panic=1 hugepagesz=1G hugepages=3844 and
> no softlockup is reported, and the hugepages are reported as
> successfully setup.
>
> Signed-off-by: Cannon Matthews <[email protected]>

My only suggestion would be to remove the mention of 2M pages in the
commit message. Thanks for adding this.

Reviewed-by: Mike Kravetz <[email protected]>
--
Mike Kravetz

> ---
> mm/hugetlb.c | 1 +
> 1 file changed, 1 insertion(+)
>
> diff --git a/mm/hugetlb.c b/mm/hugetlb.c
> index a963f2034dfc..d38273c32d3b 100644
> --- a/mm/hugetlb.c
> +++ b/mm/hugetlb.c
> @@ -2169,6 +2169,7 @@ static void __init gather_bootmem_prealloc(void)
> */
> if (hstate_is_gigantic(h))
> adjust_managed_page_count(page, 1 << h->order);
> + cond_resched();
> }
> }
>
>

2018-06-27 23:56:31

by Andrew Morton

[permalink] [raw]
Subject: Re: [PATCH] mm: hugetlb: yield when prepping struct pages

On Wed, 27 Jun 2018 16:27:24 -0700 Mike Kravetz <[email protected]> wrote:

> My only suggestion would be to remove the mention of 2M pages in the
> commit message. Thanks for adding this.

I have removed that sentence.

> Reviewed-by: Mike Kravetz <[email protected]>

Thanks again.

2018-06-28 20:26:27

by Michal Hocko

[permalink] [raw]
Subject: Re: [PATCH] mm: hugetlb: yield when prepping struct pages

On Wed 27-06-18 14:44:47, Cannon Matthews wrote:
> When booting with very large numbers of gigantic (i.e. 1G) pages, the
> operations in the loop of gather_bootmem_prealloc, and specifically
> prep_compound_gigantic_page, takes a very long time, and can cause a
> softlockup if enough pages are requested at boot.
>
> For example booting with 3844 1G pages requires prepping
> (set_compound_head, init the count) over 1 billion 4K tail pages, which
> takes considerable time. This should also apply to reserving the same
> amount of memory as 2M pages, as the same number of struct pages
> are affected in either case.
>
> Add a cond_resched() to the outer loop in gather_bootmem_prealloc() to
> prevent this lockup.
>
> Tested: Booted with softlockup_panic=1 hugepagesz=1G hugepages=3844 and
> no softlockup is reported, and the hugepages are reported as
> successfully setup.
>
> Signed-off-by: Cannon Matthews <[email protected]>

Acked-by: Michal Hocko <[email protected]>

Thanks!

> ---
> mm/hugetlb.c | 1 +
> 1 file changed, 1 insertion(+)
>
> diff --git a/mm/hugetlb.c b/mm/hugetlb.c
> index a963f2034dfc..d38273c32d3b 100644
> --- a/mm/hugetlb.c
> +++ b/mm/hugetlb.c
> @@ -2169,6 +2169,7 @@ static void __init gather_bootmem_prealloc(void)
> */
> if (hstate_is_gigantic(h))
> adjust_managed_page_count(page, 1 << h->order);
> + cond_resched();
> }
> }
>
> --
> 2.18.0.rc2.346.g013aa6912e-goog

--
Michal Hocko
SUSE Labs

2018-06-29 06:23:59

by Cannon Matthews

[permalink] [raw]
Subject: Re: [PATCH] mm: hugetlb: yield when prepping struct pages

Thanks for the quick turnaround.

Good to know about the how the 2M code path differs, I have been
trying to trace through some of this and it's easy to get lost between
which applies to which size.

Thanks!
On Thu, Jun 28, 2018 at 12:03 PM Michal Hocko <[email protected]> wrote:
>
> On Wed 27-06-18 14:44:47, Cannon Matthews wrote:
> > When booting with very large numbers of gigantic (i.e. 1G) pages, the
> > operations in the loop of gather_bootmem_prealloc, and specifically
> > prep_compound_gigantic_page, takes a very long time, and can cause a
> > softlockup if enough pages are requested at boot.
> >
> > For example booting with 3844 1G pages requires prepping
> > (set_compound_head, init the count) over 1 billion 4K tail pages, which
> > takes considerable time. This should also apply to reserving the same
> > amount of memory as 2M pages, as the same number of struct pages
> > are affected in either case.
> >
> > Add a cond_resched() to the outer loop in gather_bootmem_prealloc() to
> > prevent this lockup.
> >
> > Tested: Booted with softlockup_panic=1 hugepagesz=1G hugepages=3844 and
> > no softlockup is reported, and the hugepages are reported as
> > successfully setup.
> >
> > Signed-off-by: Cannon Matthews <[email protected]>
>
> Acked-by: Michal Hocko <[email protected]>
>
> Thanks!
>
> > ---
> > mm/hugetlb.c | 1 +
> > 1 file changed, 1 insertion(+)
> >
> > diff --git a/mm/hugetlb.c b/mm/hugetlb.c
> > index a963f2034dfc..d38273c32d3b 100644
> > --- a/mm/hugetlb.c
> > +++ b/mm/hugetlb.c
> > @@ -2169,6 +2169,7 @@ static void __init gather_bootmem_prealloc(void)
> > */
> > if (hstate_is_gigantic(h))
> > adjust_managed_page_count(page, 1 << h->order);
> > + cond_resched();
> > }
> > }
> >
> > --
> > 2.18.0.rc2.346.g013aa6912e-goog
>
> --
> Michal Hocko
> SUSE Labs

2018-06-29 07:37:15

by Michal Hocko

[permalink] [raw]
Subject: Re: [PATCH] mm: hugetlb: yield when prepping struct pages

On Thu 28-06-18 15:16:46, Cannon Matthews wrote:
> Thanks for the quick turnaround.
>
> Good to know about the how the 2M code path differs, I have been
> trying to trace through some of this and it's easy to get lost between
> which applies to which size.

Yeah, GB hugetlb pages implementation has been hacked into the existing
hugetlb code in a quite ugly way. We have done some cleanups since then
but there is still a lot of room for improvements.
--
Michal Hocko
SUSE Labs