Received: by 2002:a25:e7d8:0:0:0:0:0 with SMTP id e207csp111603ybh; Mon, 9 Mar 2020 17:26:48 -0700 (PDT) X-Google-Smtp-Source: ADFU+vuJ+PEBX8Sbl8luiiFQ3M4gZVfXh5KoMY6cSX4/MX9X1mpRosnVWCtFoJu+Ypl/UTg+1CBY X-Received: by 2002:a9d:2c69:: with SMTP id f96mr6602695otb.62.1583800008425; Mon, 09 Mar 2020 17:26:48 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1583800008; cv=none; d=google.com; s=arc-20160816; b=mDYh5zXLnunJD7lcRrXErPF8oIRYL8H4FnBUvQJSVdcxNJrv8ZRQGyuOUwMWy+nvmg S6nRzcqjU4N+nN21gQNLdKHg1ju+uXsdbLkYHkQXgHEXDrvV93dGxA0mgFMNA7tJGuQ9 sLTkboABmui48z2cf2037pTK0HFspAPOFwwbVDT7afjgtxpyg5p1GE5kJWUBqHvwcsyr NhKzJHTvytiy2X4o7RICU16y1Eq6E1AuuT4Dr9vOblKg7rmplBC3uXertWLTDqvWTHPx CjpO9YbMBwFkcClFdMzyux5dD+qAJrCOPMmsegTJ5+XuKN7h3u9IYFuQLozpb3LH10P/ 00xA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:mime-version:message-id:date:subject :smtp-origin-cluster:cc:to:smtp-origin-hostname:from :smtp-origin-hostprefix:dkim-signature; bh=q7NV9FmGQPkvn4PU4QfR75NfQSg3itTeGK7NjywUGZg=; b=EXMmAGccGI8GQ2VHwVB/DxPaB0rsZXxGHiPaFL2B+FgeYpU7ACOUBRTZWkpEIf0Oud fnSIIv1g3BGsvVWSbU6vLuFaEWP8TrJ6lFVEIT6f81r4N9tT/9S1rZW+aTX/CdZ5NYjE Z/jkG5n/4nPLBXWZHsfJVHkKkSZIG6WRqyhGVb6NJjKxumEyZ1dIz3mLjaHh/V6h8uwL KlKKPN2fxODkChJVNjrzuM+e/TTRbtBDp+mprDbUJHLBvxe28guvbJVA6mqKOoIApPSg 9qUZdu57sEBsEhSUlesay9kxvoXlPvRHn6B4KxuJEwA++ywB4HoMMZ+asvpGwGHrUcPN wUfg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@fb.com header.s=facebook header.b=agdYU2Q6; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=fb.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id u125si1650780oie.52.2020.03.09.17.26.32; Mon, 09 Mar 2020 17:26:48 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@fb.com header.s=facebook header.b=agdYU2Q6; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=fb.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727524AbgCJAZj (ORCPT + 99 others); Mon, 9 Mar 2020 20:25:39 -0400 Received: from mx0b-00082601.pphosted.com ([67.231.153.30]:28560 "EHLO mx0b-00082601.pphosted.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727322AbgCJAZj (ORCPT ); Mon, 9 Mar 2020 20:25:39 -0400 Received: from pps.filterd (m0148460.ppops.net [127.0.0.1]) by mx0a-00082601.pphosted.com (8.16.0.42/8.16.0.42) with SMTP id 02A0AH7f012566 for ; Mon, 9 Mar 2020 17:25:37 -0700 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=fb.com; h=from : to : cc : subject : date : message-id : mime-version : content-type; s=facebook; bh=q7NV9FmGQPkvn4PU4QfR75NfQSg3itTeGK7NjywUGZg=; b=agdYU2Q63QFA9LO9a1xgd5qdqpdKRn+1jfqQngHKrKutjFWKv6pE3c9DXyqozClu9eT9 tlOIIfH46jeBe1WEgt9d8Hq3pMeGAxdCsKV8y1St177tOO6OWMWMHQ1wlPhZHwvZOi1z F7l7zN0f6teDwdjsTf8YqBiP/jZyzlWvH9U= Received: from maileast.thefacebook.com ([163.114.130.16]) by mx0a-00082601.pphosted.com with ESMTP id 2yntmu9qw4-4 (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128 verify=NOT) for ; Mon, 09 Mar 2020 17:25:37 -0700 Received: from intmgw001.41.prn1.facebook.com (2620:10d:c0a8:1b::d) by mail.thefacebook.com (2620:10d:c0a8:82::e) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.1847.3; Mon, 9 Mar 2020 17:25:36 -0700 Received: by devvm2643.prn2.facebook.com (Postfix, from userid 111017) id 678113389F3DA; Mon, 9 Mar 2020 17:25:31 -0700 (PDT) Smtp-Origin-Hostprefix: devvm From: Roman Gushchin Smtp-Origin-Hostname: devvm2643.prn2.facebook.com To: Andrew Morton CC: Johannes Weiner , Michal Hocko , , , , Rik van Riel , Roman Gushchin Smtp-Origin-Cluster: prn2c23 Subject: [PATCH v2] mm: hugetlb: optionally allocate gigantic hugepages using cma Date: Mon, 9 Mar 2020 17:25:24 -0700 Message-ID: <20200310002524.2291595-1-guro@fb.com> X-Mailer: git-send-email 2.17.1 X-FB-Internal: Safe MIME-Version: 1.0 Content-Type: text/plain X-Proofpoint-Virus-Version: vendor=fsecure engine=2.50.10434:6.0.138,18.0.572 definitions=2020-03-09_13:2020-03-09,2020-03-09 signatures=0 X-Proofpoint-Spam-Details: rule=fb_default_notspam policy=fb_default score=0 lowpriorityscore=0 phishscore=0 mlxscore=0 clxscore=1015 suspectscore=2 priorityscore=1501 adultscore=0 malwarescore=0 impostorscore=0 spamscore=0 mlxlogscore=999 bulkscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2001150001 definitions=main-2003090147 X-FB-Internal: deliver Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Commit 944d9fec8d7a ("hugetlb: add support for gigantic page allocation at runtime") has added the run-time allocation of gigantic pages. However it actually works only at early stages of the system loading, when the majority of memory is free. After some time the memory gets fragmented by non-movable pages, so the chances to find a contiguous 1 GB block are getting close to zero. Even dropping caches manually doesn't help a lot. At large scale rebooting servers in order to allocate gigantic hugepages is quite expensive and complex. At the same time keeping some constant percentage of memory in reserved hugepages even if the workload isn't using it is a big waste: not all workloads can benefit from using 1 GB pages. The following solution can solve the problem: 1) On boot time a dedicated cma area* is reserved. The size is passed as a kernel argument. 2) Run-time allocations of gigantic hugepages are performed using the cma allocator and the dedicated cma area In this case gigantic hugepages can be allocated successfully with a high probability, however the memory isn't completely wasted if nobody is using 1GB hugepages: it can be used for pagecache, anon memory, THPs, etc. * On a multi-node machine a per-node cma area is allocated on each node. Following gigantic hugetlb allocation are using the first available numa node if the mask isn't specified by a user. Usage: 1) configure the kernel to allocate a cma area for hugetlb allocations: pass hugetlb_cma=10G as a kernel argument 2) allocate hugetlb pages as usual, e.g. echo 10 > /sys/kernel/mm/hugepages/hugepages-1048576kB/nr_hugepages If the option isn't enabled or the allocation of the cma area failed, the current behavior of the system is preserved. Only x86 is covered by this patch, but it's trivial to extend it to cover other architectures as well. v2: fixed !CONFIG_CMA build, suggested by Andrew Morton Signed-off-by: Roman Gushchin --- .../admin-guide/kernel-parameters.txt | 7 ++ arch/x86/kernel/setup.c | 3 + include/linux/hugetlb.h | 2 + mm/hugetlb.c | 115 ++++++++++++++++++ 4 files changed, 127 insertions(+) diff --git a/Documentation/admin-guide/kernel-parameters.txt b/Documentation/admin-guide/kernel-parameters.txt index 0c9894247015..d3349ec1dbef 100644 --- a/Documentation/admin-guide/kernel-parameters.txt +++ b/Documentation/admin-guide/kernel-parameters.txt @@ -1452,6 +1452,13 @@ hpet_mmap= [X86, HPET_MMAP] Allow userspace to mmap HPET registers. Default set by CONFIG_HPET_MMAP_DEFAULT. + hugetlb_cma= [x86-64] The size of a cma area used for allocation + of gigantic hugepages. + Format: nn[GTPE] | nn% + + If enabled, boot-time allocation of gigantic hugepages + is skipped. + hugepages= [HW,X86-32,IA-64] HugeTLB pages to allocate at boot. hugepagesz= [HW,IA-64,PPC,X86-64] The size of the HugeTLB pages. On x86-64 and powerpc, this option can be specified diff --git a/arch/x86/kernel/setup.c b/arch/x86/kernel/setup.c index a74262c71484..ceeb06ddfd41 100644 --- a/arch/x86/kernel/setup.c +++ b/arch/x86/kernel/setup.c @@ -16,6 +16,7 @@ #include #include #include +#include #include #include @@ -1158,6 +1159,8 @@ void __init setup_arch(char **cmdline_p) initmem_init(); dma_contiguous_reserve(max_pfn_mapped << PAGE_SHIFT); + hugetlb_cma_reserve(); + /* * Reserve memory for crash kernel after SRAT is parsed so that it * won't consume hotpluggable memory. diff --git a/include/linux/hugetlb.h b/include/linux/hugetlb.h index 50480d16bd33..50050c981ab9 100644 --- a/include/linux/hugetlb.h +++ b/include/linux/hugetlb.h @@ -157,6 +157,8 @@ pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud); extern int sysctl_hugetlb_shm_group; extern struct list_head huge_boot_pages; +extern void __init hugetlb_cma_reserve(void); + /* arch callbacks */ pte_t *huge_pte_alloc(struct mm_struct *mm, diff --git a/mm/hugetlb.c b/mm/hugetlb.c index 7fb31750e670..c6f58bab879c 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c @@ -28,6 +28,7 @@ #include #include #include +#include #include #include @@ -44,6 +45,9 @@ int hugetlb_max_hstate __read_mostly; unsigned int default_hstate_idx; struct hstate hstates[HUGE_MAX_HSTATE]; + +static struct cma *hugetlb_cma[MAX_NUMNODES]; + /* * Minimum page order among possible hugepage sizes, set to a proper value * at boot time. @@ -1228,6 +1232,11 @@ static void destroy_compound_gigantic_page(struct page *page, static void free_gigantic_page(struct page *page, unsigned int order) { + if (IS_ENABLED(CONFIG_CMA) && hugetlb_cma[0]) { + cma_release(hugetlb_cma[page_to_nid(page)], page, 1 << order); + return; + } + free_contig_range(page_to_pfn(page), 1 << order); } @@ -1237,6 +1246,23 @@ static struct page *alloc_gigantic_page(struct hstate *h, gfp_t gfp_mask, { unsigned long nr_pages = 1UL << huge_page_order(h); + if (IS_ENABLED(CONFIG_CMA) && hugetlb_cma[0]) { + struct page *page; + int nid; + + for_each_node_mask(nid, *nodemask) { + if (!hugetlb_cma[nid]) + break; + + page = cma_alloc(hugetlb_cma[nid], nr_pages, + huge_page_order(h), true); + if (page) + return page; + } + + return NULL; + } + return alloc_contig_pages(nr_pages, gfp_mask, nid, nodemask); } @@ -2439,6 +2465,10 @@ static void __init hugetlb_hstate_alloc_pages(struct hstate *h) for (i = 0; i < h->max_huge_pages; ++i) { if (hstate_is_gigantic(h)) { + if (IS_ENABLED(CONFIG_CMA) && hugetlb_cma[0]) { + pr_warn_once("HugeTLB: hugetlb_cma is enabled, skip boot time allocation\n"); + break; + } if (!alloc_bootmem_huge_page(h)) break; } else if (!alloc_pool_huge_page(h, @@ -5372,3 +5402,88 @@ void move_hugetlb_state(struct page *oldpage, struct page *newpage, int reason) spin_unlock(&hugetlb_lock); } } + +#ifdef CONFIG_CMA +static unsigned long hugetlb_cma_size __initdata; +static unsigned long hugetlb_cma_percent __initdata; + +static int __init cmdline_parse_hugetlb_cma(char *p) +{ + unsigned long long val; + char *endptr; + + if (!p) + return -EINVAL; + + /* Value may be a percentage of total memory, otherwise bytes */ + val = simple_strtoull(p, &endptr, 0); + if (*endptr == '%') + hugetlb_cma_percent = clamp_t(unsigned long, val, 0, 100); + else + hugetlb_cma_size = memparse(p, &p); + + return 0; +} + +early_param("hugetlb_cma", cmdline_parse_hugetlb_cma); + +void __init hugetlb_cma_reserve(void) +{ + unsigned long totalpages = 0; + unsigned long start_pfn, end_pfn; + phys_addr_t size; + int nid, i, res; + + if (!hugetlb_cma_size && !hugetlb_cma_percent) + return; + + if (hugetlb_cma_percent) { + for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, + NULL) + totalpages += end_pfn - start_pfn; + + size = PAGE_SIZE * (hugetlb_cma_percent * 100 * totalpages) / + 10000UL; + } else { + size = hugetlb_cma_size; + } + + pr_info("hugetlb_cma: reserve %llu, %llu per node\n", size, + size / nr_online_nodes); + + size /= nr_online_nodes; + + for_each_node_state(nid, N_ONLINE) { + unsigned long min_pfn = 0, max_pfn = 0; + + for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { + if (!min_pfn) + min_pfn = start_pfn; + max_pfn = end_pfn; + } + + res = cma_declare_contiguous(PFN_PHYS(min_pfn), size, + PFN_PHYS(max_pfn), (1UL << 30), + 0, false, + "hugetlb", &hugetlb_cma[nid]); + if (res) { + pr_warn("hugetlb_cma: reservation failed: err %d, node %d, [%llu, %llu)", + res, nid, PFN_PHYS(min_pfn), PFN_PHYS(max_pfn)); + + for (; nid >= 0; nid--) + hugetlb_cma[nid] = NULL; + + break; + } + + pr_info("hugetlb_cma: successfully reserved %llu on node %d\n", + size, nid); + } +} + +#else /* CONFIG_CMA */ +void __init hugetlb_cma_reserve(void) +{ +} + +#endif /* CONFIG_CMA */ -- 2.24.1