Received: by 10.223.185.116 with SMTP id b49csp1689027wrg; Thu, 22 Feb 2018 01:13:00 -0800 (PST) X-Google-Smtp-Source: AH8x227tLl263li/LXvkeDLs+QRBnZfBCKnNZz+zjbKSbBxzkV0WPhQXOeTf6aYeXM1LeVGAON6W X-Received: by 10.101.97.139 with SMTP id c11mr5033110pgv.449.1519290780405; Thu, 22 Feb 2018 01:13:00 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1519290780; cv=none; d=google.com; s=arc-20160816; b=oC2fCUAYtKDFj6YwRyAP4NHGlc6+0JH4S37nS/QY3u/yJ+36H/ZJO4N7kvnsNDaiwn rEqqUI7rHwbh6VRAG7rEc4LexLlUoCNgofEmps+nBeY5olixa2kkNL4L/d/P9gkAgMZz 72V9ORNWp2F4aCHIEtqw/uNbQ3/JxEfVZUIoBjzENPmpTNZ0YLvujGxDuAjjraYLFVKZ /NSgAXK1WZcBPVhNeDGi0y1M+rLR/q7ucPoiZvWomg6mbTJOSzJ5xO4MGFO8MmdtjPLi 1dvOc88l9ZiWd65n0rQhI8Sc2uq/i8VgMal2Gp3OF+RSaPKhJvIazDDyMkAFHwxv/2t8 Pz2g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:arc-authentication-results; bh=/uOFO+KOZnKofyN6cW2XFZZ6e7ogkuSBa+Mz+ZvhkwY=; b=m+A8Egw/bT9QRXiZ9GamPgjSxFnCCIIc6Wf06/UORL9q/zdJ0bXi5yS+Ryb3GAaUqY /s6on4YQJKiMMMIH1e7LEMga1gTGNcYvp0aQsk0FpaoLtrEe7xTBplK9epNVzxpNdXbE umCicAOZnDxKMVOIXybY2h/TVXx0sgb952berg9VPcHVP2DKGF2hfrvWvwUNJm7547H5 YUKlNr1RR+LHmyOjgkcf3KcjxDWACbXII8/QZzbO+XRUoBMN1APPMjlRIFbErBeLaR+g 1UPyT8KyLNLKNkNYIT5K2xliHlNDFAozpXtUGpNKbJCIAqd74dogofNaR+sHFNr7srJ0 0HVw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id g128si1704645pgc.207.2018.02.22.01.12.45; Thu, 22 Feb 2018 01:13:00 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753089AbeBVJL6 (ORCPT + 99 others); Thu, 22 Feb 2018 04:11:58 -0500 Received: from mx3-rdu2.redhat.com ([66.187.233.73]:57850 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1753018AbeBVJLz (ORCPT ); Thu, 22 Feb 2018 04:11:55 -0500 Received: from smtp.corp.redhat.com (int-mx04.intmail.prod.int.rdu2.redhat.com [10.11.54.4]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 25BFE40753A2; Thu, 22 Feb 2018 09:11:55 +0000 (UTC) Received: from localhost.localdomain.com (ovpn-8-20.pek2.redhat.com [10.72.8.20]) by smtp.corp.redhat.com (Postfix) with ESMTP id 378EF2024CA9; Thu, 22 Feb 2018 09:11:50 +0000 (UTC) From: Baoquan He To: linux-kernel@vger.kernel.org, dave.hansen@intel.com Cc: linux-mm@kvack.org, akpm@linux-foundation.org, kirill.shutemov@linux.intel.com, mhocko@suse.com, tglx@linutronix.de, Baoquan He Subject: [PATCH v2 3/3] mm/sparse: Optimize memmap allocation during sparse_init() Date: Thu, 22 Feb 2018 17:11:30 +0800 Message-Id: <20180222091130.32165-4-bhe@redhat.com> In-Reply-To: <20180222091130.32165-1-bhe@redhat.com> References: <20180222091130.32165-1-bhe@redhat.com> X-Scanned-By: MIMEDefang 2.78 on 10.11.54.4 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.11.55.7]); Thu, 22 Feb 2018 09:11:55 +0000 (UTC) X-Greylist: inspected by milter-greylist-4.5.16 (mx1.redhat.com [10.11.55.7]); Thu, 22 Feb 2018 09:11:55 +0000 (UTC) for IP:'10.11.54.4' DOMAIN:'int-mx04.intmail.prod.int.rdu2.redhat.com' HELO:'smtp.corp.redhat.com' FROM:'bhe@redhat.com' RCPT:'' Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org In sparse_init(), two temporary pointer arrays, usemap_map and map_map are allocated with the size of NR_MEM_SECTIONS. They are used to store each memory section's usemap and mem map if marked as present. With the help of these two arrays, continuous memory chunk is allocated for usemap and memmap for memory sections on one node. This avoids too many memory fragmentations. Like below diagram, '1' indicates the present memory section, '0' means absent one. The number 'n' could be much smaller than NR_MEM_SECTIONS on most of systems. |1|1|1|1|0|0|0|0|1|1|0|0|...|1|0||1|0|...|1||0|1|...|0| ------------------------------------------------------- 0 1 2 3 4 5 i i+1 n-1 n If fail to populate the page tables to map one section's memmap, its ->section_mem_map will be cleared finally to indicate that it's not present. After use, these two arrays will be released at the end of sparse_init(). In 4-level paging mode, each array costs 4M which can be ignorable. While in 5-level paging, they costs 256M each, 512M altogether. Kdump kernel Usually only reserves very few memory, e.g 256M. So, even thouth they are temporarily allocated, still not acceptable. In fact, there's no need to allocate them with the size of NR_MEM_SECTIONS. Since the ->section_mem_map clearing has been deferred to the last, the number of present memory sections are kept the same during sparse_init() until we finally clear out the memory section's ->section_mem_map if its usemap or memmap is not correctly handled. Thus in the middle whenever for_each_present_section_nr() loop is taken, the i-th present memory section is always the same one. Here only allocate usemap_map and map_map with the size of 'nr_present_sections'. For the i-th present memory section, install its usemap and memmap to usemap_map[i] and mam_map[i] during allocation. Then in the last for_each_present_section_nr() loop which clears the failed memory section's ->section_mem_map, fetch usemap and memmap from usemap_map[] and map_map[] array and set them into mem_section[] accordingly. Signed-off-by: Baoquan He Signed-off-by: Baoquan He --- mm/sparse-vmemmap.c | 8 +++++--- mm/sparse.c | 40 ++++++++++++++++++++++++++-------------- 2 files changed, 31 insertions(+), 17 deletions(-) diff --git a/mm/sparse-vmemmap.c b/mm/sparse-vmemmap.c index 640e68f8324b..f83723a49e47 100644 --- a/mm/sparse-vmemmap.c +++ b/mm/sparse-vmemmap.c @@ -281,6 +281,7 @@ void __init sparse_mem_maps_populate_node(struct page **map_map, unsigned long pnum; unsigned long size = sizeof(struct page) * PAGES_PER_SECTION; void *vmemmap_buf_start; + int i = 0; size = ALIGN(size, PMD_SIZE); vmemmap_buf_start = __earlyonly_bootmem_alloc(nodeid, size * map_count, @@ -291,14 +292,15 @@ void __init sparse_mem_maps_populate_node(struct page **map_map, vmemmap_buf_end = vmemmap_buf_start + size * map_count; } - for (pnum = pnum_begin; pnum < pnum_end; pnum++) { + for (pnum = pnum_begin; pnum < pnum_end && i < map_count; pnum++) { struct mem_section *ms; if (!present_section_nr(pnum)) continue; - map_map[pnum] = sparse_mem_map_populate(pnum, nodeid, NULL); - if (map_map[pnum]) + i++; + map_map[i-1] = sparse_mem_map_populate(pnum, nodeid, NULL); + if (map_map[i-1]) continue; ms = __nr_to_section(pnum); pr_err("%s: sparsemem memory map backing failed some memory will not be available\n", diff --git a/mm/sparse.c b/mm/sparse.c index e9311b44e28a..aafb6d838872 100644 --- a/mm/sparse.c +++ b/mm/sparse.c @@ -405,6 +405,7 @@ static void __init sparse_early_usemaps_alloc_node(void *data, unsigned long pnum; unsigned long **usemap_map = (unsigned long **)data; int size = usemap_size(); + int i = 0; usemap = sparse_early_usemaps_alloc_pgdat_section(NODE_DATA(nodeid), size * usemap_count); @@ -413,12 +414,13 @@ static void __init sparse_early_usemaps_alloc_node(void *data, return; } - for (pnum = pnum_begin; pnum < pnum_end; pnum++) { + for (pnum = pnum_begin; pnum < pnum_end && i < usemap_count; pnum++) { if (!present_section_nr(pnum)) continue; - usemap_map[pnum] = usemap; + usemap_map[i] = usemap; usemap += size; - check_usemap_section_nr(nodeid, usemap_map[pnum]); + check_usemap_section_nr(nodeid, usemap_map[i]); + i++; } } @@ -447,14 +449,17 @@ void __init sparse_mem_maps_populate_node(struct page **map_map, void *map; unsigned long pnum; unsigned long size = sizeof(struct page) * PAGES_PER_SECTION; + int i; map = alloc_remap(nodeid, size * map_count); if (map) { - for (pnum = pnum_begin; pnum < pnum_end; pnum++) { + i = 0; + for (pnum = pnum_begin; pnum < pnum_end && i < map_count; pnum++) { if (!present_section_nr(pnum)) continue; - map_map[pnum] = map; + map_map[i] = map; map += size; + i++; } return; } @@ -464,23 +469,27 @@ void __init sparse_mem_maps_populate_node(struct page **map_map, PAGE_SIZE, __pa(MAX_DMA_ADDRESS), BOOTMEM_ALLOC_ACCESSIBLE, nodeid); if (map) { - for (pnum = pnum_begin; pnum < pnum_end; pnum++) { + i = 0; + for (pnum = pnum_begin; pnum < pnum_end && i < map_count; pnum++) { if (!present_section_nr(pnum)) continue; - map_map[pnum] = map; + map_map[i] = map; map += size; + i++; } return; } /* fallback */ - for (pnum = pnum_begin; pnum < pnum_end; pnum++) { + i = 0; + for (pnum = pnum_begin; pnum < pnum_end && i < map_count; pnum++) { struct mem_section *ms; if (!present_section_nr(pnum)) continue; - map_map[pnum] = sparse_mem_map_populate(pnum, nodeid, NULL); - if (map_map[pnum]) + i++; + map_map[i-1] = sparse_mem_map_populate(pnum, nodeid, NULL); + if (map_map[i-1]) continue; ms = __nr_to_section(pnum); pr_err("%s: sparsemem memory map backing failed some memory will not be available\n", @@ -558,6 +567,7 @@ static void __init alloc_usemap_and_memmap(void (*alloc_func) /* new start, update count etc*/ nodeid_begin = nodeid; pnum_begin = pnum; + data += map_count; map_count = 1; } /* ok, last chunk */ @@ -576,6 +586,7 @@ void __init sparse_init(void) unsigned long *usemap; unsigned long **usemap_map; int size; + int i = 0; #ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER int size2; struct page **map_map; @@ -598,7 +609,7 @@ void __init sparse_init(void) * powerpc need to call sparse_init_one_section right after each * sparse_early_mem_map_alloc, so allocate usemap_map at first. */ - size = sizeof(unsigned long *) * NR_MEM_SECTIONS; + size = sizeof(unsigned long *) * nr_present_sections; usemap_map = memblock_virt_alloc(size, 0); if (!usemap_map) panic("can not allocate usemap_map\n"); @@ -606,7 +617,7 @@ void __init sparse_init(void) (void *)usemap_map); #ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER - size2 = sizeof(struct page *) * NR_MEM_SECTIONS; + size2 = sizeof(struct page *) * nr_present_sections; map_map = memblock_virt_alloc(size2, 0); if (!map_map) panic("can not allocate map_map\n"); @@ -617,14 +628,15 @@ void __init sparse_init(void) for_each_present_section_nr(0, pnum) { struct mem_section *ms; ms = __nr_to_section(pnum); - usemap = usemap_map[pnum]; + i++; + usemap = usemap_map[i-1]; if (!usemap) { ms->section_mem_map = 0; continue; } #ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER - map = map_map[pnum]; + map = map_map[i-1]; #else map = sparse_early_mem_map_alloc(pnum); #endif -- 2.13.6