Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1757474AbYFNMZJ (ORCPT ); Sat, 14 Jun 2008 08:25:09 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1753408AbYFNMY5 (ORCPT ); Sat, 14 Jun 2008 08:24:57 -0400 Received: from fgwmail6.fujitsu.co.jp ([192.51.44.36]:39089 "EHLO fgwmail6.fujitsu.co.jp" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752673AbYFNMY4 (ORCPT ); Sat, 14 Jun 2008 08:24:56 -0400 Date: Sat, 14 Jun 2008 21:22:44 +0900 From: Yasunori Goto To: David Miller , Badari Pulavarty , Andrew Morton , Heiko Carstens Subject: [Patch](memory hotplug)Allocate usemap on the section with pgdat (take 2) Cc: Hiroyuki KAMEZAWA , Tony Breeds , Linux Kernel ML , linux-mm X-Mailer-Plugin: BkASPil for Becky!2 Ver.2.068 Message-Id: <20080614211216.76B0.E1E9C6FF@jp.fujitsu.com> MIME-Version: 1.0 Content-Type: text/plain; charset="US-ASCII" Content-Transfer-Encoding: 7bit X-Mailer: Becky! ver. 2.45 [ja] Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 4940 Lines: 154 Hello. I update the patch which was cause of regression at bootmem. This version allows normal allocation if alloc_bootmem_section() fails. So, previous regression will not occur. I tested on my box. But it's not sparc, s390, nor powerpc. If there is any trouble by this patch, please let me know. If no trouble, please apply. Thanks. --- Usemaps are allocated on the section which has pgdat by this. Because usemap size is very small, many other sections usemaps are allocated on only one page. If a section has usemap, it can't be removed until removing other sections. This dependency is not desirable for memory removing. Pgdat has similar feature. When a section has pgdat area, it must be the last section for removing on the node. So, if section A has pgdat and section B has usemap for section A, Both sections can't be removed due to dependency each other. To solve this issue, this patch collects usemap on same section with pgdat as much as possible. If other sections doesn't have any dependency, this section will be able to be removed finally. Change log of take 2. - This feature becomes effective only when CONFIG_MEMORY_HOTREMOVE is on. If hotremove is off, this feature is not necessary. - Allow allocation on other section if alloc_bootmem_section() fails. This avoids previous regression. - Show message if allocation on same section fails. Signed-off-by: Yasunori Goto --- mm/sparse.c | 77 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++- 1 file changed, 76 insertions(+), 1 deletion(-) Index: current/mm/sparse.c =================================================================== --- current.orig/mm/sparse.c 2008-06-14 19:03:23.000000000 +0900 +++ current/mm/sparse.c 2008-06-14 20:41:39.000000000 +0900 @@ -269,16 +269,91 @@ } #endif /* CONFIG_MEMORY_HOTPLUG */ +#ifdef CONFIG_MEMORY_HOTREMOVE +static unsigned long * __init +sparse_early_usemap_alloc_section(unsigned long pnum) +{ + unsigned long section_nr; + struct mem_section *ms = __nr_to_section(pnum); + int nid = sparse_early_nid(ms); + struct pglist_data *pgdat = NODE_DATA(nid); + + /* + * Usemap's page can't be freed until freeing other sections + * which use it. And, pgdat has same feature. + * If section A has pgdat and section B has usemap for other + * sections (includes section A), both sections can't be removed, + * because there is the dependency each other. + * To solve above issue, this collects all usemap on the same section + * which has pgdat as much as possible. + */ + section_nr = pfn_to_section_nr(__pa(pgdat) >> PAGE_SHIFT); + return alloc_bootmem_section(usemap_size(), section_nr); +} + +static void __init check_usemap_section_nr(int nid, unsigned long *usemap) +{ + unsigned long usemap_snr, pgdat_snr; + static unsigned long old_usemap_snr = NR_MEM_SECTIONS; + static unsigned long old_pgdat_snr = NR_MEM_SECTIONS; + struct pglist_data *pgdat = NODE_DATA(nid); + int usemap_nid; + + usemap_snr = pfn_to_section_nr(__pa(usemap) >> PAGE_SHIFT); + pgdat_snr = pfn_to_section_nr(__pa(pgdat) >> PAGE_SHIFT); + if (usemap_snr == pgdat_snr) + return; + + if (old_usemap_snr == usemap_snr && old_pgdat_snr == pgdat_snr) + /* skip redundant message */ + return; + + old_usemap_snr = usemap_snr; + old_pgdat_snr = pgdat_snr; + + usemap_nid = sparse_early_nid(__nr_to_section(usemap_snr)); + if (usemap_nid != nid) { + printk("node %d must be removed before remove section %ld\n", + nid, usemap_snr); + return; + } + /* + * There is a dependency deadlock. + * Some platforms allow un-removable section because they will just + * gather other removable sections for dynamic partitioning. + * Just notify un-removable section's number here. + */ + printk(KERN_INFO "section %ld and %ld", usemap_snr, pgdat_snr); + printk(" can't be hotremoved due to dependency each other.\n"); +} +#else +static unsigned long * __init +sparse_early_usemap_alloc_section(unsigned long pnum) +{ + return NULL; +} + +static void __init check_usemap_section_nr(int nid, unsigned long *usemap) +{ +} +#endif /* CONFIG_MEMORY_HOTREMOVE */ + static unsigned long *__init sparse_early_usemap_alloc(unsigned long pnum) { unsigned long *usemap; struct mem_section *ms = __nr_to_section(pnum); int nid = sparse_early_nid(ms); - usemap = alloc_bootmem_node(NODE_DATA(nid), usemap_size()); + usemap = sparse_early_usemap_alloc_section(pnum); if (usemap) return usemap; + usemap = alloc_bootmem_node(NODE_DATA(nid), usemap_size()); + if (usemap) { + check_usemap_section_nr(nid, usemap); + return usemap; + } + /* Stupid: suppress gcc warning for SPARSEMEM && !NUMA */ nid = 0; -- Yasunori Goto -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/