Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1765636Ab3DDXxd (ORCPT ); Thu, 4 Apr 2013 19:53:33 -0400 Received: from aserp1040.oracle.com ([141.146.126.69]:32192 "EHLO aserp1040.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1764919Ab3DDXrz (ORCPT ); Thu, 4 Apr 2013 19:47:55 -0400 From: Yinghai Lu To: Thomas Gleixner , Ingo Molnar , "H. Peter Anvin" , Andrew Morton , Tejun Heo , Thomas Renninger , Tang Chen Cc: linux-kernel@vger.kernel.org, Yinghai Lu Subject: [PATCH v3 11/22] x86, mm, numa: Call numa_meminfo_cover_memory() checking early Date: Thu, 4 Apr 2013 16:46:15 -0700 Message-Id: <1365119186-23487-12-git-send-email-yinghai@kernel.org> X-Mailer: git-send-email 1.8.1.4 In-Reply-To: <1365119186-23487-1-git-send-email-yinghai@kernel.org> References: <1365119186-23487-1-git-send-email-yinghai@kernel.org> X-Source-IP: acsinet21.oracle.com [141.146.126.237] Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 3034 Lines: 85 For the separation, we need to set memblock nid later, as it could change memblock array, and possible doube memblock.memory array that will need to allocate buffer. We do not need to use nid in memblock to find out absent pages. So we can move that numa_meminfo_cover_memory() early. Also could change __absent_pages_in_range() to static and use absent_pages_in_range() directly. Later we can only set memblock nid one time on successful path. Signed-off-by: Yinghai Lu --- arch/x86/mm/numa.c | 7 ++++--- include/linux/mm.h | 2 -- mm/page_alloc.c | 2 +- 3 files changed, 5 insertions(+), 6 deletions(-) diff --git a/arch/x86/mm/numa.c b/arch/x86/mm/numa.c index d545638..b7173f6 100644 --- a/arch/x86/mm/numa.c +++ b/arch/x86/mm/numa.c @@ -460,7 +460,7 @@ static bool __init numa_meminfo_cover_memory(const struct numa_meminfo *mi) u64 s = mi->blk[i].start >> PAGE_SHIFT; u64 e = mi->blk[i].end >> PAGE_SHIFT; numaram += e - s; - numaram -= __absent_pages_in_range(mi->blk[i].nid, s, e); + numaram -= absent_pages_in_range(s, e); if ((s64)numaram < 0) numaram = 0; } @@ -488,6 +488,9 @@ static int __init numa_register_memblks(struct numa_meminfo *mi) if (WARN_ON(nodes_empty(node_possible_map))) return -EINVAL; + if (!numa_meminfo_cover_memory(mi)) + return -EINVAL; + for (i = 0; i < mi->nr_blks; i++) { struct numa_memblk *mb = &mi->blk[i]; memblock_set_node(mb->start, mb->end - mb->start, mb->nid); @@ -506,8 +509,6 @@ static int __init numa_register_memblks(struct numa_meminfo *mi) return -EINVAL; } #endif - if (!numa_meminfo_cover_memory(mi)) - return -EINVAL; return 0; } diff --git a/include/linux/mm.h b/include/linux/mm.h index e19ff30..192806c 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -1323,8 +1323,6 @@ extern void free_initmem(void); */ extern void free_area_init_nodes(unsigned long *max_zone_pfn); unsigned long node_map_pfn_alignment(void); -unsigned long __absent_pages_in_range(int nid, unsigned long start_pfn, - unsigned long end_pfn); extern unsigned long absent_pages_in_range(unsigned long start_pfn, unsigned long end_pfn); extern void get_pfn_range_for_nid(unsigned int nid, diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 8fcced7..580d919 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -4356,7 +4356,7 @@ static unsigned long __meminit zone_spanned_pages_in_node(int nid, * Return the number of holes in a range on a node. If nid is MAX_NUMNODES, * then all holes in the requested range will be accounted for. */ -unsigned long __meminit __absent_pages_in_range(int nid, +static unsigned long __meminit __absent_pages_in_range(int nid, unsigned long range_start_pfn, unsigned long range_end_pfn) { -- 1.8.1.4 -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/