Received: by 2002:ac0:a5a7:0:0:0:0:0 with SMTP id m36-v6csp1275969imm; Wed, 25 Jul 2018 15:03:37 -0700 (PDT) X-Google-Smtp-Source: AAOMgpesfuRAjEZbHNcJXSbopC5MYugh2o7x5cgMHpjKrVI7Da1esbUqD/fogZbR+sqTjM+YrVP5 X-Received: by 2002:a17:902:bc8b:: with SMTP id bb11-v6mr2675023plb.112.1532556216996; Wed, 25 Jul 2018 15:03:36 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1532556216; cv=none; d=google.com; s=arc-20160816; b=FgWXfEPji9QZeAGifUSzHgxBFmf/T6odXOt7uBadl1GBWsY5alT99JaVwsw85xqyYm x5feENqLv0kDeIPVRIarGN93K7N8WlfpoWlsCs0cNSlENnQL/xjrL9cKsXcbPLOARgsA i9jbZbytXnw+nDiUMDFZTjqMMRG/fFdm88MVnirGGSP+VZbU3+WF8LmHe6bA1k7sBK3T vVESEgoFTdWh8A26C/TAw3rQ3/R0/mC60pULnajcGXuaCctzw+9JJNJAq5xOj3pGlYCp IOa8VWJfdWIOMh7IdIuFMDGI0hrMwdARQz2Zk3rNjoIZaKfIX942V5IXZCAvQtTi32BS nnYA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:arc-authentication-results; bh=rskVcuLW5mTCFv880bF+K7VIucHqpnfomG/ohSkecG0=; b=jSs7D8iKvqU8gcOWX0hr3QAby231eAyuOZAZmjU5RhO52QSE45RR8BuD9RMD3YNSc8 527hzeEoarZICjds4PzkGXbZ53RZeC5x4RHD/jxt4KNIaWkvLuvDR6+VwouUj+vnhmYH xiBFAsBqOAnelS5emxHcEnbmuCIsHXm2jDelStSmL/UKtJZ3Fnwc1fuCYJ92F3hWGm7N 0v0/rlMOrHqQhmu5HNzACeBCrNYJ16M5shdyxAIwNl3vpQT88SfJOoppxhEndkK0Y+N3 Ejt5xZ+GL/c3uazwmjRd1j7ZKuKi+rg/E82MuYm7L9GCSwf1MB74iRtN522M3ny3KgXT dPvA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id x186-v6si14392405pgd.259.2018.07.25.15.03.09; Wed, 25 Jul 2018 15:03:36 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1731545AbeGYXPl (ORCPT + 99 others); Wed, 25 Jul 2018 19:15:41 -0400 Received: from mail-wr1-f67.google.com ([209.85.221.67]:35311 "EHLO mail-wr1-f67.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1731409AbeGYXPf (ORCPT ); Wed, 25 Jul 2018 19:15:35 -0400 Received: by mail-wr1-f67.google.com with SMTP id a3-v6so8768006wrt.2 for ; Wed, 25 Jul 2018 15:01:54 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=rskVcuLW5mTCFv880bF+K7VIucHqpnfomG/ohSkecG0=; b=B7qOJNxxBrFjdXl7VSSkAZO0J1uhF+b2LA1sAPiT7CjfnfU8m2qpU1jh55XvFUcs5v 0W+pwqQF1TWPCKUjie9MR6lSrNkcioq79QA2F6OJEh3yxdVXL+pJswR3H9uoXfohh6bt swKgN5WqnCqf/HgJLbOmjoUxwaDwH3WzRqJyNxgEiXRSrPCvFhAesGd8dIvYZeP6CD0w 3EKKAEgngydcbgGJ2pZrbFc4+HmmyM/Zj6ZfN4xhU0lhpguh3M6XLz7HUe8SjoyvHVD9 SDsiN0yB9WbcN/rTJGeb4Jd/Dg4SqywNy6TmdHczY7VpYa60kXVQe+9XQHnkBPwhyOhg FRCg== X-Gm-Message-State: AOUpUlFUwTjrxwEYIqrRpaoc57yqnv71CjPxMwLFvF2Y33vIdDukZMLJ ccj/YvCMp55M8bM3VFljRlU= X-Received: by 2002:adf:fac1:: with SMTP id a1-v6mr14978854wrs.74.1532556114299; Wed, 25 Jul 2018 15:01:54 -0700 (PDT) Received: from techadventures.net (techadventures.net. [62.201.165.239]) by smtp.gmail.com with ESMTPSA id d14-v6sm20991515wrr.12.2018.07.25.15.01.52 (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 25 Jul 2018 15:01:53 -0700 (PDT) Received: from d104.suse.de (charybdis-ext.suse.de [195.135.221.2]) by techadventures.net (Postfix) with ESMTPA id 6E8C6124368; Thu, 26 Jul 2018 00:01:51 +0200 (CEST) From: osalvador@techadventures.net To: akpm@linux-foundation.org Cc: mhocko@suse.com, vbabka@suse.cz, pasha.tatashin@oracle.com, mgorman@techsingularity.net, aaron.lu@intel.com, iamjoonsoo.kim@lge.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org, dan.j.williams@intel.com, Oscar Salvador Subject: [PATCH v3 5/5] mm/page_alloc: Introduce memhotplug version of free_area_init_core Date: Thu, 26 Jul 2018 00:01:44 +0200 Message-Id: <20180725220144.11531-6-osalvador@techadventures.net> X-Mailer: git-send-email 2.13.6 In-Reply-To: <20180725220144.11531-1-osalvador@techadventures.net> References: <20180725220144.11531-1-osalvador@techadventures.net> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Oscar Salvador Currently, we call free_area_init_node() from the memhotplug path. In there, we set some pgdat's fields, and call calculate_node_totalpages(). calculate_node_totalpages() calculates the # of pages the node has. Since the node is either new, or we are re-using it, the zones belonging to this node should not have any pages, so there is no point to calculate this now. Actually, we set the values to 0 later on with the calls to: reset_node_managed_pages() reset_node_present_pages() The # of pages per zone and the # of pages per zone will be calculated later on when onlining the pages: online_pages()->move_pfn_range()->move_pfn_range_to_zone()->resize_zone_range() online_pages()->move_pfn_range()->move_pfn_range_to_zone()->resize_pgdat_range() This introduces the memhotplug version of free_area_init_core and makes hotadd_new_pgdat use it. This function will only be called from memhotplug path: hotadd_new_pgdat()->free_area_init_core_hotplug(). free_area_init_core_hotplug() performs only a subset of the actions that free_area_init_core_hotplug() does. Signed-off-by: Oscar Salvador --- include/linux/mm.h | 1 + mm/memory_hotplug.c | 23 +++++++++-------------- mm/page_alloc.c | 19 +++++++++++++++++++ 3 files changed, 29 insertions(+), 14 deletions(-) diff --git a/include/linux/mm.h b/include/linux/mm.h index 6954ad183159..20430becd908 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -2001,6 +2001,7 @@ extern void free_area_init(unsigned long * zones_size); extern void free_area_init_node(int nid, unsigned long * zones_size, unsigned long zone_start_pfn, unsigned long *zholes_size); extern void free_initmem(void); +extern void free_area_init_core_hotplug(int nid); /* * Free reserved pages within range [PAGE_ALIGN(start), end & PAGE_MASK) diff --git a/mm/memory_hotplug.c b/mm/memory_hotplug.c index 4eb6e824a80c..bef8a3f7a760 100644 --- a/mm/memory_hotplug.c +++ b/mm/memory_hotplug.c @@ -978,12 +978,11 @@ static void reset_node_present_pages(pg_data_t *pgdat) pgdat->node_present_pages = 0; } + /* we are OK calling __meminit stuff here - we have CONFIG_MEMORY_HOTPLUG */ static pg_data_t __ref *hotadd_new_pgdat(int nid, u64 start) { struct pglist_data *pgdat; - unsigned long zones_size[MAX_NR_ZONES] = {0}; - unsigned long zholes_size[MAX_NR_ZONES] = {0}; unsigned long start_pfn = PFN_DOWN(start); pgdat = NODE_DATA(nid); @@ -1006,8 +1005,11 @@ static pg_data_t __ref *hotadd_new_pgdat(int nid, u64 start) /* we can use NODE_DATA(nid) from here */ + pgdat->node_id = nid; + pgdat->node_start_pfn = start_pfn; + /* init node's zones as empty zones, we don't have any present pages.*/ - free_area_init_node(nid, zones_size, start_pfn, zholes_size); + free_area_init_core_hotplug(nid); pgdat->per_cpu_nodestats = alloc_percpu(struct per_cpu_nodestat); /* @@ -1017,18 +1019,11 @@ static pg_data_t __ref *hotadd_new_pgdat(int nid, u64 start) build_all_zonelists(pgdat); /* - * zone->managed_pages is set to an approximate value in - * free_area_init_core(), which will cause - * /sys/device/system/node/nodeX/meminfo has wrong data. - * So reset it to 0 before any memory is onlined. - */ + * When memory is hot-added, all the memory is in offline state. So + * clear all zones' present_pages because they will be updated in + * online_pages() and offline_pages(). + */ reset_node_managed_pages(pgdat); - - /* - * When memory is hot-added, all the memory is in offline state. So - * clear all zones' present_pages because they will be updated in - * online_pages() and offline_pages(). - */ reset_node_present_pages(pgdat); return pgdat; diff --git a/mm/page_alloc.c b/mm/page_alloc.c index a455dc85da19..a36b4db26b50 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -6265,6 +6265,25 @@ static void zone_init_internals(struct zone *zone, enum zone_type idx, int nid, } /* + * Set up the zone data structures + * - init pgdat internals + * - init all zones belonging to this node + * + * NOTE: this function is only called during memory hotplug + */ +void __paginginit free_area_init_core_hotplug(int nid) +{ + enum zone_type j; + pg_data_t *pgdat = NODE_DATA(nid); + + pgdat_init_internals(pgdat); + for(j = 0; j < MAX_NR_ZONES; j++) { + struct zone *zone = pgdat->node_zones + j; + zone_init_internals(zone, j, nid, 0); + } +} + +/* * Set up the zone data structures: * - mark all pages reserved * - mark all memory queues empty -- 2.13.6