Received: by 2002:ac0:a5a7:0:0:0:0:0 with SMTP id m36-v6csp301392imm; Tue, 21 Aug 2018 20:10:34 -0700 (PDT) X-Google-Smtp-Source: AA+uWPzzpWcxqr3a+GXBT61f2iFk8Ci4OZs86II/k3glcJfHdRmhwPVuHgAWTWEFN8uyzwPr9o2M X-Received: by 2002:a63:4386:: with SMTP id q128-v6mr49475668pga.353.1534907434872; Tue, 21 Aug 2018 20:10:34 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1534907434; cv=none; d=google.com; s=arc-20160816; b=vNS5/gYnH0Fm8sL0Q/bWmbkEb7xSrr3X8xcgt6R5rf5kOLoioKrH/mxhqoI8UIYbPw HzGi1gNcGrjz+ml1yvXEev9FuJIRJuj+eVc5svku/824F6biDGxbMWy69yAWj8NvOG7y 8Y4twGLI2W4//7pe2pJG4qEa978SN/C63EEQCX35CLN68UNcfsC/pyqMQpKL7wwkMomZ XMdOiBd5RQvRZAb1/pfCgxvQ5XFW/avDL4RvlvheBRTFtwJyBZvYq/LHiC7UsGTD/BdL 7udHJqxDdYEmFSMBHikU72YkwCyYDyO1qRtXduoJ8iL8eBgLlzn8uW8t6hI0wArIJFzA 1ugQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:dkim-signature:arc-authentication-results; bh=FZQVcSjva4KqT1QnxQl07tGVuHDUFsqtAvdlpz1t47Q=; b=XLpR1ZNYV2H012LCl/7DGdYboECd7+muZN0VuwEYKRCiHPuRsSebbNfPv8KPLTB5BQ Y9bPV3HuItgMzYHyB6TIwtjbSREKwLGvvcREDhewYEXNUqpIJEpzIdTUGd8rfa0j2DKY UARMBtU7ldvaPymYADz5RVY3rzSsWOgCMdi030FXVcye3uZX4lDnhfv+2TCVvNniMFQr lFvjHDsrK4dGhdXjnf2BHvFoI/a3dMDwDae8unhW3oxWhJUxL0+6vsQLztsWjxMSWogH tsPrxtsKnyarvv6qmI05OwchRVpomlhAVSiBd31XxJKohr1pGjqehuzdTt5Sergf8k6H UzLQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=dUBGkRw4; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id 194-v6si633587pfz.101.2018.08.21.20.09.57; Tue, 21 Aug 2018 20:10:34 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=dUBGkRw4; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728066AbeHVGaw (ORCPT + 99 others); Wed, 22 Aug 2018 02:30:52 -0400 Received: from mail-pl0-f68.google.com ([209.85.160.68]:37952 "EHLO mail-pl0-f68.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726469AbeHVGaw (ORCPT ); Wed, 22 Aug 2018 02:30:52 -0400 Received: by mail-pl0-f68.google.com with SMTP id u11-v6so248510plq.5 for ; Tue, 21 Aug 2018 20:08:01 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=FZQVcSjva4KqT1QnxQl07tGVuHDUFsqtAvdlpz1t47Q=; b=dUBGkRw4fQaiMecEzf8cFWCp6lhVcwMzGzSJuhJ/kDZkOS/u/pcLbU9y5Xf1q1j77r dtL/gCMhC9eqDMKxzYrXtYRlA5a94kZIwjuX+W8VHjreqVs9v6Dscjei2XBBlnncnM4p n73u1ktG72mHDkQonOkyTO1/t4u/t1qqLjVMvDj7/lvdikAWO7X+SCQYDPl7U2GSVlfJ M2ft/3Qfgwnj4P1+9mUHj4676Eku6rYRBV32xYXzPBdGb0x3G/XnuydXMwFf72akSOpp 7Krekn3z/4ghm2DDp1SZzdN5QUHqpfUvUf0gya3hnJ/6JYU5XPlrKy0IN3wXa46RWS/6 MJeA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=FZQVcSjva4KqT1QnxQl07tGVuHDUFsqtAvdlpz1t47Q=; b=SSloIjYQ4y7IHK3mg6CXj/Nwv7Js77j3pKwLs0gmKhEgHW2aF0bln+nTR2JU7a846A ohqnbTDciu2dCTehGaIWlJr/4QeVSdT04XHx4ZtjQ+Ltay1msizQMYYYYu2TIQ5y1rTc LRMyWM5Jl6FOsvEU3Vn/oZgdUA8tPFcFpJPfhB3e1C29XDj0gO9wkZwHx1pO2yMV+9WH KePNV/uWwCrxNGlG6RMsq9kYWsvRK/9LIg/vG8zuAiBG2Eu8ch9q/7IdaCsykH1klk/d ZxAqpWnwK5gGj0vJN9x9oY6vC7hOa7Yl3S6w+eRZd+AphaT2zTHXclaspr1vHmDxigmw LzdA== X-Gm-Message-State: AOUpUlGARP8a6z0Q6EtvCBKrIIbeeyGFD+fCLxY/HleUt1dCAFaNEG1V REV7BYlkU77Aq/BHcx1gE48= X-Received: by 2002:a17:902:42e2:: with SMTP id h89-v6mr51827967pld.69.1534907281175; Tue, 21 Aug 2018 20:08:01 -0700 (PDT) Received: from ct7host.localdomain ([38.106.11.25]) by smtp.gmail.com with ESMTPSA id q7-v6sm440165pfq.43.2018.08.21.20.07.49 (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Tue, 21 Aug 2018 20:08:00 -0700 (PDT) From: Jia He X-Google-Original-From: Jia He To: Russell King , Catalin Marinas , Will Deacon , Mark Rutland , Ard Biesheuvel , Andrew Morton , Michal Hocko Cc: Wei Yang , Kees Cook , Laura Abbott , Vladimir Murzin , Philip Derrin , AKASHI Takahiro , James Morse , Steve Capper , Gioh Kim , Vlastimil Babka , Mel Gorman , Johannes Weiner , Kemi Wang , Petr Tesarik , YASUAKI ISHIMATSU , Andrey Ryabinin , Nikolay Borisov , Daniel Jordan , Daniel Vacek , Eugeniu Rosca , linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, Jia He Subject: [PATCH v11 2/3] mm: page_alloc: remain memblock_next_valid_pfn() on arm/arm64 Date: Wed, 22 Aug 2018 11:07:16 +0800 Message-Id: <1534907237-2982-3-git-send-email-jia.he@hxt-semitech.com> X-Mailer: git-send-email 1.8.3.1 In-Reply-To: <1534907237-2982-1-git-send-email-jia.he@hxt-semitech.com> References: <1534907237-2982-1-git-send-email-jia.he@hxt-semitech.com> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Commit b92df1de5d28 ("mm: page_alloc: skip over regions of invalid pfns where possible") optimized the loop in memmap_init_zone(). But it causes possible panic bug. So Daniel Vacek reverted it later. But as suggested by Daniel Vacek, it is fine to using memblock to skip gaps and finding next valid frame with CONFIG_HAVE_ARCH_PFN_VALID. Daniel said: "On arm and arm64, memblock is used by default. But generic version of pfn_valid() is based on mem sections and memblock_next_valid_pfn() does not always return the next valid one but skips more resulting in some valid frames to be skipped (as if they were invalid). And that's why kernel was eventually crashing on some !arm machines." About the performance consideration: As said by James in b92df1de5, "I have tested this patch on a virtual model of a Samurai CPU with a sparse memory map. The kernel boot time drops from 109 to 62 seconds." Thus it would be better if we remain memblock_next_valid_pfn on arm/arm64. Suggested-by: Daniel Vacek Signed-off-by: Jia He --- include/linux/mmzone.h | 9 +++++++++ mm/memblock.c | 30 ++++++++++++++++++++++++++++++ mm/page_alloc.c | 5 ++++- 3 files changed, 43 insertions(+), 1 deletion(-) diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h index 32699b2..8e5e20b 100644 --- a/include/linux/mmzone.h +++ b/include/linux/mmzone.h @@ -1266,6 +1266,10 @@ static inline int pfn_present(unsigned long pfn) #endif #define early_pfn_valid(pfn) pfn_valid(pfn) +#ifdef CONFIG_HAVE_MEMBLOCK_PFN_VALID +extern unsigned long memblock_next_valid_pfn(unsigned long pfn); +#define next_valid_pfn(pfn) memblock_next_valid_pfn(pfn) +#endif void sparse_init(void); #else #define sparse_init() do {} while (0) @@ -1287,6 +1291,11 @@ struct mminit_pfnnid_cache { #define early_pfn_valid(pfn) (1) #endif +/* fallback to default definitions*/ +#ifndef next_valid_pfn +#define next_valid_pfn(pfn) (pfn + 1) +#endif + void memory_present(int nid, unsigned long start, unsigned long end); /* diff --git a/mm/memblock.c b/mm/memblock.c index 3d03866..077ca62 100644 --- a/mm/memblock.c +++ b/mm/memblock.c @@ -1140,6 +1140,36 @@ int __init_memblock memblock_set_node(phys_addr_t base, phys_addr_t size, } #endif /* CONFIG_HAVE_MEMBLOCK_NODE_MAP */ +#ifdef CONFIG_HAVE_MEMBLOCK_PFN_VALID +unsigned long __init_memblock memblock_next_valid_pfn(unsigned long pfn) +{ + struct memblock_type *type = &memblock.memory; + unsigned int right = type->cnt; + unsigned int mid, left = 0; + phys_addr_t addr = PFN_PHYS(++pfn); + + do { + mid = (right + left) / 2; + + if (addr < type->regions[mid].base) + right = mid; + else if (addr >= (type->regions[mid].base + + type->regions[mid].size)) + left = mid + 1; + else { + /* addr is within the region, so pfn is valid */ + return pfn; + } + } while (left < right); + + if (right == type->cnt) + return -1UL; + else + return PHYS_PFN(type->regions[right].base); +} +EXPORT_SYMBOL(memblock_next_valid_pfn); +#endif /*CONFIG_HAVE_MEMBLOCK_PFN_VALID*/ + static phys_addr_t __init memblock_alloc_range_nid(phys_addr_t size, phys_addr_t align, phys_addr_t start, phys_addr_t end, int nid, ulong flags) diff --git a/mm/page_alloc.c b/mm/page_alloc.c index cd3c7b9..607deff 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -5485,8 +5485,11 @@ void __meminit memmap_init_zone(unsigned long size, int nid, unsigned long zone, if (context != MEMMAP_EARLY) goto not_early; - if (!early_pfn_valid(pfn)) + if (!early_pfn_valid(pfn)) { + pfn = next_valid_pfn(pfn) - 1; continue; + } + if (!early_pfn_in_nid(pfn, nid)) continue; if (!update_defer_init(pgdat, pfn, end_pfn, &nr_initialised)) -- 1.8.3.1