Received: by 2002:ac0:a582:0:0:0:0:0 with SMTP id m2-v6csp1191029imm; Wed, 10 Oct 2018 10:28:08 -0700 (PDT) X-Google-Smtp-Source: ACcGV61EYRJeclrCb4RSiOhz5ZxE52qCRfxlw/dZGYsBoWNQgFyG6m7mCbTHrteC5t1lM07SksaR X-Received: by 2002:a17:902:6bc2:: with SMTP id m2-v6mr865289plt.106.1539192488314; Wed, 10 Oct 2018 10:28:08 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1539192488; cv=none; d=google.com; s=arc-20160816; b=m72B/NMgs7FB5psuT7s0ywcXoM8GptfbDXdndDCYfHEteB6/dMtFFShnJlQYx23Gp9 8u9EYDkMTLl7APAJv8vk7BPFohyDosZ2iiDMwe8sdivhWQNA030++AveVFLzMkeuySMr fpEtqBaG6UC03tcOaOkCOLXGLHJ7EoiJ52gIs5ehU0B4WX51NbHF8NnSj/Gj6WX/3geC ZgeNEyhHb3z+e/dOoijHZF6e8eQCLNYPX7y8N+xi19RpJfEJbUGf7VtW/TUSI05fdkuK 4YqNxyM4xJL9D6qplEEvkbU5KXlOWvg1aG0Sw3WU1BLnkJvZxWtM/gCZ2zf6J/n4svng qShQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject:dkim-signature; bh=PGvn9CnLhpOFTpMI5e3BYHlOocnpI5JpBnaTYLDkkak=; b=x6n60BgOxukuA6XDbO7m7CcRUMgV/650svXwRE5NOqyh0K4XL/np8G1OukCZnXfCnj pJsa0AsrZOZiMrFxUJGl+xWEdBlwIZ0o+EhkszpDQ6fYlwYWkVEGP9JigGeY2kVPa+so cKBFNxaQkrPcUiKyKdWmMNFXZMxfrmgBw1obNCgTFeHwxL8VlXoq65q0DZgX0hNavLjp iW7N0fKRfxRWbiv1/ZSQskS/BRPYNNzmrz9Td77phqOXIjRPu0voZnF55DtkniUt/gF0 ijN3oiB7H0usF4Zuqy/WUgtI0y3I4g/NcgFfhYakxaxixkQ07ZQHy9G5cRuVtHmSQgKk n4tA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=Kv9xQMZ5; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id a3-v6si24570152plp.199.2018.10.10.10.27.53; Wed, 10 Oct 2018 10:28:08 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=Kv9xQMZ5; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727248AbeJKAuk (ORCPT + 99 others); Wed, 10 Oct 2018 20:50:40 -0400 Received: from mail-yb1-f194.google.com ([209.85.219.194]:32966 "EHLO mail-yb1-f194.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726666AbeJKAuk (ORCPT ); Wed, 10 Oct 2018 20:50:40 -0400 Received: by mail-yb1-f194.google.com with SMTP id u88-v6so2536329ybi.0 for ; Wed, 10 Oct 2018 10:27:31 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=subject:to:cc:references:from:message-id:date:user-agent :mime-version:in-reply-to:content-language:content-transfer-encoding; bh=PGvn9CnLhpOFTpMI5e3BYHlOocnpI5JpBnaTYLDkkak=; b=Kv9xQMZ5pboLUrfd8znFOPVqLLdJib3XHV5F31GLx6WxcLM899IsccntDxgdT1BElE GPpCWeldyN7KC7IuScOE1lYmayNJJ8dxnyL/WFQdSF1Fb0TXPJ2g6qaENOexfFnNg9d2 620Sv9taY7FetYaL9TILZOOC7wl/BYbsaFcX+r/Mpu7n7cJVJBG1YXn8QLIVSlHO2ywA LvE85Ikp3Q4Jv/CrORNeI3/XXkOE21oHdoV14+TTHqQcx4yuPwQ4XtlDKM6qSyYngbtD iy7/SmBMF+5+k18VvDe+OoRXbTyC301emrM8OdLsDvZSVUZRmTIKqCIX5jfYWGfbDs11 av+w== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:subject:to:cc:references:from:message-id:date :user-agent:mime-version:in-reply-to:content-language :content-transfer-encoding; bh=PGvn9CnLhpOFTpMI5e3BYHlOocnpI5JpBnaTYLDkkak=; b=uioyckctZD5jhZFnLWwU4KE22uMn1AMh+kRvjrsXqFFJHfvEH+EGYYhgu+GzAqHRlR 9KbrMCyZ5THJBSNchLAnbhbZlxHkvb8FeHdoMXQYh/F4gnoKcnvk3pkFr9XlfqBpUoxS 3Fo0pD7qFV6Ig6dB6s/ygVpSd2BQApSEoUzJd/66owL5Jf2C6u7rmiiP0KOKgYPO6PQ4 Mp9S6WZ/MOaqY5iVI0guvdAyVR+BkWkrvce3ITCg8ZqTHP1xf6JS1vpXedmU4GryOfuT t8ma6qQSvAKBWS4IQbWccGahpcVKpJzafVUs9QRuXz4YC53EyFKWjjIWPSYEQtjqha2+ 7d5A== X-Gm-Message-State: ABuFfohfZBzcrzDOl4owS6gQN99yMYBM6RxrY5ZS5FSgmKhCC66uB+Yw omhU/bF1+9sO0hJ/dlxcd6Q= X-Received: by 2002:a25:3186:: with SMTP id x128-v6mr19727755ybx.227.1539192451027; Wed, 10 Oct 2018 10:27:31 -0700 (PDT) Received: from [192.168.1.10] (c-73-69-118-222.hsd1.nh.comcast.net. [73.69.118.222]) by smtp.gmail.com with ESMTPSA id o202-v6sm40463212ywo.38.2018.10.10.10.27.29 (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 10 Oct 2018 10:27:30 -0700 (PDT) Subject: Re: [PATCH v3 1/3] mm: zero remaining unavailable struct pages To: Masayoshi Mizuma , linux-mm@kvack.org, Naoya Horiguchi , Pavel Tatashin , Michal Hocko , Thomas Gleixner , Ingo Molnar Cc: linux-kernel@vger.kernel.org, x86@kernel.org References: <20181002143821.5112-1-msys.mizuma@gmail.com> <20181002143821.5112-2-msys.mizuma@gmail.com> From: Pavel Tatashin Message-ID: <7a03eb03-89c4-8f78-b169-b6ce18e0d5a7@gmail.com> Date: Wed, 10 Oct 2018 13:27:20 -0400 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.0 MIME-Version: 1.0 In-Reply-To: <20181002143821.5112-2-msys.mizuma@gmail.com> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Reviewed-by: Pavel Tatashin On 10/2/18 10:38 AM, Masayoshi Mizuma wrote: > From: Naoya Horiguchi > > There is a kernel panic that is triggered when reading /proc/kpageflags > on the kernel booted with kernel parameter 'memmap=nn[KMG]!ss[KMG]': > > BUG: unable to handle kernel paging request at fffffffffffffffe > PGD 9b20e067 P4D 9b20e067 PUD 9b210067 PMD 0 > Oops: 0000 [#1] SMP PTI > CPU: 2 PID: 1728 Comm: page-types Not tainted 4.17.0-rc6-mm1-v4.17-rc6-180605-0816-00236-g2dfb086ef02c+ #160 > Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.11.0-2.fc28 04/01/2014 > RIP: 0010:stable_page_flags+0x27/0x3c0 > Code: 00 00 00 0f 1f 44 00 00 48 85 ff 0f 84 a0 03 00 00 41 54 55 49 89 fc 53 48 8b 57 08 48 8b 2f 48 8d 42 ff 83 e2 01 48 0f 44 c7 <48> 8b 00 f6 c4 01 0f 84 10 03 00 00 31 db 49 8b 54 24 08 4c 89 e7 > RSP: 0018:ffffbbd44111fde0 EFLAGS: 00010202 > RAX: fffffffffffffffe RBX: 00007fffffffeff9 RCX: 0000000000000000 > RDX: 0000000000000001 RSI: 0000000000000202 RDI: ffffed1182fff5c0 > RBP: ffffffffffffffff R08: 0000000000000001 R09: 0000000000000001 > R10: ffffbbd44111fed8 R11: 0000000000000000 R12: ffffed1182fff5c0 > R13: 00000000000bffd7 R14: 0000000002fff5c0 R15: ffffbbd44111ff10 > FS: 00007efc4335a500(0000) GS:ffff93a5bfc00000(0000) knlGS:0000000000000000 > CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > CR2: fffffffffffffffe CR3: 00000000b2a58000 CR4: 00000000001406e0 > Call Trace: > kpageflags_read+0xc7/0x120 > proc_reg_read+0x3c/0x60 > __vfs_read+0x36/0x170 > vfs_read+0x89/0x130 > ksys_pread64+0x71/0x90 > do_syscall_64+0x5b/0x160 > entry_SYSCALL_64_after_hwframe+0x44/0xa9 > RIP: 0033:0x7efc42e75e23 > Code: 09 00 ba 9f 01 00 00 e8 ab 81 f4 ff 66 2e 0f 1f 84 00 00 00 00 00 90 83 3d 29 0a 2d 00 00 75 13 49 89 ca b8 11 00 00 00 0f 05 <48> 3d 01 f0 ff ff 73 34 c3 48 83 ec 08 e8 db d3 01 00 48 89 04 24 > > According to kernel bisection, this problem became visible due to commit > f7f99100d8d9 which changes how struct pages are initialized. > > Memblock layout affects the pfn ranges covered by node/zone. Consider > that we have a VM with 2 NUMA nodes and each node has 4GB memory, and > the default (no memmap= given) memblock layout is like below: > > MEMBLOCK configuration: > memory size = 0x00000001fff75c00 reserved size = 0x000000000300c000 > memory.cnt = 0x4 > memory[0x0] [0x0000000000001000-0x000000000009efff], 0x000000000009e000 bytes on node 0 flags: 0x0 > memory[0x1] [0x0000000000100000-0x00000000bffd6fff], 0x00000000bfed7000 bytes on node 0 flags: 0x0 > memory[0x2] [0x0000000100000000-0x000000013fffffff], 0x0000000040000000 bytes on node 0 flags: 0x0 > memory[0x3] [0x0000000140000000-0x000000023fffffff], 0x0000000100000000 bytes on node 1 flags: 0x0 > ... > > If you give memmap=1G!4G (so it just covers memory[0x2]), > the range [0x100000000-0x13fffffff] is gone: > > MEMBLOCK configuration: > memory size = 0x00000001bff75c00 reserved size = 0x000000000300c000 > memory.cnt = 0x3 > memory[0x0] [0x0000000000001000-0x000000000009efff], 0x000000000009e000 bytes on node 0 flags: 0x0 > memory[0x1] [0x0000000000100000-0x00000000bffd6fff], 0x00000000bfed7000 bytes on node 0 flags: 0x0 > memory[0x2] [0x0000000140000000-0x000000023fffffff], 0x0000000100000000 bytes on node 1 flags: 0x0 > ... > > This causes shrinking node 0's pfn range because it is calculated by > the address range of memblock.memory. So some of struct pages in the > gap range are left uninitialized. > > We have a function zero_resv_unavail() which does zeroing the struct > pages outside memblock.memory, but currently it covers only the reserved > unavailable range (i.e. memblock.memory && !memblock.reserved). > This patch extends it to cover all unavailable range, which fixes > the reported issue. > > Fixes: f7f99100d8d9 ("mm: stop zeroing memory during allocation in vmemmap") > Signed-off-by: Naoya Horiguchi > Tested-by: Oscar Salvador > Tested-by: Masayoshi Mizuma > --- > include/linux/memblock.h | 15 --------------- > mm/page_alloc.c | 36 +++++++++++++++++++++++++----------- > 2 files changed, 25 insertions(+), 26 deletions(-) > > diff --git a/include/linux/memblock.h b/include/linux/memblock.h > index 5169205..2acdd04 100644 > --- a/include/linux/memblock.h > +++ b/include/linux/memblock.h > @@ -265,21 +265,6 @@ void __next_mem_pfn_range(int *idx, int nid, unsigned long *out_start_pfn, > for_each_mem_range_rev(i, &memblock.memory, &memblock.reserved, \ > nid, flags, p_start, p_end, p_nid) > > -/** > - * for_each_resv_unavail_range - iterate through reserved and unavailable memory > - * @i: u64 used as loop variable > - * @p_start: ptr to phys_addr_t for start address of the range, can be %NULL > - * @p_end: ptr to phys_addr_t for end address of the range, can be %NULL > - * > - * Walks over unavailable but reserved (reserved && !memory) areas of memblock. > - * Available as soon as memblock is initialized. > - * Note: because this memory does not belong to any physical node, flags and > - * nid arguments do not make sense and thus not exported as arguments. > - */ > -#define for_each_resv_unavail_range(i, p_start, p_end) \ > - for_each_mem_range(i, &memblock.reserved, &memblock.memory, \ > - NUMA_NO_NODE, MEMBLOCK_NONE, p_start, p_end, NULL) > - > static inline void memblock_set_region_flags(struct memblock_region *r, > enum memblock_flags flags) > { > diff --git a/mm/page_alloc.c b/mm/page_alloc.c > index 89d2a2a..3b9d89e 100644 > --- a/mm/page_alloc.c > +++ b/mm/page_alloc.c > @@ -6446,29 +6446,42 @@ void __init free_area_init_node(int nid, unsigned long *zones_size, > * struct pages which are reserved in memblock allocator and their fields > * may be accessed (for example page_to_pfn() on some configuration accesses > * flags). We must explicitly zero those struct pages. > + * > + * This function also addresses a similar issue where struct pages are left > + * uninitialized because the physical address range is not covered by > + * memblock.memory or memblock.reserved. That could happen when memblock > + * layout is manually configured via memmap=. > */ > void __init zero_resv_unavail(void) > { > phys_addr_t start, end; > unsigned long pfn; > u64 i, pgcnt; > + phys_addr_t next = 0; > > /* > - * Loop through ranges that are reserved, but do not have reported > - * physical memory backing. > + * Loop through unavailable ranges not covered by memblock.memory. > */ > pgcnt = 0; > - for_each_resv_unavail_range(i, &start, &end) { > - for (pfn = PFN_DOWN(start); pfn < PFN_UP(end); pfn++) { > - if (!pfn_valid(ALIGN_DOWN(pfn, pageblock_nr_pages))) { > - pfn = ALIGN_DOWN(pfn, pageblock_nr_pages) > - + pageblock_nr_pages - 1; > - continue; > + for_each_mem_range(i, &memblock.memory, NULL, > + NUMA_NO_NODE, MEMBLOCK_NONE, &start, &end, NULL) { > + if (next < start) { > + for (pfn = PFN_DOWN(next); pfn < PFN_UP(start); pfn++) { > + if (!pfn_valid(ALIGN_DOWN(pfn, pageblock_nr_pages))) > + continue; > + mm_zero_struct_page(pfn_to_page(pfn)); > + pgcnt++; > } > - mm_zero_struct_page(pfn_to_page(pfn)); > - pgcnt++; > } > + next = end; > } > + for (pfn = PFN_DOWN(next); pfn < max_pfn; pfn++) { > + if (!pfn_valid(ALIGN_DOWN(pfn, pageblock_nr_pages))) > + continue; > + mm_zero_struct_page(pfn_to_page(pfn)); > + pgcnt++; > + } > + > > /* > * Struct pages that do not have backing memory. This could be because > @@ -6478,7 +6491,8 @@ void __init zero_resv_unavail(void) > * this code can be removed. > */ > if (pgcnt) > - pr_info("Reserved but unavailable: %lld pages", pgcnt); > + pr_info("Zeroed struct page in unavailable ranges: %lld pages", pgcnt); > + > } > #endif /* CONFIG_HAVE_MEMBLOCK && !CONFIG_FLAT_NODE_MEM_MAP */ > >