Received: by 2002:ac0:98c7:0:0:0:0:0 with SMTP id g7-v6csp3610076imd; Mon, 29 Oct 2018 09:37:00 -0700 (PDT) X-Google-Smtp-Source: AJdET5e+Kw95qppLlnw+BwaebuXZJ2JlS8uPm1M2HI8mmGnlr/4omwnL+BSrWrkY4Xh2J7n1v7AY X-Received: by 2002:a63:981:: with SMTP id 123mr14600936pgj.444.1540831020660; Mon, 29 Oct 2018 09:37:00 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1540831020; cv=none; d=google.com; s=arc-20160816; b=TtN+DfYTlHWOQ7xRvnbkGNOtN5nrNbMjwU16jqsona3QOBnQXoQildgWZTBMr3X14z DI1tv2rgYtdKSUXPAZn2eay+9Px89Y5YGeJNs/ZcK7TzuTt13f9/7KyXEF/6yPRNpnZR qVX2o7UXQtWUD8oAo3aiMlATLOKxhHMox64DNtUi3dvqOC++JepBRtNgMD50bM41Z1m9 LNUf9IrIQG59d8mWxG7Bs/LFrPtOO9fm+JAaMkRpQCXBg3H/iIOn7UwcK8wXKfvlmzg5 zCVQjfbCZVfPl/M7f24GBCjvVPA6rWt48+OIJhRs2cWuuGr/rkE+K08Luon3QxERlH2p BXvQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date; bh=YDXL6363cC+bkaVsO8fMGPW354VeEgERwKlHELJy8MA=; b=aP/N1moPims1skxUsybc3AtsBN/se3ZZVKh8BGqQGLbV/gg32T7nXHSJdbZd2n4JOA /BQbnpY9N2ss2QR0xxHJCMc5hXBmFpJePTJkTbJ83g8xpMTnbr9k6MaPqH/dXp061z09 qzfrnPBOoLWOiu+SfzS56b6bRTybcuQWdbG1YVnPRtYvv7vNJyypbKp9c+C9xNJwRAO7 WKJ/4gpE50SazyPFyNNWCMKOV0DGmnDOKMLhUp9XJsXUiDFSL0KQQ/+mENc5IfIMG5s0 4Q7lEdwTvCG2leZv2iwrCsi0uJsx3dNvBKbgL5yOq5zDHflh29qBBYq0iYG3WE2asDFs Shmw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id n5-v6si21068101pgg.186.2018.10.29.09.36.43; Mon, 29 Oct 2018 09:37:00 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728062AbeJ3BYw (ORCPT + 99 others); Mon, 29 Oct 2018 21:24:52 -0400 Received: from mx2.suse.de ([195.135.220.15]:40100 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1727872AbeJ3BYw (ORCPT ); Mon, 29 Oct 2018 21:24:52 -0400 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay1.suse.de (unknown [195.135.220.254]) by mx1.suse.de (Postfix) with ESMTP id A3941ACDE; Mon, 29 Oct 2018 16:35:29 +0000 (UTC) Date: Mon, 29 Oct 2018 17:35:28 +0100 From: Michal Hocko To: Alexander Duyck Cc: Dan Williams , Linux MM , Andrew Morton , Linux Kernel Mailing List , linux-nvdimm , Pasha Tatashin , Dave Hansen , =?iso-8859-1?B?Suly9G1l?= Glisse , rppt@linux.vnet.ibm.com, Ingo Molnar , "Kirill A. Shutemov" , yi.z.zhang@linux.intel.com Subject: Re: [PATCH v5 4/4] mm: Defer ZONE_DEVICE page initialization to the point where we init pgmap Message-ID: <20181029163528.GL32673@dhcp22.suse.cz> References: <20181010172451.GK5873@dhcp22.suse.cz> <98c35e19-13b9-0913-87d9-b3f1ab738b61@linux.intel.com> <20181010185242.GP5873@dhcp22.suse.cz> <20181011085509.GS5873@dhcp22.suse.cz> <6f32f23c-c21c-9d42-7dda-a1d18613cd3c@linux.intel.com> <20181017075257.GF18839@dhcp22.suse.cz> <971729e6-bcfe-a386-361b-d662951e69a7@linux.intel.com> <20181029141210.GJ32673@dhcp22.suse.cz> <84f09883c16608ddd2ba88103f43ec6a1c649e97.camel@linux.intel.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <84f09883c16608ddd2ba88103f43ec6a1c649e97.camel@linux.intel.com> User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon 29-10-18 08:59:34, Alexander Duyck wrote: > On Mon, 2018-10-29 at 15:12 +0100, Michal Hocko wrote: > > On Wed 17-10-18 08:02:20, Alexander Duyck wrote: > > > On 10/17/2018 12:52 AM, Michal Hocko wrote: > > > > On Thu 11-10-18 10:38:39, Alexander Duyck wrote: > > > > > On 10/11/2018 1:55 AM, Michal Hocko wrote: > > > > > > On Wed 10-10-18 20:52:42, Michal Hocko wrote: > > > > > > [...] > > > > > > > My recollection was that we do clear the reserved bit in > > > > > > > move_pfn_range_to_zone and we indeed do in __init_single_page. But then > > > > > > > we set the bit back right afterwards. This seems to be the case since > > > > > > > d0dc12e86b319 which reorganized the code. I have to study this some more > > > > > > > obviously. > > > > > > > > > > > > so my recollection was wrong and d0dc12e86b319 hasn't really changed > > > > > > much because __init_single_page wouldn't zero out the struct page for > > > > > > the hotplug contex. A comment in move_pfn_range_to_zone explains that we > > > > > > want the reserved bit because pfn walkers already do see the pfn range > > > > > > and the page is not fully associated with the zone until it is onlined. > > > > > > > > > > > > I am thinking that we might be overzealous here. With the full state > > > > > > initialized we shouldn't actually care. pfn_to_online_page should return > > > > > > NULL regardless of the reserved bit and normal pfn walkers shouldn't > > > > > > touch pages they do not recognize and a plain page with ref. count 1 > > > > > > doesn't tell much to anybody. So I _suspect_ that we can simply drop the > > > > > > reserved bit setting here. > > > > > > > > > > So this has me a bit hesitant to want to just drop the bit entirely. If > > > > > nothing else I think I may wan to make that a patch onto itself so that if > > > > > we aren't going to set it we just drop it there. That way if it does cause > > > > > issues we can bisect it to that patch and pinpoint the cause. > > > > > > > > Yes a patch on its own make sense for bisectability. > > > > > > For now I think I am going to back off of this. There is a bunch of other > > > changes that need to happen in order for us to make this work. As far as I > > > can tell there are several places that are relying on this reserved bit. > > > > Please be more specific. Unless I misremember, I have added this > > PageReserved just to be sure (f1dd2cd13c4bb) because pages where just > > half initialized back then. I am not aware anybody is depending on this. > > If there is somebody then be explicit about that. The last thing I want > > to see is to preserve a cargo cult and build a design around it. > > It is mostly just a matter of going through and auditing all the > places that are using PageReserved to identify pages that they aren't > supposed to touch for whatever reason. > > From what I can tell the issue appears to be the fact that the reserved > bit is used to identify if a region of memory is "online" or "offline". No, this is wrong. pfn_to_online_page does that. PageReserved has nothing to do with online vs. offline status. It merely says that you shouldn't touch the page unless you own it. Sure we might have few places relying on it but nothing should really depend the reserved bit check from the MM hotplug POV. > So for example the call "online_pages_range" doesn't invoke the > online_page_callback unless the first pfn in the range is marked as > reserved. Yes and there is no fundamental reason to do that. We can easily check the online status without that. > Another example Dan had pointed out was the saveable_page function in > kernel/power/snapshot.c. Use pfn_to_online_page there. > > > > > > Regarding the post initialization required by devm_memremap_pages and > > > > > > potentially others. Can we update the altmap which is already a way how > > > > > > to get alternative struct pages by a constructor which we could call > > > > > > from memmap_init_zone and do the post initialization? This would reduce > > > > > > the additional loop in the caller while it would still fit the overall > > > > > > design of the altmap and the core hotplug doesn't have to know anything > > > > > > about DAX or whatever needs a special treatment. > > > > > > > > > > > > Does that make any sense? > > > > > > > > > > I think the only thing that is currently using the altmap is the ZONE_DEVICE > > > > > memory init. Specifically I think it is only really used by the > > > > > devm_memremap_pages version of things, and then only under certain > > > > > circumstances. Also the HMM driver doesn't pass an altmap. What we would > > > > > really need is a non-ZONE_DEVICE users of the altmap to really justify > > > > > sticking with that as the preferred argument to pass. > > > > > > > > I am not aware of any upstream HMM user so I am not sure what are the > > > > expectations there. But I thought that ZONE_DEVICE users use altmap. If > > > > that is not generally true then we certainly have to think about a > > > > better interface. > > > > > > I'm just basing my statement on the use of the move_pfn_range_to_zone call. > > > The only caller that is actually passing the altmap is devm_memremap_pages > > > and if I understand things correctly that is only used when we want to stare > > > the vmmemmap on the same memory that we just hotplugged. > > > > Yes, and that is what I've called as allocator callback earlier. > > I am really not a fan of the callback approach. It just means we will > have to do everything multiple times in terms of initialization. I do not follow. Could you elaborate? > > > That is why it made more sense to me to just create a ZONE_DEVICE specific > > > function for handling the page initialization because the one value I do > > > have to pass is the dev_pagemap in both HMM and memremap case, and that has > > > the altmap already embedded inside of it. > > > > And I have argued that this is a wrong approach to the problem. If you > > need a very specific struct page initialization then create a init > > (constructor) callback. > > The callback solution just ends up being more expensive because we lose > multiple layers of possible optimization. By putting everything into on > initization function we are able to let the compiler go through and > optimize things to the point where we are essentially just doing > something akin to one bit memcpy/memset where we are able to construct > one set of page values and write that to every single page we have to > initialize within a given page block. You are already doing per-page initialization so I fail to see a larger unit to operate on. > My concern is that we are going to see a 2-4x regression if I were to > update the current patches I have to improve init performance in order > to achieve the purity of the page initilization functions that you are > looking for. I feel we are much better off having one function that can > handle all cases and do so with high performance, than trying to > construct a set of functions that end up having to reinitialize the > same memory from the previous step and end up with us wasting cycles > and duplicating overhead in multiple spots. The memory hotplug is just one pile of unmaintainable mess mostly because of this kind of attitude. You just care about _your_ particular usecase and not a wee bit beyond that. -- Michal Hocko SUSE Labs