Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S965684AbbFJPpa (ORCPT ); Wed, 10 Jun 2015 11:45:30 -0400 Received: from mail-qg0-f41.google.com ([209.85.192.41]:35406 "EHLO mail-qg0-f41.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S964801AbbFJPmt (ORCPT ); Wed, 10 Jun 2015 11:42:49 -0400 Date: Wed, 10 Jun 2015 11:42:38 -0400 From: Jerome Glisse To: Mark Hairgrove Cc: "akpm@linux-foundation.org" , "linux-kernel@vger.kernel.org" , "linux-mm@kvack.org" , Linus Torvalds , "joro@8bytes.org" , Mel Gorman , "H. Peter Anvin" , Peter Zijlstra , Andrea Arcangeli , Johannes Weiner , Larry Woodman , Rik van Riel , Dave Airlie , Brendan Conoboy , Joe Donohue , Duncan Poole , Sherry Cheung , Subhash Gutti , John Hubbard , Lucien Dunning , Cameron Buschardt , Arvind Gopalakrishnan , Haggai Eran , Shachar Raindel , Liran Liss , Roland Dreier , Ben Sander , Greg Stoner , John Bridgman , Michael Mantor , Paul Blinzer , Laurent Morichetti , Alexander Deucher , Oded Gabbay , =?iso-8859-1?B?Suly9G1l?= Glisse , Jatin Kumar , "linux-rdma@vger.kernel.org" Subject: Re: [PATCH 05/36] HMM: introduce heterogeneous memory management v3. Message-ID: <20150610154237.GA13465@gmail.com> References: <1432236705-4209-1-git-send-email-j.glisse@gmail.com> <1432236705-4209-6-git-send-email-j.glisse@gmail.com> <20150608211740.GA5241@gmail.com> <20150609155601.GA3101@gmail.com> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: User-Agent: Mutt/1.5.23 (2014-03-12) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 9704 Lines: 203 On Tue, Jun 09, 2015 at 08:33:12PM -0700, Mark Hairgrove wrote: > > > On Tue, 9 Jun 2015, Jerome Glisse wrote: > > > On Mon, Jun 08, 2015 at 06:54:29PM -0700, Mark Hairgrove wrote: > > > Can you clarify how that's different from mmu_notifiers? Those are also > > > embedded into a driver-owned struct. > > > > For HMM you want to be able to kill a mirror from HMM, you might have kernel > > thread call inside HMM with a mirror (outside any device file lifetime) ... > > The mirror is not only use at register & unregister, there is a lot more thing > > you can call using the HMM mirror struct. > > > > So the HMM mirror lifetime as a result is more complex, it can not simply be > > free from the mmu_notifier_release callback or randomly. It needs to be > > refcounted. > > Sure, there are driver -> HMM calls like hmm_mirror_fault that > mmu_notifiers don't have, but I don't understand why that fundamentally > makes HMM mirror lifetimes more complex. Decoupling hmm_mirror lifetime > from mm lifetime adds complexity too. Driver->HMM calls can happen from random kernel thread thus you need to garanty that hmm_mirror can not go away. More over you can have CPU MM call into HMM outside of mmu_notifier. Basicly you can get to HMM code by many different code path, unlike any of the current mmu_notifier. So refcounting is necessary as otherwise the device driver might decide to unregister and free the mirror while some other kernel thread is about to dereference the exact same mirror. Synchronization with mmu notifier srcu will not be enough in the case of page fault on remote memory for instance. Other case too. > > > The mmu_notifier code assume that the mmu_notifier struct is > > embedded inside a struct that has a lifetime properly synchronize with the > > mm. For HMM mirror this is not something that sounds like a good idea as there > > is too many way to get it wrong. > > What kind of synchronization with the mm are you referring to here? > Clients of mmu_notifiers don't have to do anything as far as I know. > They're guaranteed that the mm won't go away because each registered > notifier bumps mm_count. So for all current user afaict (kvm, xen, intel, radeon) tie to a file, (sgi gru) tie to vma, (iommu) tie to mm. Which means this is all properly synchronize with lifetime of mm (ignoring the fork case). The struct that is tie to mmu_notifier for all of them can be accessed only through one code path (ioctl for most of them). > > > So idea of HMM mirror is that it can out last the mm lifetime but the HMM > > struct can not. So you have hmm_mirror <~> hmm <-> mm and the mirror can be > > "unlink" and have different lifetime from the hmm that itself has same life > > time as mm. > > Per the earlier discussion hmm_mirror_destroy is missing a call to > hmm_unref. If that's added back I don't understand how the mirror can > persist past the hmm struct. The mirror can be unlinked from hmm's list, > yes, but that doesn't mean that hmm/mm can be torn down. The hmm/mm > structs will stick around until hmm_destroy since that does the > mmu_notifier_unregister. hmm_destroy can't be called until the last > hmm_mirror_destroy. > > Doesn't that mean that hmm/mm are guaranteed to be allocated until the > last hmm_mirror_unregister? That sounds like a good guarantee to make. Like said, just ignore current code it is utterly broken in so many way when it comes to lifetime. I screw that part badly when reworking the patchset, i was focusing on other part. I fixed that in my tree, i am waiting for more review on other part as anyway the lifetime thing is easy to rework/fix. http://cgit.freedesktop.org/~glisse/linux/log/?h=hmm [...] > > > > > If so, I think there's a race here in the case of mm teardown happening > > > > > concurrently with hmm_mirror_unregister. > > > > > > > > > > [...] > > > > > > > > > > Do you agree that this sequence can happen, or am I missing something > > > > > which prevents it? > > > > > > > > Can't happen because child have mm->hmm = NULL ie only one hmm per mm > > > > and hmm is tie to only one mm. It is the responsability of the device > > > > driver to make sure same apply to private reference to the hmm mirror > > > > struct ie hmm_mirror should never be tie to a private file struct. > > > > > > It's useful for the driver to have some association between files and > > > mirrors. If the file is closed prior to process exit we would like to > > > unregister the mirror, otherwise it will persist until process teardown. > > > The association doesn't have to be 1:1 but having the files ref count the > > > mirror or something would be useful. > > > > This is allowed, i might have put strong word here, but you can associate > > with a file struct. What you can not do is use the mirror from a different > > process ie one with a different mm struct as mirror is linked to a single > > mm. So on fork there is no callback to update the private file struct, when > > the device file is duplicated (well just refcount inc) against a different > > process. This is something you need to be carefull in your driver. Inside > > the dummy driver i abuse that to actually test proper behavior of HMM but > > it should not be use as an example. > > So to confirm, on all file operations from user space the driver is > expected to check that current->mm matches the mm associated with the > struct file's hmm_mirror? Well you might have a valid usecase for that, just be aware that anything your driver do with the hmm_mirror will actually impact the mm of the parent. Which i assume is not what you want. I would actualy thought that what you want is having a way to find hmm_mirror using both device file & mm as a key. Otherwise you can not really use HMM with process that like to fork themself. Which is a valid usecase to me. For instance process start using HMM through your driver, decide to fork itself and to also use HMM through your driver inside its child. > > On file->release the driver still ought to call hmm_mirror_unregister > regardless of whether the mms match, otherwise we'll never tear down the > mirror. That means we're not saved from the race condition because > hmm_mirror_unregister can happen in one thread while hmm_notifier_release > might be happening in another thread. Again there is no race the mirror list is the synchronization point and it is protected by a lock. So either hmm_mirror_unregister() wins or the other thread hmm_notifier_release() > > > But even if we assume no association at all between files and mirrors, are > > > you sure that prevents the race? The driver may choose to unregister the > > > hmm_device at any point once its files are closed. In the case of module > > > unload the device unregister can't be prevented. If mm teardown hasn't > > > happened yet mirrors may still be active and registered on that > > > hmm_device. The driver thus has to first call hmm_mirror_unregister on all > > > active mirrors, then call hmm_device_unregister. mm teardown of those > > > mirrors may trigger at any point in this sequence, so we're right back to > > > that race. > > > > So when device driver unload the first thing it needs to do is kill all of > > its context ie all of its HMM mirror (unregister them) by doing so it will > > make sure that there can be no more call to any of its functions. > > When is the driver expected to call hmm_mirror_unregister? Is it file > close, module unload, or some other time? > > If it's file close, there's no need to unregister anything on module > unload because the files were all closed already. > > If it's module unload, then the mirrors and mms all get leaked until that > point. > > We're exposed to the race in both cases. You unregister as soon as you want, it is up to your driver to do it, i do not enforce anything. The only thing i enforce is that you can not unregister the hmm device driver before all mirror are unregistered and free. So yes for device driver you want to unregister when device file is close (which happens when the process exit). In both cases there is no race as explain above. > > > > > The race with mm teardown does not exist as what matter for mm teardown is > > the fact that the mirror is on the struct hmm mirrors list or not. Either > > the device driver is first to remove the mirror from the list or it is the > > mm teardown but this is lock protected so only one thread can do it. > > > > Agreed, removing the mirror from the list is not a "race" in the classical > sense. The true race is between hmm_notifier_release's device mutex_unlock > (process exit) and post-hmm_device_unregister device mutex free (driver > close/unload). What I meant is that in order to expose that race you first > need one thread to call hmm_mirror_unregister while another thread is in > hmm_notifier_release. > > Regardless of where hmm_mirror_unregister is called (file close, module > unload, etc) it can happen concurrently with hmm_notifier_release so we're > exposed to this race. There is no race here, the mirror struct will only be freed once as again the list is a synchronization point. Whoever remove the mirror from the list is responsible to drop the list reference. In the fixed code the only thing that will happen twice is the ->release() callback. Even that can be work around to garanty it is call only once. Anyway i do not see anyrace here. Cheers, J?r?me -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/