Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753699AbbG2N72 (ORCPT ); Wed, 29 Jul 2015 09:59:28 -0400 Received: from relay.parallels.com ([195.214.232.42]:46089 "EHLO relay.parallels.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751445AbbG2N70 (ORCPT ); Wed, 29 Jul 2015 09:59:26 -0400 Date: Wed, 29 Jul 2015 16:59:07 +0300 From: Vladimir Davydov To: Michal Hocko CC: Andrew Morton , Andres Lagar-Cavilla , Minchan Kim , Raghavendra K T , Johannes Weiner , Greg Thelen , Michel Lespinasse , David Rientjes , Pavel Emelyanov , Cyrill Gorcunov , Jonathan Corbet , , , , , Subject: Re: [PATCH -mm v9 0/8] idle memory tracking Message-ID: <20150729135907.GT8100@esperanza> References: <20150729123629.GI15801@dhcp22.suse.cz> MIME-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Disposition: inline In-Reply-To: <20150729123629.GI15801@dhcp22.suse.cz> X-ClientProxiedBy: US-EXCH2.sw.swsoft.com (10.255.249.46) To US-EXCH2.sw.swsoft.com (10.255.249.46) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2630 Lines: 64 On Wed, Jul 29, 2015 at 02:36:30PM +0200, Michal Hocko wrote: > On Sun 19-07-15 15:31:09, Vladimir Davydov wrote: > [...] > > ---- USER API ---- > > > > The user API consists of two new proc files: > > I was thinking about this for a while. I dislike the interface. It is > quite awkward to use - e.g. you have to read the full memory to check a > single memcg idleness. This might turn out being a problem especially on > large machines. Yes, with this API estimating the wss of a single memory cgroup will cost almost as much as doing this for the whole system. Come to think of it, does anyone really need to estimate idleness of one particular cgroup? If we are doing this for finding an optimal memcg limits configuration or while considering a load move within a cluster (which I think are the primary use cases for the feature), we must do it system-wide to see the whole picture. > It also provides a very low level information (per-pfn idleness) which > is inherently racy. Does anybody really require this level of detail? Well, one might want to do it per-process, obtaining PFNs from /proc/pid/pagemap. > > I would assume that most users are interested only in a single number > which tells the idleness of the system/memcg. Yes, that's what I need it for - estimating containers' wss for setting their limits accordingly. > Well, you have mentioned a per-process reclaim but I am quite > skeptical about this. This is what Minchan mentioned initially. Personally, I'm not going to use it per-process, but I wouldn't rule out this use case either. > > I guess the primary reason to rely on the pfn rather than the LRU walk, > which would be more targeted (especially for memcg cases), is that we > cannot hold lru lock for the whole LRU walk and we cannot continue > walking after the lock is dropped. Maybe we can try to address that > instead? I do not think this is easy to achieve but have you considered > that as an option? Yes, I have, and I've come to a conclusion it's not doable, because LRU lists can be constantly rotating at an arbitrary rate. If you have an idea in mind how this could be done, please share. Speaking of LRU-vs-PFN walk, iterating over PFNs has its own advantages: - You can distribute a walk in time to avoid CPU bursts. - You are free to parallelize the scanner as you wish to decrease the scan time. Thanks, Vladimir -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/