Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754282Ab2JRIvQ (ORCPT ); Thu, 18 Oct 2012 04:51:16 -0400 Received: from mx2.parallels.com ([64.131.90.16]:58781 "EHLO mx2.parallels.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1753413Ab2JRIvN (ORCPT ); Thu, 18 Oct 2012 04:51:13 -0400 Message-ID: <50803379.8000808@parallels.com> Date: Thu, 18 Oct 2012 20:51:05 +0400 From: Glauber Costa User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:16.0) Gecko/20121009 Thunderbird/16.0 MIME-Version: 1.0 To: Andrew Morton CC: , , Mel Gorman , Tejun Heo , Michal Hocko , Johannes Weiner , , Christoph Lameter , David Rientjes , Pekka Enberg , , Subject: Re: [PATCH v5 00/14] kmem controller for memcg. References: <1350382611-20579-1-git-send-email-glommer@parallels.com> <20121017151142.71e1f3c5.akpm@linux-foundation.org> In-Reply-To: <20121017151142.71e1f3c5.akpm@linux-foundation.org> Content-Type: text/plain; charset="ISO-8859-1" Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 4260 Lines: 102 On 10/18/2012 02:11 AM, Andrew Morton wrote: > On Tue, 16 Oct 2012 14:16:37 +0400 > Glauber Costa wrote: > >> ... >> >> A general explanation of what this is all about follows: >> >> The kernel memory limitation mechanism for memcg concerns itself with >> disallowing potentially non-reclaimable allocations to happen in exaggerate >> quantities by a particular set of processes (cgroup). Those allocations could >> create pressure that affects the behavior of a different and unrelated set of >> processes. >> >> Its basic working mechanism is to annotate some allocations with the >> _GFP_KMEMCG flag. When this flag is set, the current process allocating will >> have its memcg identified and charged against. When reaching a specific limit, >> further allocations will be denied. > > The need to set _GFP_KMEMCG is rather unpleasing, and makes one wonder > "why didn't it just track all allocations". > This was raised as well by Peter Zijlstra during the memcg summit. The answer I gave to him still stands: There is a cost associated with it. We believe it comes down to a trade off situation. How much tracking a particular kind of allocation help vs how much does it cost. The free path is specially more expensive, since it will always incur in a page_cgroup lookup. > Does this mean that over time we can expect more sites to get the > _GFP_KMEMCG tagging? We have being doing kernel memory limitation for OpenVZ for a lot of times, using a quite different mechanism. What we do in this work (with slab included), allows us to achieve feature parity with that. It means it is good enough for production environments. Whether or not more people will want other allocations to be tracked, I can't predict. What I do can say is that stack + slab is a very significant part of the memory one potentially cares about, and if anyone else ever have the need for more, it will come down to a trade-off calculation. > If so, are there any special implications, or do > we just go in, do the one-line patch and expect everything to work? With the infrastructure in place, it shouldn't be hard. But it's not necessarily a one-liner either. It depends on what are the pratical considerations for having that specific kind of allocation tied to a memcg. The slab, for instance, that follows this series, is far away from a one-liner: it is in fact, a 19-patch patch series. > > And how *accurate* is the proposed code? What percentage of kernel > memory allocations are unaccounted, typical case and worst case? With both patchsets applied, all memory used for the stack and most of the memory used for slab objects allocated in userspace process contexts are accounted. I honestly don't know which percentage of the total kernel memory this represents. The accuracy for stack pages is very high: In this series, we don't move stack pages around when moving a task to other cgroups (for stack, it could be done), but other than that, all processes that pops up in a cgroup and stay there will have its memory accurately accounted. The slab is more complicated, and depends on the workload. It will be more accurate in workloads in which the level of object-sharing among cgroups is low. A container, for instance, is the perfect example of where this happen. > > All sorts of questions come to mind over this decision, but it was > unexplained. It should be, please. A lot! > >> >> ... >> >> Limits lower than >> the user limit effectively means there is a separate kernel memory limit that >> may be reached independently than the user limit. Values equal or greater than >> the user limit implies only that kernel memory is tracked. This provides a >> unified vision of "maximum memory", be it kernel or user memory. >> > > I'm struggling to understand that text much at all. Reading the > Documentation/cgroups/memory.txt patch helped. > Great. If you have any specific suggestions I can change that. Maybe I should just paste the documentation bit in here... -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/