Received: by 2002:a25:8b91:0:0:0:0:0 with SMTP id j17csp4165320ybl; Mon, 3 Feb 2020 13:53:18 -0800 (PST) X-Google-Smtp-Source: APXvYqzU4FLhra7FMWSUC7TPuk4q3a2Rxv8vvhNaX0c+89l/jT/e01hTJw9uioSRAHWfgf1xz092 X-Received: by 2002:a9d:20c1:: with SMTP id x59mr7950465ota.286.1580766798609; Mon, 03 Feb 2020 13:53:18 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1580766798; cv=none; d=google.com; s=arc-20160816; b=bAq5611Aiw/XCwp0jsx5CablDiNJFkNKty/+kzhMGzqvEkCbkAyQX6kucYwvwxIDvt iLMfYmOtA1omtBaamC+cgjsSYQXjvwlACwwGSge3cVhXE56Qqy4LB0zexw7SeuEx41G0 fCj2wEKcnkD0SEP8LehmKaJtc58N0XWHjPuox5qPvYTzhXqRoEMjNG/7oYFXtu3sqbda 3upZI1Hh66HzC1zd9P86urm4fldF56sbLM2KwTCXxET1/Y3hbCOfB6CWuF+Rq5v4/KPc Nc35yaT5rfmKrUgDt5Kv3XgGWCFC+0qa1VVKtITY8GmbzmbELh3tK5q4UR35+Kjpuhr1 9bzA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:in-reply-to:content-disposition :mime-version:references:message-id:subject:cc:to:from:date :dkim-signature; bh=V51SDVF0mKXCbTNKSToYu7ZilLaxAU/vCwK4/RCrZA4=; b=1AWYH7+QFZFQi6CjPFsDQm3k5oO5HaCbTlS1DXempEIOd6VpAKiWwxoFsp0aWJB+jj Bxq3i1xW/Y1h5jXiSurUXvh7X804Px/BkC3YWN7hb9TBGKoVm3qJ/hjppDApoKcX5sF8 Lq41PwJJR9FCBmxKCqxa3SBvw8RGlEGTAHRxnmwhb0SfpbWWa8CnwB40PHIRpKHaOUIR HAC7GK60FCt+txQWiLBGtx54bAIY+2CjQdnkL0o7yocVRE29gvdmq+bhwTnMidK3p9Po kEW4LDQeiL0Ugok+U62+5ZbkijMLM3XZmwjwJznw8bzb6A8lvbJNi4tgxKnszPstQnb6 5byQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@cmpxchg-org.20150623.gappssmtp.com header.s=20150623 header.b=giRevXzD; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=cmpxchg.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id b18si2957253otp.17.2020.02.03.13.53.06; Mon, 03 Feb 2020 13:53:18 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@cmpxchg-org.20150623.gappssmtp.com header.s=20150623 header.b=giRevXzD; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=cmpxchg.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727004AbgBCVwE (ORCPT + 99 others); Mon, 3 Feb 2020 16:52:04 -0500 Received: from mail-qv1-f66.google.com ([209.85.219.66]:45098 "EHLO mail-qv1-f66.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726369AbgBCVwE (ORCPT ); Mon, 3 Feb 2020 16:52:04 -0500 Received: by mail-qv1-f66.google.com with SMTP id l14so7558906qvu.12 for ; Mon, 03 Feb 2020 13:52:03 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=cmpxchg-org.20150623.gappssmtp.com; s=20150623; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to; bh=V51SDVF0mKXCbTNKSToYu7ZilLaxAU/vCwK4/RCrZA4=; b=giRevXzDJvug1jZpCW5DGCFjfiH6/eaG6UsakDYaGiO7cfEKkOaBQ6TNSAKlW9XAwG 4YNHeU90opOhqWzmFCgFfeNg1w8W6TKjPCzhkwBMJU1c9713Ky7nj03p4pioBcnIo/eY Uf6aOI5dUz5IE5D49KI7or77gL0ex18cpdaMI9cyvKg5Uqn4+6/7az2y1E2YsYog/Uox ISxPU/FtGg4aWoSb2335LPuVR7I///3vnMUSCgQ5+XFuan+0Ih6u5eZI1rGNIAozfRm+ P26rmvDM830Xm1RHTJylmglELqOIczAw52rIBGAH1L22d2ck1R7xRoizPugv/ZXXGpnn BdjA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=V51SDVF0mKXCbTNKSToYu7ZilLaxAU/vCwK4/RCrZA4=; b=ADGho2Oh1HPRaS593vv1nmZ6crQ4gHRWDQS41TfGmx7AWZkvNQYBu1kIeM/GEXnzSz pbkte6alsfbRt/zbR9+kiC3qYUyv4eAM+Z7+NK1gSr7kAQ2Nnw1kftgXoqBuyswMZUWg XFxZucb8/mr1sOmiZAq1A1uS21yJBPesZkH2dD61bEwHRsM/dXbV6FoLFsK4AAsdhdxt j7A3Hur2U6k92MAQYMnRmBHVGhObv3hMxZL6w+w8ZyMpk4T2HYpDI8d1cXAVDt/qUe8w 1TOlITAGEfbm5g//K+DkgR18UZKP+FmtwsNcPh/zRdkLx7PbGbCHjnhJjHzddHlOXTNt /E5A== X-Gm-Message-State: APjAAAU1+IGzAtXyYxTNt18rQw/BMJl8GGLvrw9+PnCkitXq/ILQIEjv UE3RWSSz24LVdjqZ7jyiroLF7g== X-Received: by 2002:a05:6214:11a8:: with SMTP id u8mr24526626qvv.16.1580766722728; Mon, 03 Feb 2020 13:52:02 -0800 (PST) Received: from localhost ([2620:10d:c091:500::2:6320]) by smtp.gmail.com with ESMTPSA id s11sm9892149qkg.99.2020.02.03.13.52.01 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 03 Feb 2020 13:52:02 -0800 (PST) Date: Mon, 3 Feb 2020 16:52:01 -0500 From: Johannes Weiner To: Michal Hocko Cc: Andrew Morton , Roman Gushchin , Tejun Heo , linux-mm@kvack.org, cgroups@vger.kernel.org, linux-kernel@vger.kernel.org, kernel-team@fb.com Subject: Re: [PATCH v2 3/3] mm: memcontrol: recursive memory.low protection Message-ID: <20200203215201.GD6380@cmpxchg.org> References: <20191219200718.15696-1-hannes@cmpxchg.org> <20191219200718.15696-4-hannes@cmpxchg.org> <20200130170020.GZ24244@dhcp22.suse.cz> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20200130170020.GZ24244@dhcp22.suse.cz> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Jan 30, 2020 at 06:00:20PM +0100, Michal Hocko wrote: > On Thu 19-12-19 15:07:18, Johannes Weiner wrote: > > Right now, the effective protection of any given cgroup is capped by > > its own explicit memory.low setting, regardless of what the parent > > says. The reasons for this are mostly historical and ease of > > implementation: to make delegation of memory.low safe, effective > > protection is the min() of all memory.low up the tree. > > > > Unfortunately, this limitation makes it impossible to protect an > > entire subtree from another without forcing the user to make explicit > > protection allocations all the way to the leaf cgroups - something > > that is highly undesirable in real life scenarios. > > > > Consider memory in a data center host. At the cgroup top level, we > > have a distinction between system management software and the actual > > workload the system is executing. Both branches are further subdivided > > into individual services, job components etc. > > > > We want to protect the workload as a whole from the system management > > software, but that doesn't mean we want to protect and prioritize > > individual workload wrt each other. Their memory demand can vary over > > time, and we'd want the VM to simply cache the hottest data within the > > workload subtree. Yet, the current memory.low limitations force us to > > allocate a fixed amount of protection to each workload component in > > order to get protection from system management software in > > general. This results in very inefficient resource distribution. > > I do agree that configuring the reclaim protection is not an easy task. > Especially in a deeper reclaim hierarchy. systemd tends to create a deep > and commonly shared subtrees. So having a protected workload really > requires to be put directly into a new first level cgroup in practice > AFAICT. That is a simpler example though. Just imagine you want to > protect a certain user slice. Can you elaborate a bit on this? I don't quite understand the two usecases you are contrasting here. > You seem to be facing a different problem though IIUC. You know how much > memory you want to protect and you do not have to care about the cgroup > hierarchy up but you do not know/care how to distribute that protection > among workloads running under that protection. I agree that this is a > reasonable usecase. I'm not sure I'm parsing this right, but the use case is this: When I'm running a multi-component workload on a host without any cgrouping, the individual components compete over the host's memory based on rate of allocation, how often they reference their memory and so forth. It's a need-based distribution of pages, and the weight can change as demand changes throughout the life of the workload. If I now stick several of such workloads into a containerized environment, I want to use memory.low to assign each workload as a whole a chunk of memory it can use - I don't want to assign fixed-size subchunks to each individual component of each workload! I want the same free competition *within* the workload while setting clear rules for competition *between* the different workloads. [ What I can do today to achieve this is disable the memory controller for the subgroups. When I do this, all pages of the workload are on one single LRU that is protected by one single memory.low. But obviously I lose any detailed accounting as well. This patch allows me to have the same recursive protection semantics while retaining accounting. ] > Those both problems however show that we have a more general > configurability problem for both leaf and intermediate nodes. They are > both a result of strong requirements imposed by delegation as you have > noted above. I am thinking didn't we just go too rigid here? The requirement for delegation is that child groups cannot claim more than the parent affords. Is that the restriction you are referring to? > Delegation points are certainly a security boundary and they should > be treated like that but do we really need a strong containment when > the reclaim protection is under admin full control? Does the admin > really have to reconfigure a large part of the hierarchy to protect a > particular subtree? > > I do not have a great answer on how to implement this unfortunately. The > best I could come up with was to add a "$inherited_protection" magic > value to distinguish from an explicit >=0 protection. What's the > difference? $inherited_protection would be a default and it would always > refer to the closest explicit protection up the hierarchy (with 0 as a > default if there is none defined). > A > / \ > B C (low=10G) > / \ > D E (low = 5G) > > A, B don't get any protection (low=0). C gets protection (10G) and > distributes the pressure to D, E when in excess. D inherits (low=10G) > and E overrides the protection to 5G. > > That would help both usecases AFAICS while the delegation should be > still possible (configure the delegation point with an explicit > value). I have very likely not thought that through completely. Does > that sound like a completely insane idea? > > Or do you think that the two usecases are simply impossible to handle > at the same time? Doesn't my patch accomplish this? Any cgroup or group of cgroups still cannot claim more than the ancestral protection for the subtree. If a cgroup says 10G, the sum of all children's protection will never exceed that. This ensures delegation is safe. But once an ancestor does declare protection, that value automatically applies recursively to the entire subtree. The subtree can *chose* to assign fixed shares of that to its cgroups. But it doesn't have to. While designing this, I also thought about a new magic token for the memory.low to distinguish "inherit" from 0 (or other values smaller than what the parent affords). But the only thing it buys you is that you allow children to *opt out* of protection the parent prescribes. And I'm not sure that's a valid usecase: If the parent specifies that a subtree shall receive a protection of 10G compared to other subtrees at that level, does it really make sense for children to opt out and reduce the subtree's protection? IMO, no. A cgroup should be able to influence competition among its children, it should not be able to change the rules of competition among its ancestors.