Received: by 2002:a25:1985:0:0:0:0:0 with SMTP id 127csp3732391ybz; Mon, 4 May 2020 08:41:29 -0700 (PDT) X-Google-Smtp-Source: APiQypLKjGdE9V3+1oiDSQndHrtsd8iD/RWMCNCkGv680axd9t9FkkqRfRjNztTZ3nJj/pRmxDh4 X-Received: by 2002:a17:906:add7:: with SMTP id lb23mr16177145ejb.6.1588606889211; Mon, 04 May 2020 08:41:29 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1588606889; cv=none; d=google.com; s=arc-20160816; b=QNjBtPQQQwAL1G6NMOpkm2vZMlmo1sqj2zrA4SWAn/u74wlpF6p7uMp0EJo997CADD spnByyupP03+ZrTY3DTTOGTBMX1rIKY/NOskG34zJj7Kd4XnnJolFq9qy6sjGI/MD2Ns 3W3ChyfJDYwLflJIsjaGqG5zmSX8E3919C0/YuFguCr41NXvifQJ1EAM3xHR2LAgXvAS Vgynro75Q2c2hr3oPJMiitAzuADN4Ime/FDLC6Yup0t6NXNh18FYVA78t8w1jHg/a9Oz 13Ngy375P8Zq0pjDqDu9KwNmMWeQCXTX7nNdxPZ/WA3k3S1GMsN+iLVuykEilZHH/feG Yl1A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=qFsRyyCs1CNjwXLV7fTk66+/Kw4cVk2SMY32klLrVvY=; b=jF5et2X2mND8hvXr2zsOvYCjqfANV91ZkpcBrxEVxP2UiU+0GVaEb6xhYtTmT1xsk7 Bla9GLfDq+sk3Y4LkqLvhCCyJVCJozB6982q8FfFH9LnghgpldX2bdPRDWiE2btzYvIL 1VoWzeJl9PaIYSvyX6o+69YgcprVfDj4Jn/AnrvWRiys9nA3TVLO/rDyOdsNBFAvjcXE hqaVc2c93Ta+xi8if0tjzIMkLgQn4NJOrvVZcRWHEXVrtkyZ3ubSsCcp96y5DMvxvweR RW4VTtI5hgGMO5tU2B+4KYkVNmRih+daWDMVmyMYo90SyjrITvoRI8usTgYUOZO+hPmR V+tA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=kI2P1QAT; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id qn23si7061129ejb.111.2020.05.04.08.41.05; Mon, 04 May 2020 08:41:29 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=kI2P1QAT; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728873AbgEDPgM (ORCPT + 99 others); Mon, 4 May 2020 11:36:12 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:34252 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-FAIL-OK-FAIL) by vger.kernel.org with ESMTP id S1725941AbgEDPgL (ORCPT ); Mon, 4 May 2020 11:36:11 -0400 Received: from mail-il1-x141.google.com (mail-il1-x141.google.com [IPv6:2607:f8b0:4864:20::141]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 6B7A4C061A0F for ; Mon, 4 May 2020 08:36:10 -0700 (PDT) Received: by mail-il1-x141.google.com with SMTP id i16so11645629ils.12 for ; Mon, 04 May 2020 08:36:10 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=qFsRyyCs1CNjwXLV7fTk66+/Kw4cVk2SMY32klLrVvY=; b=kI2P1QATu4fMvFXgsjZaHWkW8Y3XJGqJkCH4VH4S2hAg9+Vf3AvornlrHqlVTarDb+ EeZu5F+xjbst6eGmuwij20/VCcnhWF1dciottRi7qPgLl/9SMkVACFCg+R/+LzAN9eJt dXKpQWeXwlRZiFyghDThzUDLG/kqUKYN7TU0lsFBUVBE/+fexBbfhqcgKcNwFQUbeZ0z g7Xc95Ow3xr7zwzCLZ/8WkH0u9iyCiatwX1ogMrRo07yD/YlqVoguTXJ4TVw8tywH3vK 5axni/ZTLBoKU0pep98AOq6XRYFLGKe2DyOkbo8hRcpffcqW8QZhkXVT8ClK1sR/2rS5 pbZw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=qFsRyyCs1CNjwXLV7fTk66+/Kw4cVk2SMY32klLrVvY=; b=mfFNlu59Q0OjmbB/Ok3nanZrREQE4n7YgIh2/KojYKJ6GBdaiiz6bfaTwa8rCvYrnW 8FaFeB3gbliwa2iftErxqb+mjBycjUeSQxCk/ASFeh1yvuiLGtc4TqzP41aUbg4k50xz b2XdA1wYqpbSkMP9DTM1lQwEct1r6H8VksoWsblxaTooN+zRNlDPEMmyFQTin79Ds81O euAuDW1WqFbiLJn0D5UB6qT19TOQrft81YEMgdIkrR1HZ1Wly18nHvdAsmXk4aW+0rFm WtCpzSXPNiuxmt1hBBYFY+3E5/WWjocTGYco9Cxgh5Yu7c5Mmd5xQmEWweOe2dWTOBe9 srQA== X-Gm-Message-State: AGi0PubOBRE3H/cynO9WhGJ7q+hkQuUNHxZhXcCQKZ/Luvi2QkX+s1/2 IhP15RsyvaPeZaJK+LjMD+9OCcm+rSmIsQMz8JlSmQ== X-Received: by 2002:a92:7e4a:: with SMTP id z71mr17584200ilc.83.1588606569319; Mon, 04 May 2020 08:36:09 -0700 (PDT) MIME-Version: 1.0 References: <20200430182712.237526-1-shakeelb@google.com> <20200504065600.GA22838@dhcp22.suse.cz> <20200504141136.GR22838@dhcp22.suse.cz> <20200504150052.GT22838@dhcp22.suse.cz> In-Reply-To: <20200504150052.GT22838@dhcp22.suse.cz> From: Shakeel Butt Date: Mon, 4 May 2020 08:35:57 -0700 Message-ID: Subject: Re: [PATCH] memcg: oom: ignore oom warnings from memory.max To: Michal Hocko Cc: Johannes Weiner , Roman Gushchin , Greg Thelen , Andrew Morton , Linux MM , Cgroups , LKML Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, May 4, 2020 at 8:00 AM Michal Hocko wrote: > > On Mon 04-05-20 07:53:01, Shakeel Butt wrote: > > On Mon, May 4, 2020 at 7:11 AM Michal Hocko wrote: > > > > > > On Mon 04-05-20 06:54:40, Shakeel Butt wrote: > > > > On Sun, May 3, 2020 at 11:56 PM Michal Hocko wrote: > > > > > > > > > > On Thu 30-04-20 11:27:12, Shakeel Butt wrote: > > > > > > Lowering memory.max can trigger an oom-kill if the reclaim does not > > > > > > succeed. However if oom-killer does not find a process for killing, it > > > > > > dumps a lot of warnings. > > > > > > > > > > It shouldn't dump much more than the regular OOM report AFAICS. Sure > > > > > there is "Out of memory and no killable processes..." message printed as > > > > > well but is that a real problem? > > > > > > > > > > > Deleting a memcg does not reclaim memory from it and the memory can > > > > > > linger till there is a memory pressure. One normal way to proactively > > > > > > reclaim such memory is to set memory.max to 0 just before deleting the > > > > > > memcg. However if some of the memcg's memory is pinned by others, this > > > > > > operation can trigger an oom-kill without any process and thus can log a > > > > > > lot un-needed warnings. So, ignore all such warnings from memory.max. > > > > > > > > > > OK, I can see why you might want to use memory.max for that purpose but > > > > > I do not really understand why the oom report is a problem here. > > > > > > > > It may not be a problem for an individual or small scale deployment > > > > but when "sweep before tear down" is the part of the workflow for > > > > thousands of machines cycling through hundreds of thousands of cgroups > > > > then we can potentially flood the logs with not useful dumps and may > > > > hide (or overflow) any useful information in the logs. > > > > > > If you are doing this in a large scale and the oom report is really a > > > problem then you shouldn't be resetting hard limit to 0 in the first > > > place. > > > > > > > I think I have pretty clearly described why we want to reset the hard > > limit to 0, so, unless there is an alternative I don't see why we > > should not be doing this. > > I am not saying you shouldn't be doing that. I am just saying that if > you do then you have to live with oom reports. > > > > > > memory.max can trigger the oom kill and user should be expecting the oom > > > > > report under that condition. Why is "no eligible task" so special? Is it > > > > > because you know that there won't be any tasks for your particular case? > > > > > What about other use cases where memory.max is not used as a "sweep > > > > > before tear down"? > > > > > > > > What other such use-cases would be? The only use-case I can envision > > > > of adjusting limits dynamically of a live cgroup are resource > > > > managers. However for cgroup v2, memory.high is the recommended way to > > > > limit the usage, so, why would resource managers be changing > > > > memory.max instead of memory.high? I am not sure. What do you think? > > > > > > There are different reasons to use the hard limit. Mostly to contain > > > potential runaways. While high limit might be a sufficient measure to > > > achieve that as well the hard limit is the last resort. And it clearly > > > has the oom killer semantic so I am not really sure why you are > > > comparing the two. > > > > > > > I am trying to see if "no eligible task" is really an issue and should > > be warned for the "other use cases". The only real use-case I can > > think of are resource managers adjusting the limit dynamically. I > > don't see "no eligible task" a concerning reason for such use-case. > > It is very much a concerning reason to notify about like any other OOM > situation due to hard limit breach. In this case it is worse in some > sense because the limit cannot be trimmed down because there is no > directly reclaimable memory at all. Such an oom situation is > effectivelly conserved. > -- Let me make a more precise statement and tell me if you agree. The "no eligible task" is concerning for the charging path but not for the writer of memory.max. The writer can read the usage and cgroup.[procs|events] to figure out the situation if needed. Usually such writers (i.e. resource managers) use memory.high in addition to memory.max. First set memory.high and once the usage is below the high then set max to not induce the oom-kills. Shakeel