Received: by 2002:a25:ad19:0:0:0:0:0 with SMTP id y25csp1849799ybi; Thu, 4 Jul 2019 00:38:25 -0700 (PDT) X-Google-Smtp-Source: APXvYqxLYQFx+nC+yY8eAgE8WuvAF+l7tKK2uh1HGKqwM/oXlFg0TZ9m2sui1dkuz/IOL8pjBH5+ X-Received: by 2002:a17:90a:338b:: with SMTP id n11mr17864010pjb.21.1562225905267; Thu, 04 Jul 2019 00:38:25 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1562225905; cv=none; d=google.com; s=arc-20160816; b=nPu62smttyZvRHG915WveHJThHeyZAs/lBCJRL5dKkNhAv3j+zaQ0UiXXJZWWvJu2V kjzFlqHdXBD6ppkRo8kYkYCTy7nck7Xhke1AP8p6hWhGlxiRC97qXqK5sThw8j4qhJy2 kOt+mtUi4Ev15uX0VcJvmeh2LMVpPnChjkjOSayMA91Ma7Uxxftg7zFqEC9+X8jJ6Jmm YnZI0EflR/VMwB/sdyU/BGzSl4daHJyfmPivrDVDNQwrFxjVwmNKX4CdMVZeDRhE3gjy in8VimP56qIXn9tmKR/NErAO8j7SYqcgU4UtGa0PDWKsNHgEtDY2ki7b61LPStDweuP2 RqUQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-transfer-encoding:content-disposition:mime-version :references:message-id:subject:cc:to:from:date; bh=NPbcT7fNZKxE9QIwn5p/+MnGQ2etLcnd4JIXTlUCZm4=; b=RxgoZAAXSE/UOcv+GipIviASIRyn93EBpv1Rm38vKjwS53J87J2Av1lVqxX9nZX/Vj xldIjQ95WMNO4oOCHzfDoXk2wtylJuF3Vm3539hUcG+ELWp6A4db2JiLsoVbnwPpX49u UB39DngY2bm/NIOyTBmDM1PjcRwS+0ElxV3oEGorwZxVGyp/gk8PB0BIesoHdFRxBAW1 VS2LKxpGlRUgWnuFBPvIoFT4rYYRkw8rGziIvolp8N/Q8fZcyMtmEWr67ZJU3unk23z1 ElzwjzVs6qbP8MXvhrbRUpuHSA7ZFFIGxf6XprRGwZcoc9cyF2q7GZNy28srh8BgAZgr vcIQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id y3si4838556pgi.125.2019.07.04.00.38.10; Thu, 04 Jul 2019 00:38:25 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727464AbfGDHhf (ORCPT + 99 others); Thu, 4 Jul 2019 03:37:35 -0400 Received: from mx2.suse.de ([195.135.220.15]:46180 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1725945AbfGDHhf (ORCPT ); Thu, 4 Jul 2019 03:37:35 -0400 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay2.suse.de (unknown [195.135.220.254]) by mx1.suse.de (Postfix) with ESMTP id 5CBB9AD7F; Thu, 4 Jul 2019 07:37:32 +0000 (UTC) Date: Thu, 4 Jul 2019 09:37:30 +0200 From: Michal Hocko To: Waiman Long Cc: Andrew Morton , Christoph Lameter , Pekka Enberg , David Rientjes , Joonsoo Kim , Alexander Viro , Jonathan Corbet , Luis Chamberlain , Kees Cook , Johannes Weiner , Vladimir Davydov , linux-mm@kvack.org, linux-doc@vger.kernel.org, linux-fsdevel@vger.kernel.org, cgroups@vger.kernel.org, linux-kernel@vger.kernel.org, Roman Gushchin , Shakeel Butt , Andrea Arcangeli Subject: Re: [PATCH] mm, slab: Extend slab/shrink to shrink all the memcg caches Message-ID: <20190704073730.GA5620@dhcp22.suse.cz> References: <20190702183730.14461-1-longman@redhat.com> <20190702130318.39d187dc27dbdd9267788165@linux-foundation.org> <78879b79-1b8f-cdfd-d4fa-610afe5e5d48@redhat.com> <20190702143340.715f771192721f60de1699d7@linux-foundation.org> <20190703155314.GT978@dhcp22.suse.cz> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed 03-07-19 12:16:09, Waiman Long wrote: > On 7/3/19 11:53 AM, Michal Hocko wrote: > > On Wed 03-07-19 11:21:16, Waiman Long wrote: > >> On 7/2/19 5:33 PM, Andrew Morton wrote: > >>> On Tue, 2 Jul 2019 16:44:24 -0400 Waiman Long wrote: > >>> > >>>> On 7/2/19 4:03 PM, Andrew Morton wrote: > >>>>> On Tue, 2 Jul 2019 14:37:30 -0400 Waiman Long wrote: > >>>>> > >>>>>> Currently, a value of '1" is written to /sys/kernel/slab//shrink > >>>>>> file to shrink the slab by flushing all the per-cpu slabs and free > >>>>>> slabs in partial lists. This applies only to the root caches, though. > >>>>>> > >>>>>> Extends this capability by shrinking all the child memcg caches and > >>>>>> the root cache when a value of '2' is written to the shrink sysfs file. > >>>>> Why? > >>>>> > >>>>> Please fully describe the value of the proposed feature to or users. > >>>>> Always. > >>>> Sure. Essentially, the sysfs shrink interface is not complete. It allows > >>>> the root cache to be shrunk, but not any of the memcg caches.? > >>> But that doesn't describe anything of value. Who wants to use this, > >>> and why? How will it be used? What are the use-cases? > >>> > >> For me, the primary motivation of posting this patch is to have a way to > >> make the number of active objects reported in /proc/slabinfo more > >> accurately reflect the number of objects that are actually being used by > >> the kernel. > > I believe we have been through that. If the number is inexact due to > > caching then lets fix slabinfo rather than trick around it and teach > > people to do a magic write to some file that will "solve" a problem. > > This is exactly what drop_caches turned out to be in fact. People just > > got used to drop caches because they were told so by $random web page. > > So really, think about the underlying problem and try to fix it. > > > > It is true that you could argue that this patch is actually fixing the > > existing interface because it doesn't really do what it is documented to > > do and on those grounds I would agree with the change. > > I do think that we should correct the shrink file to do what it is > designed to do to include the memcg caches as well. > > > > But do not teach > > people that they have to write to some file to get proper numbers. > > Because that is just a bad idea and it will kick back the same way > > drop_caches. > > The /proc/slabinfo file is a well-known file that is probably used > relatively extensively. Making it to scan through all the per-cpu > structures will probably cause performance issues as the slab_mutex has > to be taken during the whole duration of the scan. That could have > undesirable side effect. Please be more specific with some numbers ideally. Also if collecting data is too expensive, why cannot we simply account cached objects count in pcp manner? > Instead, I am thinking about extending the slab/objects sysfs file to > also show the number of objects hold up by the per-cpu structures and > thus we can get an accurate count by subtracting it from the reported > active objects. That will have a more limited performance impact as it > is just one kmem cache instead of all the kmem caches in the system. > Also the sysfs files are not as commonly used as slabinfo. That will be > another patch in the near future. Both are root only and once it is widespread that slabinfo doesn't provide precise data you can expect tools will try to fix that by adding another file(s) and we are back to square one, no? In other words slabinfo -- Michal Hocko SUSE Labs