Received: by 2002:a25:ad19:0:0:0:0:0 with SMTP id y25csp3682538ybi; Tue, 2 Jul 2019 11:38:50 -0700 (PDT) X-Google-Smtp-Source: APXvYqxrCrPr96dXI30Z9yCY5KZS8N7dN/+YFTVy7FS85YkVu9QoCjl/OeYfgoakvkKdrlPQUpyy X-Received: by 2002:a17:902:8b82:: with SMTP id ay2mr34475049plb.164.1562092730861; Tue, 02 Jul 2019 11:38:50 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1562092730; cv=none; d=google.com; s=arc-20160816; b=MwuhlCfk07nKXK4FV2lv1zah8NA6Jm56USi4u1h1U3lypMtG5NfS+FVKZKGpkUoKOa gh9yqJPnfTeCbCMY5O5Y1u3zP1sLsKD08S+MSSJtrhTDbKjRtOHY+lERceLw5zClRVID TLuBuPyKbgVKeodQ12HShj/Z+qZAS7hbC7sHfS2Rg++0we+K9UZNqi1yxD3ToLm+GP9/ Q5MGUYoPTswjDBoINUotArGKPZeKy83ZHRTls0j+YpdMT7e7foxiGhE6wU//wRvrZHZm OniV5HXyVSVUVUKbQtd9AZnbVF4ybAtrbwCKM7dNZF18TsqcURZcoxrkjf4mSqzHHfzp HHdQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:message-id:date:subject:cc:to:from; bh=AEW55TgtVi2BKt6wgfN3muW8kwr4Eo/kHBzKeJce5lA=; b=O4iP5sMWwUbxSVkJ/4IxAY0VJwH4cxagC4/PXcWXn4GTL4KeH6sa1HVUVXVoipMaY0 a2HuWqh7a7QPqHkFYNtkMHi3nzsrdnhyJF48H4b5TjKCbaCNPRoFCGlKhd8zgK6ZGI8Q Nissscr6pFtAcTTGRV3FwZPIHjBGzdmQw1CBBHgwI/ZUHzME/DD54W8N3LjnIZrtYFjC bj9JT0yTfxFrV9Lr5IAkv/lWORJxd0FTEj2SG9vlwH9Jh+TuryRbmOB4fW0bDUP40ROa ZzDbNvCBK7yJfdHzKxn9L3YSOtVXKUVn1yN++G+1udZdh2rl6wLGP66aTYPJlppSwPuD AgDg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id 142si14168676pgf.35.2019.07.02.11.38.35; Tue, 02 Jul 2019 11:38:50 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726821AbfGBSiO (ORCPT + 99 others); Tue, 2 Jul 2019 14:38:14 -0400 Received: from mx1.redhat.com ([209.132.183.28]:57250 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726150AbfGBSiN (ORCPT ); Tue, 2 Jul 2019 14:38:13 -0400 Received: from smtp.corp.redhat.com (int-mx04.intmail.prod.int.phx2.redhat.com [10.5.11.14]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id F1A6B81F31; Tue, 2 Jul 2019 18:37:54 +0000 (UTC) Received: from llong.com (dhcp-17-160.bos.redhat.com [10.18.17.160]) by smtp.corp.redhat.com (Postfix) with ESMTP id 599AA5D968; Tue, 2 Jul 2019 18:37:44 +0000 (UTC) From: Waiman Long To: Christoph Lameter , Pekka Enberg , David Rientjes , Joonsoo Kim , Andrew Morton , Alexander Viro , Jonathan Corbet , Luis Chamberlain , Kees Cook , Johannes Weiner , Michal Hocko , Vladimir Davydov Cc: linux-mm@kvack.org, linux-doc@vger.kernel.org, linux-fsdevel@vger.kernel.org, cgroups@vger.kernel.org, linux-kernel@vger.kernel.org, Roman Gushchin , Shakeel Butt , Andrea Arcangeli , Waiman Long Subject: [PATCH] mm, slab: Extend slab/shrink to shrink all the memcg caches Date: Tue, 2 Jul 2019 14:37:30 -0400 Message-Id: <20190702183730.14461-1-longman@redhat.com> X-Scanned-By: MIMEDefang 2.79 on 10.5.11.14 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.25]); Tue, 02 Jul 2019 18:38:13 +0000 (UTC) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Currently, a value of '1" is written to /sys/kernel/slab//shrink file to shrink the slab by flushing all the per-cpu slabs and free slabs in partial lists. This applies only to the root caches, though. Extends this capability by shrinking all the child memcg caches and the root cache when a value of '2' is written to the shrink sysfs file. On a 4-socket 112-core 224-thread x86-64 system after a parallel kernel build, the the amount of memory occupied by slabs before shrinking slabs were: # grep task_struct /proc/slabinfo task_struct 7114 7296 7744 4 8 : tunables 0 0 0 : slabdata 1824 1824 0 # grep "^S[lRU]" /proc/meminfo Slab: 1310444 kB SReclaimable: 377604 kB SUnreclaim: 932840 kB After shrinking slabs: # grep "^S[lRU]" /proc/meminfo Slab: 695652 kB SReclaimable: 322796 kB SUnreclaim: 372856 kB # grep task_struct /proc/slabinfo task_struct 2262 2572 7744 4 8 : tunables 0 0 0 : slabdata 643 643 0 Signed-off-by: Waiman Long --- Documentation/ABI/testing/sysfs-kernel-slab | 10 +++-- mm/slab.h | 1 + mm/slab_common.c | 43 +++++++++++++++++++++ mm/slub.c | 2 + 4 files changed, 52 insertions(+), 4 deletions(-) diff --git a/Documentation/ABI/testing/sysfs-kernel-slab b/Documentation/ABI/testing/sysfs-kernel-slab index 29601d93a1c2..2a3d0fc4b4ac 100644 --- a/Documentation/ABI/testing/sysfs-kernel-slab +++ b/Documentation/ABI/testing/sysfs-kernel-slab @@ -429,10 +429,12 @@ KernelVersion: 2.6.22 Contact: Pekka Enberg , Christoph Lameter Description: - The shrink file is written when memory should be reclaimed from - a cache. Empty partial slabs are freed and the partial list is - sorted so the slabs with the fewest available objects are used - first. + A value of '1' is written to the shrink file when memory should + be reclaimed from a cache. Empty partial slabs are freed and + the partial list is sorted so the slabs with the fewest + available objects are used first. When a value of '2' is + written, all the corresponding child memory cgroup caches + should be shrunk as well. All other values are invalid. What: /sys/kernel/slab/cache/slab_size Date: May 2007 diff --git a/mm/slab.h b/mm/slab.h index 3b22931bb557..a16b2c7ff4dd 100644 --- a/mm/slab.h +++ b/mm/slab.h @@ -174,6 +174,7 @@ int __kmem_cache_shrink(struct kmem_cache *); void __kmemcg_cache_deactivate(struct kmem_cache *s); void __kmemcg_cache_deactivate_after_rcu(struct kmem_cache *s); void slab_kmem_cache_release(struct kmem_cache *); +int kmem_cache_shrink_all(struct kmem_cache *s); struct seq_file; struct file; diff --git a/mm/slab_common.c b/mm/slab_common.c index 464faaa9fd81..493697ba1da5 100644 --- a/mm/slab_common.c +++ b/mm/slab_common.c @@ -981,6 +981,49 @@ int kmem_cache_shrink(struct kmem_cache *cachep) } EXPORT_SYMBOL(kmem_cache_shrink); +/** + * kmem_cache_shrink_all - shrink a cache and all its memcg children + * @s: The root cache to shrink. + * + * Return: 0 if successful, -EINVAL if not a root cache + */ +int kmem_cache_shrink_all(struct kmem_cache *s) +{ + struct kmem_cache *c; + + if (!IS_ENABLED(CONFIG_MEMCG_KMEM)) { + kmem_cache_shrink(s); + return 0; + } + if (!is_root_cache(s)) + return -EINVAL; + + /* + * The caller should have a reference to the root cache and so + * we don't need to take the slab_mutex. We have to take the + * slab_mutex, however, to iterate the memcg caches. + */ + get_online_cpus(); + get_online_mems(); + kasan_cache_shrink(s); + __kmem_cache_shrink(s); + + mutex_lock(&slab_mutex); + for_each_memcg_cache(c, s) { + /* + * Don't need to shrink deactivated memcg caches. + */ + if (s->flags & SLAB_DEACTIVATED) + continue; + kasan_cache_shrink(c); + __kmem_cache_shrink(c); + } + mutex_unlock(&slab_mutex); + put_online_mems(); + put_online_cpus(); + return 0; +} + bool slab_is_available(void) { return slab_state >= UP; diff --git a/mm/slub.c b/mm/slub.c index a384228ff6d3..5d7b0004c51f 100644 --- a/mm/slub.c +++ b/mm/slub.c @@ -5298,6 +5298,8 @@ static ssize_t shrink_store(struct kmem_cache *s, { if (buf[0] == '1') kmem_cache_shrink(s); + else if (buf[0] == '2') + kmem_cache_shrink_all(s); else return -EINVAL; return length; -- 2.18.1