Received: by 2002:a25:1985:0:0:0:0:0 with SMTP id 127csp1112405ybz; Wed, 22 Apr 2020 13:49:48 -0700 (PDT) X-Google-Smtp-Source: APiQypJ+O5qZ5JRA6POZ7P/XWZOp0vRePPs4Zu+TPsePmzF0t975xMSlkFcwx1aFQ0M/gFYnUBUb X-Received: by 2002:a17:906:38c:: with SMTP id b12mr183949eja.332.1587588588103; Wed, 22 Apr 2020 13:49:48 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1587588588; cv=none; d=google.com; s=arc-20160816; b=zOTR13OQ5gnJEg1MJaOGRUMGYeeGQKKrT3HJgiVFRSfXNli967NIp5HHcYWuRyj6nG HaNRSmlzfdjbNjwH8fFAfMa/DkgaZY+tQRXTQoHhdgC1b8jfN2bnTq+lB5DBQ4VJ3I/8 nDvNybvQg4vgFcLTvsVIPdcmexdw2OU0E5ySEYir3WZ+ZFpK91FOdbPRRFH4uYLfQw34 Mj2uFrlJiCpPoAoV9s1RiJ2W43r9ZSeQ4vHqTaqtHE1w8ucSGxDKuEAsOT1eMex/SSoA DBjLHSJpXJWW3J/MWt/7fKKDgbfCCKokvgiyygKBjfj5BFazW2Jy7/vWgrktFLokk2d9 UXeA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :message-id:date:subject:smtp-origin-cluster:cc:to :smtp-origin-hostname:from:smtp-origin-hostprefix:dkim-signature; bh=bYXySgtA0glsWw22mN546qhTFN/cHgQM7XvI05NPNl0=; b=aVYtnOZFUWK6uthHhbMcvRHS1LhlFb76GPv+pYFa8DZ5q/vKclgwl6wzWGeljiJAbm ZJuZ4SBZgrakinI0+Ud8zLRLX5wdg1olNomKllzl6BLWnywOH/9Cp4352wl/e9ac7/jU jQi3h3UFHINXpf7e736o9R7PDnb4FyvpMXAMtIV+TWBB2EQ3oW/BheUC2DrnZkbUF1ow 7EKdBucBpgoCzrwLin1rNDmCxuIHe3jLPwx9uIgdoC7dm5RMDxXIEsE/HPlGihen6yGW LJE2EKr9LTOtHbuJ6R5hZabND7Xbk8TCwLkIxPmRuDo5PzvTvBgoxnM8sIgJOmPIzXwD QWuQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@fb.com header.s=facebook header.b=Hs0BTxGL; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=fb.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id s8si139560edj.463.2020.04.22.13.49.25; Wed, 22 Apr 2020 13:49:48 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@fb.com header.s=facebook header.b=Hs0BTxGL; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=fb.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726625AbgDVUrb (ORCPT + 99 others); Wed, 22 Apr 2020 16:47:31 -0400 Received: from mx0a-00082601.pphosted.com ([67.231.145.42]:32870 "EHLO mx0a-00082601.pphosted.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726117AbgDVUrV (ORCPT ); Wed, 22 Apr 2020 16:47:21 -0400 Received: from pps.filterd (m0148461.ppops.net [127.0.0.1]) by mx0a-00082601.pphosted.com (8.16.0.42/8.16.0.42) with SMTP id 03MKhRKC028348 for ; Wed, 22 Apr 2020 13:47:21 -0700 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=fb.com; h=from : to : cc : subject : date : message-id : mime-version : content-transfer-encoding : content-type; s=facebook; bh=bYXySgtA0glsWw22mN546qhTFN/cHgQM7XvI05NPNl0=; b=Hs0BTxGLRAmGPqW4n8qUX47lQDmBBx2ISyhzRACb6/YWvcHgXsNzOQOPlMkO2HnfbR0l Okx/IYcub3tHBHaVp9Hp1KTprjqr9Bxc/GkGENFl0dNsz9e0pMl4IqeO6qSzELDZPNOb CWrvQSVpKyAQ+CH2F+b8p1LPSIuoyPT7trY= Received: from mail.thefacebook.com ([163.114.132.120]) by mx0a-00082601.pphosted.com with ESMTP id 30ghfebmvp-9 (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128 verify=NOT) for ; Wed, 22 Apr 2020 13:47:21 -0700 Received: from intmgw002.41.prn1.facebook.com (2620:10d:c085:108::8) by mail.thefacebook.com (2620:10d:c085:21d::4) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.1847.3; Wed, 22 Apr 2020 13:47:19 -0700 Received: by devvm4439.prn2.facebook.com (Postfix, from userid 111017) id BB5BB229357F7; Wed, 22 Apr 2020 13:47:16 -0700 (PDT) Smtp-Origin-Hostprefix: devvm From: Roman Gushchin Smtp-Origin-Hostname: devvm4439.prn2.facebook.com To: Andrew Morton CC: Johannes Weiner , Michal Hocko , , , , Roman Gushchin Smtp-Origin-Cluster: prn2c23 Subject: [PATCH v3 00/19] The new cgroup slab memory controller Date: Wed, 22 Apr 2020 13:46:49 -0700 Message-ID: <20200422204708.2176080-1-guro@fb.com> X-Mailer: git-send-email 2.24.1 MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-FB-Internal: Safe Content-Type: text/plain X-Proofpoint-Virus-Version: vendor=fsecure engine=2.50.10434:6.0.138,18.0.676 definitions=2020-04-22_07:2020-04-22,2020-04-22 signatures=0 X-Proofpoint-Spam-Details: rule=fb_default_notspam policy=fb_default score=0 bulkscore=0 priorityscore=1501 mlxlogscore=999 lowpriorityscore=0 phishscore=0 mlxscore=0 adultscore=0 clxscore=1015 malwarescore=0 spamscore=0 suspectscore=0 impostorscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2003020000 definitions=main-2004220158 X-FB-Internal: deliver Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org This is a third version of the slab cgroup controller rework. The patchset moves the accounting from the page level to the object level. It allows to share slab pages between memory cgroups. This leads to a significant win in the slab utilization (up to 45%) and the corresponding drop in the total kernel memory footprint. The reduced number of unmovable slab pages should also have a positive effect on the memory fragmentation. The patchset makes the slab accounting code simpler: there is no more need in the complicated dynamic creation and destruction of per-cgroup slab caches, all memory cgroups use a global set of shared slab caches. The lifetime of slab caches is not more connected to the lifetime of memory cgroups. The more precise accounting does require more CPU, however in practice the difference seems to be negligible. We've been using the new slab controller in Facebook production for several months with different workloads and haven't seen any noticeable regressions. What we've seen were memory savings in order of 1 GB per host (it varied heavily dependin= g on the actual workload, size of RAM, number of CPUs, memory pressure, etc= ). The third version of the patchset added yet another step towards the simplification of the code: sharing of slab caches between accounted and non-accounted allocations. It comes with significant upsides (most noticeable, a complete elimination of dynamic slab caches creation) but not without some regression risks, so this change sits on top of the patchset and is not completely merged in. So in the unlikel= y event of a noticeable performance regression it can be reverted separatel= y. v3: 1) added a patch that switches to a global single set of kmem_caches 2) kmem API clean up dropped, because if has been already merged 3) byte-sized slab vmstat API over page-sized global counters and bytes-sized memcg/lruvec counters 3) obj_cgroup refcounting simplifications and other minor fixes 4) other minor changes v2: 1) implemented re-layering and renaming suggested by Johannes, added his patch to the set. Thanks! 2) fixed the issue discovered by Bharata B Rao. Thanks! 3) added kmem API clean up part 4) added slab/memcg follow-up clean up part 5) fixed a couple of issues discovered by internal testing on FB fleet. 6) added kselftests 7) included metadata into the charge calculation 8) refreshed commit logs, regrouped patches, rebased onto mm tree, etc v1: 1) fixed a bug in zoneinfo_show_print() 2) added some comments to the subpage charging API, a minor fix 3) separated memory.kmem.slabinfo deprecation into a separate patch, provided a drgn-based replacement 4) rebased on top of the current mm tree RFC: https://lwn.net/Articles/798605/ Johannes Weiner (1): mm: memcontrol: decouple reference counting from page accounting Roman Gushchin (18): mm: memcg: factor out memcg- and lruvec-level changes out of __mod_lruvec_state() mm: memcg: prepare for byte-sized vmstat items mm: memcg: convert vmstat slab counters to bytes mm: slub: implement SLUB version of obj_to_index() mm: memcg/slab: obj_cgroup API mm: memcg/slab: allocate obj_cgroups for non-root slab pages mm: memcg/slab: save obj_cgroup for non-root slab objects mm: memcg/slab: charge individual slab objects instead of pages mm: memcg/slab: deprecate memory.kmem.slabinfo mm: memcg/slab: move memcg_kmem_bypass() to memcontrol.h mm: memcg/slab: use a single set of kmem_caches for all accounted allocations mm: memcg/slab: simplify memcg cache creation mm: memcg/slab: deprecate memcg_kmem_get_cache() mm: memcg/slab: deprecate slab_root_caches mm: memcg/slab: remove redundant check in memcg_accumulate_slabinfo() mm: memcg/slab: use a single set of kmem_caches for all allocations kselftests: cgroup: add kernel memory accounting tests tools/cgroup: add memcg_slabinfo.py tool drivers/base/node.c | 6 +- fs/proc/meminfo.c | 4 +- include/linux/memcontrol.h | 80 ++- include/linux/mm_types.h | 5 +- include/linux/mmzone.h | 19 +- include/linux/slab.h | 5 - include/linux/slab_def.h | 8 +- include/linux/slub_def.h | 20 +- include/linux/vmstat.h | 16 +- kernel/power/snapshot.c | 2 +- mm/memcontrol.c | 569 ++++++++++-------- mm/oom_kill.c | 2 +- mm/page_alloc.c | 8 +- mm/slab.c | 39 +- mm/slab.h | 365 +++++------- mm/slab_common.c | 643 +-------------------- mm/slob.c | 12 +- mm/slub.c | 183 +----- mm/vmscan.c | 3 +- mm/vmstat.c | 33 +- mm/workingset.c | 6 +- tools/cgroup/memcg_slabinfo.py | 226 ++++++++ tools/testing/selftests/cgroup/.gitignore | 1 + tools/testing/selftests/cgroup/Makefile | 2 + tools/testing/selftests/cgroup/test_kmem.c | 382 ++++++++++++ 25 files changed, 1322 insertions(+), 1317 deletions(-) create mode 100755 tools/cgroup/memcg_slabinfo.py create mode 100644 tools/testing/selftests/cgroup/test_kmem.c --=20 2.25.3