Received: by 10.192.165.148 with SMTP id m20csp2633876imm; Sun, 22 Apr 2018 11:21:28 -0700 (PDT) X-Google-Smtp-Source: AIpwx49MHbTsuWG5pmHLgzCHhsrRgpF2Qv61PeW7TUg3RTH++VUsp//w3nuAhL0O2YKHi5tespMW X-Received: by 2002:a17:902:5c6:: with SMTP id f64-v6mr17944274plf.77.1524421288505; Sun, 22 Apr 2018 11:21:28 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1524421288; cv=none; d=google.com; s=arc-20160816; b=z5li3D1iKmNzDxPpp/Ajwf5/1ogSWGwiad0suyuM+dFsFf536g3gpRs3JFjg82QO89 ZI892S5HaKLKeS1tYXA8v+ERiU7OrLerORDXy3sqYwbuxg1k4JS8SWwFiCeeTT5Hwbpm mv78rtKqb40qBiC98McF9qJHfz+bf9BBYhw009ZztyLSTPrQaGg1qmcKy/jNRrL5V6/F peiAs+9AwxKSHzG/JljCtCUv9bXMYEGC1MmKw3FWvEbCcT7Vs+ulzDEd8l/CCZYPDWNg 2T3ZJ39FJt/M34EkrtDDUdKqEQHm6ZbqqwStL1fzRf5SKhhkF+057JMkht94YJtUa1aA WtNw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:in-reply-to:content-disposition :mime-version:references:message-id:subject:cc:to:from:date :dkim-signature:arc-authentication-results; bh=ycM3aFJvgkGBA2aqAcuyHiRVLCy27gg5xEj+XbrE1is=; b=RYV2YQiQT7Ca5aRChnbN0xXGmvgAMnF6KPMxwYo54CmZbCGBA4HHxNz/Dc8eh9yYP4 AVj036/CMjlKYWZR5YmwY4j8QzlykYSRsszLQhtReG/tQWjTVAfjDTaukd2Kzrp9Q44I qHc2/y2hOUJ9GbtdjLtdSsmFc9t5BJQhKgXZ3YHh1yuyBVoNoIVH/jD/Uxgr22zzCrGL QDNDELpYRA/edm54XdVVG2C96raZMm2mXSeVQtZPtjQLyJOVTVO8nQg4NeJW19sNnD7U /vK/0UrRv9Jalt8d9U69RHGtQhMSKPUwCCasUZCW6I9er+B3F2gYYnrFsLtQ0DDIrKRJ Qawg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=iZT9XgoG; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id p1si9712032pfp.72.2018.04.22.11.21.02; Sun, 22 Apr 2018 11:21:28 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=iZT9XgoG; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753423AbeDVSTS (ORCPT + 99 others); Sun, 22 Apr 2018 14:19:18 -0400 Received: from mail-lf0-f66.google.com ([209.85.215.66]:35066 "EHLO mail-lf0-f66.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751683AbeDVSTQ (ORCPT ); Sun, 22 Apr 2018 14:19:16 -0400 Received: by mail-lf0-f66.google.com with SMTP id r125-v6so11722060lfe.2 for ; Sun, 22 Apr 2018 11:19:15 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to; bh=ycM3aFJvgkGBA2aqAcuyHiRVLCy27gg5xEj+XbrE1is=; b=iZT9XgoGLRuPJk9TIoV1OaK37HmUlNLDcYruDuZrA9uacQN/tV3ROLTVMETbw1C0Pf juEMUCiQd076cZ/cGFYXfMrj9SyTY8lxTYCGnB0YTq5B2mADIH+d/ZSPErvdlw0DfXqA gDL5a4OTL1K/sAF2YkkKL7ZRowPkjWMVSMs1Y2L97MRPmEJ4I3WozSo1tQOxE11g1jIe Ujpvmg/NDA0LxVGPw9PY87dx+A2couwUgu/zCUJoUnmVDQzblgX6VkDjWieJ69hgfeQ/ iQIb8YzBUnq9YIwYqqTBlL8g5L+97qRYkS67oa3YDfnj96jDvhg//ktSlH0/pxw9Pgkp BNgg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=ycM3aFJvgkGBA2aqAcuyHiRVLCy27gg5xEj+XbrE1is=; b=gGgxWsLgTuP2olvKQwcnr30effI240z5h4rjewrB3Y5CvN/hD4J8xZXqFqmrrnUtsQ RAntDQz0+3KEJCt8A5kzK5sn5VKbtSKJbd8SP0HdZWZDQ6h8WkSpcq8TSCe2Cy+zXaRI A11CRNNtD40IweTWTv6yBTVzJ9Jd2yUW+TPITvSur7xg61HgLgfo1r1XQq/HAmXy47Bk 6shYeaZb1o3oxAkYJxcw2cjss5qmr6CP37z3wVFkR30j0gAbrAkyIVuSMS9dmNVgPc7m S7n2KmWLbGoKq3w2+0WN6w1khtqGwi1ps7P8q4fuZnwKgYRSc/p6Ya8s4GOzNywBkW1r tqaw== X-Gm-Message-State: ALQs6tBz5lRS8hwMT8t5hzwNvcnmEjlNu4kLHfRoB/cJJGmzyCb8VFYF +nJ0XcqX3m6wKu8IdeDqfDk= X-Received: by 10.46.82.210 with SMTP id n79mr11637778lje.146.1524421154915; Sun, 22 Apr 2018 11:19:14 -0700 (PDT) Received: from esperanza (81.5.110.211.dhcp.mipt-telecom.ru. [81.5.110.211]) by smtp.gmail.com with ESMTPSA id f26-v6sm2420198lfl.90.2018.04.22.11.19.13 (version=TLS1_2 cipher=ECDHE-RSA-CHACHA20-POLY1305 bits=256/256); Sun, 22 Apr 2018 11:19:14 -0700 (PDT) Date: Sun, 22 Apr 2018 21:19:11 +0300 From: Vladimir Davydov To: Kirill Tkhai Cc: akpm@linux-foundation.org, shakeelb@google.com, viro@zeniv.linux.org.uk, hannes@cmpxchg.org, mhocko@kernel.org, tglx@linutronix.de, pombredanne@nexb.com, stummala@codeaurora.org, gregkh@linuxfoundation.org, sfr@canb.auug.org.au, guro@fb.com, mka@chromium.org, penguin-kernel@I-love.SAKURA.ne.jp, chris@chris-wilson.co.uk, longman@redhat.com, minchan@kernel.org, hillf.zj@alibaba-inc.com, ying.huang@intel.com, mgorman@techsingularity.net, jbacik@fb.com, linux@roeck-us.net, linux-kernel@vger.kernel.org, linux-mm@kvack.org, willy@infradead.org, lirongqing@baidu.com, aryabinin@virtuozzo.com Subject: Re: [PATCH v2 10/12] mm: Iterate only over charged shrinkers during memcg shrink_slab() Message-ID: <20180422181911.axqiabv3cl7qtrpc@esperanza> References: <152397794111.3456.1281420602140818725.stgit@localhost.localdomain> <152399127400.3456.6644633244163904030.stgit@localhost.localdomain> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <152399127400.3456.6644633244163904030.stgit@localhost.localdomain> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Apr 17, 2018 at 09:54:34PM +0300, Kirill Tkhai wrote: > Using the preparations made in previous patches, in case of memcg > shrink, we may avoid shrinkers, which are not set in memcg's shrinkers > bitmap. To do that, we separate iterations over memcg-aware and > !memcg-aware shrinkers, and memcg-aware shrinkers are chosen > via for_each_set_bit() from the bitmap. In case of big nodes, > having many isolated environments, this gives significant > performance growth. See next patches for the details. > > Note, that the patch does not respect to empty memcg shrinkers, > since we never clear the bitmap bits after we set it once. > Their shrinkers will be called again, with no shrinked objects > as result. This functionality is provided by next patches. > > Signed-off-by: Kirill Tkhai > --- > mm/vmscan.c | 88 ++++++++++++++++++++++++++++++++++++++++++++++++----------- > 1 file changed, 72 insertions(+), 16 deletions(-) > > diff --git a/mm/vmscan.c b/mm/vmscan.c > index 34cd1d9b8b22..b81b8a7727b5 100644 > --- a/mm/vmscan.c > +++ b/mm/vmscan.c > @@ -169,6 +169,20 @@ unsigned long vm_total_pages; > static LIST_HEAD(shrinker_list); > static DECLARE_RWSEM(shrinker_rwsem); > > +static void link_shrinker(struct shrinker *shrinker) > +{ > + down_write(&shrinker_rwsem); > + list_add_tail(&shrinker->list, &shrinker_list); > + up_write(&shrinker_rwsem); > +} > + > +static void unlink_shrinker(struct shrinker *shrinker) > +{ > + down_write(&shrinker_rwsem); > + list_del(&shrinker->list); > + up_write(&shrinker_rwsem); > +} > + > #if defined(CONFIG_MEMCG) && !defined(CONFIG_SLOB) > static DEFINE_IDR(shrinkers_id_idr); > > @@ -221,11 +235,13 @@ static void del_memcg_shrinker(struct shrinker *shrinker) > #else /* CONFIG_MEMCG && !CONFIG_SLOB */ > static int add_memcg_shrinker(struct shrinker *shrinker, int nr, va_list args) > { > + link_shrinker(shrinker); > return 0; > } > > static void del_memcg_shrinker(struct shrinker *shrinker) > { > + unlink_shrinker(shrinker); > } > #endif /* CONFIG_MEMCG && !CONFIG_SLOB */ > > @@ -382,11 +398,9 @@ int __register_shrinker(struct shrinker *shrinker, int nr, ...) > va_end(args); > if (ret) > goto free_deferred; > - } > + } else > + link_shrinker(shrinker); > > - down_write(&shrinker_rwsem); > - list_add_tail(&shrinker->list, &shrinker_list); > - up_write(&shrinker_rwsem); > return 0; > > free_deferred: > @@ -405,9 +419,8 @@ void unregister_shrinker(struct shrinker *shrinker) > return; > if (shrinker->flags & SHRINKER_MEMCG_AWARE) > del_memcg_shrinker(shrinker); > - down_write(&shrinker_rwsem); > - list_del(&shrinker->list); > - up_write(&shrinker_rwsem); > + else > + unlink_shrinker(shrinker); I really don't like that depending on the config, the shrinker_list stores either all shrinkers or only memcg-unaware ones. I think it should always store all shrinkers and it should be used in case of global reclaim. That is IMO shrink_slab should look like this: shrink_slab(memcg) { if (!mem_cgroup_is_root(memcg)) return shrink_slab_memcg() list_for_each(shrinker, shrinker_list, link) do_shrink_slab() } Yeah, that means that for the root mem cgroup we will always call all shrinkers, but IMO it is OK as there's the only root mem cgroup out there and it is visited only on global reclaim so it shouldn't degrade performance. > kfree(shrinker->nr_deferred); > shrinker->nr_deferred = NULL; > } > @@ -532,6 +545,53 @@ static unsigned long do_shrink_slab(struct shrink_control *shrinkctl, > return freed; > } > > +#if defined(CONFIG_MEMCG) && !defined(CONFIG_SLOB) > +static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid, > + struct mem_cgroup *memcg, > + int priority) > +{ > + struct memcg_shrinker_map *map; > + unsigned long freed = 0; > + int ret, i; > + > + if (!down_read_trylock(&shrinker_rwsem)) > + return 0; > + > + /* > + * 1)Caller passes only alive memcg, so map can't be NULL. > + * 2)shrinker_rwsem protects from maps expanding. > + */ > + map = rcu_dereference_protected(SHRINKERS_MAP(memcg, nid), true); > + BUG_ON(!map); > + > + for_each_set_bit(i, map->map, shrinkers_max_nr) { > + struct shrink_control sc = { > + .gfp_mask = gfp_mask, > + .nid = nid, > + .memcg = memcg, > + }; > + struct shrinker *shrinker; > + > + shrinker = idr_find(&shrinkers_id_idr, i); > + if (!shrinker) { > + clear_bit(i, map->map); > + continue; > + } > + > + ret = do_shrink_slab(&sc, shrinker, priority); > + freed += ret; > + > + if (rwsem_is_contended(&shrinker_rwsem)) { > + freed = freed ? : 1; > + break; > + } > + } > + > + up_read(&shrinker_rwsem); > + return freed; > +} > +#endif > + > /** > * shrink_slab - shrink slab caches > * @gfp_mask: allocation context > @@ -564,6 +624,11 @@ static unsigned long shrink_slab(gfp_t gfp_mask, int nid, > if (memcg && (!memcg_kmem_enabled() || !mem_cgroup_online(memcg))) > return 0; The check above should be moved to shrink_slab_memcg. > > +#if defined(CONFIG_MEMCG) && !defined(CONFIG_SLOB) Please don't use ifdef here - define a stub function for no-memcg case. > + if (memcg) > + return shrink_slab_memcg(gfp_mask, nid, memcg, priority); > +#endif > + > if (!down_read_trylock(&shrinker_rwsem)) > goto out; > > @@ -574,15 +639,6 @@ static unsigned long shrink_slab(gfp_t gfp_mask, int nid, > .memcg = memcg, > }; > > - /* > - * If kernel memory accounting is disabled, we ignore > - * SHRINKER_MEMCG_AWARE flag and call all shrinkers > - * passing NULL for memcg. > - */ > - if (memcg_kmem_enabled() && > - !!memcg != !!(shrinker->flags & SHRINKER_MEMCG_AWARE)) > - continue; > - > if (!(shrinker->flags & SHRINKER_NUMA_AWARE)) > sc.nid = 0; > >