Received: by 10.213.65.68 with SMTP id h4csp440799imn; Tue, 27 Mar 2018 02:10:42 -0700 (PDT) X-Google-Smtp-Source: AG47ELsD4Kvo1O7pgNyHqjjs06ALqgfNDcoJwZmc5qU/69FKO4dHsQ/JtQUYaWek5Qe5/ekczv6E X-Received: by 2002:a17:902:bc04:: with SMTP id n4-v6mr44159124pls.97.1522141842607; Tue, 27 Mar 2018 02:10:42 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1522141842; cv=none; d=google.com; s=arc-20160816; b=jbWTvmoUIJ4XQhD4ylU9/hBrCxFmPyYVXnCdvxTGPqS/lRMA++tpCQ3tidqXxQTRkB bgk9yByVzZbFTA0pOKknn8Ziql+9bQYpEDAHn4HZeBjKcE6zgbb+7RG5HtKWHSu5gDJe xTVf6rSvbaUBOCwIBSBD7WJZyKJmOLEn1ZNKxThEMXfJuMsih+khinGo4houNtWYXH98 /AljbkPDbBeJTrKH4PdaviZCStFzCIVRHtdfjQ1XuudUG+xi/kYGfWd90XQbfaTYSISA 9uhGVfXt3GuzxM49IVRKc7azmq3HG3Ka9xUZTXbp0GGhqP2XqXd0ZlP6Qq5cy54Faun4 4kcw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:in-reply-to:content-disposition :mime-version:references:message-id:subject:cc:to:from:date :dkim-signature:arc-authentication-results; bh=EcX1PF/5XS37HJJ9uq6v8kGo43zwTwSBHk8dNiM+PA4=; b=QR7z8LYLdFpZKfez3uwf2NGpmUbF7SquJg0txwm4BrjWTiiv0eKlOKiBqPLbMIDLCt qkcwlLbIsgtSAtcxKdLhXBoRdPY7l+O7ijLBZUTeranwjgnwfZ10hVwrWG2Iu7aQrVYW RIy9Z7rhDZO+RZKcvi0SjLr7HvAzRwRcqzwOdiZMjKHxjLqmtVl2RqXVmd0V1LuyZmqM gpLMQAF4pj7E2HwaIOUI4ltZP2cUT+nteGU/7Np6ooNevtXuZcZlVqsZkYWwc6SSob7k uxQpCHyEbUcovs4kFACu3l8vnnHCKKLbpk8sbk+kPX1ZFZicYpa/fyHoHtSe+KfCX7yU uNFg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=EvUF4rDw; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id g67si694603pfk.90.2018.03.27.02.10.28; Tue, 27 Mar 2018 02:10:42 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=EvUF4rDw; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752180AbeC0JIr (ORCPT + 99 others); Tue, 27 Mar 2018 05:08:47 -0400 Received: from mail-lf0-f67.google.com ([209.85.215.67]:35025 "EHLO mail-lf0-f67.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751246AbeC0JIp (ORCPT ); Tue, 27 Mar 2018 05:08:45 -0400 Received: by mail-lf0-f67.google.com with SMTP id t132-v6so32234720lfe.2 for ; Tue, 27 Mar 2018 02:08:44 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to; bh=EcX1PF/5XS37HJJ9uq6v8kGo43zwTwSBHk8dNiM+PA4=; b=EvUF4rDw9siC6SWxbIbiorH4t8L49ll7g/Dj3cJMel5TRuYMcQcmZgCQ/a7kDFE247 23jFSgXNm8xxqVV2tRTYfOh35HaZE7DJebEGu7X3hw6PYErukWQOqzWxnLs9K77Xwal/ j5W8ZchDh5vitJXhEaiubrzs6E3xk4NN/c9ypqdsnol5+V40WutQazBkxG5AFj7MkFWp 5JfscVfHCm5d4PyeGo3bf6XGB7JYiu8Vqm+Nrfi/I6OyladqvCqhgiFuyN5gFdcRf7xP DtSAFoPAfLyIVb8OZHCc8ckMmdlKV3Oh2qgJwkwcG5tGLACjoArP73uCYwaWaKFvO2ZR mDqg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=EcX1PF/5XS37HJJ9uq6v8kGo43zwTwSBHk8dNiM+PA4=; b=cZVajx7krMJuaCedDFqBEJ4uc4MpTTE+OPI6nbDyHuCCGTwDykvMUmQwyzgtxCT4LN e3Co7LIP5prSGXjrBTS9vEFFBVJoWe4VMtWsdfbKeqb/2X4Nm/o0a6Vil+7H/tiz6nwi SaDI4hTImEir2jlUcmh6mKXQIGTBdysN1KJk5xrtQ74gNUJEhE7CNb4De36TEYIFxPZ7 usOvwmxUhvBVKJ2bQJQIXmwq+AUCm86USDGVFzYv7tIsOWvMbjVB1kHgsFhrQ6TE/TyF Kt0A3ngpscEf6/hCSXRZycJDcpNO4VxLXVM09wTM7eAj5ynDmQ5XWQMCH7Umi7dcrnO+ xEbA== X-Gm-Message-State: AElRT7F6X5enDuVBs1AYElLcYGIIfmpkdMtbEfQt/bNyXmzuDJdc9IKr HsXg547SFgqbAj4iVsW44o4= X-Received: by 10.46.133.217 with SMTP id h25mr10947487ljj.109.1522141724005; Tue, 27 Mar 2018 02:08:44 -0700 (PDT) Received: from esperanza ([185.6.245.156]) by smtp.gmail.com with ESMTPSA id g64sm138206ljf.69.2018.03.27.02.08.42 (version=TLS1_2 cipher=ECDHE-RSA-CHACHA20-POLY1305 bits=256/256); Tue, 27 Mar 2018 02:08:43 -0700 (PDT) Date: Tue, 27 Mar 2018 12:08:41 +0300 From: Vladimir Davydov To: Michal Hocko Cc: Li RongQing , linux-kernel@vger.kernel.org, linux-mm@kvack.org, Andrew Morton , Johannes Weiner , Dave Chinner , Kirill Tkhai Subject: Re: [PATCH] mm/list_lru: replace spinlock with RCU in __list_lru_count_one Message-ID: <20180327090841.ujscbnb54cepencf@esperanza> References: <1522137544-27496-1-git-send-email-lirongqing@baidu.com> <20180327081546.GZ5652@dhcp22.suse.cz> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20180327081546.GZ5652@dhcp22.suse.cz> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org [Cc Kirill] AFAIU this has already been fixed in exactly the same fashion by Kirill (mmotm commit 8e7d1201ec71 "mm: make counting of list_lru_one::nr_items lockless"). Kirill is working on further optimizations right now, see https://lkml.kernel.org/r/152163840790.21546.980703278415599202.stgit@localhost.localdomain On Tue, Mar 27, 2018 at 10:15:46AM +0200, Michal Hocko wrote: > [CC Dave] > > On Tue 27-03-18 15:59:04, Li RongQing wrote: > > when reclaim memory, shink_slab will take lots of time even if > > no memory is reclaimed, since list_lru_count_one called by it > > needs to take a spinlock > > > > try to optimize it by replacing spinlock with RCU in > > __list_lru_count_one > > Isn't the RCU overkill here? Why cannot we simply do an optimistic > lockless check for nr_items? It would be racy but does it actually > matter? We should be able to tolerate occasional 0 to non-zero and vice > versa transitions AFAICS. > > > > > $dd if=aaa of=bbb bs=1k count=3886080 > > $rm -f bbb > > $time echo 100000000 >/cgroup/memory/test/memory.limit_in_bytes > > > > Before: 0m0.415s ===> after: 0m0.395s > > > > Signed-off-by: Li RongQing > > --- > > include/linux/list_lru.h | 2 ++ > > mm/list_lru.c | 69 ++++++++++++++++++++++++++++++++++-------------- > > 2 files changed, 51 insertions(+), 20 deletions(-) > > > > diff --git a/include/linux/list_lru.h b/include/linux/list_lru.h > > index bb8129a3474d..ae472538038e 100644 > > --- a/include/linux/list_lru.h > > +++ b/include/linux/list_lru.h > > @@ -29,6 +29,7 @@ struct list_lru_one { > > struct list_head list; > > /* may become negative during memcg reparenting */ > > long nr_items; > > + struct rcu_head rcu; > > }; > > > > struct list_lru_memcg { > > @@ -46,6 +47,7 @@ struct list_lru_node { > > struct list_lru_memcg *memcg_lrus; > > #endif > > long nr_items; > > + struct rcu_head rcu; > > } ____cacheline_aligned_in_smp; > > > > struct list_lru { > > diff --git a/mm/list_lru.c b/mm/list_lru.c > > index fd41e969ede5..4c58ed861729 100644 > > --- a/mm/list_lru.c > > +++ b/mm/list_lru.c > > @@ -52,13 +52,13 @@ static inline bool list_lru_memcg_aware(struct list_lru *lru) > > static inline struct list_lru_one * > > list_lru_from_memcg_idx(struct list_lru_node *nlru, int idx) > > { > > - /* > > - * The lock protects the array of per cgroup lists from relocation > > - * (see memcg_update_list_lru_node). > > - */ > > - lockdep_assert_held(&nlru->lock); > > - if (nlru->memcg_lrus && idx >= 0) > > - return nlru->memcg_lrus->lru[idx]; > > + struct list_lru_memcg *tmp; > > + > > + WARN_ON_ONCE(!rcu_read_lock_held()); > > + > > + tmp = rcu_dereference(nlru->memcg_lrus); > > + if (tmp && idx >= 0) > > + return rcu_dereference(tmp->lru[idx]); > > > > return &nlru->lru; > > } > > @@ -113,14 +113,17 @@ bool list_lru_add(struct list_lru *lru, struct list_head *item) > > struct list_lru_one *l; > > > > spin_lock(&nlru->lock); > > + rcu_read_lock(); > > if (list_empty(item)) { > > l = list_lru_from_kmem(nlru, item); > > list_add_tail(item, &l->list); > > l->nr_items++; > > nlru->nr_items++; > > + rcu_read_unlock(); > > spin_unlock(&nlru->lock); > > return true; > > } > > + rcu_read_unlock(); > > spin_unlock(&nlru->lock); > > return false; > > } > > @@ -133,14 +136,17 @@ bool list_lru_del(struct list_lru *lru, struct list_head *item) > > struct list_lru_one *l; > > > > spin_lock(&nlru->lock); > > + rcu_read_lock(); > > if (!list_empty(item)) { > > l = list_lru_from_kmem(nlru, item); > > list_del_init(item); > > l->nr_items--; > > nlru->nr_items--; > > + rcu_read_unlock(); > > spin_unlock(&nlru->lock); > > return true; > > } > > + rcu_read_unlock(); > > spin_unlock(&nlru->lock); > > return false; > > } > > @@ -166,12 +172,13 @@ static unsigned long __list_lru_count_one(struct list_lru *lru, > > { > > struct list_lru_node *nlru = &lru->node[nid]; > > struct list_lru_one *l; > > - unsigned long count; > > + unsigned long count = 0; > > > > - spin_lock(&nlru->lock); > > + rcu_read_lock(); > > l = list_lru_from_memcg_idx(nlru, memcg_idx); > > - count = l->nr_items; > > - spin_unlock(&nlru->lock); > > + if (l) > > + count = l->nr_items; > > + rcu_read_unlock(); > > > > return count; > > } > > @@ -204,6 +211,7 @@ __list_lru_walk_one(struct list_lru *lru, int nid, int memcg_idx, > > unsigned long isolated = 0; > > > > spin_lock(&nlru->lock); > > + rcu_read_lock(); > > l = list_lru_from_memcg_idx(nlru, memcg_idx); > > restart: > > list_for_each_safe(item, n, &l->list) { > > @@ -250,6 +258,7 @@ __list_lru_walk_one(struct list_lru *lru, int nid, int memcg_idx, > > } > > } > > > > + rcu_read_unlock(); > > spin_unlock(&nlru->lock); > > return isolated; > > } > > @@ -296,9 +305,14 @@ static void __memcg_destroy_list_lru_node(struct list_lru_memcg *memcg_lrus, > > int begin, int end) > > { > > int i; > > + struct list_lru_one *tmp; > > > > - for (i = begin; i < end; i++) > > - kfree(memcg_lrus->lru[i]); > > + for (i = begin; i < end; i++) { > > + tmp = memcg_lrus->lru[i]; > > + rcu_assign_pointer(memcg_lrus->lru[i], NULL); > > + if (tmp) > > + kfree_rcu(tmp, rcu); > > + } > > } > > > > static int __memcg_init_list_lru_node(struct list_lru_memcg *memcg_lrus, > > @@ -314,7 +328,7 @@ static int __memcg_init_list_lru_node(struct list_lru_memcg *memcg_lrus, > > goto fail; > > > > init_one_lru(l); > > - memcg_lrus->lru[i] = l; > > + rcu_assign_pointer(memcg_lrus->lru[i], l); > > } > > return 0; > > fail: > > @@ -325,25 +339,37 @@ static int __memcg_init_list_lru_node(struct list_lru_memcg *memcg_lrus, > > static int memcg_init_list_lru_node(struct list_lru_node *nlru) > > { > > int size = memcg_nr_cache_ids; > > + struct list_lru_memcg *tmp; > > > > - nlru->memcg_lrus = kvmalloc(size * sizeof(void *), GFP_KERNEL); > > - if (!nlru->memcg_lrus) > > + tmp = kvmalloc(size * sizeof(void *), GFP_KERNEL); > > + if (!tmp) > > return -ENOMEM; > > > > - if (__memcg_init_list_lru_node(nlru->memcg_lrus, 0, size)) { > > - kvfree(nlru->memcg_lrus); > > + if (__memcg_init_list_lru_node(tmp, 0, size)) { > > + kvfree(tmp); > > return -ENOMEM; > > } > > > > + rcu_assign_pointer(nlru->memcg_lrus, tmp); > > + > > return 0; > > } > > > > -static void memcg_destroy_list_lru_node(struct list_lru_node *nlru) > > +static void memcg_destroy_list_lru_node_rcu(struct rcu_head *rcu) > > { > > + struct list_lru_node *nlru; > > + > > + nlru = container_of(rcu, struct list_lru_node, rcu); > > + > > __memcg_destroy_list_lru_node(nlru->memcg_lrus, 0, memcg_nr_cache_ids); > > kvfree(nlru->memcg_lrus); > > } > > > > +static void memcg_destroy_list_lru_node(struct list_lru_node *nlru) > > +{ > > + call_rcu(&nlru->rcu, memcg_destroy_list_lru_node_rcu); > > +} > > + > > static int memcg_update_list_lru_node(struct list_lru_node *nlru, > > int old_size, int new_size) > > { > > @@ -371,9 +397,10 @@ static int memcg_update_list_lru_node(struct list_lru_node *nlru, > > * we have to use IRQ-safe primitives here to avoid deadlock. > > */ > > spin_lock_irq(&nlru->lock); > > - nlru->memcg_lrus = new; > > + rcu_assign_pointer(nlru->memcg_lrus, new); > > spin_unlock_irq(&nlru->lock); > > > > + synchronize_rcu(); > > kvfree(old); > > return 0; > > } > > @@ -487,6 +514,7 @@ static void memcg_drain_list_lru_node(struct list_lru_node *nlru, > > * we have to use IRQ-safe primitives here to avoid deadlock. > > */ > > spin_lock_irq(&nlru->lock); > > + rcu_read_lock(); > > > > src = list_lru_from_memcg_idx(nlru, src_idx); > > dst = list_lru_from_memcg_idx(nlru, dst_idx); > > @@ -495,6 +523,7 @@ static void memcg_drain_list_lru_node(struct list_lru_node *nlru, > > dst->nr_items += src->nr_items; > > src->nr_items = 0; > > > > + rcu_read_unlock(); > > spin_unlock_irq(&nlru->lock); > > } > > > > -- > > 2.11.0 > > -- > Michal Hocko > SUSE Labs >