Received: by 2002:a05:6358:11c7:b0:104:8066:f915 with SMTP id i7csp56001rwl; Thu, 23 Mar 2023 12:36:31 -0700 (PDT) X-Google-Smtp-Source: AKy350bkhlQa6tDPIlFwwhQ6LdNxWKHdH24oML4fKLOnZIry+jLiZzeLT0D+Gs+IUn5E7ULDl77h X-Received: by 2002:a17:903:1c2:b0:19a:e96a:58b3 with SMTP id e2-20020a17090301c200b0019ae96a58b3mr52632plh.22.1679600191355; Thu, 23 Mar 2023 12:36:31 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1679600191; cv=none; d=google.com; s=arc-20160816; b=rxnKwqftXKTdJRf2uW47fqO6/1ApCDpNuUPGSzL0vD4pNJxRE9BReqYha0fUqtezBS eVcLd8FB5x4PjJn6gWReueoEBD67R0ISziNy9d7FFrPTzaya7mV60+KIsA2IGMa1bN5T heLeTIgXPGF0LagnI0FkdX+IA00FwcpNfUe28Gm/4+j74xwxyuFDPFsZr3gtf5Fcsl++ j2oaadiR6CezBtnXmiQaEcKiIHHkOw2AOiwphMqT6WT7Ppgzxp8d92d7IJEs4n6sC48I JKRuqIF+wWHLMqZbYnPjcxh8jPwqtTNnR7XnFJadDVvDzOPIUBcMZKCtaZ5/KL4UOMKa ea3g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:cc:to:subject :message-id:date:from:in-reply-to:references:mime-version :dkim-signature; bh=7VcmchQqt9DaMxSLZ78hVIFyFO9YofibofDFlcn/C/A=; b=kzYnR7kK32eUT39JTvHsgPVCjOgIgGLV/H3Q+XVZXCiE5BP9RYbwfT8yrfgUFy2J1g 6woqP+x+DjKPdUQMshvzhX6lsYPI8Q9R59OvN+MLfSpT7S+sGoe2r1yDpyfpr28g0wbp H7kWQicLvB1EDgK0MraP9WbD3vgZkSC9awqTuDH6lcK2RV/XldCN1ppBuOCE1TM+NSGx FQbv8Q+e6JR3RaHFYlQqR5QyX/i90gG5fUyIH5md6HGjYd7tyjcVMJUilPuimusRykQL UIUX7HZr4D9efAe0Tx0NmTSa4SQy7MAEmTv3dsc+mxD9r/w9l/FBjdmOxUg5cSELYMc3 m/cQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20210112 header.b=qcmPuY4U; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id g14-20020a63e60e000000b004fb40593898si20113047pgh.374.2023.03.23.12.36.15; Thu, 23 Mar 2023 12:36:31 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20210112 header.b=qcmPuY4U; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231530AbjCWTfi (ORCPT + 99 others); Thu, 23 Mar 2023 15:35:38 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:36448 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231245AbjCWTff (ORCPT ); Thu, 23 Mar 2023 15:35:35 -0400 Received: from mail-yw1-x112e.google.com (mail-yw1-x112e.google.com [IPv6:2607:f8b0:4864:20::112e]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id AB3AE1A65E for ; Thu, 23 Mar 2023 12:35:33 -0700 (PDT) Received: by mail-yw1-x112e.google.com with SMTP id 00721157ae682-544b959a971so378236777b3.3 for ; Thu, 23 Mar 2023 12:35:33 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20210112; t=1679600133; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:from:to:cc:subject:date :message-id:reply-to; bh=7VcmchQqt9DaMxSLZ78hVIFyFO9YofibofDFlcn/C/A=; b=qcmPuY4UNpBr+P2ZKeYHu1R051ec4TvA2Clkq4ip2880X7aXBn2kafw99H0CzcQE8Y 55btBqMovU8tiALDLxofgtxE/bZ0HgUAqQ7/uL9ivj3gfCrk6Yuh0oGKI0h3hhLcEY7P dflHLX+NDoqrZpncNEKvk51KUcHVmknk2JEAaqvkhAXJHpK54fIOZMMCA7hah18AZ/bd JkZhELoYu+zcuogS32m/qQmFm811rpAtxS23YU+81bGhLBg7rvz3YmoDRQyp8dCCDNhb lUPfwH5qF/cf+E5RouEWAGBRgF9zRDRmzCb9ikWcAdK0VJGHDe0vWwWNlwRMGXpXzXAs LaXw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; t=1679600133; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=7VcmchQqt9DaMxSLZ78hVIFyFO9YofibofDFlcn/C/A=; b=KNcaIR4XlMU9TIyNso4bJzeZtnQS/ecStSYbw+fAgRPG/fw4+RXDITXg9dz6BQnKYT YxSge9W0mJBAwIqhbpmrqyyttlKiZjAAIYSlZkcgWpKazHsDd2uXnwdd4jxSPO9PPCEY CarP0sPKvS0RrMdlcqzoBlWJGUbZELB9zYi0Fc/LJKk44piT+/Qg0yEjGOZiIHS8WdHK KKq831UE0WAOhEBGBZw+1UNQ1vAEFFFRzAwUXRcyD1zyKugAnvQJHcjYf9R72g1wJCl7 2wRloAJwvoSPMrKjCbWQoeoRZqqNBIZM/xBTYrN/6GXtTDJWJSlD5HqV7R6D1MBRq76K D2AA== X-Gm-Message-State: AAQBX9edMJ6rZKMZ7RlQaH/Nk66HP1isCoeZd5tnPh9cLeApZ7oupI5g 4BIRfrCOFo8T2Cn+RSBIjJWp+ZzNBBkHvYlbOXzShw== X-Received: by 2002:a81:ef08:0:b0:545:883a:542c with SMTP id o8-20020a81ef08000000b00545883a542cmr1610295ywm.0.1679600131909; Thu, 23 Mar 2023 12:35:31 -0700 (PDT) MIME-Version: 1.0 References: <20230323040037.2389095-1-yosryahmed@google.com> <20230323040037.2389095-5-yosryahmed@google.com> <20230323155613.GC739026@cmpxchg.org> <20230323172732.GE739026@cmpxchg.org> In-Reply-To: From: Shakeel Butt Date: Thu, 23 Mar 2023 12:35:20 -0700 Message-ID: Subject: Re: [RFC PATCH 4/7] memcg: sleep during flushing stats in safe contexts To: Yosry Ahmed Cc: Johannes Weiner , Tejun Heo , Josef Bacik , Jens Axboe , Zefan Li , Michal Hocko , Roman Gushchin , Muchun Song , Andrew Morton , Vasily Averin , cgroups@vger.kernel.org, linux-block@vger.kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, bpf@vger.kernel.org Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable X-Spam-Status: No, score=-15.7 required=5.0 tests=DKIMWL_WL_MED,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,ENV_AND_HDR_SPF_MATCH, RCVD_IN_DNSWL_NONE,SPF_HELO_NONE,SPF_PASS,USER_IN_DEF_DKIM_WL, USER_IN_DEF_SPF_WL autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Mar 23, 2023 at 11:08=E2=80=AFAM Yosry Ahmed wrote: > > On Thu, Mar 23, 2023 at 10:27=E2=80=AFAM Johannes Weiner wrote: > > > > On Thu, Mar 23, 2023 at 09:01:12AM -0700, Yosry Ahmed wrote: > > > On Thu, Mar 23, 2023 at 8:56=E2=80=AFAM Johannes Weiner wrote: > > > > > > > > On Thu, Mar 23, 2023 at 04:00:34AM +0000, Yosry Ahmed wrote: > > > > > @@ -644,26 +644,26 @@ static void __mem_cgroup_flush_stats(void) > > > > > return; > > > > > > > > > > flush_next_time =3D jiffies_64 + 2*FLUSH_TIME; > > > > > - cgroup_rstat_flush(root_mem_cgroup->css.cgroup, false); > > > > > + cgroup_rstat_flush(root_mem_cgroup->css.cgroup, may_sleep); > > > > > > > > How is it safe to call this with may_sleep=3Dtrue when it's holding= the > > > > stats_flush_lock? > > > > > > stats_flush_lock is always called with trylock, it is only used today > > > so that we can skip flushing if another cpu is already doing a flush > > > (which is not 100% correct as they may have not finished flushing yet= , > > > but that's orthogonal here). So I think it should be safe to sleep as > > > no one can be blocked waiting for this spinlock. > > > > I see. It still cannot sleep while the lock is held, though, because > > preemption is disabled. Make sure you have all lock debugging on while > > testing this. > > Thanks for pointing this out, will do. > > > > > > Perhaps it would be better semantically to replace the spinlock with > > > an atomic test and set, instead of having a lock that can only be use= d > > > with trylock? > > > > It could be helpful to clarify what stats_flush_lock is protecting > > first. Keep in mind that locks should protect data, not code paths. > > > > Right now it's doing multiple things: > > > > 1. It protects updates to stats_flush_threshold > > 2. It protects updates to flush_next_time > > 3. It serializes calls to cgroup_rstat_flush() based on those ratelimit= s > > > > However, > > > > 1. stats_flush_threshold is already an atomic > > > > 2. flush_next_time is not atomic. The writer is locked, but the reader > > is lockless. If the reader races with a flush, you could see this: > > > > if (time_after(jiffies, flush_n= ext_time)) > > spin_trylock() > > flush_next_time =3D now + delay > > flush() > > spin_unlock() > > spin_trylock() > > flush_next_time =3D now + delay > > flush() > > spin_unlock() > > > > which means we already can get flushes at a higher frequency than > > FLUSH_TIME during races. But it isn't really a problem. > > > > The reader could also see garbled partial updates, so it needs at > > least READ_ONCE and WRITE_ONCE protection. > > > > 3. Serializing cgroup_rstat_flush() calls against the ratelimit > > factors is currently broken because of the race in 2. But the race > > is actually harmless, all we might get is the occasional earlier > > flush. If there is no delta, the flush won't do much. And if there > > is, the flush is justified. > > > > In summary, it seems to me the lock can be ditched altogether. All the > > code needs is READ_ONCE/WRITE_ONCE around flush_next_time. > > Thanks a lot for this analysis. I agree that the lock can be removed > with proper READ_ONCE/WRITE_ONCE, but I think there is another purpose > of the lock that we are missing here. > > I think one other purpose of the lock is avoiding a thundering herd > problem on cgroup_rstat_lock, particularly from reclaim context, as > mentioned by the log of commit aa48e47e3906 ("memcg: infrastructure > to flush memcg stats"). > > While testing, I did notice that removing this lock indeed causes a > thundering herd problem if we have a lot of concurrent reclaimers. The > trylock makes sure we abort immediately if someone else is flushing -- > which is not ideal because that flusher might have just started, and > we may end up reading stale data anyway. > > This is why I suggested replacing the lock by an atomic, and do > something like this if we want to maintain the current behavior: > > static void __mem_cgroup_flush_stats(void) > { > ... > if (atomic_xchg(&ongoing_flush, 1)) > return; > ... > atomic_set(&ongoing_flush, 0) > } > > Alternatively, if we want to change the behavior and wait for the > concurrent flusher to finish flushing, we can maybe spin until > ongoing_flush goes back to 0 and then return: > > static void __mem_cgroup_flush_stats(void) > { > ... > if (atomic_xchg(&ongoing_flush, 1)) { > /* wait until the ongoing flusher finishes to get updated stats *= / > while (atomic_read(&ongoing_flush) {}; > return; > } > /* flush the stats ourselves */ > ... > atomic_set(&ongoing_flush, 0) > } > > WDYT? I would go with your first approach i.e. no spinning.