Received: by 2002:a05:6358:11c7:b0:104:8066:f915 with SMTP id i7csp1182828rwl; Fri, 24 Mar 2023 07:18:57 -0700 (PDT) X-Google-Smtp-Source: AK7set+wbSRAT8zl0XClkVTMu5C7YuAM7RyJ5SiBQOw4DNtiM/D7LSdvzOqvg82AGfPPnBP5GEax X-Received: by 2002:a05:6402:c99:b0:4bc:edde:150d with SMTP id cm25-20020a0564020c9900b004bcedde150dmr10889832edb.0.1679667537105; Fri, 24 Mar 2023 07:18:57 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1679667537; cv=none; d=google.com; s=arc-20160816; b=J1MQ32I8HfUBuqk3LqeO1PKMkM5aeCPykXqR5iaHUDaGqzf0pWxhlWO+fw7316yLfG We58RRWccuXXQqTORv65LSGboNl+FgQ57L/6YWVWsrLR4KUE2R6V9SnlzDpCnCbDN/f+ BBK/BlrfSCSjMS6Qmy9nLJOFBCcFR+2sfXIAf4Z9xtp6XENekTDulsUnIdFDd16D5FrD 5EWj0pU9fV8SknLZjtyI4M6uz46y/l28nCNWf2ujPLNCMKkffWFaCf288t7mUqfG5mwt lLiSKwtVuXOOZFJLz1u/Ifji/yzU2dh9QQSy7o9SBu5eF0n+4uLqA/2shfnBExjfWB1T bGcA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:from :references:cc:to:content-language:subject:user-agent:mime-version :date:message-id:dkim-signature; bh=ADiIDdAPnjqVAbV/udpxZCAiMLd8rfZn004mVe2ENYg=; b=b0W7embyeIkfJlhNqbviD06EZmzVy90PxwgTnhAnSJTd6SPeu8QAxk4rkKLsD0xfds qDgf3d09kgUXBvZuMEoZ7reWCIssJw53HmfkPSyhP6W90vYrO1YbddrEwwriGjUUwj7p Ptxf5tfGRIUHQpvCFqpnNVwA6RswlmqEwVlSa5k9CgR1RXTrGJM6KuJa8KYaY6ZTVOo8 RrOqgRJ7z4+8Ux7TkGESfnACI+aR4IQBtJpLHs6fa27tda1u3dyeDKKahDdZAXfqHGUI kV5COwdSF9t/4XJPlWSmSPiUfQ+6OCeUS5+SAQJNFpMtIAhNXohZ4/lp1D9Kzg+4BnKX FeDw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=ToO3WvzO; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id d25-20020aa7d699000000b004fc8943f1ffsi236033edr.211.2023.03.24.07.18.25; Fri, 24 Mar 2023 07:18:57 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=ToO3WvzO; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231987AbjCXONd (ORCPT + 99 others); Fri, 24 Mar 2023 10:13:33 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:48118 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231823AbjCXONb (ORCPT ); Fri, 24 Mar 2023 10:13:31 -0400 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D03F51CAE9 for ; Fri, 24 Mar 2023 07:12:49 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1679667168; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=ADiIDdAPnjqVAbV/udpxZCAiMLd8rfZn004mVe2ENYg=; b=ToO3WvzOcpNA6POY7Rl5QDYxTX0ssWQb+5juObBFhHeheNcIRs8cQR1kjq0J5aPP6DX88N 1x4gBg/w9uLq7OQNd2WQMA4tAyn/ChUwGIAL/UAypMn22eSvNJdu0734FGu7yARjKq0iOI 4W/m+hYVeGL+x22Jm9VxxJZyHJ3NH60= Received: from mimecast-mx02.redhat.com (mx3-rdu2.redhat.com [66.187.233.73]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-638-1m-g8VaUM1S0ZM79i3uKWA-1; Fri, 24 Mar 2023 10:12:45 -0400 X-MC-Unique: 1m-g8VaUM1S0ZM79i3uKWA-1 Received: from smtp.corp.redhat.com (int-mx07.intmail.prod.int.rdu2.redhat.com [10.11.54.7]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id E7BAE29AA3B5; Fri, 24 Mar 2023 14:12:43 +0000 (UTC) Received: from [10.22.33.184] (unknown [10.22.33.184]) by smtp.corp.redhat.com (Postfix) with ESMTP id E174D140EBF4; Fri, 24 Mar 2023 14:12:42 +0000 (UTC) Message-ID: <53582a07-81c6-35eb-10bf-7920b27483be@redhat.com> Date: Fri, 24 Mar 2023 10:12:42 -0400 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:102.0) Gecko/20100101 Thunderbird/102.7.1 Subject: Re: [RFC PATCH 1/7] cgroup: rstat: only disable interrupts for the percpu lock Content-Language: en-US To: Yosry Ahmed , Tejun Heo Cc: Josef Bacik , Jens Axboe , Zefan Li , Johannes Weiner , Michal Hocko , Roman Gushchin , Shakeel Butt , Muchun Song , Andrew Morton , Vasily Averin , cgroups@vger.kernel.org, linux-block@vger.kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, bpf@vger.kernel.org References: <20230323040037.2389095-1-yosryahmed@google.com> <20230323040037.2389095-2-yosryahmed@google.com> From: Waiman Long In-Reply-To: Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit X-Scanned-By: MIMEDefang 3.1 on 10.11.54.7 X-Spam-Status: No, score=-0.2 required=5.0 tests=DKIMWL_WL_HIGH,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,NICE_REPLY_A,RCVD_IN_DNSWL_NONE, RCVD_IN_MSPIKE_H2,SPF_HELO_NONE,SPF_NONE autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 3/24/23 03:22, Yosry Ahmed wrote: > On Thu, Mar 23, 2023 at 6:39 PM Tejun Heo wrote: >> Hello, >> >> On Thu, Mar 23, 2023 at 04:00:31AM +0000, Yosry Ahmed wrote: >>> Currently, when sleeping is not allowed during rstat flushing, we hold >>> the global rstat lock with interrupts disabled throughout the entire >>> flush operation. Flushing in an O(# cgroups * # cpus) operation, and >>> having interrupts disabled throughout is dangerous. >>> >>> For some contexts, we may not want to sleep, but can be interrupted >>> (e.g. while holding a spinlock or RCU read lock). As such, do not >>> disable interrupts throughout rstat flushing, only when holding the >>> percpu lock. This breaks down the O(# cgroups * # cpus) duration with >>> interrupts disabled to a series of O(# cgroups) durations. >>> >>> Furthermore, if a cpu spinning waiting for the global rstat lock, it >>> doesn't need to spin with interrupts disabled anymore. >> I'm generally not a fan of big spin locks w/o irq protection. They too often >> become a source of unpredictable latency spikes. As you said, the global >> rstat lock can be held for quite a while. Removing _irq makes irq latency >> better on the CPU but on the other hand it makes a lot more likely that the >> lock is gonna be held even longer, possibly significantly so depending on >> the configuration and workload which will in turn stall other CPUs waiting >> for the lock. Sure, irqs are being serviced quicker but if the cost is more >> and longer !irq context multi-cpu stalls, what's the point? >> >> I don't think there's anything which requires the global lock to be held >> throughout the entire flushing sequence and irq needs to be disabled when >> grabbing the percpu lock anyway, so why not just release the global lock on >> CPU boundaries instead? We don't really lose anything significant that way. >> The durations of irq disabled sections are still about the same as in the >> currently proposed solution at O(# cgroups) and we avoid the risk of holding >> the global lock for too long unexpectedly from getting hit repeatedly by >> irqs while holding the global lock. > Thanks for taking a look! > > I think a problem with this approach is that we risk having to contend > for the global lock at every CPU boundary in atomic contexts. Right Isn't it the plan to just do a trylock in atomic contexts so that it won't get stuck spinning for the lock for an indeterminate amount of time? > now we contend for the global lock once, and once we have it we go > through all CPUs to flush, only having to contend with updates taking > the percpu locks at this point. If we unconditionally release & > reacquire the global lock at every CPU boundary then we may contend > for it much more frequently with concurrent flushers. Note that with the use of qspinlock in all the major arches, the impact of thundering herds of lockers are much less serious than before. There are certainly some overhead in doing multiple lock acquires and releases, but that shouldn't been too excessive. I am all in for reducing lock hold time as much as possible as it will improve the response time. Cheers, Longman