Received: by 2002:a05:6358:11c7:b0:104:8066:f915 with SMTP id i7csp1408990rwl; Fri, 24 Mar 2023 10:06:28 -0700 (PDT) X-Google-Smtp-Source: AKy350YV0ZfSBZsaw3DncU0l5u2NqbxGYdbF9LQDeNjpX3aggNMUFXWaTnIj1737wfE8pfbkfDN5 X-Received: by 2002:a17:902:db0e:b0:1a1:ab92:5c88 with SMTP id m14-20020a170902db0e00b001a1ab925c88mr3770614plx.13.1679677587882; Fri, 24 Mar 2023 10:06:27 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1679677587; cv=none; d=google.com; s=arc-20160816; b=rPW14vcQy3RWg2adU3naB9BZe8aZNdXOvuVufIVvRWPMo+zg4L7s7EGHZQCnixGGqJ Q/Nz0UxSCdw5kaFdO5u2Zgiafe4VQbPxm5kbuxNg+fNxHBtzhRNPIToSfYve9HGQ+Yyx aBVwszB6VFoSjQ0HT2yFpsJm8U/dC63vYYx7WADMj+JVZ1OOOy47dXJo8nnBFO/8GUlM KhpTqBHerStYCz6i0NQIHor8pA+4oDojeipwgm704R50WJIJO+N9TpR49iSK0s72oyt0 3ODx6okbuKCzHGTxGrUMA/VMAzR+7SgCgF6Z1MUDdWpuTHHYgeAdemsh9brqVz3DicPO Ysyw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date; bh=RxXBw0bo2ltl6/A6B38WBI2ESgr5g1GiIMzsArnvtqU=; b=Z54KO+SWVBJ06gjVW6es3jiCcXlRQk7IRS0U/RKGT3bG1GUCNi8DKxGFykKvzhESja rTpvl3QCeJ9Bkj1YmxRHbFR9+nIVfehQWbqQOgSetfpZ+0eC2ySPOQs2vjwE4Q7nLGIG JlhYTBkbCKS0q0r6/GQwRE3jq+FjI8X7LWPNMPPWz1TA4M5aBKTssRB0s3nAcAGsreKV ntN5CQIG9fcrtANbDVU/KClexyqs6EOwj8FMW+P491UI5Xgqf5QvTgNZFnAZswq++xjm 7z/Xe6pGIbf8kJFBccuciyKZ2jCn7sNYJj68KDC36rh97jiAsxhlQqwmRu4TJ9Dww8L1 2dgw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id s17-20020a170902ea1100b001a1ca4d1210si15771405plg.374.2023.03.24.10.06.07; Fri, 24 Mar 2023 10:06:27 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231873AbjCXRBl (ORCPT + 99 others); Fri, 24 Mar 2023 13:01:41 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:48842 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229508AbjCXRBi (ORCPT ); Fri, 24 Mar 2023 13:01:38 -0400 Received: from sin.source.kernel.org (sin.source.kernel.org [145.40.73.55]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 54E67E3A0; Fri, 24 Mar 2023 10:01:36 -0700 (PDT) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by sin.source.kernel.org (Postfix) with ESMTPS id 6768CCE2486; Fri, 24 Mar 2023 17:01:34 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id 297EEC433EF; Fri, 24 Mar 2023 17:01:31 +0000 (UTC) Date: Fri, 24 Mar 2023 17:01:28 +0000 From: Catalin Marinas To: "Bouska, Zdenek" Cc: Thomas Gleixner , Will Deacon , "linux-arm-kernel@lists.infradead.org" , "linux-kernel@vger.kernel.org" , "Kiszka, Jan" , "linux-rt-users@vger.kernel.org" , Nishanth Menon , Puranjay Mohan Subject: Re: Unfair qspinlocks on ARM64 without LSE atomics => 3ms delay in interrupt handling Message-ID: References: MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: X-Spam-Status: No, score=-4.8 required=5.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, RCVD_IN_DNSWL_HI,SPF_HELO_NONE,SPF_PASS autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Mar 24, 2023 at 08:43:38AM +0000, Bouska, Zdenek wrote: > I have seen ~3 ms delay in interrupt handling on ARM64. > > I have traced it down to raw_spin_lock() call in handle_irq_event() in > kernel/irq/handle.c: > > irqreturn_t handle_irq_event(struct irq_desc *desc) > { > irqreturn_t ret; > > desc->istate &= ~IRQS_PENDING; > irqd_set(&desc->irq_data, IRQD_IRQ_INPROGRESS); > raw_spin_unlock(&desc->lock); > > ret = handle_irq_event_percpu(desc); > > --> raw_spin_lock(&desc->lock); > irqd_clear(&desc->irq_data, IRQD_IRQ_INPROGRESS); > return ret; > } > > It took ~3 ms for this raw_spin_lock() to lock. That's quite a large indeed. > During this time irq_finalize_oneshot() from kernel/irq/manage.c locks and > unlocks the same raw spin lock more than 1000 times: > > static void irq_finalize_oneshot(struct irq_desc *desc, > struct irqaction *action) > { > if (!(desc->istate & IRQS_ONESHOT) || > action->handler == irq_forced_secondary_handler) > return; > again: > chip_bus_lock(desc); > --> raw_spin_lock_irq(&desc->lock); > > /* > * Implausible though it may be we need to protect us against > * the following scenario: > * > * The thread is faster done than the hard interrupt handler > * on the other CPU. If we unmask the irq line then the > * interrupt can come in again and masks the line, leaves due > * to IRQS_INPROGRESS and the irq line is masked forever. > * > * This also serializes the state of shared oneshot handlers > * versus "desc->threads_oneshot |= action->thread_mask;" in > * irq_wake_thread(). See the comment there which explains the > * serialization. > */ > if (unlikely(irqd_irq_inprogress(&desc->irq_data))) { > --> raw_spin_unlock_irq(&desc->lock); > chip_bus_sync_unlock(desc); > cpu_relax(); > goto again; > } So this path is hammering the desc->lock location and another CPU cannot change it. As you found, the problem is not the spinlock algorithm but the atomic primitives. The LDXR/STXR constructs on arm64 are known to have this issue with STXR failing indefinitely. raw_spin_unlock() simply does an STLR and this clears the exclusive monitor that the other CPU may have set with LDXR but before the STXR. The queued spinlock only provides fairness if the CPU manages to get in the queue. > So I confirmed that atomic operations from > arch/arm64/include/asm/atomic_ll_sc.h can be quite slow when they are > contested from second CPU. > > Do you think that it is possible to create fair qspinlock implementation > on top of atomic instructions supported by ARM64 version 8 (no LSE atomic > instructions) without compromising performance in the uncontested case? > For example ARM64 could have custom queued_fetch_set_pending_acquire > implementation same as x86 has in arch/x86/include/asm/qspinlock.h. Is the > retry loop in irq_finalize_oneshot() ok together with the current ARM64 > cpu_relax() implementation for processor with no LSE atomic instructions? So is the queued_fetch_set_pending_acquire() where it gets stuck or the earlier atomic_try_cmpxchg_acquire() before entering on the slow path? I guess both can fail in a similar way. A longer cpu_relax() here would improve things (on arm64 this function is a no-op) but maybe Thomas or Will have a better idea. -- Catalin