Received: by 2002:a05:6358:d09b:b0:dc:cd0c:909e with SMTP id jc27csp5479226rwb; Mon, 14 Nov 2022 05:23:54 -0800 (PST) X-Google-Smtp-Source: AA0mqf4r1Rq1O4aPSYCfTtwMdKn0ZERdxEO3loUNRswW2gSAFSfGRIWC4W05xjKK2PuswOstiJdT X-Received: by 2002:a63:1948:0:b0:46e:d655:60ab with SMTP id 8-20020a631948000000b0046ed65560abmr11738447pgz.533.1668432234497; Mon, 14 Nov 2022 05:23:54 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1668432234; cv=none; d=google.com; s=arc-20160816; b=ggGyPbG5GZL3fNJrmaopjEM720NaRF8gVcDpSZV6TVNjO4dbGuVcV8OCKXUrTE7fXG usj+XnSvbkBOZACRSVJjpsR8NHZE+S2RvuGZiExdlpmGFWIPYoS7ogmvE16ZqixVAohj jMD5Wv70Xk/Qt/Dyiok9bOW2113AtL48YYVEyhkrV/9J/4e+EF41yURJBAbDKGhMIbL8 /auBnjCW9fT8+FYwD/5z4e06diptteX4APJVyGeTVcxQYuVENxMMt8opp6BkP3ZA9Uk9 jSXEzwAGG9VwXoxvQ8N/pbbi5OnX5nnHz4LjSQHlLlyAq3KMLcMuuIvn+fU7MfTWYhTp qObA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:user-agent:in-reply-to:content-disposition :mime-version:references:message-id:subject:cc:to:from:date :dkim-signature; bh=RiK/oQi7Q+aoSc1+VQpsOu1ppDcGNbAPyB0OS6bJJvo=; b=Y1jNdmsAkjZvMPsDH9bzDHOAM1EdCiYcPLFBV2lDhbvwwVzvRpKpfY5JVV5z/dpWDd /20nytI5gsKPRE2MyM/Basvradr+5QnxMeiMNGTNLQDeNtalXiZGn5ntSb0gQ7IisP7R 1oIQpl2UWxuu6mPN6GNc6/zf3TLkSDyhJPdaCk0B/z0zQ+rTvk83EeHqMN0moffvO9aa HCCc4ewK+oFx+u6M3CrAwv1t5mEBy2FQcWmumTJZW5Y34YTyL3uEhGBpu25/gViMQJ0s OdaeHBDdQIbYidWGVKqKWfJQHp5aWfxzdx3ljKlioAl50OMv4sxgEY9Chv+kL6u5uQoP 1oTA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=pRbwAOb4; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id g184-20020a636bc1000000b0047060a43798si8797376pgc.461.2022.11.14.05.23.41; Mon, 14 Nov 2022 05:23:54 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=pRbwAOb4; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S237387AbiKNMmj (ORCPT + 88 others); Mon, 14 Nov 2022 07:42:39 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:59090 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S237307AbiKNMmL (ORCPT ); Mon, 14 Nov 2022 07:42:11 -0500 Received: from dfw.source.kernel.org (dfw.source.kernel.org [139.178.84.217]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 979DC252A6 for ; Mon, 14 Nov 2022 04:41:55 -0800 (PST) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by dfw.source.kernel.org (Postfix) with ESMTPS id 32A8561146 for ; Mon, 14 Nov 2022 12:41:55 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id BC72EC4347C; Mon, 14 Nov 2022 12:41:52 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1668429714; bh=ROHcpC/nKLxTl8eOQBxhF5rqrx9+B1Spb6cN+ewIOYM=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=pRbwAOb4tOZ4brHaU+bnOXX6UlNTXkV85DPjU8WeLz86Zg+bja92UBVWz1qPvrel1 mA0XhVeRTDHS+fw4BkToX5sQjQqSfKunk2ZVNfeiyNY8lRoN8k6PPNM6zruBeYk9iF WFVUMv+Q7LawCIo+vqOFkQ22kaORIWHSdwJSHXgAi/mOEM/6IAefpqlGxIOyBv9f48 gypkvQanr9jOqgp5VDWf7F3XwHlIhk655vpNu6y2azzoVO2Jp2p99FTuVijn5isjdY lKrs9u07Bh9XPki4RJvkdFAGCoNzibESPWBycvTWqkgb3IBK786oEgDRT60TBG9tYu 0eJYzCUR5Md0A== Date: Mon, 14 Nov 2022 12:41:48 +0000 From: Will Deacon To: Jan Kara Cc: Waiman Long , Sebastian Andrzej Siewior , LKML , Thomas Gleixner , Steven Rostedt , Mel Gorman , Peter Zijlstra , Ingo Molnar , Catalin Marinas Subject: Re: Crash with PREEMPT_RT on aarch64 machine Message-ID: <20221114124147.GA30263@willie-the-truck> References: <20221103115444.m2rjglbkubydidts@quack3> <20221107135636.biouna36osqc4rik@quack3> <359cc93a-fce0-5af2-0fd5-81999fad186b@redhat.com> <20221109125756.GA24388@willie-the-truck> <20221109154023.cx2d4y3e7zqnuo35@quack3> <20221111142742.rh677sdwu55aeeno@quack3> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20221111142742.rh677sdwu55aeeno@quack3> User-Agent: Mutt/1.10.1 (2018-07-13) X-Spam-Status: No, score=-7.1 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_HI, SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Nov 11, 2022 at 03:27:42PM +0100, Jan Kara wrote: > On Wed 09-11-22 16:40:23, Jan Kara wrote: > > On Wed 09-11-22 12:57:57, Will Deacon wrote: > > > On Mon, Nov 07, 2022 at 11:49:01AM -0500, Waiman Long wrote: > > > > On 11/7/22 10:10, Sebastian Andrzej Siewior wrote: > > > > > + locking, arm64 > > > > > > > > > > On 2022-11-07 14:56:36 [+0100], Jan Kara wrote: > > > > > > > spinlock_t and raw_spinlock_t differ slightly in terms of locking. > > > > > > > rt_spin_lock() has the fast path via try_cmpxchg_acquire(). If you > > > > > > > enable CONFIG_DEBUG_RT_MUTEXES then you would force the slow path which > > > > > > > always acquires the rt_mutex_base::wait_lock (which is a raw_spinlock_t) > > > > > > > while the actual lock is modified via cmpxchg. > > > > > > So I've tried enabling CONFIG_DEBUG_RT_MUTEXES and indeed the corruption > > > > > > stops happening as well. So do you suspect some bug in the CPU itself? > > > > > If it is only enabling CONFIG_DEBUG_RT_MUTEXES (and not whole lockdep) > > > > > then it looks very suspicious. > > > > > CONFIG_DEBUG_RT_MUTEXES enables a few additional checks but the main > > > > > part is that rt_mutex_cmpxchg_acquire() + rt_mutex_cmpxchg_release() > > > > > always fail (and so the slowpath under a raw_spinlock_t is done). > > > > > > > > > > So if it is really the fast path (rt_mutex_cmpxchg_acquire()) then it > > > > > somehow smells like the CPU is misbehaving. > > > > > > > > > > Could someone from the locking/arm64 department check if the locking in > > > > > RT-mutex (rtlock_lock()) is correct? > > > > > > > > > > rtmutex locking uses try_cmpxchg_acquire(, ptr, ptr) for the fastpath > > > > > (and try_cmpxchg_release(, ptr, ptr) for unlock). > > > > > Now looking at it again, I don't see much difference compared to what > > > > > queued_spin_trylock() does except the latter always operates on 32bit > > > > > value instead a pointer. > > > > > > > > Both the fast path of queued spinlock and rt_spin_lock are using > > > > try_cmpxchg_acquire(), the only difference I saw is the size of the data to > > > > be cmpxchg'ed. qspinlock uses 32-bit integer whereas rt_spin_lock uses > > > > 64-bit pointer. So I believe it is more on how the arm64 does cmpxchg. I > > > > believe there are two different ways of doing it depending on whether LSE > > > > atomics is available in the platform. So exactly what arm64 system is being > > > > used here and what hardware capability does it have? > > > > > > I'd be more inclined to be suspicious of the slowpath tbh, as we need to > > > make sure that we have acquire semantics on all paths where the lock can > > > be taken. Looking at the rtmutex code, this really isn't obvious to me -- > > > for example, try_to_take_rt_mutex() appears to be able to return via the > > > 'takeit' label without acquire semantics and it looks like we might be > > > relying on the caller's subsequent _unlock_ of the wait_lock for ordering, > > > but that will give us release semantics which aren't correct. > > > > > > As a quick hack, can you try chucking a barrier into rt_mutex_set_owner()? > > > > Bingo! This patch fixes the crashes for me. > > So I suppose this is not an official fix, is it? Sebastian, it appears to > be a bug in rtmutex implementation in the end AFAIU ;) Right, somebody needs to go audit all the acquisition paths on the slow-path and make sure they all have acquire semantics. The trick is doing that without incurring unnecessary overhead, e.g. by making use of dependency ordering where it already exists. Will > > > > diff --git a/kernel/locking/rtmutex.c b/kernel/locking/rtmutex.c > > > index 7779ee8abc2a..dd6a66c90f53 100644 > > > --- a/kernel/locking/rtmutex.c > > > +++ b/kernel/locking/rtmutex.c > > > @@ -98,6 +98,7 @@ rt_mutex_set_owner(struct rt_mutex_base *lock, struct task_struct *owner) > > > val |= RT_MUTEX_HAS_WAITERS; > > > > > > WRITE_ONCE(lock->owner, (struct task_struct *)val); > > > + smp_mb(); > > > } > > > > > > static __always_inline void clear_rt_mutex_waiters(struct rt_mutex_base *lock) > > Honza > -- > Jan Kara > SUSE Labs, CR