Received: by 2002:a05:6a10:17d3:0:0:0:0 with SMTP id hz19csp2658653pxb; Tue, 13 Apr 2021 07:14:01 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzTA41/SxS6ckVz42yipcnhyRN0GjaNBeegU5+u6FiqENwTXrl359AyZ/mimKJ2leZesxku X-Received: by 2002:a17:906:151a:: with SMTP id b26mr32569043ejd.492.1618323241690; Tue, 13 Apr 2021 07:14:01 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1618323241; cv=none; d=google.com; s=arc-20160816; b=QTyARllk2tm6ZYJJwgsIUqTHYZ0FkfRfLwIfRYIyM1dKWoSJGxVJmD5iYeUlUdKekC 3soZ4W+SpskCCcAbjTrINH4qFKx2q7QdSGsaC6z6SCb56Yc/Xxm4KAsMZsGitPypQMpY TLRVkO3hRuH6QyFWxd57LjPyIijfz0Pu21M7qsQnTViRpWB0XsyH8SMihPc/pSfvsGT8 0usoOmtNbtt9C3YLzqlOYgjXBIGswVM+zlwk/bll2XajykQ/zlgH4cwmcL+b7ewsPpw5 336Ro6hC31QjsOEj+VVBUHv4fYav2MtG5wK49vdPfSLtqAlPlAnFUG9A8mqWGDUH6K1E wHTg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:user-agent:in-reply-to:content-transfer-encoding :content-disposition:mime-version:references:message-id:subject:cc :to:from:date; bh=cUHex+FzSBdWYnFrFeQCHLHCGaVq+XLMvFksXbckuyI=; b=HZVOAr+1qlH/w6arBAtH9S9YMekGynuHT6uZ0OccghjeqLNisKuPX62MBsHnIMZ0zl ylHLMv0zkRgUKHnMZyPVCBwCS48Fgrq6d+9wmH5VfWO5JstiNY3N2FRCj/c08ZlsDMhL S25NFDrWeqlyfOUJXybwtxAKTV874x30vospqSC8ZEBkP0fOjVHL8BnZd4tCEdQnW8m5 ApSVfW2iu75oHG7VARZTcnSApKVa8P8EdGv5ay26dP/ydhZohegPqmoVs9yBCTobbeMJ 1OIpAsa1qAxWc81d7MALQLiTc7VUjIFiHESGBcTp3YmmyAjgG9EV7ZeQMXNTXgn0TLSE oQnQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id b19si9960283ede.447.2021.04.13.07.13.38; Tue, 13 Apr 2021 07:14:01 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S241931AbhDMJbZ (ORCPT + 99 others); Tue, 13 Apr 2021 05:31:25 -0400 Received: from mail.kernel.org ([198.145.29.99]:49494 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S241801AbhDMJbW (ORCPT ); Tue, 13 Apr 2021 05:31:22 -0400 Received: by mail.kernel.org (Postfix) with ESMTPSA id D94A6613B2; Tue, 13 Apr 2021 09:31:01 +0000 (UTC) Date: Tue, 13 Apr 2021 10:30:59 +0100 From: Catalin Marinas To: Christoph =?iso-8859-1?Q?M=FCllner?= Cc: Peter Zijlstra , Palmer Dabbelt , Anup Patel , Guo Ren , linux-riscv , Linux Kernel Mailing List , Guo Ren , will.deacon@arm.com, Arnd Bergmann Subject: Re: [PATCH] riscv: locks: introduce ticket-based spinlock implementation Message-ID: <20210413093059.GB15806@arm.com> References: MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: User-Agent: Mutt/1.10.1 (2018-07-13) Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Apr 13, 2021 at 11:22:40AM +0200, Christoph M?llner wrote: > On Tue, Apr 13, 2021 at 10:03 AM Peter Zijlstra wrote: > > On Mon, Apr 12, 2021 at 11:54:55PM +0200, Christoph M?llner wrote: > > > On Mon, Apr 12, 2021 at 7:33 PM Palmer Dabbelt wrote: > > > > My plan is to add a generic ticket-based lock, which can be selected at > > > > compile time. It'll have no architecture dependencies (though it'll > > > > likely have some hooks for architectures that can make this go faster). > > > > Users can then just pick which spinlock flavor they want, with the idea > > > > being that smaller systems will perform better with ticket locks and > > > > larger systems will perform better with queued locks. The main goal > > > > here is to give the less widely used architectures an easy way to have > > > > fair locks, as right now we've got a lot of code duplication because any > > > > architecture that wants ticket locks has to do it themselves. > > > > > > In the case of LL/SC sequences, we have a maximum of 16 instructions > > > on RISC-V. My concern with a pure-C implementation would be that > > > we cannot guarantee this (e.g. somebody wants to compile with -O0) > > > and I don't know of a way to abort the build in case this limit exceeds. > > > Therefore I have preferred inline assembly for OpenSBI (my initial idea > > > was to use closure-like LL/SC macros, where you can write the loop > > > in form of C code). > > > > For ticket locks you really only needs atomic_fetch_add() and > > smp_store_release() and an architectural guarantees that the > > atomic_fetch_add() has fwd progress under contention and that a sub-word > > store (through smp_store_release()) will fail the SC. > > > > Then you can do something like: > > > > void lock(atomic_t *lock) > > { > > u32 val = atomic_fetch_add(1<<16, lock); /* SC, gives us RCsc */ > > u16 ticket = val >> 16; > > > > for (;;) { > > if (ticket == (u16)val) > > break; > > cpu_relax(); > > val = atomic_read_acquire(lock); > > } > > } > > > > void unlock(atomic_t *lock) > > { > > u16 *ptr = (u16 *)lock + (!!__BIG_ENDIAN__); > > u32 val = atomic_read(lock); > > > > smp_store_release(ptr, (u16)val + 1); > > } > > > > That's _almost_ as simple as a test-and-set :-) It isn't quite optimal > > on x86 for not being allowed to use a memop on unlock, since its being > > forced into a load-store because of all the volatile, but whatever. > > What about trylock()? > I.e. one could implement trylock() without a loop, by letting > trylock() fail if the SC fails. > That looks safe on first view, but nobody does this right now. Not familiar with RISC-V but I'd recommend that a trylock only fails if the lock is locked (after LR). A SC may fail for other reasons (cacheline eviction; depending on the microarchitecture) even if the lock is unlocked. At least on arm64 we had this issue with an implementation having a tendency to always fail the first STXR. -- Catalin