Received: by 2002:ac0:a581:0:0:0:0:0 with SMTP id m1-v6csp609444imm; Fri, 22 Jun 2018 02:08:22 -0700 (PDT) X-Google-Smtp-Source: ADUXVKIXm+9A9u0/r+iDDcXDufTl8p+1/IwIwZHmWh1R42L91Mnob5QovHHbTMnYEkuG3352fyP3 X-Received: by 2002:a65:58cc:: with SMTP id e12-v6mr648409pgu.445.1529658501995; Fri, 22 Jun 2018 02:08:21 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1529658501; cv=none; d=google.com; s=arc-20160816; b=DJaTgJn3pyRNPdQObH8EnmIG8Bbs92gQHb19Bnkqb1HLH7ibTAhkhOVoor3xTXcJxE BeN9yAeJBnGObHqYf5MrMEnRVEiasgszvlDr/37//f4YEVCBSLnv2HqbBUGlRA4RE9se GJyNRMaRT1QVIv7V6m3H/Dkn0egoK9ZnU9qLbSJA0S04UWHPMCUXsu1yFrpegueMAKz9 vkfuULvEmrAmu2npezR4NHloD42TUBvCdn4ghDh84j54S/zpHACDsqP0WNvxjlGdgCty IRDFGqdH70MpcrNdYpkJXnrquT7XLk0FVy32wFkkXB4/FcGHpkYNLb7qrDcEjLUEcdbD ksDg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dkim-signature:arc-authentication-results; bh=nFlv8k/+t0Nepf0ALOPGnXEleR+ZvXyKWeWsaTY82wI=; b=V5S1iftAUDPqUjq0d3/9WgrQP2HyEjanAb+g+D4U+M4fj8bH9vlnDrjh4G8oiXD/na 6HpxH2UiDARZlIQFx8sdghPJYfairbgwGO2MBc82rd8xKVBK7n12MFB0cQWDEc9kjPps S6FxccIrc2FiNg31M7OJNPDdfZbNHdtbme6SzMbJvhpYsD8uw+19BMpWHANvoeHVrqBU uOsJi+Shj4HBgXQOKvgUt6HOAqs+JITRtzyCSpoDnMdGhfpypGFAQ/oGQH7C7CkkxawX R72HF/7QVG2+c/c1LjHqQuau0Xs7C1C01LWUlEzxAhZZcD2kypCCZ/sJMtimVA+/RL/H OaEA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@amarulasolutions.com header.s=google header.b=rsTXH4Kb; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id j7-v6si6896043plk.334.2018.06.22.02.07.59; Fri, 22 Jun 2018 02:08:21 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@amarulasolutions.com header.s=google header.b=rsTXH4Kb; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S932719AbeFVJG5 (ORCPT + 99 others); Fri, 22 Jun 2018 05:06:57 -0400 Received: from mail-wm0-f68.google.com ([74.125.82.68]:53730 "EHLO mail-wm0-f68.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751086AbeFVJGz (ORCPT ); Fri, 22 Jun 2018 05:06:55 -0400 Received: by mail-wm0-f68.google.com with SMTP id x6-v6so1452687wmc.3 for ; Fri, 22 Jun 2018 02:06:55 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=amarulasolutions.com; s=google; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to:user-agent; bh=nFlv8k/+t0Nepf0ALOPGnXEleR+ZvXyKWeWsaTY82wI=; b=rsTXH4KbZENFy6n/ddcJ6huxfEAgSPSGxoXRIYL9IilF0mTbZJ9uYucrFNOXvnj+Yp FMy6GPx2edbektnjr2vhbbLOiGJ8H6Ci5dc9pu96ASroH+tk/X/c7pR2GL5Ec8mJ8eTw AWkonclX+nf1OkTR8dciEB0nrHAPhEUoOp4ug= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=nFlv8k/+t0Nepf0ALOPGnXEleR+ZvXyKWeWsaTY82wI=; b=I1Z32c/SZoU3Fp3JQRBzeA6dGlchEXfSdM/9LCFXMssxmA4XI00Ebb6q+KlKCgei76 Tby5nUER1olR8E1uHB0v3dEoycwoE7DafdDtuvSjj6SE9Hnr9rsus0+7ajeCnF7UxIOl +XU+blIL+m8puk7eaYllBR3sXqPqANLHACoXTcq4ory9sV8QEJ+fHJAHFkQ+0sq/NGhN 0OWXCofDJcRh/2+TD+yFAK8gsCnr2gAg85lFoy3HoLHWydSarxgJh3quhdoUlQ8AKwqz 6jXSgR/8H8dRc0Z3Op0c2ASqv87w3mK8XNH3VKPxlISWYa265vIMqdCU5M8FBhoZeSpw wsig== X-Gm-Message-State: APt69E3t8xg5emgLR6u+0wh2zRW32Yahc3DWLgqxNDt4YXc2JqG4mzQa Rtpr4nJ44kpKuvoBzs3ZiuUzeg== X-Received: by 2002:a1c:d884:: with SMTP id p126-v6mr949518wmg.110.1529658414232; Fri, 22 Jun 2018 02:06:54 -0700 (PDT) Received: from andrea (85.100.broadband17.iol.cz. [109.80.100.85]) by smtp.gmail.com with ESMTPSA id v138-v6sm1857590wmd.47.2018.06.22.02.06.53 (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Fri, 22 Jun 2018 02:06:53 -0700 (PDT) Date: Fri, 22 Jun 2018 11:06:51 +0200 From: Andrea Parri To: Alan Stern Cc: LKMM Maintainers -- Akira Yokosawa , Boqun Feng , David Howells , Jade Alglave , Luc Maranget , Nicholas Piggin , "Paul E. McKenney" , Peter Zijlstra , Will Deacon , Kernel development list Subject: Re: [PATCH 2/2] tools/memory-model: Add write ordering by release-acquire and by locks Message-ID: <20180622090651.GB6933@andrea> References: MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.5.24 (2015-08-30) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Jun 21, 2018 at 01:27:12PM -0400, Alan Stern wrote: > More than one kernel developer has expressed the opinion that the LKMM > should enforce ordering of writes by release-acquire chains and by > locking. In other words, given the following code: > > WRITE_ONCE(x, 1); > spin_unlock(&s): > spin_lock(&s); > WRITE_ONCE(y, 1); > > or the following: > > smp_store_release(&x, 1); > r1 = smp_load_acquire(&x); // r1 = 1 > WRITE_ONCE(y, 1); > > the stores to x and y should be propagated in order to all other CPUs, > even though those other CPUs might not access the lock s or be part of > the release-acquire chain. In terms of the memory model, this means > that rel-rf-acq-po should be part of the cumul-fence relation. > > All the architectures supported by the Linux kernel (including RISC-V) > do behave this way, albeit for varying reasons. Therefore this patch > changes the model in accordance with the developers' wishes. > > Signed-off-by: Alan Stern This patch changes the "Result" for ISA2+pooncelock+pooncelock+pombonce, so it should update the corresponding comment/README. Reviewed-and-Tested-by: Andrea Parri Andrea > > --- > > > [as1871] > > > tools/memory-model/Documentation/explanation.txt | 81 +++++++++++++++++++++++ > tools/memory-model/linux-kernel.cat | 2 > 2 files changed, 82 insertions(+), 1 deletion(-) > > Index: usb-4.x/tools/memory-model/linux-kernel.cat > =================================================================== > --- usb-4.x.orig/tools/memory-model/linux-kernel.cat > +++ usb-4.x/tools/memory-model/linux-kernel.cat > @@ -66,7 +66,7 @@ let ppo = to-r | to-w | fence > > (* Propagation: Ordering from release operations and strong fences. *) > let A-cumul(r) = rfe? ; r > -let cumul-fence = A-cumul(strong-fence | po-rel) | wmb > +let cumul-fence = A-cumul(strong-fence | po-rel) | wmb | rel-rf-acq-po > let prop = (overwrite & ext)? ; cumul-fence* ; rfe? > > (* > Index: usb-4.x/tools/memory-model/Documentation/explanation.txt > =================================================================== > --- usb-4.x.orig/tools/memory-model/Documentation/explanation.txt > +++ usb-4.x/tools/memory-model/Documentation/explanation.txt > @@ -1897,3 +1897,84 @@ non-deadlocking executions. For example > Is it possible to end up with r0 = 36 at the end? The LKMM will tell > you it is not, but the model won't mention that this is because P1 > will self-deadlock in the executions where it stores 36 in y. > + > +In the LKMM, locks and release-acquire chains cause stores to > +propagate in order. For example: > + > + int x, y, z; > + > + P0() > + { > + WRITE_ONCE(x, 1); > + smp_store_release(&y, 1); > + } > + > + P1() > + { > + int r1; > + > + r1 = smp_load_acquire(&y); > + WRITE_ONCE(z, 1); > + } > + > + P2() > + { > + int r2, r3, r4; > + > + r2 = READ_ONCE(z); > + smp_rmb(); > + r3 = READ_ONCE(x); > + r4 = READ_ONCE(y); > + } > + > +If r1 = 1 and r2 = 1 at the end, then both r3 and r4 must also be 1. > +In other words, the smp_store_release() read by the smp_load_acquire() > +together act as a sort of inter-processor fence, forcing the stores to > +x and y to propagate to P2 before the store to z does, regardless of > +the fact that P2 doesn't execute any release or acquire instructions. > +This conclusion would hold even if P0 and P1 were on the same CPU, so > +long as r1 = 1. > + > +We have mentioned that the LKMM treats locks as acquires and unlocks > +as releases. Therefore it should not be surprising that something > +analogous to this ordering also holds for locks: > + > + int x, y; > + spinlock_t s; > + > + P0() > + { > + spin_lock(&s); > + WRITE_ONCE(x, 1); > + spin_unlock(&s); > + } > + > + P1() > + { > + int r1; > + > + spin_lock(&s); > + r1 = READ_ONCE(x): > + WRITE_ONCE(y, 1); > + spin_unlock(&s); > + } > + > + P2() > + { > + int r2, r3; > + > + r2 = READ_ONCE(y); > + smp_rmb(); > + r3 = READ_ONCE(x); > + } > + > +If r1 = 1 at the end (implying that P1's critical section executes > +after P0's) and r2 = 1, then r3 must be 1; the ordering of the > +critical sections forces the store to x to propagate to P2 before the > +store to y does. > + > +In both versions of this scenario, the store-propagation ordering is > +not required by the operational model. However, it does happen on all > +the architectures supporting the Linux kernel, and kernel developers > +seem to expect it; they have requested that this behavior be included > +in the LKMM. >