Received: by 2002:ac0:a581:0:0:0:0:0 with SMTP id m1-v6csp3839283imm; Mon, 25 Jun 2018 05:34:12 -0700 (PDT) X-Google-Smtp-Source: ADUXVKKD+Ei5GBXb5lUtpBQveHwoM49gkB8OppzR77Du2Qm7C1MHxtzC7bvsKMNAzb+h6H0wUPA1 X-Received: by 2002:a62:c6d8:: with SMTP id x85-v6mr12520253pfk.249.1529930052150; Mon, 25 Jun 2018 05:34:12 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1529930052; cv=none; d=google.com; s=arc-20160816; b=W/CNxF+YRYdnUw/t/41lK9wcEzSCr/GvfAaOs3W68SEgQ9o3NZWPi8YNOy0zohHfZR qOcTjq7jZXVPHEabA8WdtTc+8mnhWUvtNg9d+AkjNLWhulCoVOkkX0Ie/pusPmkm7geD 5LYDY2Ff6sWoTzk99MAiRg7ypNQzRIDxJkM+A8uhjH+QKgTnMdVyjQ0pfw1AeuuQbrxf 0FhybSm5T75cQZs3qWB9tCNIJB9fnPNq3jG7hIsYh9XqdUk5NmiPM5dYrO82K0A0KvNw s0TEMJaDjoLrmryaGgNgVR5U9S9hiFCzVDW+vPPSf63lCuESO5mlqqZqUUkk/ZbMHr/4 kU+w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dkim-signature:arc-authentication-results; bh=f9KiNDVz1b5Pz6C608BpVjCnxur/Rw1mf8e1suOXpV0=; b=rJYsf0uRLVy5pq3rX4Og7dH+5NSEfmHDtr6tI9Of9z/8H3skzF+v3zzojz87J7OxAb dkeAGbxd7eAHMfyTbKqchQZGUItVrTjMmBVACcSX/toqH6lXDpXKlakVNRcBd0tfP5uD pus1M0PlAODqidu2UKLMigjyQ8zHGCUWRhCUtIEZe0K9Mp+rrMvIQQPnRM0zQv9afbGl 3Z6x8HOS3TMhqsDNQIQdze0K0GxzCd5PkxZnNLJgoq/sdLvFYqrd5ZMIho4/dqGICdw5 CjLkwDS8hjyzQ3vDEIpemgBCutUmaX6dir+Pa08md/KyHTbllMqy8ye3fyzf+9q9Cb4F Hb+A== ARC-Authentication-Results: i=1; mx.google.com; dkim=fail header.i=@infradead.org header.s=merlin.20170209 header.b=n+VFJFk+; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id h2-v6si11677115pgf.334.2018.06.25.05.33.57; Mon, 25 Jun 2018 05:34:12 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=fail header.i=@infradead.org header.s=merlin.20170209 header.b=n+VFJFk+; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S933904AbeFYMbq (ORCPT + 99 others); Mon, 25 Jun 2018 08:31:46 -0400 Received: from merlin.infradead.org ([205.233.59.134]:51932 "EHLO merlin.infradead.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S933288AbeFYMbm (ORCPT ); Mon, 25 Jun 2018 08:31:42 -0400 DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=merlin.20170209; h=In-Reply-To:Content-Type:MIME-Version: References:Message-ID:Subject:Cc:To:From:Date:Sender:Reply-To: Content-Transfer-Encoding:Content-ID:Content-Description:Resent-Date: Resent-From:Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:List-Id: List-Help:List-Unsubscribe:List-Subscribe:List-Post:List-Owner:List-Archive; bh=f9KiNDVz1b5Pz6C608BpVjCnxur/Rw1mf8e1suOXpV0=; b=n+VFJFk+REYVTyw4a5TAyIpY5 vgEJtdw2DC1EGRmkMPU94rDq+VxzHdN7sCE8GYuf3Z7OxPIBLlCUe0k4Z+8kpH/GQubLAcrNo9A9A v4czAo6q7jsdCPt4Pwb2uez1GW7E73pfsfUK76f1qU+ZHGJQ1A2VqSXsVMjZA4+0RljeUJAdqQjXb T+UdJkLkPHJNDInoQ3J2gu9tyo1X9qKFt0nDms6mv9rdCQUVsd65vtDEUNyfou8UUy7XMN45Os1Jl kAieU2+DIS+rtGXdTYwy72sCev1URs5SqA2PlNasGx0ujBFyIuMw2mbdTkliI9LODZ2RyRMPvrX4e o2brt+nMA==; Received: from j217100.upc-j.chello.nl ([24.132.217.100] helo=hirez.programming.kicks-ass.net) by merlin.infradead.org with esmtpsa (Exim 4.90_1 #2 (Red Hat Linux)) id 1fXQeR-0008HC-V2; Mon, 25 Jun 2018 12:31:24 +0000 Received: by hirez.programming.kicks-ass.net (Postfix, from userid 1000) id 7DF122029FA0A; Mon, 25 Jun 2018 14:31:21 +0200 (CEST) Date: Mon, 25 Jun 2018 14:31:21 +0200 From: Peter Zijlstra To: Andrea Parri Cc: linux-kernel@vger.kernel.org, linux-doc@vger.kernel.org, Alan Stern , Will Deacon , Boqun Feng , Nicholas Piggin , David Howells , Jade Alglave , Luc Maranget , "Paul E. McKenney" , Akira Yokosawa , Daniel Lustig , Jonathan Corbet , Ingo Molnar , Randy Dunlap Subject: Re: [PATCH] doc: Update wake_up() & co. memory-barrier guarantees Message-ID: <20180625123121.GY2494@hirez.programming.kicks-ass.net> References: <1529918258-7295-1-git-send-email-andrea.parri@amarulasolutions.com> <20180625095031.GX2494@hirez.programming.kicks-ass.net> <20180625105618.GA12676@andrea> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20180625105618.GA12676@andrea> User-Agent: Mutt/1.10.0 (2018-05-17) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Jun 25, 2018 at 12:56:18PM +0200, Andrea Parri wrote: > On Mon, Jun 25, 2018 at 11:50:31AM +0200, Peter Zijlstra wrote: > > On Mon, Jun 25, 2018 at 11:17:38AM +0200, Andrea Parri wrote: > > > Both the implementation and the users' expectation [1] for the various > > > wakeup primitives have evolved over time, but the documentation has not > > > kept up with these changes: brings it into 2018. > > > > I wanted to reply to this saying that I'm not aware of anything relying > > on this actually being a smp_mb() and that I've been treating it as an > > RELEASE. > > > > But then I found my own comment that goes with smp_mb__after_spinlock(), > > which explains why we do in fact need the transitive thing if I'm not > > mistaken. > > A concrete example being the store-buffering pattern reported in [1]. Well, that example only needs a store->load barrier. It so happens smp_mb() is the only one actually doing that, but imagine we had a weaker barrier that did just that, one that did not imply the full transitivity smp_mb() does. Then the example from [1] could use that weaker thing. > > So yes, I suppose we're entirely suck with the full memory barrier > > semantics like that. But I still find it easier to think of it like a > > RELEASE that pairs with the ACQUIRE of waking up, such that the task > > is guaranteed to observe it's own wake condition. > > > > And maybe that is the thing I'm missing here. These comments only state > > that it does in fact imply a full memory barrier, but do not explain > > why, should it? > > "code (people) is relying on it" is really the only "why" I can think > of. With this patch, that same/SB pattern is also reported in memory > -barriers.txt. Other ideas? So I'm not actually sure how many people rely on the RCsc transitive smp_mb() here. People certainly rely on the RELEASE semantics, and the code itself requires the store->load ordering, together that gives us the smp_mb() because that's simply the only barrier we have. And looking at smp_mb__after_spinlock() again, we really only need the RCsc thing for rq->lock, not for the wakeups. The wakeups really only need that RCpc RELEASE + store->load thing (which we don't have). So yes, smp_mb(), however the below still makes more sense to me, or am I just being obtuse again? --- kernel/sched/core.c | 19 +++++++++++++------ 1 file changed, 13 insertions(+), 6 deletions(-) diff --git a/kernel/sched/core.c b/kernel/sched/core.c index a98d54cd5535..8374d01b2820 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -1879,7 +1879,9 @@ static void ttwu_queue(struct task_struct *p, int cpu, int wake_flags) * C) LOCK of the rq(c1)->lock scheduling in task * * Transitivity guarantees that B happens after A and C after B. - * Note: we only require RCpc transitivity. + * Note: we only require RCpc transitivity for these cases, + * but see smp_mb__after_spinlock() for why rq->lock is required + * to be RCsc. * Note: the CPU doing B need not be c0 or c1 * * Example: @@ -1944,13 +1946,14 @@ static void ttwu_queue(struct task_struct *p, int cpu, int wake_flags) * However; for wakeups there is a second guarantee we must provide, namely we * must observe the state that lead to our wakeup. That is, not only must our * task observe its own prior state, it must also observe the stores prior to - * its wakeup. + * its wakeup, see set_current_state(). * * This means that any means of doing remote wakeups must order the CPU doing - * the wakeup against the CPU the task is going to end up running on. This, - * however, is already required for the regular Program-Order guarantee above, - * since the waking CPU is the one issueing the ACQUIRE (smp_cond_load_acquire). - * + * the wakeup against the CPU the task is going to end up running on. This + * means two things: firstly that try_to_wake_up() must (at least) imply a + * RELEASE (smp_mb__after_spinlock()), and secondly, as is already required + * for the regular Program-Order guarantee above, that waking implies an ACQUIRE + * (see smp_cond_load_acquire() above). */ /** @@ -1966,6 +1969,10 @@ static void ttwu_queue(struct task_struct *p, int cpu, int wake_flags) * Atomic against schedule() which would dequeue a task, also see * set_current_state(). * + * Implies at least a RELEASE such that the waking task is guaranteed to + * observe the stores to the wait-condition; see set_task_state() and the + * Program-Order constraints. + * * Return: %true if @p->state changes (an actual wakeup was done), * %false otherwise. */