Received: by 2002:a05:6a10:16a7:0:0:0:0 with SMTP id gp39csp3721979pxb; Tue, 17 Nov 2020 01:32:37 -0800 (PST) X-Google-Smtp-Source: ABdhPJyRaMdvxBcXaT3V/ysbDQzTQ5QQBUwgxYwgUgvehohkIQcQqWL+BvJ9dw2VCXdoZqb8tbZD X-Received: by 2002:a17:906:6414:: with SMTP id d20mr17879462ejm.82.1605605557283; Tue, 17 Nov 2020 01:32:37 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1605605557; cv=none; d=google.com; s=arc-20160816; b=S8f3CTt+89yK+pyyuwZMun2UAIrVpL5c5q+TE1g7Oke7BMX6P1J7kkPgFjtoTysHnh NWlybbUJv5Q5g1gW8vpFhD140Dq9AFPPL4nCRiLgRr+TWuNM42sJV+31EUui31qHXAUW YsRReodF3JyEq4hvMDvz8RqwffEnoZh+TdYe87oih+NS3aPhuAL45R2ulzmZ1KbfmUOc GGfC94QdJHG5SKQqB1PrYdff+G5OeCpCDRSMOX2i0ehjClQJzX1Tjm5nA9sRZ41dVu8P XVizqYEjLDrRLktAkfQ50oyJT8RjVMraED2ZaqjB7IVVMkxqiBLR7mBlo3HN9EhR9zPI IW3Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=/mGspyDrUE9cLixCRZMRH+ZDRs8K51mv90gNUMNpFR4=; b=FPa+9uXCJV7oubSFkRCEFk65kt73unY0WzTvT63cBUdFIa1sqVccgi9CU4qGzA6Lvv qGZ7k8Mzr1MOOpGcpR7Xt25hQcFe8Xg4z254LihIQB639Py2KVjy86VkF811GKsc8zzP ECWNs6+Yz0DIHcNjvdpeDyW6m7P3otyWXDt5LO3Ncjj8d2pseeQTqbY6Mgr2T5vb4j03 0pfuPuUzS9va+c6YWV5LFTKGIk9V7/awqUeGMnkSVsUq16wMELfBtUiyrictxGz3UDwn g05kVd9OieTDBPxmAXC2DLtgScfK7BfomeQhS4Rt2EiZRIhkwAQz9cRIbMQFxISWxgd9 qXTA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@infradead.org header.s=casper.20170209 header.b=jy4Z5odY; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id be17si9311901edb.595.2020.11.17.01.32.13; Tue, 17 Nov 2020 01:32:37 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@infradead.org header.s=casper.20170209 header.b=jy4Z5odY; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727218AbgKQJ3r (ORCPT + 99 others); Tue, 17 Nov 2020 04:29:47 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:51964 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725747AbgKQJ3q (ORCPT ); Tue, 17 Nov 2020 04:29:46 -0500 Received: from casper.infradead.org (casper.infradead.org [IPv6:2001:8b0:10b:1236::1]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id A7F2AC0613CF for ; Tue, 17 Nov 2020 01:29:46 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=casper.20170209; h=In-Reply-To:Content-Type:MIME-Version: References:Message-ID:Subject:Cc:To:From:Date:Sender:Reply-To: Content-Transfer-Encoding:Content-ID:Content-Description; bh=/mGspyDrUE9cLixCRZMRH+ZDRs8K51mv90gNUMNpFR4=; b=jy4Z5odYqv8DhSI+/PSRBa85wX rvInEFPfcMMTaM1e1C08rsjC41c/UEC9HduBUpE6gaWftm/z/TA+n/FHhrhHodeU/cWudoJ/QvVVG nkq0+apjqw1kxT1pdIVkUbubXbspMRvQEqq3x5oOTfAYXd0ciPNxfktwTL0JB1xBMxITWMBYrcBLG 1FC6LnVk+pW5hgXWyiEnnO9ZVqNJfcAa1KTUizN6W0xIFikLZyW4YJu+S0uSgcLEe5KUYwNz9QKOH yjo3p0X1W9cNj58tTt1WnOvd1d8uwyOPIjR68IbIgyM9Cxh2ncu3h1spaIz0WArnevv8WmC213+hB 8/4BvdOg==; Received: from j217100.upc-j.chello.nl ([24.132.217.100] helo=noisy.programming.kicks-ass.net) by casper.infradead.org with esmtpsa (Exim 4.92.3 #3 (Red Hat Linux)) id 1kexIx-0005X4-6i; Tue, 17 Nov 2020 09:29:39 +0000 Received: from hirez.programming.kicks-ass.net (hirez.programming.kicks-ass.net [192.168.1.225]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (Client did not present a certificate) by noisy.programming.kicks-ass.net (Postfix) with ESMTPS id 26D2D3019CE; Tue, 17 Nov 2020 10:29:36 +0100 (CET) Received: by hirez.programming.kicks-ass.net (Postfix, from userid 1000) id 12D382012638E; Tue, 17 Nov 2020 10:29:36 +0100 (CET) Date: Tue, 17 Nov 2020 10:29:36 +0100 From: Peter Zijlstra To: Will Deacon Cc: Mel Gorman , Davidlohr Bueso , linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org Subject: Re: [PATCH] sched: Fix data-race in wakeup Message-ID: <20201117092936.GA3121406@hirez.programming.kicks-ass.net> References: <20201116091054.GL3371@techsingularity.net> <20201116131102.GA29992@willie-the-truck> <20201116133721.GQ3371@techsingularity.net> <20201116142005.GE3121392@hirez.programming.kicks-ass.net> <20201116193149.GW3371@techsingularity.net> <20201117083016.GK3121392@hirez.programming.kicks-ass.net> <20201117091545.GA31837@willie-the-truck> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20201117091545.GA31837@willie-the-truck> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Nov 17, 2020 at 09:15:46AM +0000, Will Deacon wrote: > On Tue, Nov 17, 2020 at 09:30:16AM +0100, Peter Zijlstra wrote: > > Subject: sched: Fix data-race in wakeup > > From: Peter Zijlstra > > Date: Tue Nov 17 09:08:41 CET 2020 > > > > Mel reported that on some ARM64 platforms loadavg goes bananas and > > tracked it down to the following data race: > > > > CPU0 CPU1 > > > > schedule() > > prev->sched_contributes_to_load = X; > > deactivate_task(prev); > > > > try_to_wake_up() > > if (p->on_rq &&) // false > > if (smp_load_acquire(&p->on_cpu) && // true > > ttwu_queue_wakelist()) > > p->sched_remote_wakeup = Y; > > > > smp_store_release(prev->on_cpu, 0); > > (nit: I suggested this race over at [1] ;) Ah, I'll ammend and get you a Debugged-by line or something ;-) > > where both p->sched_contributes_to_load and p->sched_remote_wakeup are > > in the same word, and thus the stores X and Y race (and can clobber > > one another's data). > > > > Whereas prior to commit c6e7bd7afaeb ("sched/core: Optimize ttwu() > > spinning on p->on_cpu") the p->on_cpu handoff serialized access to > > p->sched_remote_wakeup (just as it still does with > > p->sched_contributes_to_load) that commit broke that by calling > > ttwu_queue_wakelist() with p->on_cpu != 0. > > > > However, due to > > > > p->XXX ttwu() > > schedule() if (p->on_rq && ...) // false > > smp_mb__after_spinlock() if (smp_load_acquire(&p->on_cpu) && > > deactivate_task() ttwu_queue_wakelist()) > > p->on_rq = 0; p->sched_remote_wakeup = X; > > > > We can be sure any 'current' store is complete and 'current' is > > guaranteed asleep. Therefore we can move p->sched_remote_wakeup into > > the current flags word. > > > > Note: while the observed failure was loadavg accounting gone wrong due > > to ttwu() cobbering p->sched_contributes_to_load, the reverse problem > > is also possible where schedule() clobbers p->sched_remote_wakeup, > > this could result in enqueue_entity() wrecking ->vruntime and causing > > scheduling artifacts. > > > > Fixes: c6e7bd7afaeb ("sched/core: Optimize ttwu() spinning on p->on_cpu") > > Reported-by: Mel Gorman > > Signed-off-by: Peter Zijlstra (Intel) > > --- > > include/linux/sched.h | 13 ++++++++++++- > > 1 file changed, 12 insertions(+), 1 deletion(-) > > > > --- a/include/linux/sched.h > > +++ b/include/linux/sched.h > > @@ -775,7 +775,6 @@ struct task_struct { > > unsigned sched_reset_on_fork:1; > > unsigned sched_contributes_to_load:1; > > unsigned sched_migrated:1; > > - unsigned sched_remote_wakeup:1; > > #ifdef CONFIG_PSI > > unsigned sched_psi_wake_requeue:1; > > #endif > > @@ -785,6 +784,18 @@ struct task_struct { > > > > /* Unserialized, strictly 'current' */ > > > > + /* > > + * p->in_iowait = 1; ttwu() > > + * schedule() if (p->on_rq && ..) // false > > + * smp_mb__after_spinlock(); if (smp_load_acquire(&p->on_cpu) && //true > > + * deactivate_task() ttwu_queue_wakelist()) > > + * p->on_rq = 0; p->sched_remote_wakeup = X; > > + * > > + * Guarantees all stores of 'current' are visible before > > + * ->sched_remote_wakeup gets used. > > I'm still not sure this is particularly clear -- don't we want to highlight > that the store of p->on_rq is unordered wrt the update to > p->sched_contributes_to_load() in deactivate_task()? I can explicitly call that out I suppose. > I dislike bitfields with a passion, but the fix looks good: I don't particularly hate them, they're just a flag field with names on (in this case). > Acked-by: Will Deacon Thanks! > Now the million dollar question is why KCSAN hasn't run into this. Hrmph. kernel/sched/Makefile:KCSAN_SANITIZE := n might have something to do with that, I suppose.