Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1755009AbbBOQTx (ORCPT ); Sun, 15 Feb 2015 11:19:53 -0500 Received: from mx1.redhat.com ([209.132.183.28]:46417 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754550AbbBOQTw (ORCPT ); Sun, 15 Feb 2015 11:19:52 -0500 Date: Sun, 15 Feb 2015 17:17:33 +0100 From: Oleg Nesterov To: Raghavendra K T Cc: tglx@linutronix.de, mingo@redhat.com, hpa@zytor.com, peterz@infradead.org, torvalds@linux-foundation.org, konrad.wilk@oracle.com, pbonzini@redhat.com, paulmck@linux.vnet.ibm.com, waiman.long@hp.com, davej@redhat.com, x86@kernel.org, jeremy@goop.org, paul.gortmaker@windriver.com, ak@linux.intel.com, jasowang@redhat.com, linux-kernel@vger.kernel.org, kvm@vger.kernel.org, virtualization@lists.linux-foundation.org, xen-devel@lists.xenproject.org, riel@redhat.com, borntraeger@de.ibm.com, akpm@linux-foundation.org, a.ryabinin@samsung.com, sasha.levin@oracle.com, dave@stgolabs.net Subject: Re: [PATCH V5] x86 spinlock: Fix memory corruption on completing completions Message-ID: <20150215161733.GB27608@redhat.com> References: <1423979744-18320-1-git-send-email-raghavendra.kt@linux.vnet.ibm.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <1423979744-18320-1-git-send-email-raghavendra.kt@linux.vnet.ibm.com> User-Agent: Mutt/1.5.18 (2008-05-17) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 1083 Lines: 33 Well, I regret I mentioned the lack of barrier after enter_slowpath ;) On 02/15, Raghavendra K T wrote: > > @@ -46,7 +46,8 @@ static __always_inline bool static_key_false(struct static_key *key); > > static inline void __ticket_enter_slowpath(arch_spinlock_t *lock) > { > - set_bit(0, (volatile unsigned long *)&lock->tickets.tail); > + set_bit(0, (volatile unsigned long *)&lock->tickets.head); > + barrier(); > } Because this barrier() looks really confusing. Firsty, it is equally unneeded on x86. At the same time, it can not help. We need a memory barrier() between set_bit(SLOWPATH) and READ_ONCE(head) to avoid the race with spin_unlock(). So I think you should replace it with smp_mb__after_atomic() or remove it. Other than that I believe this version is correct. So I won't insist, this is cosmetic after all. Oleg. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/