Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752631AbaBGRzL (ORCPT ); Fri, 7 Feb 2014 12:55:11 -0500 Received: from verein.lst.de ([213.95.11.211]:44773 "EHLO newverein.lst.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751680AbaBGRzK (ORCPT ); Fri, 7 Feb 2014 12:55:10 -0500 Date: Fri, 7 Feb 2014 18:55:05 +0100 From: Torsten Duwe To: Peter Zijlstra Cc: Benjamin Herrenschmidt , Paul Mackerras , Anton Blanchard , "Paul E. McKenney" , Scott Wood , Tom Musta , Ingo Molnar , linux-kernel@vger.kernel.org, linuxppc-dev@lists.ozlabs.org Subject: Re: [PATCH v2] powerpc ticket locks Message-ID: <20140207175505.GE2107@lst.de> References: <20140207165801.GC2107@lst.de> <20140207171224.GR5002@laptop.programming.kicks-ass.net> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20140207171224.GR5002@laptop.programming.kicks-ass.net> User-Agent: Mutt/1.5.17 (2007-11-01) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Feb 07, 2014 at 06:12:24PM +0100, Peter Zijlstra wrote: > On Fri, Feb 07, 2014 at 05:58:01PM +0100, Torsten Duwe wrote: > > +static __always_inline void arch_spin_lock(arch_spinlock_t *lock) > > { > > + register struct __raw_tickets old, tmp, > > + inc = { .tail = TICKET_LOCK_INC }; > > + > > CLEAR_IO_SYNC; > > + __asm__ __volatile__( > > +"1: lwarx %0,0,%4 # arch_spin_lock\n" > > +" add %1,%3,%0\n" > > + PPC405_ERR77(0, "%4") > > +" stwcx. %1,0,%4\n" > > +" bne- 1b" > > + : "=&r" (old), "=&r" (tmp), "+m" (lock->tickets) > > + : "r" (inc), "r" (&lock->tickets) > > + : "cc"); > > + > > + if (likely(old.head == old.tail)) > > + goto out; > > I would have expected an lwsync someplace hereabouts. Let me reconsider this. The v1 code worked on an 8 core, maybe I didn't beat it enough. > > static inline void arch_spin_unlock(arch_spinlock_t *lock) > > { > > + arch_spinlock_t old, new; > > + > > +#if defined(CONFIG_PPC_SPLPAR) > > + lock->holder = 0; > > +#endif > > + do { > > + old.tickets = ACCESS_ONCE(lock->tickets); > > + new.tickets.head = old.tickets.head + TICKET_LOCK_INC; > > + new.tickets.tail = old.tickets.tail; > > + } while (unlikely(__arch_spin_cmpxchg_eq(lock, > > + old.head_tail, > > + new.head_tail))); > > SYNC_IO; > > __asm__ __volatile__("# arch_spin_unlock\n\t" > > PPC_RELEASE_BARRIER: : :"memory"); > > Doens't your cmpxchg_eq not already imply a lwsync? Right. > > - lock->slock = 0; > > } > > I'm still failing to see why you need an ll/sc pair for unlock. Like so: static inline void arch_spin_unlock(arch_spinlock_t *lock) { arch_spinlock_t tmp; #if defined(CONFIG_PPC_SPLPAR) lock->holder = 0; #endif tmp.tickets = ACCESS_ONCE(lock->tickets); tmp.tickets.head += TICKET_LOCK_INC; lock->tickets.head = tmp.tickets.head; SYNC_IO; __asm__ __volatile__("# arch_spin_unlock\n\t" PPC_RELEASE_BARRIER: : :"memory"); } ? I'll wrap it all up next week. I only wanted to post an updated v2 with the agreed-upon changes for BenH. Thanks so far! Torsten -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/