Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1760284Ab1FXBTb (ORCPT ); Thu, 23 Jun 2011 21:19:31 -0400 Received: from claw.goop.org ([74.207.240.146]:41617 "EHLO claw.goop.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1760175Ab1FXBT0 (ORCPT ); Thu, 23 Jun 2011 21:19:26 -0400 From: Jeremy Fitzhardinge To: Peter Zijlstra Cc: Linus Torvalds , "H. Peter Anvin" , Ingo Molnar , the arch/x86 maintainers , Linux Kernel Mailing List , Nick Piggin , Jeremy Fitzhardinge Subject: [PATCH 5/8] x86/ticketlock: make __ticket_spin_lock common Date: Thu, 23 Jun 2011 18:19:17 -0700 Message-Id: <11b6893703912054024bdb60251ed944f31cbafe.1308878118.git.jeremy.fitzhardinge@citrix.com> X-Mailer: git-send-email 1.7.5.4 In-Reply-To: References: In-Reply-To: <05552e1c97739f589cb76c20b5a6565ccd506636.1308878118.git.jeremy.fitzhardinge@citrix.com> References: <05552e1c97739f589cb76c20b5a6565ccd506636.1308878118.git.jeremy.fitzhardinge@citrix.com> Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2993 Lines: 87 From: Jeremy Fitzhardinge Aside from the particular form of the xadd instruction, they're identical. So factor out the xadd and use common code for the rest. Signed-off-by: Jeremy Fitzhardinge --- arch/x86/include/asm/spinlock.h | 42 ++++++++++++++++++-------------------- 1 files changed, 20 insertions(+), 22 deletions(-) diff --git a/arch/x86/include/asm/spinlock.h b/arch/x86/include/asm/spinlock.h index 694b8a0..0b8d2b2 100644 --- a/arch/x86/include/asm/spinlock.h +++ b/arch/x86/include/asm/spinlock.h @@ -67,13 +67,27 @@ static __always_inline void __ticket_unlock_release(struct arch_spinlock *lock) * save some instructions and make the code more elegant. There really isn't * much between them in performance though, especially as locks are out of line. */ -#if (NR_CPUS < 256) -static __always_inline void __ticket_spin_lock(arch_spinlock_t *lock) +static __always_inline struct __raw_tickets __ticket_spin_claim(struct arch_spinlock *lock) { - register struct __raw_tickets inc = { .tail = 1 }; + register struct __raw_tickets tickets = { .tail = 1 }; + + if (sizeof(lock->tickets.head) == sizeof(u8)) + asm volatile (LOCK_PREFIX "xaddw %w0, %1\n" + : "+r" (tickets), "+m" (lock->tickets) + : : "memory", "cc"); + else + asm volatile (LOCK_PREFIX "xaddl %0, %1\n" + : "+r" (tickets), "+m" (lock->tickets) + : : "memory", "cc"); - asm volatile (LOCK_PREFIX "xaddw %w0, %1\n" - : "+r" (inc), "+m" (lock->tickets) : : "memory", "cc"); + return tickets; +} + +static __always_inline void __ticket_spin_lock(struct arch_spinlock *lock) +{ + register struct __raw_tickets inc; + + inc = __ticket_spin_claim(lock); for (;;) { if (inc.head == inc.tail) @@ -84,6 +98,7 @@ static __always_inline void __ticket_spin_lock(arch_spinlock_t *lock) out: barrier(); /* make sure nothing creeps before the lock is taken */ } +#if (NR_CPUS < 256) static __always_inline int __ticket_spin_trylock(arch_spinlock_t *lock) { unsigned int tmp, new; @@ -103,23 +118,6 @@ static __always_inline int __ticket_spin_trylock(arch_spinlock_t *lock) return tmp; } #else -static __always_inline void __ticket_spin_lock(arch_spinlock_t *lock) -{ - register struct __raw_tickets inc = { .tail = 1 }; - - asm volatile(LOCK_PREFIX "xaddl %0, %1\n\t" - : "+r" (inc), "+m" (lock->tickets) - : : "memory", "cc"); - - for (;;) { - if (inc.head == inc.tail) - goto out; - cpu_relax(); - inc.head = ACCESS_ONCE(lock->tickets.head); - } -out: barrier(); /* make sure nothing creeps before the lock is taken */ -} - static __always_inline int __ticket_spin_trylock(arch_spinlock_t *lock) { unsigned tmp; -- 1.7.5.4 -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/