Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754182AbcDTPGZ (ORCPT ); Wed, 20 Apr 2016 11:06:25 -0400 Received: from e23smtp09.au.ibm.com ([202.81.31.142]:56373 "EHLO e23smtp09.au.ibm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1753505AbcDTPGY (ORCPT ); Wed, 20 Apr 2016 11:06:24 -0400 X-IBM-Helo: d23dlp03.au.ibm.com X-IBM-MailFrom: xinhui@linux.vnet.ibm.com X-IBM-RcptTo: linux-kernel@vger.kernel.org Message-ID: <57179AB1.3080403@linux.vnet.ibm.com> Date: Wed, 20 Apr 2016 23:05:21 +0800 From: Pan Xinhui User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:31.0) Gecko/20100101 Thunderbird/31.8.0 MIME-Version: 1.0 To: Peter Zijlstra CC: Waiman Long , Ingo Molnar , linux-kernel@vger.kernel.org, Scott J Norton , Douglas Hatch Subject: Re: [PATCH v2] locking/pvqspinlock: Add lock holder CPU argument to pv_wait() References: <1460659318-53312-1-git-send-email-Waiman.Long@hpe.com> <20160420120805.GB3408@twins.programming.kicks-ass.net> <57178EED.1060207@linux.vnet.ibm.com> <20160420141819.GD3430@twins.programming.kicks-ass.net> In-Reply-To: <20160420141819.GD3430@twins.programming.kicks-ass.net> Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-TM-AS-MML: disable X-Content-Scanned: Fidelis XPS MAILER x-cbid: 16042015-0033-0000-0000-0000059A7674 Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 1662 Lines: 50 On 2016年04月20日 22:18, Peter Zijlstra wrote: > On Wed, Apr 20, 2016 at 10:15:09PM +0800, Pan Xinhui wrote: >>>> +static struct pv_node *pv_lookup_hash(struct qspinlock *lock) >>>> +{ >>>> + unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits); >>>> + struct pv_hash_entry *he; >>>> + >>>> + for_each_hash_entry(he, offset, hash) { >>>> + struct qspinlock *l = READ_ONCE(he->lock); >>>> + >>>> + if (l == lock) >>> >>> The other loop writes: >>> >>> if (READ_ONCE(he->lock) == lock) >>> >> Maybe because we check l is NULL or not later. So save one load. > > Ah duh, yes. > >>>> + return READ_ONCE(he->node); >>>> + /* >>>> + * Presence of an empty slot signal the end of search. We >>>> + * may miss the entry, but that will limit the amount of >>>> + * time doing the search when the desired entry isn't there. >>>> + */ >>>> + else if (!l) >>>> + break; >>> >>> That 'else' is entirely pointless. Also, why isn't this: return NULL; >>> >>>> + } >>>> + return NULL; >>> >>> and this BUG() ? >>> >> It's not a bug, the lock might not be stored in the hashtable. in unlock function, we will unhash the lock, then what will happen is: > > It should be if the above becomes a return NULL, no? > no, the lock might not be there, even if we search the whole hashtable. Only pv_kick_node and pv_wait_head_or_lock will hash the lock. if both vcpu's state is vcpu_running, who will hash the lock on behalf of us? Can pv_wait return without anyone kicking it? If yes, then this not a bug. > If we can iterate the _entire_ hashtable, this lookup can be immensely > expensive and we should not be doing it inside of a wait-loop. >