Received: by 2002:ac0:946b:0:0:0:0:0 with SMTP id j40csp749352imj; Thu, 7 Feb 2019 11:12:39 -0800 (PST) X-Google-Smtp-Source: AHgI3IbzqS+e/sVkqa37clnfPagIWhosHkaKWF7QUdBvI0//j/oMPMhFTVhUQg6WpSzxGB7NB2Z2 X-Received: by 2002:a17:902:2e01:: with SMTP id q1mr17756543plb.97.1549566759755; Thu, 07 Feb 2019 11:12:39 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1549566759; cv=none; d=google.com; s=arc-20160816; b=OTbUI+6XqLE5w4UfJXfZlsQ11myKNd2fNxWHqP+xgx1rQllYaMpFxIVuX0YwktewAf TYORE+kkMjfWmjq7prSnwdPRwmYGcqGd/Gi/8ThVRpcpnxVIXRg2EI3vFBksM8W4wz0L m4hoG62Uqn+inbBd/cfO/IkmBFOW02su2Qe5ORijrTKDdL//Fcof1ifawOZkhk5hwu4W S6y8lZGn0x9BXPOVAvdWKtxrGOgO8zvsPq9z6RAmux/HZtGumyk04FXh6CfmgM7lfQhH 3hl8yTlFLgFdHywP8Sy52eFqXr82dVtiCVGRrXFvYzFZ4IvvQNTc8IJvv1CWJ3dIV2pE eOeg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from; bh=eHhQhinaswkSZ3tDxgVYX2eml2+lVrUAEK+ZSfFmm8k=; b=taju8SfOxi0XXH6Dwbii/nQmsOQ/+0yMBJAKMBU5vJUFPvsrl589Vt+V5WHkEoFR9X pDXRAjQvX8pRILxbKoJK0b5CjFZ3S5jdupG1qJJPg+wlBPa0EpfUebunLg0XMoF9YV0p jq79Vo7nCqix40WfKzyyhSe+l+sQsH/OP6+mF+BvGQyDoXaAXwlz9GBdw9aHY7aOz2UF Hrr5wrqZ1k1G2u496gXACONXKQXqHTG94K1Wvq9ffleToXs+LMqH14WtsZ6oUYj3I+mk jd+oo1fDGE0cwqKozwDbqIFrVt1qEZELQIqDwndbJiddhoftXHwkv3pRYcxSJB3OKlon FIdw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id k22si741532pls.14.2019.02.07.11.12.22; Thu, 07 Feb 2019 11:12:39 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727645AbfBGTKa (ORCPT + 99 others); Thu, 7 Feb 2019 14:10:30 -0500 Received: from mx1.redhat.com ([209.132.183.28]:49796 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727612AbfBGTKY (ORCPT ); Thu, 7 Feb 2019 14:10:24 -0500 Received: from smtp.corp.redhat.com (int-mx02.intmail.prod.int.phx2.redhat.com [10.5.11.12]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id BB5252551; Thu, 7 Feb 2019 19:10:23 +0000 (UTC) Received: from llong.com (dhcp-17-35.bos.redhat.com [10.18.17.35]) by smtp.corp.redhat.com (Postfix) with ESMTP id 9488961146; Thu, 7 Feb 2019 19:10:21 +0000 (UTC) From: Waiman Long To: Peter Zijlstra , Ingo Molnar , Will Deacon , Thomas Gleixner Cc: linux-kernel@vger.kernel.org, linux-alpha@vger.kernel.org, linux-arm-kernel@lists.infradead.org, linux-hexagon@vger.kernel.org, linux-ia64@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-sh@vger.kernel.org, sparclinux@vger.kernel.org, linux-xtensa@linux-xtensa.org, linux-arch@vger.kernel.org, x86@kernel.org, Arnd Bergmann , Borislav Petkov , "H. Peter Anvin" , Davidlohr Bueso , Linus Torvalds , Andrew Morton , Tim Chen , Waiman Long Subject: [PATCH-tip 21/22] locking/rwsem: Wake up all readers in wait queue Date: Thu, 7 Feb 2019 14:07:25 -0500 Message-Id: <1549566446-27967-22-git-send-email-longman@redhat.com> In-Reply-To: <1549566446-27967-1-git-send-email-longman@redhat.com> References: <1549566446-27967-1-git-send-email-longman@redhat.com> X-Scanned-By: MIMEDefang 2.79 on 10.5.11.12 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.30]); Thu, 07 Feb 2019 19:10:24 +0000 (UTC) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org When the front of the wait queue is a reader, other readers immediately following the first reader will also be woken up at the same time. However, if there is a writer in between. Those readers behind the writer will not be woken up. Because of optimistic spinning, the lock acquisition order is not FIFO anyway. The lock handoff mechanism will ensure that lock starvation will not happen. Assuming that the lock hold times of the other readers still in the queue will be about the same as the readers that are being woken up, there is really not much additional cost other than the additional latency due to the wakeup of additional tasks by the waker. Therefore all the readers in the queue are woken up when the first waiter is a reader to improve reader throughput. With a locking microbenchmark running on 5.0 based kernel, the total locking rates (in kops/s) of the benchmark on a 4-socket 56-core x86-64 system with equal numbers of readers and writers before all the reader spining patches, before this patch and after this patch were as follows: # of Threads Pre-rspin Pre-Patch Post-patch ------------ --------- --------- ---------- 2 1,926 8,057 7,397 4 1,391 7,680 6,161 8 716 7,284 6,405 16 618 6,542 6,768 32 501 1,449 6,550 64 61 480 5,548 112 75 769 5,216 At low contention level, there is a slight drop in performance. At high contention level, however, this patch gives a big performance boost. Signed-off-by: Waiman Long --- kernel/locking/rwsem-xadd.c | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/kernel/locking/rwsem-xadd.c b/kernel/locking/rwsem-xadd.c index 3beb942..3cf2e84 100644 --- a/kernel/locking/rwsem-xadd.c +++ b/kernel/locking/rwsem-xadd.c @@ -180,16 +180,16 @@ static void __rwsem_mark_wake(struct rw_semaphore *sem, } /* - * Grant an infinite number of read locks to the readers at the front - * of the queue. We know that woken will be at least 1 as we accounted - * for above. Note we increment the 'active part' of the count by the + * Grant an infinite number of read locks to all the readers in the + * queue. We know that woken will be at least 1 as we accounted for + * above. Note we increment the 'active part' of the count by the * number of readers before waking any processes up. */ list_for_each_entry_safe(waiter, tmp, &sem->wait_list, list) { struct task_struct *tsk; if (waiter->type == RWSEM_WAITING_FOR_WRITE) - break; + continue; woken++; tsk = waiter->task; -- 1.8.3.1