Received: by 2002:ac0:a5a7:0:0:0:0:0 with SMTP id m36-v6csp1144315imm; Wed, 8 Aug 2018 11:31:05 -0700 (PDT) X-Google-Smtp-Source: AA+uWPxoPhjaJwkQ13nGLblKV+2LDUS1BfJq7TWgLxaXpYw4wL4SyjhFLqTroKp9OZEbgWQfs52H X-Received: by 2002:a63:b00f:: with SMTP id h15-v6mr3638773pgf.442.1533753065516; Wed, 08 Aug 2018 11:31:05 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1533753065; cv=none; d=google.com; s=arc-20160816; b=dAsZgdA4VozL+28bg+0eTukhMH5mvoWJ5hSh5GFUSQyjPe7WMpub9Et0f/y+iNDHZV G9ApBJor2b2BXcEnPxvn1ZejKTk/0Gusq0FBkQzurw+LMaJ4jqoholPXJpo1tj1o1d1V 58nNMTm81cWAnitC9hu/eIiLxQ5pQFcw0lxw/4PaZJpHQEGf3F/2OpLBDbWQ3lfLuyez jYJGtRwbB9a1jYx1V8sQmJxCpBPMDgF1TGjUn8w1IOX/W/dh/hV42Zri7Z66glVW9f1T U4bWv70rlIZLb0PHSgQQpqIJVnkPRbwZ2sbmTsBhreu2vaC7sAJQ1VCaCWIuGfGXgOYF 1h3w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:arc-authentication-results; bh=qRcfMCtfwyQgesB3VAdTLOtk6ZvQg+TeoSOryVQ55r0=; b=OHaCfQvTBbvMeejGp/hrjpJXteub/D64gcNE4gqRTFrOar0eI17BPQNlb4VRrTZZWB LVojW12g+2P1oHb0g83avm8UdT1lPdmYd7uzaiOi3NX+aK+1y6aEAzwxeHpwOJNJtd3z sqVNs90enjxKUlcBvuor8B3a8OiQxK7ARArXQHS1RvklZDOpEt2SbhjL0T582La1L/dH d3VZ9x72sQibfbrPdJ5d7Kd9bPQ0EEaubT7a20oxgEm2S39GEsH0WmXH5T9OjLJmcFc3 XYs8Bz13mA2UZv/5SBjnCVmnhJhli9JFyIIGWapOMi/mJoSPYyj2qUa9nIvee1xD9ilS gjrA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id p35-v6si4502893pgl.202.2018.08.08.11.30.49; Wed, 08 Aug 2018 11:31:05 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1729679AbeHHUuy (ORCPT + 99 others); Wed, 8 Aug 2018 16:50:54 -0400 Received: from fieldses.org ([173.255.197.46]:46912 "EHLO fieldses.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727295AbeHHUux (ORCPT ); Wed, 8 Aug 2018 16:50:53 -0400 Received: by fieldses.org (Postfix, from userid 2815) id 40A098BF; Wed, 8 Aug 2018 14:29:59 -0400 (EDT) Date: Wed, 8 Aug 2018 14:29:59 -0400 From: "J. Bruce Fields" To: Jeff Layton Cc: NeilBrown , Alexander Viro , linux-fsdevel@vger.kernel.org, linux-kernel@vger.kernel.org, Martin Wilck Subject: Re: [PATCH 0/4] locks: avoid thundering-herd wake-ups Message-ID: <20180808182959.GB23873@fieldses.org> References: <153369219467.12605.13472423449508444601.stgit@noble> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.5.21 (2010-09-15) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, Aug 08, 2018 at 12:47:22PM -0400, Jeff Layton wrote: > On Wed, 2018-08-08 at 11:51 +1000, NeilBrown wrote: > > If you have a many-core machine, and have many threads all wanting to > > briefly lock a give file (udev is known to do this), you can get quite > > poor performance. > > > > When one thread releases a lock, it wakes up all other threads that > > are waiting (classic thundering-herd) - one will get the lock and the > > others go to sleep. > > When you have few cores, this is not very noticeable: by the time the > > 4th or 5th thread gets enough CPU time to try to claim the lock, the > > earlier threads have claimed it, done what was needed, and released. > > With 50+ cores, the contention can easily be measured. > > > > This patchset creates a tree of pending lock request in which siblings > > don't conflict and each lock request does conflict with its parent. > > When a lock is released, only requests which don't conflict with each > > other a woken. > > > > Testing shows that lock-acquisitions-per-second is now fairly stable even > > as number of contending process goes to 1000. Without this patch, > > locks-per-second drops off steeply after a few 10s of processes. > > > > There is a small cost to this extra complexity. > > At 20 processes running a particular test on 72 cores, the lock > > acquisitions per second drops from 1.8 million to 1.4 million with > > this patch. For 100 processes, this patch still provides 1.4 million > > while without this patch there are about 700,000. > > > > NeilBrown > > > > --- > > > > NeilBrown (4): > > fs/locks: rename some lists and pointers. > > fs/locks: allow a lock request to block other requests. > > fs/locks: change all *_conflict() functions to return bool. > > fs/locks: create a tree of dependent requests. > > > > > > fs/cifs/file.c | 2 - > > fs/locks.c | 142 +++++++++++++++++++++++++-------------- > > include/linux/fs.h | 5 + > > include/trace/events/filelock.h | 16 ++-- > > 4 files changed, 103 insertions(+), 62 deletions(-) > > > > Nice work! I looked over this and I think it looks good. > > I made an attempt to fix this issue several years ago, but my method > sucked as it ended up penalizing the unlocking task too much. This is > much cleaner and should scale well overall, I think. I think I also took a crack at this at one point while I was at UM/CITI and never got anything I was happy with. Looks like good work! I remember one main obstacle that I felt like I never had a good benchmark.... How did you choose this workload and hardware? Was it in fact udev (booting a large machine?), or was there some other motivation? Not that I'm likely to do it any time soon, but could you share sufficient details for someone else to reproduce your results? --b.