Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753034AbaBJWlv (ORCPT ); Mon, 10 Feb 2014 17:41:51 -0500 Received: from mail-pb0-f42.google.com ([209.85.160.42]:42803 "EHLO mail-pb0-f42.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752111AbaBJWlu (ORCPT ); Mon, 10 Feb 2014 17:41:50 -0500 Date: Mon, 10 Feb 2014 14:41:45 -0800 From: Kent Overstreet To: Jens Axboe Cc: Christoph Hellwig , Alexander Gordeev , Shaohua Li , linux-kernel@vger.kernel.org Subject: Re: [patch 1/2]percpu_ida: fix a live lock Message-ID: <20140210224145.GB2362@kmo> References: <20131231033827.GA31994@kernel.org> <20140104210804.GA24199@kmo-pixel> <20140105131300.GB4186@kernel.org> <20140106204641.GB9037@kmo> <52CB1783.4050205@kernel.dk> <20140106214726.GD9037@kmo> <20140209155006.GA16149@dhcp-26-207.brq.redhat.com> <20140210103211.GA28396@infradead.org> <52F8FDA7.7070809@kernel.dk> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <52F8FDA7.7070809@kernel.dk> User-Agent: Mutt/1.5.21 (2010-09-15) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Feb 10, 2014 at 09:26:15AM -0700, Jens Axboe wrote: > > > On 02/10/2014 03:32 AM, Christoph Hellwig wrote: > >On Sun, Feb 09, 2014 at 04:50:07PM +0100, Alexander Gordeev wrote: > >>Yeah, that was my first thought when I posted "percpu_ida: Allow variable > >>maximum number of cached tags" patch some few months ago. But I am back- > >>pedalling as it does not appear solves the fundamental problem - what is the > >>best threshold? > >> > >>May be we can walk off with a per-cpu timeout that flushes batch nr of tags > >>from local caches to the pool? Each local allocation would restart the timer, > >>but once allocation requests stopped coming on a CPU the tags would not gather > >>dust in local caches. > > > >We'll defintively need a fix to be able to allow the whole tag space. > > Certainly. The current situation of effectively only allowing half > the tags (if spread) is pretty crappy with (by far) most hardware. > > >For large numbers of tags per device the flush might work, but for > >devices with low number of tags we need something more efficient. The > >case of less tags than CPUs isn't that unusual either and we probably > >want to switch to an allocator without per cpu allocations for them to > >avoid all this. E.g. for many ATA devices we just have a single tag, > >and many scsi drivers also only want single digit outstanding commands > >per LUN. > > Even for cases where you have as many (or more) CPUs than tags, > per-cpu allocation is not necessarily a bad idea. It's a rare case > where you have all the CPUs touching the device at the same time, > after all. You do still need to have enough tags to shard across the number of cpus _currently_ touching the device. I think I'm with Christoph here, I'm not sure how percpu tag allocation would be helpful when we have single digits/low double digits of tags available. I would expect that in that case we're better off with just a well implemented atomic bit vector and waitlist. However, I don't know where the crossover point is and I think Jens has done by far the most and most relevant benchmarking here. How about we just make the number of tags that are allowed to be stranded an explicit parameter (somehow) - then it can be up to device drivers to do something sensible with it. Half is probably an ideal default for devices where that works, but this way more constrained devices will be able to futz with it however they want. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/