Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1760565AbZAWESO (ORCPT ); Thu, 22 Jan 2009 23:18:14 -0500 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1755482AbZAWER7 (ORCPT ); Thu, 22 Jan 2009 23:17:59 -0500 Received: from mail.suse.de ([195.135.220.2]:51183 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755416AbZAWER6 (ORCPT ); Thu, 22 Jan 2009 23:17:58 -0500 Date: Fri, 23 Jan 2009 05:17:56 +0100 From: Nick Piggin To: Christoph Lameter Cc: Pekka Enberg , "Zhang, Yanmin" , Lin Ming , linux-mm@kvack.org, linux-kernel@vger.kernel.org, Andrew Morton , Linus Torvalds Subject: Re: [patch] SLQB slab allocator Message-ID: <20090123041756.GH20098@wotan.suse.de> References: <20090114142200.GB25401@wotan.suse.de> <84144f020901140645o68328e01ne0e10ace47555e19@mail.gmail.com> <20090114150900.GC25401@wotan.suse.de> <20090115060330.GB17810@wotan.suse.de> <20090116031940.GL17810@wotan.suse.de> <20090119054730.GA22584@wotan.suse.de> Mime-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.5.9i Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 3802 Lines: 85 On Wed, Jan 21, 2009 at 07:19:08PM -0500, Christoph Lameter wrote: > On Mon, 19 Jan 2009, Nick Piggin wrote: > > > The thing IMO you forget with all these doomsday scenarios about SGI's peta > > scale systems is that no matter what you do, you can't avoid the fact that > > computing is about locality. Even if you totally take the TLB out of the > > equation, you still have the small detail of other caches. Code that jumps > > all over that 1024 TB of memory with no locality is going to suck regardless > > of what the kernel ever does, due to physical limitations of hardware. > > Typically we traverse lists of objects that are in the same slab cache. Very often that is not the case. And the price you pay for that is that you have to drain and switch freelists whenever you encounter an object that is not on the same page. This gives your freelists a chaotic and unpredictable behaviour IMO in a running system where pages succumb to fragmentation so your freelist maximum sizes are limited. It also means you can lose track of cache hot objects when you switch to different "fast" pages. I don't consider this to be "queueing done right". > > > Sorry not at all. SLAB and SLQB queue objects from different pages in the > > > same queue. > > > > The last sentence is what I was replying to. Ie. "simplification of > > numa handling" does not follow from the SLUB implementation of per-page > > freelists. > > If all objects are from the same page then you need not check > the NUMA locality of any object on that queue. In SLAB and SLQB, all objects on the freelist are on the same node. So tell me how does same-page objects simplify numa handling? > > > As I sad it pins a single page in the per cpu page and uses that in a way > > > that you call a queue and I call a freelist. > > > > And you found you have to increase the size of your pages because you > > need bigger queues. (must we argue semantics? it is a list of free > > objects) > > Right. That may be the case and its a similar tuning to what SLAB does. SLAB and SLQB doesn't need bigger pages to do that. > > > SLAB and SLUB can have large quantities of objects in their queues that > > > each can keep a single page out of circulation if its the last > > > object in that page. This is per queue thing and you have at least two > > > > And if that were a problem, SLQB can easily be runtime tuned to keep no > > objects in its object lists. But as I said, queueing is good, so why > > would anybody want to get rid of it? > > Queing is sometimes good.... > > > Again, this doesn't really go anywhere while we disagree on the > > fundamental goodliness of queueing. This is just describing the > > implementation. > > I am not sure that you understand the fine points of queuing in slub. I am > not a fundamentalist: Queues are good if used the right way and as you say > SLUB has "queues" designed in a particular fashion that solves issus that > we had with SLAB queues. OK, and I juts don't think they solved all the problems and they added other worse ones. And if you would tell me what the problems are and how to reproduce them (or point to someone who might be able to help with reproducing them), then I'm confident that I can solve those problems in SLQB, which has fewer downsides than SLUB. At least I will try my best. So can you please give a better idea of the problems? "latency sensitive HPC applications" is about as much help to me solving that as telling you that "OLTP applications slow down" helps solve one of the problems in SLUB. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/