Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751093AbWHRIYd (ORCPT ); Fri, 18 Aug 2006 04:24:33 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1751111AbWHRIYd (ORCPT ); Fri, 18 Aug 2006 04:24:33 -0400 Received: from cantor2.suse.de ([195.135.220.15]:30870 "EHLO mx2.suse.de") by vger.kernel.org with ESMTP id S1751072AbWHRIYc (ORCPT ); Fri, 18 Aug 2006 04:24:32 -0400 From: Andi Kleen To: Christoph Lameter Subject: Re: [PATCH 1/1] network memory allocator. Date: Fri, 18 Aug 2006 11:29:14 +0200 User-Agent: KMail/1.9.1 Cc: Christoph Hellwig , Evgeniy Polyakov , Arnd Bergmann , David Miller , netdev@vger.kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org References: <20060814110359.GA27704@2ka.mipt.ru> <20060816142557.acccdfcf.ak@suse.de> In-Reply-To: MIME-Version: 1.0 Content-Type: text/plain; charset="iso-8859-1" Content-Transfer-Encoding: 7bit Content-Disposition: inline Message-Id: <200608181129.15075.ak@suse.de> Sender: linux-kernel-owner@vger.kernel.org X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2376 Lines: 48 On Friday 18 August 2006 04:25, Christoph Lameter wrote: > On Wed, 16 Aug 2006, Andi Kleen wrote: > > That's not true on all NUMA systems (that they have a slow interconnect) > > I think on x86-64 I would prefer if it was distributed evenly or maybe > > even on the CPU who is finally going to process it. > > > > -Andi "not all NUMA is an Altix" > > The Altix NUMA interconnect has the same speed as far as I can recall as > Hypertransport. It is the distance (real physical cable length) that > creates latencies for huge systems. Sadly the Hypertransport is designed > to stay on the motherboard. Hypertransport can only be said to be fast > because its only used for tinzy winzy systems of a few processors. Are > you saying that the design limitations of Hypertransport are an > advantage? Sorry, didn't want to state anything particular about advantages or disadvantages of different interconnects. I just wanted to say that there are a lot of NUMA systems out there which have a very low NUMA factor (for whatever reason, including them being quite small) and that they should be considered for NUMA optimizations too. So if you really want strict IO placement at least allow an easy way to turn it off even when CONFIG_NUMA is defined. BTW there are large x86-64 NUMA systems that don't use HyperTransport and have a varying NUMA factor, and also even HyperTransport based systems have a widely varying NUMA factor depending on machine size and hop distance (2-8 sockets and larger systems are in development) So ideal would be something dynamic to turn on/off io placement, maybe based on node_distance() again, with the threshold tweakable per architecture? Also I must say it's still not quite clear to me if it's better to place network packets on the node the device is connected to or on the node which contains the CPU who processes the packet data For RX this can be three different nodes in the worst case (CPU processing is often split on different CPUs between softirq and user context), for TX two. Do you have some experience that shows that a particular placement is better than the other? -Andi - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/