Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S264296AbTHVW6V (ORCPT ); Fri, 22 Aug 2003 18:58:21 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S263705AbTHVW6V (ORCPT ); Fri, 22 Aug 2003 18:58:21 -0400 Received: from dyn-ctb-210-9-245-87.webone.com.au ([210.9.245.87]:15876 "EHLO chimp.local.net") by vger.kernel.org with ESMTP id S264296AbTHVW6N (ORCPT ); Fri, 22 Aug 2003 18:58:13 -0400 Message-ID: <3F469FA4.6020203@cyberone.com.au> Date: Sat, 23 Aug 2003 08:56:36 +1000 From: Nick Piggin User-Agent: Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.4) Gecko/20030714 Debian/1.4-2 X-Accept-Language: en MIME-Version: 1.0 To: Andrew Theurer CC: Bill Davidsen , "Martin J. Bligh" , Erich Focht , linux-kernel , LSE , Andi Kleen , torvalds@osdl.org Subject: Re: [Lse-tech] Re: [patch] scheduler fix for 1cpu/node case References: <200308221046.31662.habanero@us.ibm.com> In-Reply-To: <200308221046.31662.habanero@us.ibm.com> Content-Type: text/plain; charset=us-ascii; format=flowed Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 1864 Lines: 46 Andrew Theurer wrote: >On Wednesday 13 August 2003 15:49, Bill Davidsen wrote: > >>On Mon, 28 Jul 2003, Andrew Theurer wrote: >> >>>Personally, I'd like to see all systems use NUMA sched, non NUMA systems >>>being a single node (no policy difference from non-numa sched), allowing >>>us to remove all NUMA ifdefs. I think the code would be much more >>>readable. >>> >>That sounds like a great idea, but I'm not sure it could be realized short >>of a major rewrite. Look how hard Ingo and Con are working just to get a >>single node doing a good job with interactive and throughput tradeoffs. >> > >Actually it's not too bad. Attached is a patch to do it. It also does >multi-level node support and makes all the load balance routines >runqueue-centric instead of cpu-centric, so adding something like shared >runqueues (for HT) should be really easy. Hmm, other things: inter-node >balance intervals are now arch specific (AMD is "1"). The default busy/idle >balance timers of 200/1 are not arch specific, but I'm thinking they should >be. And for non-numa, the scheduling policy is the same as it was with >vanilla O(1). > I'm not saying you're wrong, but do you have some numbers where this helps? ie. two architectures that need very different balance numbers. And what is the reason for making AMD's balance interval 1? Also, things like nr_running_inc are supposed to be very fast. I am a bit worried to see a loop and CPU shared atomics in there. node_2_node is an odd sounding conversion too ;) BTW. you should be CC'ing Ingo if you have any intention of scheduler stuff getting into 2.6. - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/