Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1758422AbZGCSyT (ORCPT ); Fri, 3 Jul 2009 14:54:19 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1756264AbZGCSyM (ORCPT ); Fri, 3 Jul 2009 14:54:12 -0400 Received: from brick.kernel.dk ([93.163.65.50]:47301 "EHLO kernel.dk" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755842AbZGCSyL (ORCPT ); Fri, 3 Jul 2009 14:54:11 -0400 Date: Fri, 3 Jul 2009 20:54:14 +0200 From: Jens Axboe To: Andi Kleen Cc: Matthew Wilcox , linux-kernel@vger.kernel.org, "Styner, Douglas W" , Chinang Ma , "Prickett, Terry O" , Matthew Wilcox , Eric.Moore@lsi.com, DL-MPTFusionLinux@lsi.com Subject: Re: >10% performance degradation since 2.6.18 Message-ID: <20090703185414.GP23611@kernel.dk> References: <20090703025607.GK5480@parisc-linux.org> <87skhdaaub.fsf@basil.nowhere.org> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <87skhdaaub.fsf@basil.nowhere.org> Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 1643 Lines: 44 On Fri, Jul 03 2009, Andi Kleen wrote: > > Matthew Wilcox writes: > > > > ======oprofile CPU_CLK_UNHALTED for top 30 functions > > Cycles% 2.6.18-92.el5-op Cycles% 2.6.30 > > 70.1409 67.0207 > > 1.3556 mpt_interrupt 1.7029 mpt_interrupt > > It's strange that mpt_interrupt is that more costly in 2.6.30 > than in 2.6.18. I diffed 2.6.30's drivers/message/fusion/mptbase.c > to a rhel 5.3s and they seem to be about the same. > > So why does it cost 0.5% more in 2.6.30? > > [adding MPT maintainers] Look at the irqs/sec rate, it's higher by about the same percentage. So it's likely not a more costly irq handler, it's likely just called that much more. It could be IO pattern, causing more commands to be issued (which leads to more interrupts, etc). > > 1.1622 __blockdev_direct_IO 1.1443 kmem_cache_alloc > > It would be interesting to find out why kmem_cache_alloc > is that more expensive. Either it is called more or the allocator > is slower. Any chance of a callgraph profile run so we > can see the callers? Could be more IO as well, that hits the allocate often. I agree with some callgraph data, that would at least eliminate the guessing here. And some detailed IO statistics, amount of data transferred as well as iostat info to see if the pattern is significantly worse. -- Jens Axboe -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/