Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1758644Ab3HOPgv (ORCPT ); Thu, 15 Aug 2013 11:36:51 -0400 Received: from mga01.intel.com ([192.55.52.88]:51979 "EHLO mga01.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1757806Ab3HOPgt (ORCPT ); Thu, 15 Aug 2013 11:36:49 -0400 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="4.89,886,1367996400"; d="scan'208";a="386822362" Message-ID: <520CF588.7090800@intel.com> Date: Thu, 15 Aug 2013 08:36:40 -0700 From: Dave Hansen User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:17.0) Gecko/20130803 Thunderbird/17.0.8 MIME-Version: 1.0 To: Dave Chinner CC: Andi Kleen , Dave Hansen , linux-fsdevel@vger.kernel.org, xfs@oss.sgi.com, linux-ext4@vger.kernel.org, Jan Kara , LKML , Tim Chen , Andy Lutomirski Subject: Re: page fault scalability (ext3, ext4, xfs) References: <520BB9EF.5020308@linux.intel.com> <20130815002436.GI6023@dastard> <20130815022401.GQ23412@tassilo.jf.intel.com> <20130815042930.GO6023@dastard> In-Reply-To: <20130815042930.GO6023@dastard> Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2380 Lines: 53 On 08/14/2013 09:29 PM, Dave Chinner wrote: > On Wed, Aug 14, 2013 at 07:24:01PM -0700, Andi Kleen wrote: >>> And FWIW, it's no secret that XFS has more per-operation overhead >>> than ext4 through the write path when it comes to allocation, so >>> it's no surprise that on a workload that is highly dependent on >>> allocation overhead that ext4 is a bit faster.... >> >> This cannot explain a worse scaling curve though? > > The scaling curve is pretty much identical. The difference in > performance will be the overhead of timestamp updates through > the transaction subsystems of the filesystems. I guess how you read it is in the eye of the beholder. I see xfs being slower than ext3 or ext4. Nobody sits and does this in a loop in real life (it's a microbenchbark), but I'd be willing to bet that this is a real *component* of real-life workloads. It's a component where I think it's pretty clear xfs and ext4 lag behind ext3, and it _looks_ to me like it gets worse on larger systems. Maybe that's because of design decisions in the filesystem, or because of the enhanced integrity guarantees that xfs/ext4 provide. >> w-i-s is all about scaling. > > Sure, but scaling *what*? It's spending all it's time in the > filesystem through the .page_mkwrite path. It's not a page fault > scaling test - it's a filesystem overwrite test that uses mmap. will-it-scale tests a bunch of different scenarios. This is just one of at least 6 tests that we do which beat on the page fault path. It was the only one of those 6 that showed any kind of bottleneck being in the fs code. > Indeed, I bet if you replace the mmap() with a write(fd, buf, 4096) > loop, you'd get almost identical behaviour from the filesystems. In a quick 60-second test: xfs went from ~70M writes/sec (doing faults) to ~18M/sec (using write()). ext4 went down to 0.5M/sec. I didn't take the mmap()/munmap() out: lseek(fd, 0, SEEK_SET); for (i = 0; i < MEMSIZE; i += pgsize) { write(fd, xxx, 4096); //c[i] = 0; (*iterations)++; } -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/