Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1759114Ab2HHRs7 (ORCPT ); Wed, 8 Aug 2012 13:48:59 -0400 Received: from rcsinet15.oracle.com ([148.87.113.117]:29292 "EHLO rcsinet15.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1759043Ab2HHRs2 convert rfc822-to-8bit (ORCPT ); Wed, 8 Aug 2012 13:48:28 -0400 MIME-Version: 1.0 Message-ID: Date: Wed, 8 Aug 2012 10:47:48 -0700 (PDT) From: Dan Magenheimer To: Seth Jennings Cc: Greg Kroah-Hartman , Andrew Morton , Nitin Gupta , Minchan Kim , Konrad Wilk , Robert Jennings , linux-mm@kvack.org, linux-kernel@vger.kernel.org, devel@driverdev.osuosl.org, Kurt Hackel Subject: RE: [PATCH 0/4] promote zcache from staging References: <1343413117-1989-1-git-send-email-sjenning@linux.vnet.ibm.com> <5021795A.5000509@linux.vnet.ibm.com> <3f8dfac9-2b92-442c-800a-f0bfef8a90cb@default> <502293E2.8010505@linux.vnet.ibm.com> In-Reply-To: <502293E2.8010505@linux.vnet.ibm.com> X-Priority: 3 X-Mailer: Oracle Beehive Extensions for Outlook 2.0.1.7 (607090) [OL 12.0.6661.5003 (x86)] Content-Type: text/plain; charset=us-ascii Content-Transfer-Encoding: 8BIT X-Source-IP: ucsinet22.oracle.com [156.151.31.94] Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 4584 Lines: 104 > From: Seth Jennings [mailto:sjenning@linux.vnet.ibm.com] Hi Seth -- Good discussion. Even though we disagree, I appreciate your enthusiasm and your good work on the kernel! > Subject: Re: [PATCH 0/4] promote zcache from staging > > On 08/07/2012 04:47 PM, Dan Magenheimer wrote: > > I notice your original published benchmarks [1] include > > N=24, N=28, and N=32, but these updated results do not. Are you planning > > on completing the runs? Second, I now see the numbers I originally > > published for what I thought was the same benchmark as yours are actually > > an order of magnitude larger (in sec) than yours. I didn't notice > > this in March because we were focused on the percent improvement, not > > the raw measurements. Since the hardware is highly similar, I suspect > > it is not a hardware difference but instead that you are compiling > > a much smaller kernel. In other words, your test case is much > > smaller, and so exercises zcache much less. My test case compiles > > a full enterprise kernel... what is yours doing? > > I am doing a minimal kernel build for my local hardware > configuration. > > With the reduction in RAM, 1GB to 512MB, I didn't need to do > test runs with >20 threads to find the peak of the benefit > curve at 16 threads. Past that, zcache is saturated and I'd > just be burning up my disk. I think that's exactly what I said in a snippet of my response that you deleted. A cache needs to work well both when it is non-full and when it is full. You are only demonstrating that it works well when it is non-full. When it is "saturated", bad things can happen. Finding the "peak of the benefit" is only half the work of benchmarking. So it appears you are trying to prove your point by showing the workloads that look good, while _not_ showing the workloads that look bad, and then claiming you don't care about those bad workloads anyway. > Also, I provide the magnitude numbers (pages, seconds) just > to show my source data. The %change numbers are the real > results as they remove build size as a factor. You'll have to explain what you mean because, if I understand correctly, this is just not true. Different build sizes definitely affect memory management differently, just as different values of N (for make -jN) have an effect. > > At LSFMM, Andrea > > Arcangeli pointed out that zcache, for frontswap pages, has no "writeback" > > capabilities and, when it is full, it simply rejects further attempts > > to put data in its cache. He said this is unacceptable for KVM and I > > agreed that it was a flaw that needed to be fixed before zcache should > > be promoted. > > KVM (in-tree) is not a current user of zcache. While the > use cases of possible future zcache users should be > considered, I don't think they can be used to prevent promotion. That wasn't my point. Andrea identified the flaw as an issue of zcache. > > A second flaw is that the "demo" zcache has no concept of LRU for > > either cleancache or frontswap pages, or ability to reclaim pageframes > > at all for frontswap pages. > ... > > > > A third flaw is that the "demo" version has a very poor policy to > > determine what pages are "admitted". > ... > > > > I can add more issues to the list, but will stop here. > > All of the flaws you list do not prevent zcache from being > beneficial right now, as my results demonstrate. Therefore, > the flaws listed are really potential improvements and can > be done in mainline after promotion. Even if large changes > are required to make these improvements, they can be made in > mainline in an incremental and public way. Your results only demonstrate that zcache is beneficial on the workloads that you chose to present. But using the same workload with slightly different parameters (-jN or compiling a larger kernel), zcache can be _detrimental_, and you've chosen to not measure or present those cases, even though you did measure and present some of those cases in your first benchmark runs posted in March (on an earlier kernel). I can only speak for myself, but this appears disingenuous to me. Sorry, but FWIW my vote is still a NACK. IMHO zcache needs major work before it should be promoted, and I think we should be spending the time fixing the known flaws rather than arguing about promoting "demo" code. Dan -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/