Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1757982AbZFJIMK (ORCPT ); Wed, 10 Jun 2009 04:12:10 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1757799AbZFJILi (ORCPT ); Wed, 10 Jun 2009 04:11:38 -0400 Received: from mga03.intel.com ([143.182.124.21]:53772 "EHLO mga03.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754853AbZFJILf (ORCPT ); Wed, 10 Jun 2009 04:11:35 -0400 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="4.41,339,1241420400"; d="scan'208";a="152628147" Date: Wed, 10 Jun 2009 16:11:32 +0800 From: Wu Fengguang To: Johannes Weiner Cc: Andrew Morton , Rik van Riel , Hugh Dickins , Andi Kleen , KAMEZAWA Hiroyuki , Minchan Kim , "linux-mm@kvack.org" , "linux-kernel@vger.kernel.org" Subject: Re: [patch v3] swap: virtual swap readahead Message-ID: <20090610081132.GA27519@localhost> References: <20090609190128.GA1785@cmpxchg.org> <20090609193702.GA2017@cmpxchg.org> <20090610050342.GA8867@localhost> <20090610074508.GA1960@cmpxchg.org> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20090610074508.GA1960@cmpxchg.org> User-Agent: Mutt/1.5.18 (2008-05-17) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 3833 Lines: 109 On Wed, Jun 10, 2009 at 03:45:08PM +0800, Johannes Weiner wrote: > Hi Fengguang, > > On Wed, Jun 10, 2009 at 01:03:42PM +0800, Wu Fengguang wrote: > > On Wed, Jun 10, 2009 at 03:37:02AM +0800, Johannes Weiner wrote: > > > On Tue, Jun 09, 2009 at 09:01:28PM +0200, Johannes Weiner wrote: > > > > [resend with lists cc'd, sorry] > > > > > > [and fixed Hugh's email. crap] > > > > > > > Hi, > > > > > > > > here is a new iteration of the virtual swap readahead. Per Hugh's > > > > suggestion, I moved the pte collecting to the callsite and thus out > > > > ouf swap code. Unfortunately, I had to bound page_cluster due to an > > > > array of that many swap entries on the stack, but I think it is better > > > > to limit the cluster size to a sane maximum than using dynamic > > > > allocation for this purpose. > > > > Hi Johannes, > > > > When stress testing your patch, I found it triggered many OOM kills. > > Around the time of last OOMs, the memory usage is: > > > > total used free shared buffers cached > > Mem: 474 468 5 0 0 239 > > -/+ buffers/cache: 229 244 > > Swap: 1023 221 802 > > Wow, that really confused me for a second as we shouldn't read more > pages ahead than without the patch, probably even less under stress. Yup - swap readahead is much more challenging than sequential readahead, in that it must be accurate enough given some really obscure patterns. > So the problem has to be a runaway reading. And indeed, severe > stupidity here: > > + window = cluster << PAGE_SHIFT; > + min = addr & ~(window - 1); > + max = min + cluster; > + /* > + * To keep the locking/highpte mapping simple, stay > + * within the PTE range of one PMD entry. > + */ > + limit = addr & PMD_MASK; > + if (limit > min) > + min = limit; > + limit = pmd_addr_end(addr, max); > + if (limit < max) > + max = limit; > + limit = max - min; > > The mistake is at the initial calculation of max. It should be > > max = min + window; > > The resulting problem is that min could get bigger than max when > cluster is bigger than PMD_SHIFT. Did you use page_cluster == 5? No I use the default 3. btw, the mistake reflects bad named variables. How about rename cluster => pages window => bytes ? > The initial min is aligned to a value below the PMD boundary and max > based on it with a too small offset, staying below the PMD boundary as > well. When min is rounded up, this becomes a bit large: > > limit = max - min; > > So if my brain is already functioning, fixing the initial max should > be enough because either > > o window is smaller than PMD_SIZE, than we won't round down > below a PMD boundary in the first place or > > o window is bigger than PMD_SIZE, than we can round down below > a PMD boundary but adding window to that is garuanteed to > cross the boundary again > > and thus max is always bigger than min. > > Fengguang, does this make sense? If so, the patch below should fix > it. Too bad, a quick test of the below patch freezes the box.. Thanks, Fengguang > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -2467,7 +2467,7 @@ static int swap_readahead_ptes(struct mm > > window = cluster << PAGE_SHIFT; > min = addr & ~(window - 1); > - max = min + cluster; > + max = min + window; > /* > * To keep the locking/highpte mapping simple, stay > * within the PTE range of one PMD entry. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/