Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1756343Ab0GEBkA (ORCPT ); Sun, 4 Jul 2010 21:40:00 -0400 Received: from fgwmail6.fujitsu.co.jp ([192.51.44.36]:52651 "EHLO fgwmail6.fujitsu.co.jp" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755016Ab0GEBj6 (ORCPT ); Sun, 4 Jul 2010 21:39:58 -0400 X-SecurityPolicyCheck-FJ: OK by FujitsuOutboundMailChecker v1.3.1 Date: Mon, 5 Jul 2010 10:35:06 +0900 From: KAMEZAWA Hiroyuki To: Andrew Morton Cc: Mel Gorman , linux-kernel@vger.kernel.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, Dave Chinner , Chris Mason , Nick Piggin , Rik van Riel , Johannes Weiner , Christoph Hellwig , KOSAKI Motohiro , Andrea Arcangeli Subject: Re: [PATCH 0/14] Avoid overflowing of stack during page reclaim V3 Message-Id: <20100705103506.8fbd1509.kamezawa.hiroyu@jp.fujitsu.com> In-Reply-To: <20100702123315.667c6eac.akpm@linux-foundation.org> References: <1277811288-5195-1-git-send-email-mel@csn.ul.ie> <20100702123315.667c6eac.akpm@linux-foundation.org> Organization: FUJITSU Co. LTD. X-Mailer: Sylpheed 3.0.3 (GTK+ 2.10.14; i686-pc-mingw32) Mime-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2016 Lines: 54 On Fri, 2 Jul 2010 12:33:15 -0700 Andrew Morton wrote: > On Tue, 29 Jun 2010 12:34:34 +0100 > Mel Gorman wrote: > > > Here is V3 that depends again on flusher threads to do writeback in > > direct reclaim rather than stack switching which is not something I'm > > likely to get done before xfs/btrfs are ignoring writeback in mainline > > (phd sucking up time). > > IMO, implemetning stack switching for this is not a good idea. We > _already_ have a way of doing stack-switching. It's called > "schedule()". > Sure. > The only reason I can see for implementing an in-place stack switch > would be if schedule() is too expensive. And if we were to see > excessive context-switch overheads in this code path (and we won't) > then we should get in there and try to reduce the contect switch rate > first. > Maybe a concern of in-place stack exchange lovers is that it's difficult to guarantee when the pageout() will be issued. I'd like to try to add a call as - pageout_request(page) .... request to pageout a page to a daemon(kswapd). - pageout_barrier(zone? node?) .... wait until all writebacks ends. Implementation dilemmna: Because page->lru is very useful link to implement calls like above, but there is a concern that using it will hide pages from vmscan unnecessarily. Avoding to use of page->lru means to use another structure like pagevec, but it means page_count()+1 and pins pages unnecessarily. I'm now considering how to implement safe&scalable way to pageout-in-another-stack(thread)... I wonder it will require some throttling method for pageout, anyway. And my own problem is that I should add per-memcg threads or using some thread-pool ;( But it will be an another topic. Thanks, -Kame -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/