Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1764847AbZLQOoF (ORCPT ); Thu, 17 Dec 2009 09:44:05 -0500 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1764831AbZLQOn7 (ORCPT ); Thu, 17 Dec 2009 09:43:59 -0500 Received: from mx1.redhat.com ([209.132.183.28]:16852 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1758619AbZLQOn6 (ORCPT ); Thu, 17 Dec 2009 09:43:58 -0500 Message-ID: <4B2A438A.6000908@redhat.com> Date: Thu, 17 Dec 2009 09:43:22 -0500 From: Rik van Riel Organization: Red Hat, Inc User-Agent: Mozilla/5.0 (X11; U; Linux x86_64; en-US; rv:1.9.1.4pre) Gecko/20090922 Fedora/3.0-3.9.b4.fc12 Lightning/1.0pre Thunderbird/3.0b4 MIME-Version: 1.0 To: lwoodman@redhat.com CC: KOSAKI Motohiro , LKML , akpm@linux-foundation.org, linux-mm Subject: Re: FWD: [PATCH v2] vmscan: limit concurrent reclaimers in shrink_zone References: <20091211164651.036f5340@annuminas.surriel.com> <1260810481.6666.13.camel@dhcp-100-19-198.bos.redhat.com> <20091217193818.9FA9.A69D9226@jp.fujitsu.com> <4B2A22C0.8080001@redhat.com> In-Reply-To: <4B2A22C0.8080001@redhat.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 1537 Lines: 40 On 12/17/2009 07:23 AM, Larry Woodman wrote: >>> The system would not respond so I dont know whats going on yet. I'll >>> add debug code to figure out why its in that state as soon as I get >>> access to the hardware. > > This was in response to Rik's first patch and seems to be fixed by the > latest path set. > > Finally, having said all that, the system still struggles reclaiming > memory with > ~10000 processes trying at the same time, you fix one bottleneck and it > moves > somewhere else. The latest run showed all but one running process > spinning in > page_lock_anon_vma() trying for the anon_vma_lock. I noticed that there are > ~5000 vma's linked to one anon_vma, this seems excessive!!! I have some ideas on how to keep processes waiting better on the per zone reclaim_wait waitqueue. For one, we should probably only do the lots-free wakeup if we have more than zone->pages_high free pages in the zone - having each of the waiters free some memory one after another should not be a problem as long as we do not have too much free memory in the zone. Currently it's a hair trigger, with the threshold for processes going into the page reclaim path and processes exiting it "plenty free" being exactly the same. Some hysteresis there could help. -- All rights reversed. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/