Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1762907Ab3ECJL4 (ORCPT ); Fri, 3 May 2013 05:11:56 -0400 Received: from cantor2.suse.de ([195.135.220.15]:51870 "EHLO mx2.suse.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755368Ab3ECJLx (ORCPT ); Fri, 3 May 2013 05:11:53 -0400 Date: Fri, 3 May 2013 11:11:49 +0200 From: Michal Hocko To: Sha Zhengju Cc: linux-kernel@vger.kernel.org, cgroups@vger.kernel.org, linux-mm@kvack.org, linux-fsdevel@vger.kernel.org, akpm@linux-foundation.org, kamezawa.hiroyu@jp.fujitsu.com, gthelen@google.com, fengguang.wu@intel.com, glommer@parallels.com, dchinner@redhat.com, Sha Zhengju Subject: Re: [PATCH V3 4/8] memcg: add per cgroup dirty pages accounting Message-ID: <20130503091149.GA17496@dhcp22.suse.cz> References: <1356455919-14445-1-git-send-email-handai.szj@taobao.com> <1356456367-14660-1-git-send-email-handai.szj@taobao.com> <20130102104421.GC22160@dhcp22.suse.cz> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20130102104421.GC22160@dhcp22.suse.cz> User-Agent: Mutt/1.5.21 (2010-09-15) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2040 Lines: 46 On Wed 02-01-13 11:44:21, Michal Hocko wrote: > On Wed 26-12-12 01:26:07, Sha Zhengju wrote: > > From: Sha Zhengju > > > > This patch adds memcg routines to count dirty pages, which allows memory controller > > to maintain an accurate view of the amount of its dirty memory and can provide some > > info for users while cgroup's direct reclaim is working. > > I guess you meant targeted resp. (hard/soft) limit reclaim here, > right? It is true that this is direct reclaim but it is not clear to me > why the usefulnes should be limitted to the reclaim for users. I would > understand this if the users was in fact in-kernel users. > > [...] > > To prevent AB/BA deadlock mentioned by Greg Thelen in previous version > > (https://lkml.org/lkml/2012/7/30/227), we adjust the lock order: > > ->private_lock --> mapping->tree_lock --> memcg->move_lock. > > So we need to make mapping->tree_lock ahead of TestSetPageDirty in __set_page_dirty() > > and __set_page_dirty_nobuffers(). But in order to avoiding useless spinlock contention, > > a prepare PageDirty() checking is added. > > But there is another AA deadlock here I believe. > page_remove_rmap > mem_cgroup_begin_update_page_stat <<< 1 > set_page_dirty > __set_page_dirty_buffers > __set_page_dirty > mem_cgroup_begin_update_page_stat <<< 2 > move_lock_mem_cgroup > spin_lock_irqsave(&memcg->move_lock, *flags); JFYI since abf09bed (s390/mm: implement software dirty bits) this is no longer possible. I haven't checked wheter there are other cases like this one and it should be better if mem_cgroup_begin_update_page_stat was recursive safe if that can be done without too many hacks. I will have a look at this (hopefully) sometimes next week. [...] -- Michal Hocko SUSE Labs -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/