Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753681Ab1CQMoR (ORCPT ); Thu, 17 Mar 2011 08:44:17 -0400 Received: from zene.cmpxchg.org ([85.214.230.12]:45457 "EHLO zene.cmpxchg.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752016Ab1CQMoO (ORCPT ); Thu, 17 Mar 2011 08:44:14 -0400 Date: Thu, 17 Mar 2011 13:43:50 +0100 From: Johannes Weiner To: Greg Thelen Cc: Jan Kara , Vivek Goyal , Andrew Morton , linux-kernel@vger.kernel.org, linux-mm@kvack.org, containers@lists.osdl.org, linux-fsdevel@vger.kernel.org, Andrea Righi , Balbir Singh , KAMEZAWA Hiroyuki , Daisuke Nishimura , Minchan Kim , Ciju Rajan K , David Rientjes , Wu Fengguang , Chad Talbott , Justin TerAvest , Curt Wohlgemuth Subject: Re: [PATCH v6 0/9] memcg: per cgroup dirty page accounting Message-ID: <20110317124350.GQ2140@cmpxchg.org> References: <1299869011-26152-1-git-send-email-gthelen@google.com> <20110311171006.ec0d9c37.akpm@linux-foundation.org> <20110314202324.GG31120@redhat.com> <20110315184839.GB5740@redhat.com> <20110316131324.GM2140@cmpxchg.org> <20110316215214.GO2140@cmpxchg.org> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 3676 Lines: 91 On Wed, Mar 16, 2011 at 09:41:48PM -0700, Greg Thelen wrote: > In '[PATCH v6 8/9] memcg: check memcg dirty limits in page writeback' Jan and > Vivek have had some discussion around how memcg and writeback mesh. > In my mind, the > discussions in 8/9 are starting to blend with this thread. > > I have been thinking about Johannes' struct memcg_mapping. I think the idea > may address several of the issues being discussed, especially > interaction between > IO-less balance_dirty_pages() and memcg writeback. > > Here is my thinking. Feedback is most welcome! > > The data structures: > - struct memcg_mapping { > struct address_space *mapping; > struct mem_cgroup *memcg; > int refcnt; > }; > - each memcg contains a (radix, hash_table, etc.) mapping from bdi to memcg_bdi. > - each memcg_bdi contains a mapping from inode to memcg_mapping. This may be a > very large set representing many cached inodes. > - each memcg_mapping represents all pages within an bdi,inode,memcg. All > corresponding cached inode pages point to the same memcg_mapping via > pc->mapping. I assume that all pages of inode belong to no more than one bdi. > - manage a global list of memcg that are over their respective background dirty > limit. > - i_mapping continues to point to a traditional non-memcg mapping (no change > here). > - none of these memcg_* structures affect root cgroup or kernels with memcg > configured out. So structures roughly like this: struct mem_cgroup { ... /* key is struct backing_dev_info * */ struct rb_root memcg_bdis; }; struct memcg_bdi { /* key is struct address_space * */ struct rb_root memcg_mappings; struct rb_node node; }; struct memcg_mapping { struct address_space *mapping; struct mem_cgroup *memcg; struct rb_node node; atomic_t count; }; struct page_cgroup { ... struct memcg_mapping *memcg_mapping; }; > The routines under discussion: > - memcg charging a new inode page to a memcg: will use inode->mapping and inode > to walk memcg -> memcg_bdi -> memcg_mapping and lazily allocating missing > levels in data structure. > > - Uncharging a inode page from a memcg: will use pc->mapping->memcg to locate > memcg. If refcnt drops to zero, then remove memcg_mapping from the memcg_bdi. > Also delete memcg_bdi if last memcg_mapping is removed. > > - account_page_dirtied(): nothing new here, continue to set the per-page flags > and increment the memcg per-cpu dirty page counter. Same goes for routines > that mark pages in writeback and clean states. We may want to remember the dirty memcg_mappings so that on writeback we don't have to go through every single one that the memcg refers to? > - mem_cgroup_balance_dirty_pages(): if memcg dirty memory usage if above > background limit, then add memcg to global memcg_over_bg_limit list and use > memcg's set of memcg_bdi to wakeup each(?) corresponding bdi flusher. If over > fg limit, then use IO-less style foreground throttling with per-memcg per-bdi > (aka memcg_bdi) accounting structure. I wonder if we could just schedule a for_background work manually in the memcg case that writes back the corresponding memcg_bdi set (and e.g. having it continue until either the memcg is below bg thresh OR the global bg thresh is exceeded OR there is other work scheduled)? Then we would get away without the extra list, and it doesn't sound overly complex to implement. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/