Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1755487AbWKUWwj (ORCPT ); Tue, 21 Nov 2006 17:52:39 -0500 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1755483AbWKUWwi (ORCPT ); Tue, 21 Nov 2006 17:52:38 -0500 Received: from calculon.skynet.ie ([193.1.99.88]:13479 "EHLO calculon.skynet.ie") by vger.kernel.org with ESMTP id S1031274AbWKUWwY (ORCPT ); Tue, 21 Nov 2006 17:52:24 -0500 From: Mel Gorman To: linux-mm@kvack.org Cc: Mel Gorman , linux-kernel@vger.kernel.org, clameter@sgi.com Message-Id: <20061121225223.11710.73233.sendpatchset@skynet.skynet.ie> In-Reply-To: <20061121225022.11710.72178.sendpatchset@skynet.skynet.ie> References: <20061121225022.11710.72178.sendpatchset@skynet.skynet.ie> Subject: [PATCH 6/11] Move free pages between lists on steal Date: Tue, 21 Nov 2006 22:52:23 +0000 (GMT) Sender: linux-kernel-owner@vger.kernel.org X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 3701 Lines: 124 When a fallback occurs, there will be free pages for one allocation type stored on the list for another. When a large steal occurs, this patch will move all the free pages within one list to the another. Signed-off-by: Mel Gorman --- page_alloc.c | 80 +++++++++++++++++++++++++++++++++++++++++++++++++----- 1 files changed, 73 insertions(+), 7 deletions(-) diff -rup -X /usr/src/patchset-0.6/bin//dontdiff linux-2.6.19-rc5-mm2-006_drainpercpu/mm/page_alloc.c linux-2.6.19-rc5-mm2-007_movefree/mm/page_alloc.c --- linux-2.6.19-rc5-mm2-006_drainpercpu/mm/page_alloc.c 2006-11-21 10:54:11.000000000 +0000 +++ linux-2.6.19-rc5-mm2-007_movefree/mm/page_alloc.c 2006-11-21 10:56:06.000000000 +0000 @@ -661,6 +661,62 @@ static int prep_new_page(struct page *pa } #ifdef CONFIG_PAGE_CLUSTERING +/* + * Move the free pages in a range to the free lists of the requested type. + * Note that start_page and end_pages are not aligned in a MAX_ORDER_NR_PAGES + * boundary. If alignment is required, use move_freepages_block() + */ +int move_freepages(struct zone *zone, + struct page *start_page, struct page *end_page, + int migratetype) +{ + struct page *page; + unsigned long order; + int blocks_moved = 0; + + BUG_ON(page_zone(start_page) != page_zone(end_page)); + + for (page = start_page; page < end_page;) { + if (!PageBuddy(page)) { + page++; + continue; + } +#ifdef CONFIG_HOLES_IN_ZONE + if (!pfn_valid(page_to_pfn(page))) { + page++; + continue; + } +#endif + + order = page_order(page); + list_del(&page->lru); + list_add(&page->lru, + &zone->free_area[order].free_list[migratetype]); + page += 1 << order; + blocks_moved++; + } + + return blocks_moved; +} + +int move_freepages_block(struct zone *zone, struct page *page, int migratetype) +{ + unsigned long start_pfn; + struct page *start_page, *end_page; + + start_pfn = page_to_pfn(page); + start_pfn = start_pfn & ~(MAX_ORDER_NR_PAGES-1); + start_page = pfn_to_page(start_pfn); + end_page = start_page + MAX_ORDER_NR_PAGES; + + if (page_zone(page) != page_zone(start_page)) + start_page = page; + if (page_zone(page) != page_zone(end_page)) + return 0; + + return move_freepages(zone, start_page, end_page, migratetype); +} + /* Remove an element from the buddy allocator from the fallback list */ static struct page *__rmqueue_fallback(struct zone *zone, int order, int start_migratetype) @@ -681,24 +737,34 @@ static struct page *__rmqueue_fallback(s struct page, lru); area->nr_free--; - /* - * If breaking a large block of pages, place the buddies - * on the preferred allocation list - */ - if (unlikely(current_order >= MAX_ORDER / 2)) - migratetype = !migratetype; - /* Remove the page from the freelists */ list_del(&page->lru); rmv_page_order(page); zone->free_pages -= 1UL << order; expand(zone, page, order, current_order, area, migratetype); + + /* Move free pages between lists if stealing a large block */ + if (current_order > MAX_ORDER / 2) + move_freepages_block(zone, page, start_migratetype); + return page; } return NULL; } #else +int move_freepages(struct zone *zone, + struct page *start_page, struct page *end_page, + int migratetype) +{ + return 0; +} + +int move_freepages_block(struct zone *zone, struct page *page, int migratetype) +{ + return 0; +} + static struct page *__rmqueue_fallback(struct zone *zone, unsigned int order, int migratetype) { - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/