Received: by 10.213.65.68 with SMTP id h4csp261400imn; Tue, 20 Mar 2018 02:53:59 -0700 (PDT) X-Google-Smtp-Source: AG47ELsw26QaLuZECxAvjBKi1WRjvsk0sh7jH6JEoGJzPoYZN5bH//vbFtgrLtB+dGqcyWwKrUfm X-Received: by 10.99.188.9 with SMTP id q9mr8736709pge.381.1521539639274; Tue, 20 Mar 2018 02:53:59 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1521539639; cv=none; d=google.com; s=arc-20160816; b=xP4bm/ncX1JmluPoQLgL4GhXnPoYat9xjC5eW2laPGiKDUy2j+0hhpyKLmiq17Ufrx 2qAAH/TydAUzx/o8KSv5eWkxi1VH42EbdF5TWOqwHI51QvuYsdERbUVPHWZbBM3vBjsX K2Sn+t9NwF/BYbUcI6VJ5Y0dER1U52RNyNBEPcXeyd3iF1dQkBEZz/yVTEPzEzmvBqmJ ijylHgVXT4rrersIzEnGAjk6Zhr0E91rSz7147J5YQH7ef8WP/oM8afg6KwTV2Sy1CWu WXxSq9fymzPesdjLGFaCbL5Hev8EXerTWpsb7mO/fhdrlbhkqeYnZi7AeJYOVccmXFQB cTGA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject:arc-authentication-results; bh=agDtEvdBtj195HeUMuzagWOigDJnq7oiiYIGwUpt3CM=; b=U/uNWPIAJGd733Nk2+q/Y2BtKi6/cGKKyoglXEXlZW+m/WUW0FCRmtVNuSEPbQuU/d lZVaB5QJu+WS2VgljukKVmmymNIQyk8CUNo9Fg4DR4bO4J3cfY+h6odcjIzuWhhq3Av+ qF1T+mJPwXklj6Jc1YhwMEfpWCqG3bX+gUkuSOYNGPE477BJn9MFhRdtDCLlNz6XR/Zt YnYzxrxgodS5HGLPJnGvmnI+7TJmIXdw8Snnavstm7ElhpIlhKh4X2O7zm7o5uL3DLSX uPkKGMK1k5QgdfB/AiJAxTD9cj68P9MHjP2BKIt+kW3QBTSGd9fByq+t0l1fj5l3Uao2 Saew== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id z10si934936pgz.781.2018.03.20.02.53.41; Tue, 20 Mar 2018 02:53:59 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752449AbeCTJwO (ORCPT + 99 others); Tue, 20 Mar 2018 05:52:14 -0400 Received: from mx2.suse.de ([195.135.220.15]:44634 "EHLO mx2.suse.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751554AbeCTJwK (ORCPT ); Tue, 20 Mar 2018 05:52:10 -0400 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay1.suse.de (charybdis-ext.suse.de [195.135.220.254]) by mx2.suse.de (Postfix) with ESMTP id B4F79ADB7; Tue, 20 Mar 2018 09:52:08 +0000 (UTC) Subject: Re: [PATCH v4 3/3 update] mm/free_pcppages_bulk: prefetch buddy while not holding lock To: Aaron Lu , Dave Hansen Cc: Andrew Morton , linux-mm@kvack.org, linux-kernel@vger.kernel.org, Huang Ying , Kemi Wang , Tim Chen , Andi Kleen , Michal Hocko , Mel Gorman , Matthew Wilcox , David Rientjes References: <20180301062845.26038-1-aaron.lu@intel.com> <20180301062845.26038-4-aaron.lu@intel.com> <20180301160950.b561d6b8b561217bad511229@linux-foundation.org> <20180302082756.GC6356@intel.com> <20180309082431.GB30868@intel.com> <988ce376-bdc4-0989-5133-612bfa3f7c45@intel.com> <20180313033519.GC13782@intel.com> <20180313070404.GA7501@intel.com> From: Vlastimil Babka Message-ID: <5600c827-d22b-136c-6b90-a4b52f40af31@suse.cz> Date: Tue, 20 Mar 2018 10:50:18 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:52.0) Gecko/20100101 Thunderbird/52.6.0 MIME-Version: 1.0 In-Reply-To: <20180313070404.GA7501@intel.com> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 03/13/2018 08:04 AM, Aaron Lu wrote: > On Tue, Mar 13, 2018 at 11:35:19AM +0800, Aaron Lu wrote: >> On Mon, Mar 12, 2018 at 10:32:32AM -0700, Dave Hansen wrote: >>> On 03/09/2018 12:24 AM, Aaron Lu wrote: >>>> + /* >>>> + * We are going to put the page back to the global >>>> + * pool, prefetch its buddy to speed up later access >>>> + * under zone->lock. It is believed the overhead of >>>> + * an additional test and calculating buddy_pfn here >>>> + * can be offset by reduced memory latency later. To >>>> + * avoid excessive prefetching due to large count, only >>>> + * prefetch buddy for the last pcp->batch nr of pages. >>>> + */ >>>> + if (count > pcp->batch) >>>> + continue; >>>> + pfn = page_to_pfn(page); >>>> + buddy_pfn = __find_buddy_pfn(pfn, 0); >>>> + buddy = page + (buddy_pfn - pfn); >>>> + prefetch(buddy); >>> >>> FWIW, I think this needs to go into a helper function. Is that possible? >> >> I'll give it a try. >> >>> >>> There's too much logic happening here. Also, 'count' going from >>> batch_size->0 is totally non-obvious from the patch context. It makes >>> this hunk look totally wrong by itself. > > I tried to avoid adding one more local variable but looks like it caused > a lot of pain. What about the following? It doesn't use count any more > but prefetch_nr to indicate how many prefetches have happened. > > Also, I think it's not worth the risk of disordering pages in free_list > by changing list_add_tail() to list_add() as Andrew reminded so I > dropped that change too. Looks fine, you can add Acked-by: Vlastimil Babka > > diff --git a/mm/page_alloc.c b/mm/page_alloc.c > index dafdcdec9c1f..00ea4483f679 100644 > --- a/mm/page_alloc.c > +++ b/mm/page_alloc.c > @@ -1099,6 +1099,15 @@ static bool bulkfree_pcp_prepare(struct page *page) > } > #endif /* CONFIG_DEBUG_VM */ > > +static inline void prefetch_buddy(struct page *page) > +{ > + unsigned long pfn = page_to_pfn(page); > + unsigned long buddy_pfn = __find_buddy_pfn(pfn, 0); > + struct page *buddy = page + (buddy_pfn - pfn); > + > + prefetch(buddy); > +} > + > /* > * Frees a number of pages from the PCP lists > * Assumes all pages on list are in same zone, and of same order. > @@ -1115,6 +1124,7 @@ static void free_pcppages_bulk(struct zone *zone, int count, > { > int migratetype = 0; > int batch_free = 0; > + int prefetch_nr = 0; > bool isolated_pageblocks; > struct page *page, *tmp; > LIST_HEAD(head); > @@ -1150,6 +1160,18 @@ static void free_pcppages_bulk(struct zone *zone, int count, > continue; > > list_add_tail(&page->lru, &head); > + > + /* > + * We are going to put the page back to the global > + * pool, prefetch its buddy to speed up later access > + * under zone->lock. It is believed the overhead of > + * an additional test and calculating buddy_pfn here > + * can be offset by reduced memory latency later. To > + * avoid excessive prefetching due to large count, only > + * prefetch buddy for the first pcp->batch nr of pages. > + */ > + if (prefetch_nr++ < pcp->batch) > + prefetch_buddy(page); > } while (--count && --batch_free && !list_empty(list)); > } > >