Received: by 10.223.185.116 with SMTP id b49csp3750447wrg; Tue, 6 Mar 2018 04:28:21 -0800 (PST) X-Google-Smtp-Source: AG47ELtYjzlChESPaYsKrRdXphVuzMaEfCsZ804vHnwrXoFJZPPlNfENbKhUdX/Dy6AwO1RpMa1D X-Received: by 10.99.123.74 with SMTP id k10mr15064744pgn.217.1520339301110; Tue, 06 Mar 2018 04:28:21 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1520339301; cv=none; d=google.com; s=arc-20160816; b=JpUVL7oWRBsMPAXUH6tPTQ/9GGE91NJJucK0b5VKJ7VjbTd/eJqOH2N0B37V+Q/eRx rxnbOAgpK1nHHPa7/vrqSUDlDmZACYaToGN5/MgKdQY7Lk+dklYxyhHWNrmzB6lad4eE GOOvtsWhMiTJVO2iAbO+M5HjfVpQ84T5szn3TZnthLhSHTgFtGeENr/BUV6qNbgO0HVA cISNGf9LeLdSLkwtJxaxCwCCIOZiScQ0iYBckmjb4yqxOPCa324KQKDaEdP+eVfHWB++ q9YhJ2jso5T1mL1CeFULmQgDEzmIQPZXh9TpmKtmTq/vuduusa4jTprIzA2PdP1C8FXx iX3A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-transfer-encoding:content-disposition:mime-version :references:message-id:subject:cc:to:from:date :arc-authentication-results; bh=WOjhBa73NGbUcl44LXzf1bAKKbrYHxEFau1izwxBxFY=; b=rh+XSqbocKtnscysa61DY58yHOH1k0msE3H++EqgKmLOc8H1uvYGCm+u5SiQNJrCwE 055NgpzawG5HLLsobn6CQtlsbcLdMmZlNusV0c/7CyhrXY6Rp0vMzGi5dRjdhkkAE6Eq v3YcRX2OuTC1rmq+uWbUjOhM+J6NH4UVBTbSaOhRiGHc0pX0wJXx2pndsxRFHq9UE9a0 Crx0efoIAqjc3z4s+SXVyBQWTu0+bExnPwqqOwQ75jrqhhKK6Qm2YNB581O6Jd5vTMm+ qiw14hSI7MFkZh4lvhBOOJsYqFvZoPm7cc/qoEaqlLMQXrkVWHsNgoIOQ6NM7TF/FnWP i7Yw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id n18si11777426pfi.262.2018.03.06.04.28.07; Tue, 06 Mar 2018 04:28:21 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753687AbeCFM0k (ORCPT + 99 others); Tue, 6 Mar 2018 07:26:40 -0500 Received: from mga18.intel.com ([134.134.136.126]:13330 "EHLO mga18.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1753672AbeCFM0i (ORCPT ); Tue, 6 Mar 2018 07:26:38 -0500 X-Amp-Result: UNSCANNABLE X-Amp-File-Uploaded: False Received: from fmsmga004.fm.intel.com ([10.253.24.48]) by orsmga106.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 06 Mar 2018 04:26:37 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.47,431,1515484800"; d="scan'208";a="34998970" Received: from aaronlu.sh.intel.com (HELO intel.com) ([10.239.159.135]) by fmsmga004.fm.intel.com with ESMTP; 06 Mar 2018 04:26:35 -0800 Date: Tue, 6 Mar 2018 20:27:33 +0800 From: Aaron Lu To: Vlastimil Babka Cc: Michal Hocko , linux-mm@kvack.org, linux-kernel@vger.kernel.org, Andrew Morton , Huang Ying , Dave Hansen , Kemi Wang , Tim Chen , Andi Kleen , Mel Gorman , Matthew Wilcox , David Rientjes Subject: Re: [PATCH v4 3/3] mm/free_pcppages_bulk: prefetch buddy while not holding lock Message-ID: <20180306122733.GA9664@intel.com> References: <20180301062845.26038-1-aaron.lu@intel.com> <20180301062845.26038-4-aaron.lu@intel.com> <20180301140044.GK15057@dhcp22.suse.cz> <20180305114159.GA32573@intel.com> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: User-Agent: Mutt/1.9.2 (2017-12-15) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Mar 06, 2018 at 08:55:57AM +0100, Vlastimil Babka wrote: > On 03/05/2018 12:41 PM, Aaron Lu wrote: > > On Fri, Mar 02, 2018 at 06:55:25PM +0100, Vlastimil Babka wrote: > >> On 03/01/2018 03:00 PM, Michal Hocko wrote: > >>> > >>> I am really surprised that this has such a big impact. > >> > >> It's even stranger to me. Struct page is 64 bytes these days, exactly a > >> a cache line. Unless that changed, Intel CPUs prefetched a "buddy" cache > >> line (that forms an aligned 128 bytes block with the one we touch). > >> Which is exactly a order-0 buddy struct page! Maybe that implicit > >> prefetching stopped at L2 and explicit goes all the way to L1, can't > > > > The Intel Architecture Optimization Manual section 7.3.2 says: > > > > prefetchT0 - fetch data into all cache levels > > Intel Xeon Processors based on Nehalem, Westmere, Sandy Bridge and newer > > microarchitectures: 1st, 2nd and 3rd level cache. > > > > prefetchT2 - fetch data into 2nd and 3rd level caches (identical to > > prefetchT1) > > Intel Xeon Processors based on Nehalem, Westmere, Sandy Bridge and newer > > microarchitectures: 2nd and 3rd level cache. > > > > prefetchNTA - fetch data into non-temporal cache close to the processor, > > minimizing cache pollution > > Intel Xeon Processors based on Nehalem, Westmere, Sandy Bridge and newer > > microarchitectures: must fetch into 3rd level cache with fast replacement. > > > > I tried 'prefetcht0' and 'prefetcht2' instead of the default > > 'prefetchNTA' on a 2 sockets Intel Skylake, the two ended up with about > > the same performance number as prefetchNTA. I had expected prefetchT0 to > > deliver a better score if it was indeed due to L1D since prefetchT2 will > > not place data into L1 while prefetchT0 will, but looks like it is not > > the case here. > > > > It feels more like the buddy cacheline isn't in any level of the caches > > without prefetch for some reason. > > So the adjacent line prefetch might be disabled? Could you check bios or > the MSR mentioned in > https://software.intel.com/en-us/articles/disclosure-of-hw-prefetcher-control-on-some-intel-processors root@lkp-bdw-ep2 ~# rdmsr 0x1a4 0 Looks like this feature isn't disabled(the doc you linked says value 1 means disable). > >> remember. Would that make such a difference? It would be nice to do some > >> perf tests with cache counters to see what is really going on... > > > > Compare prefetchT2 to no-prefetch, I saw these metrics change: > > > > no-prefetch change prefetchT2 metrics > > \ \ > > stddev stddev > > ------------------------------------------------------------------------ > > 0.18 +0.0 0.18 perf-stat.branch-miss-rate% > > 8.268e+09 +3.8% 8.585e+09 perf-stat.branch-misses > > 2.333e+10 +4.7% 2.443e+10 perf-stat.cache-misses > > 2.402e+11 +5.0% 2.522e+11 perf-stat.cache-references > > 3.52 -1.1% 3.48 perf-stat.cpi > > 0.02 -0.0 0.01 ?3% perf-stat.dTLB-load-miss-rate% > > 8.677e+08 -7.3% 8.048e+08 ?3% perf-stat.dTLB-load-misses > > 1.18 +0.0 1.19 perf-stat.dTLB-store-miss-rate% > > 2.359e+10 +6.0% 2.502e+10 perf-stat.dTLB-store-misses > > 1.979e+12 +5.0% 2.078e+12 perf-stat.dTLB-stores > > 6.126e+09 +10.1% 6.745e+09 ?3% perf-stat.iTLB-load-misses > > 3464 -8.4% 3172 ?3% perf-stat.instructions-per-iTLB-miss > > 0.28 +1.1% 0.29 perf-stat.ipc > > 2.929e+09 +5.1% 3.077e+09 perf-stat.minor-faults > > 9.244e+09 +4.7% 9.681e+09 perf-stat.node-loads > > 2.491e+08 +5.8% 2.634e+08 perf-stat.node-store-misses > > 6.472e+09 +6.1% 6.869e+09 perf-stat.node-stores > > 2.929e+09 +5.1% 3.077e+09 perf-stat.page-faults > > 2182469 -4.2% 2090977 perf-stat.path-length > > > > Not sure if this is useful though... > > Looks like most stats increased in absolute values as the work done > increased and this is a time-limited benchmark? Although number of Yes it is. > instructions (calculated from itlb misses and insns-per-itlb-miss) shows > less than 1% increase, so dunno. And the improvement comes from reduced > dTLB-load-misses? That makes no sense for order-0 buddy struct pages > which always share a page. And the memmap mapping should use huge pages. THP is disabled to stress order 0 pages(should have mentioned this in patch's description, sorry about this). > BTW what is path-length? It's the instruction path length: the number of machine code instructions required to execute a section of a computer program.