Received: by 10.213.65.68 with SMTP id h4csp2125617imn; Thu, 29 Mar 2018 18:42:47 -0700 (PDT) X-Google-Smtp-Source: AIpwx480eVjpMXDeFy5AdphdLTYDm88wqXkuS6qEoxDFT53JYClmVmYGarCdoJFvNjn9tSskuwrp X-Received: by 10.98.200.9 with SMTP id z9mr8326150pff.128.1522374167512; Thu, 29 Mar 2018 18:42:47 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1522374167; cv=none; d=google.com; s=arc-20160816; b=UF8/nMgepdqSIK5n8l7p+asifC9aAc6mjydPvrirolw7Pupy6Bidf4qmXct7upJRjo 273RnaOoI3EHNCT5barMSYVL6Z1LQst/coOK2H+s/w96P+bVMaOJHN4zmrsOc+3QjOMU bvXvoas4DxiST5enUpR0objuTWo6g35oZw5H2v/5yuMQC4J3lPwbFsbT0cUP/SwmQ4q7 iH8iWblf8hknmvtx8V3wTAL9yJV5cUVLbqjxYdMdzo9xz7K9efBPFXTrMldgre06y86e z9Y5U/Mkqf22w70lPPLvSrH7ieV669PYLQP9a3zLuSJneK9drm6nPsKES/xP0v5ARmPq S71A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:arc-authentication-results; bh=KjQYvVqmHDGKNV9Hmk8NKjYzpxFUNeDjW1xZz4F8yDE=; b=wM/A699MQOeWjslL8Hi/FcYZ8PJ+ohrFqzFXEjc4r1Q3wp+wLYoKltTRuQPUYx7tQV Q2wzLjsgAD35CslbfeYNG7B0PWRDkhhKETz+Tn9VdUR77mfPZR8hyS/fFFrqnX/Z7Nta D3pVuZRJmMrGF5Vm3MEnfVyIYuSvcl/5eBLJHaPvW34ikpJA+XSwjynv8hZvnJPveFng WjxSFoAY7M5hjs+p7HgyDPnPVte6EMATsExHzA1tEZy4BJn7JavAWteQ0SqwPB0vDwtW tG1Ave13bKJg8SJoZ854U8YMZVq1KeYM5PJ9GKQRAGl5jaAEqZILxOQNdpVq/8FhtOuJ Nk8g== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id 4-v6si7480292pld.371.2018.03.29.18.42.33; Thu, 29 Mar 2018 18:42:47 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752561AbeC3Bl0 (ORCPT + 99 others); Thu, 29 Mar 2018 21:41:26 -0400 Received: from mga06.intel.com ([134.134.136.31]:32235 "EHLO mga06.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752289AbeC3BlX (ORCPT ); Thu, 29 Mar 2018 21:41:23 -0400 X-Amp-Result: UNSCANNABLE X-Amp-File-Uploaded: False Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by orsmga104.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 29 Mar 2018 18:41:23 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.48,378,1517904000"; d="scan'208";a="32653901" Received: from aaronlu.sh.intel.com (HELO intel.com) ([10.239.159.135]) by fmsmga002.fm.intel.com with ESMTP; 29 Mar 2018 18:41:20 -0700 Date: Fri, 30 Mar 2018 09:42:17 +0800 From: Aaron Lu To: Daniel Jordan Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, Andrew Morton , Huang Ying , Dave Hansen , Kemi Wang , Tim Chen , Andi Kleen , Michal Hocko , Vlastimil Babka , Mel Gorman , Matthew Wilcox Subject: Re: [RFC PATCH v2 0/4] Eliminate zone->lock contention for will-it-scale/page_fault1 and parallel free Message-ID: <20180330014217.GA28440@intel.com> References: <20180320085452.24641-1-aaron.lu@intel.com> <2606b76f-be64-4cef-b1f7-055732d09251@oracle.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <2606b76f-be64-4cef-b1f7-055732d09251@oracle.com> User-Agent: Mutt/1.9.2 (2017-12-15) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Mar 29, 2018 at 03:19:46PM -0400, Daniel Jordan wrote: > On 03/20/2018 04:54 AM, Aaron Lu wrote: > > This series is meant to improve zone->lock scalability for order 0 pages. > > With will-it-scale/page_fault1 workload, on a 2 sockets Intel Skylake > > server with 112 CPUs, CPU spend 80% of its time spinning on zone->lock. > > Perf profile shows the most time consuming part under zone->lock is the > > cache miss on "struct page", so here I'm trying to avoid those cache > > misses. > > I ran page_fault1 comparing 4.16-rc5 to your recent work, these four patches > plus the three others from your github branch zone_lock_rfc_v2. Out of > curiosity I also threw in another 4.16-rc5 with the pcp batch size adjusted > so high (10922 pages) that we always stay in the pcp lists and out of buddy > completely. I used your patch[*] in this last kernel. > > This was on a 2-socket, 20-core broadwell server. > > There were some small regressions a bit outside the noise at low process > counts (2-5) but I'm not sure they're repeatable. Anyway, it does improve > the microbenchmark across the board. Thanks for the result. The limited improvement is expected since lock contention only shifts, not entirely gone. So what is interesting to see is how it performs with v4.16-rc5 + my_zone_lock_patchset + your_lru_lock_patchset