Received: by 2002:ac0:a5b6:0:0:0:0:0 with SMTP id m51-v6csp437890imm; Mon, 4 Jun 2018 21:30:53 -0700 (PDT) X-Google-Smtp-Source: ADUXVKKBUzeBu+CN3AAEjHJm9C0JrL0MZyi5C2ov0Yy3wVL7PX1yAXNHjtTIifbHxJCA2pH4lQIM X-Received: by 2002:a62:d388:: with SMTP id z8-v6mr24262916pfk.8.1528173053270; Mon, 04 Jun 2018 21:30:53 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1528173053; cv=none; d=google.com; s=arc-20160816; b=SuBh4GhIIUtt18HN9L/zBsqFny5YD8RFUeMWDEE8moniVSRVbFb64Byp3mu0MhJnIf QOAbiDAdgcI9K3mWlc4rJwNBdVSSiMmrbp9FLLf3gFHbPrr5jBTVPrL0DQkCN7zoi6hN U7EKbCPSOz+pauRrnzPcC+Jpwkz2g8c7A2PDt1dTkFej97do7AMrPhwVLeK/9sJsuwvK aPoxTl9rj+hXjT2BhPQwPifys4STl1LSzbni6+GzRHByZtOBcTGl+c3XJzWW5SixT38V HXJ5OtpbA7YVzf2h2sevi4MpeB4KYOyl/fpiibQMIxRxGgVzD5be/7y8W0s5TTzG+lwo mRJw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:mime-version:user-agent:message-id :in-reply-to:date:references:subject:cc:to:from :arc-authentication-results; bh=uT2xND54PqkDageZoKqPoqAfcj56T/jXNObyzlHRgwM=; b=E0Lwb5VdZ+PhxvwFJNitBtLORjsVBD12UWutfH9En56SrmxbzliN+PET1nd0ymwFwV 5ojwXmQGeaIpRrhZHAcTwFPt2tazBmciQBiiEj9ZEeeXlnL8urM3sG36hCVmg3mZOTVY jZpMWUXzPkZTWJeKmNTmtJniMGdtWTU7MozNqy9eN+2t+cQhL16m3KTgLkuaQ52u+c16 GjpD75h6LoNr4wQ3aIdbdL1fxMhu3elFznzl9UUIZdMTL+89aP9ps/wKmqSDGw0POsc1 1huILmd6BsLsaIf2lac+KBR8lT/tSIRTTQwzVtf3uANvZFAWwqDu+z8YVjUSCd29ObZe d92Q== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id v3-v6si18503744pgr.44.2018.06.04.21.30.38; Mon, 04 Jun 2018 21:30:53 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751395AbeFEEaP (ORCPT + 99 others); Tue, 5 Jun 2018 00:30:15 -0400 Received: from mga12.intel.com ([192.55.52.136]:3411 "EHLO mga12.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1750828AbeFEEaO (ORCPT ); Tue, 5 Jun 2018 00:30:14 -0400 X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga007.fm.intel.com ([10.253.24.52]) by fmsmga106.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 04 Jun 2018 21:30:14 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.49,477,1520924400"; d="scan'208";a="44688588" Received: from yhuang-dev.sh.intel.com (HELO yhuang-dev) ([10.239.13.118]) by fmsmga007.fm.intel.com with ESMTP; 04 Jun 2018 21:30:13 -0700 From: "Huang\, Ying" To: Daniel Jordan Cc: Andrew Morton , , Subject: Re: [PATCH -mm -V3 00/21] mm, THP, swap: Swapout/swapin THP in one piece References: <20180523082625.6897-1-ying.huang@intel.com> <20180604180642.qexvwe5dqvkgraij@ca-dmjordan1.us.oracle.com> Date: Tue, 05 Jun 2018 12:30:13 +0800 In-Reply-To: <20180604180642.qexvwe5dqvkgraij@ca-dmjordan1.us.oracle.com> (Daniel Jordan's message of "Mon, 4 Jun 2018 11:06:42 -0700") Message-ID: <87lgbt3ley.fsf@yhuang-dev.intel.com> User-Agent: Gnus/5.13 (Gnus v5.13) Emacs/25.2 (gnu/linux) MIME-Version: 1.0 Content-Type: text/plain; charset=ascii Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Daniel Jordan writes: > On Wed, May 23, 2018 at 04:26:04PM +0800, Huang, Ying wrote: >> And for all, Any comment is welcome! >> >> This patchset is based on the 2018-05-18 head of mmotm/master. > > Trying to review this and it doesn't apply to mmotm-2018-05-18-16-44. git > fails on patch 10: > > Applying: mm, THP, swap: Support to count THP swapin and its fallback > error: Documentation/vm/transhuge.rst: does not exist in index > Patch failed at 0010 mm, THP, swap: Support to count THP swapin and its fallback > > Sure enough, this tag has Documentation/vm/transhuge.txt but not the .rst > version. Was this the tag you meant? If so did you pull in some of Mike > Rapoport's doc changes on top? I use the mmotm tree at git://git.cmpxchg.org/linux-mmotm.git Maybe you are using the other one? >> base optimized >> ---------------- -------------------------- >> %stddev %change %stddev >> \ | \ >> 1417897 2% +992.8% 15494673 vm-scalability.throughput >> 1020489 4% +1091.2% 12156349 vmstat.swap.si >> 1255093 3% +940.3% 13056114 vmstat.swap.so >> 1259769 7% +1818.3% 24166779 meminfo.AnonHugePages >> 28021761 -10.7% 25018848 2% meminfo.AnonPages >> 64080064 4% -95.6% 2787565 33% interrupts.CAL:Function_call_interrupts >> 13.91 5% -13.8 0.10 27% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath >> > ...snip... >> test, while in optimized kernel, that is 96.6%. The TLB flushing IPI >> (represented as interrupts.CAL:Function_call_interrupts) reduced >> 95.6%, while cycles for spinlock reduced from 13.9% to 0.1%. These >> are performance benefit of THP swapout/swapin too. > > Which spinlocks are we spending less time on? "perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.mem_cgroup_commit_charge.do_swap_page.__handle_mm_fault": 4.39, "perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_pcppages_bulk.drain_pages_zone.drain_pages": 1.53, "perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.get_page_from_freelist.__alloc_pages_slowpath.__alloc_pages_nodemask": 1.34, "perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.swapcache_free_entries.free_swap_slot.do_swap_page": 1.02, "perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.shrink_inactive_list.shrink_node_memcg.shrink_node": 0.61, "perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.shrink_active_list.shrink_node_memcg.shrink_node": 0.54, Best Regards, Huang, Ying