Received: by 2002:a05:6a10:16a7:0:0:0:0 with SMTP id gp39csp1073887pxb; Thu, 5 Nov 2020 23:32:09 -0800 (PST) X-Google-Smtp-Source: ABdhPJzlFBxsJu3isvcV9kSCLg4MF+13dYDFST+KZaswjvuTT0wXoiUtkBjExyOkQ5i6os7D4ylK X-Received: by 2002:a50:ff02:: with SMTP id a2mr686210edu.364.1604647929120; Thu, 05 Nov 2020 23:32:09 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1604647929; cv=none; d=google.com; s=arc-20160816; b=0VkYLMJES7xxP0bjB/J8+C3+SD6+cC9gcfTMIWnz5OsseCrNJCw/MM+21y3e7fVD3M na74L3evcbbWGqZj+MtZO18GwoCepNtarGgYSzK/OoW7fyuJHYM7dWZ60sPj1wIcpdya 5vBXKrD+TDtu4En9jxH9qV7f8OE1iZ88RAFJMwluysT28BNBhU25U8ixQ7BTjL7FrlBz xCRDTzvdrvxMOnvnsa1zk3NF9FWpwvAoMa0y4o/Y7bwywA1vc2A72asA6PbxeKzDqaGB +a1YLf/RV4v1/yBOmwEED3U5WK2rSOszIgR3VTwc9vyq6BhTzO5eyqREfZgZJVW/SF3p UyCg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:mime-version:user-agent:message-id:in-reply-to :date:references:subject:cc:to:from:ironport-sdr:ironport-sdr; bh=QIcQknv37a5fyq4h9ZYnl4PQcQJIq1dr7rT31c4W72c=; b=K+abrbqt+JuneiBIEWgzr2H6puY6dZ8wBGErwwv9QDEY0F8Z8HMi6gga8aVqdvN8Zq 83I0OlH0N/DwEvv9dxbS6LQ2IR7wdAFlXFJt698bcq7j7aHMhfTt8At2RY2yAnaRFxlR IKXP04PORps5OpU/mO+qkYzhPLxghGKlCzxvotH9hYix9AFUllYE5BUBQv0JsIpqFfgu qYA7on958EsSEGF1Fs2xCwrnyfEmqoYR2yi/TQCvGQSNU8F3L2m7WjNVT9OTyBRat1h+ VcilEalmZObyhP6dd/Nw7/Uh0nhl8LsRCpLc3mxmOmswTrAZOY0PlrGu6HxbU2araS0X RKFg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id y10si280598ejq.401.2020.11.05.23.31.45; Thu, 05 Nov 2020 23:32:09 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726438AbgKFH3E (ORCPT + 99 others); Fri, 6 Nov 2020 02:29:04 -0500 Received: from mga18.intel.com ([134.134.136.126]:30669 "EHLO mga18.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725830AbgKFH3E (ORCPT ); Fri, 6 Nov 2020 02:29:04 -0500 IronPort-SDR: QUl+KxmImdBiN9AZAriuCLrFgp60TwRXhDaXI1PuHYzIf9skOEoVEvA4TgoLmg3hLsZs/8Df9l BiBUbgL0BAQA== X-IronPort-AV: E=McAfee;i="6000,8403,9796"; a="157292853" X-IronPort-AV: E=Sophos;i="5.77,455,1596524400"; d="scan'208";a="157292853" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga004.fm.intel.com ([10.253.24.48]) by orsmga106.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 05 Nov 2020 23:29:03 -0800 IronPort-SDR: 7V1IJ6pjk8oQLtLEcGW9OjlhbuwWfMdcaCDbIGN/97HD6vuY+I3SKCm+h2OvjHDbOkruujbqKH s65cIa9Mlwkg== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.77,455,1596524400"; d="scan'208";a="354616998" Received: from yhuang-dev.sh.intel.com (HELO yhuang-dev) ([10.239.159.65]) by fmsmga004.fm.intel.com with ESMTP; 05 Nov 2020 23:29:00 -0800 From: "Huang\, Ying" To: Mel Gorman Cc: Peter Zijlstra , , , Andrew Morton , Ingo Molnar , Rik van Riel , "Johannes Weiner" , "Matthew Wilcox \(Oracle\)" , Dave Hansen , Andi Kleen , Michal Hocko , David Rientjes , ben.widawsky@intel.com Subject: Re: [PATCH -V2 2/2] autonuma: Migrate on fault among multiple bound nodes References: <20201028023411.15045-1-ying.huang@intel.com> <20201028023411.15045-3-ying.huang@intel.com> <20201102111717.GB3306@suse.de> <87eel9wumd.fsf@yhuang-dev.intel.com> <20201105112523.GQ3306@suse.de> Date: Fri, 06 Nov 2020 15:28:59 +0800 In-Reply-To: <20201105112523.GQ3306@suse.de> (Mel Gorman's message of "Thu, 5 Nov 2020 11:25:23 +0000") Message-ID: <87v9ejosec.fsf@yhuang-dev.intel.com> User-Agent: Gnus/5.13 (Gnus v5.13) Emacs/26.3 (gnu/linux) MIME-Version: 1.0 Content-Type: text/plain; charset=ascii Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Mel Gorman writes: > On Wed, Nov 04, 2020 at 01:36:58PM +0800, Huang, Ying wrote: >> But from another point of view, I suggest to remove the constraints of >> MPOL_F_MOF in the future. If the overhead of AutoNUMA isn't acceptable, >> why not just disable AutoNUMA globally via sysctl knob? >> > > Because it's a double edged sword. NUMA Balancing can make a workload > faster while still incurring more overhead than it should -- particularly > when threads are involved rescanning the same or unrelated regions. > Global disabling only really should happen when an application is running > that is the only application on the machine and has full NUMA awareness. Got it. So NUMA Balancing may in generally benefit some workloads but hurt some other workloads on one machine. So we need a method to enable/disable NUMA Balancing for one workload. Previously, this is done via the explicit NUMA policy. If some explicit NUMA policy is specified, NUMA Balancing is disabled for the memory region or the thread. And this can be reverted again for a memory region via MPOL_MF_LAZY. It appears that we lacks MPOL_MF_LAZY for the thread yet. >> > It might still end up being better but I was not aware of a >> > *realistic* workload that binds to multiple nodes >> > deliberately. Generally I expect if an application is binding, it's >> > binding to one local node. >> >> Yes. It's not popular configuration for now. But for the memory >> tiering system with both DRAM and PMEM, the DRAM and PMEM in one socket >> will become 2 NUMA nodes. To avoid too much cross-socket memory >> accessing, but take advantage of both the DRAM and PMEM, the workload >> can be bound to 2 NUMA nodes (DRAM and PMEM). >> > > Ok, that may lead to unpredictable performance as it'll have variable > performance with limited control of the "important" applications that > should use DRAM over PMEM. That's a long road but the step is not > incompatible with the long-term goal. Yes. Ben Widawsky is working on a patchset to make it possible to prefer the remote DRAM instead of the local PMEM as follows, https://lore.kernel.org/linux-mm/20200630212517.308045-1-ben.widawsky@intel.com/ Best Regards, Huang, Ying