Received: by 2002:a25:683:0:0:0:0:0 with SMTP id 125csp365508ybg; Wed, 3 Jun 2020 02:50:30 -0700 (PDT) X-Google-Smtp-Source: ABdhPJy1FCsY85DEBX13Ewi5lNjIu39laK1buxySZsOwVdpGxiL0fDc0uHqfF5uHVz4MNSxlYAl/ X-Received: by 2002:a17:906:4356:: with SMTP id z22mr26815403ejm.334.1591177829874; Wed, 03 Jun 2020 02:50:29 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1591177829; cv=none; d=google.com; s=arc-20160816; b=WRbAEeOiKmrTpZoeuOtxR4DhLHNOX+NwKts3ULxxM/7iBKYonTmKHtsblrPoS28LfU VQ4iLFZi2URKAIA0TkzuzTFK6QNPzcmrAhNF2tPZkkhWSvjAKzQAqTLbzcwqK9ZZrQDJ w+c+IVoooLV2MfF0MfbQWn0O1ruWXrUKwcR33B62y6yUOQ8XBWsrpolQJXk0GpHhA4Xb miasuxJD4VxtfKAvtiS9AIRDPMTJButc0v84ghQoZ4fbVi5x9gzjlLxV217ncd0lleBY ut72g4pW2zIIANBNX/Q6hgp+tvV3UTyobrFDUIgx8yWEPBpPVNqrrwnshfSbZtfbtPD4 PApw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-transfer-encoding:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:ironport-sdr :ironport-sdr; bh=CAsIsw0AHVRKP0GKV6BcFWW336su1vZo4sQv3Bwlp/U=; b=oWIsPMDON/ihlj7lTF1JPeRChNCb2y2vH/25C1TkdJCug4eE5uQ4Sp5d8/O+EzQwls 9NvwgLV7hYli940NQmoxGvf25qkGcXDsxOAieP4o9/e90yswZ9Y2pVeivsfuUoUMTWwL ysS+O8tqnNfwqGpJBpON/qoRiIA2GAPAEdaCeHBxKxSyxFBI+Zi26wVZGOAmQYLVFy9O LgAkf5CxjnOITb9LYUq7i9e5BRgth7GEnY2I2UI2+gIVYAgrN64/mqQmy1SW2W5lZaTx JtX+ecLEVryp8izl5hQuqyWAmNZqs8imMrSIOtc4giBey0gBQOCMRZZdkZc7uTGZwxqo CeqA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id y92si847665edy.117.2020.06.03.02.50.06; Wed, 03 Jun 2020 02:50:29 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726295AbgFCJsJ (ORCPT + 99 others); Wed, 3 Jun 2020 05:48:09 -0400 Received: from mga11.intel.com ([192.55.52.93]:12471 "EHLO mga11.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725854AbgFCJsJ (ORCPT ); Wed, 3 Jun 2020 05:48:09 -0400 IronPort-SDR: vRUeNr66CzYmMLVANaRq0+HxetS9c7KwSVlWVVLI0kYRKFRYF9iI3mBJrrbFb5zGb/Z7Qc0nRX 6h0rak4bR2yQ== X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga001.fm.intel.com ([10.253.24.23]) by fmsmga102.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 03 Jun 2020 02:48:08 -0700 IronPort-SDR: BkwBQ/BHqe1tonku3ksMYgxt4OD8StANvjvo9vJtQJn4ihY0f+YLmjOugXm+LeJNDB1EUNFCrz wef8VnH/rkew== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.73,467,1583222400"; d="scan'208";a="378049486" Received: from shbuild999.sh.intel.com (HELO localhost) ([10.239.146.107]) by fmsmga001.fm.intel.com with ESMTP; 03 Jun 2020 02:48:04 -0700 Date: Wed, 3 Jun 2020 17:48:04 +0800 From: Feng Tang To: Qian Cai Cc: Andrew Morton , Michal Hocko , Johannes Weiner , Matthew Wilcox , Mel Gorman , Kees Cook , andi.kleen@intel.com, tim.c.chen@intel.com, dave.hansen@intel.com, ying.huang@intel.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org, Konstantin Khlebnikov Subject: Re: [PATCH v4 3/4] mm/util.c: remove the VM_WARN_ONCE for vm_committed_as underflow check Message-ID: <20200603094804.GB89848@shbuild999.sh.intel.com> References: MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: User-Agent: Mutt/1.5.24 (2015-08-30) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Jun 02, 2020 at 12:02:22AM -0400, Qian Cai wrote: > > > > On Jun 1, 2020, at 11:37 PM, Feng Tang wrote: > > > > I re-run the same benchmark with v5.7 and 5.7+remove_warning kernels, > > the overall performance change is trivial (which is expected) > > > > 1330147 +0.1% 1331032 will-it-scale.72.processes > > > > But the perf stats of "self" shows big change for __vm_enough_memory() > > > > 0.27 -0.3 0.00 pp.self.__vm_enough_memory > > > > I post the full compare result in the end. > > I don’t really see what that means exactly, but I suppose the warning is there for so long and no one seems notice much trouble (or benefit) because of it, so I think you will probably need to come up with a proper justification to explain why it is a trouble now, and how your patchset suddenly start to trigger the warning as well as why it is no better way but to suffer this debuggability regression (probably tiny but still). Thanks for the suggestion, and I updated the commit log. From 1633da8228bd3d0dcbbd8df982977ad4594962a1 Mon Sep 17 00:00:00 2001 From: Feng Tang Date: Fri, 29 May 2020 08:48:48 +0800 Subject: [PATCH] mm/util.c: remove the VM_WARN_ONCE for vm_committed_as underflow check This check was added by 82f71ae4a2b8 ("mm: catch memory commitment underflow") in 2014 to have a safety check for issues which have been fixed. And there has been few report caught by it, as described in its commit log: : This shouldn't happen any more - the previous two patches fixed : the committed_as underflow issues. But it was really found by Qian Cai when he used the LTP memory stress suite to test a RFC patchset, which tries to improve scalability of per-cpu counter 'vm_committed_as', by chosing a bigger 'batch' number for loose overcommit policies (OVERCOMMIT_ALWAYS and OVERCOMMIT_GUESS), while keeping current number for OVERCOMMIT_NEVER. With that patchset, when system firstly uses a loose policy, the 'vm_committed_as' count could be a big negative value, as its big 'batch' number allows a big deviation, then when the policy is changed to OVERCOMMIT_NEVER, the 'batch' will be decreased to a much smaller value, thus hits this WARN check. To mitigate this, one proposed solution is to queue work on all online CPUs to do a local sync for 'vm_committed_as' when changing policy to OVERCOMMIT_NEVER, plus some global syncing to garante the case won't be hit. But this solution is costy and slow, given this check hasn't shown real trouble or benefit, simply drop it from one hot path of MM. And perf stats does show some tiny saving for removing it. Reported-by: Qian Cai Signed-off-by: Feng Tang Cc: Konstantin Khlebnikov Cc: Michal Hocko Cc: Andi Kleen --- mm/util.c | 8 -------- 1 file changed, 8 deletions(-) diff --git a/mm/util.c b/mm/util.c index 9b3be03..c63c8e4 100644 --- a/mm/util.c +++ b/mm/util.c @@ -814,14 +814,6 @@ int __vm_enough_memory(struct mm_struct *mm, long pages, int cap_sys_admin) { long allowed; - /* - * A transient decrease in the value is unlikely, so no need - * READ_ONCE() for vm_committed_as.count. - */ - VM_WARN_ONCE(data_race(percpu_counter_read(&vm_committed_as) < - -(s64)vm_committed_as_batch * num_online_cpus()), - "memory commitment underflow"); - vm_acct_memory(pages); /* -- 2.7.4