Received: by 2002:ad5:474a:0:0:0:0:0 with SMTP id i10csp9918830imu; Wed, 5 Dec 2018 12:35:09 -0800 (PST) X-Google-Smtp-Source: AFSGD/XlsAgdA/ujk+aVPH+6qn1g7iLMPUmLH8yCzTBPxR2/rlSRJtioDTegg3ek+Y/yLIi/zPCh X-Received: by 2002:a65:62da:: with SMTP id m26mr21841264pgv.278.1544042109485; Wed, 05 Dec 2018 12:35:09 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1544042109; cv=none; d=google.com; s=arc-20160816; b=EOlOUThM1b6DcIgZ1e2o1iN9XOpuRtOydMA43HORHhEf9OdAi8i/01cNZwIuonGTV2 AaWOLBQGXAvXJVrdnLR6TviQzhVwM/z2OCKl+9f2B6d7u0S/HZJFYsiXGT1T4U6RuFMk UJYi7+REI/7E7lc4Pi/VGaMUBvwuMMrwUP/Uheyn+vp0uWYtJp8VZaqgV8U6zfXJX42B NUswT8mhk20itADCNNEjj+36L8kxlGMsudVqoFaFpTekm3uXAFqpI+eSMAM2LnNUso62 okhVBMBkkmz+jSkb9FosbjEq+/Q0XdvE+8IV+uAerSV+b5Rn65Unl95BB+POr0pEzROk zf7A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date; bh=ZJDLSqtkKAtt7hlopjg9XH0ZAHHZr+8Bd+z+zwP4OcE=; b=yHpWOH2DZghZPlNyLIG7tiBr2sUe71unVRZn8edL8meQpP2mg5TWI3PRwjsU4+3Gfj fiVG237QzamBBjVkwEzctD/l+CTs2Pj2N40ohqBGNcco7Tub/klX3Dft94Yhwma4uxkx ZvFKTS6hnQQ5jDeSt4ZNuYxBxQSmDXro+AcSfwFyIyMGq0Kds+qbNHXXPuBf1gbK3zbx uPBtbZYQy8JYiPo2H6bsMeLNEpk+meKfKr8OzZQyU694Ykki7EbkKQPZyuYs5Snyepez Gh3NJgOv3OQoGXVyymiA/NB2MsV8XyDVp5Iq6Syutb7FRdiFldwG+s3s+0Il45wfWI+3 l76Q== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id a19si18604216pgj.429.2018.12.05.12.34.54; Wed, 05 Dec 2018 12:35:09 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728553AbeLEUcs (ORCPT + 99 others); Wed, 5 Dec 2018 15:32:48 -0500 Received: from mx2.suse.de ([195.135.220.15]:54326 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1727388AbeLEUcs (ORCPT ); Wed, 5 Dec 2018 15:32:48 -0500 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay2.suse.de (unknown [195.135.220.254]) by mx1.suse.de (Postfix) with ESMTP id CBC3EAEFC; Wed, 5 Dec 2018 20:32:45 +0000 (UTC) Date: Wed, 5 Dec 2018 21:32:43 +0100 From: Michal Hocko To: David Rientjes Cc: Vlastimil Babka , Linus Torvalds , Andrea Arcangeli , ying.huang@intel.com, s.priebe@profihost.ag, mgorman@techsingularity.net, Linux List Kernel Mailing , alex.williamson@redhat.com, lkp@01.org, kirill@shutemov.name, Andrew Morton , zi.yan@cs.rutgers.edu Subject: Re: [patch 0/2 for-4.20] mm, thp: fix remote access and allocation regressions Message-ID: <20181205203243.GX1286@dhcp22.suse.cz> References: <20181205090554.GX1286@dhcp22.suse.cz> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed 05-12-18 11:49:26, David Rientjes wrote: > On Wed, 5 Dec 2018, Michal Hocko wrote: > > > > The revert is certainly needed to prevent the regression, yes, but I > > > anticipate that Andrea will report back that patch 2 at least improves the > > > situation for the problem that he was addressing, specifically that it is > > > pointless to thrash any node or reclaim unnecessarily when compaction has > > > already failed. This is what setting __GFP_NORETRY for all thp fault > > > allocations fixes. > > > > Yes but earlier numbers from Mel and repeated again [1] simply show > > that the swap storms are only handled in favor of an absolute drop of > > THP success rate. > > > > As we've been over countless times, this is the desired effect for > workloads that fit on a single node. We want local pages of the native > page size because they (1) are accessed faster than remote hugepages and > (2) are candidates for collapse by khugepaged. > > For applications that do not fit in a single node, we have discussed > possible ways to extend the API to allow remote faulting of hugepages, > absent remote fragmentation as well, then the long-standing behavior is > preserved and large applications can use the API to increase their thp > success rate. OK, I just give up. This doesn't lead anywhere. You keep repeating the same stuff over and over, neglect other usecases and actually force them to do something special just to keep your very specific usecase which you clearly refuse to abstract into a form other people can experiment with or at least provide more detailed broken down numbers for a more serious analyses. Fault latency is only a part of the picture which is much more complex. Look at Mel's report to get an impression of what might be really useful for a _productive_ discussion. -- Michal Hocko SUSE Labs