Received: by 2002:ac0:a5b6:0:0:0:0:0 with SMTP id m51-v6csp718559imm; Fri, 8 Jun 2018 04:15:47 -0700 (PDT) X-Google-Smtp-Source: ADUXVKL4vTMAWkh+na53PjyO0643EazU2iSDPC3lpJxLVRQynZ8ksj0CWmVAKBselX9dc2fZxkfM X-Received: by 2002:a63:770b:: with SMTP id s11-v6mr4934430pgc.339.1528456547222; Fri, 08 Jun 2018 04:15:47 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1528456547; cv=none; d=google.com; s=arc-20160816; b=OxfTtLtPGNxumxR/mLnh0Z3zhfpLG0z3Q5fhOsAPurZVlWSy/npp0pQjtscHv5aPza LyXHTx5aKNyk5SgJ6CsoEodlYD3kYy02i0fqd+u+ZT8cczROrTTgxv0n0I9c2Xw70BMz XmjbbpAZV0f5YOcw46ipEXFvcoImITGVNxJwMQXZbuRV+TVRWmb6YmbPCoEtDF45oVa0 alBAd8ubHdxakKdf9834tRdUpxAzGi9NohgraWJLwd8jIRLamwLJFs6aQ8ibQ1xLC3sX dQWLLkuxYEg3H/bafkSz6u96FKZExlCqJkWqIzyKC/wohKfeHw8F6geGlxh8FsaVG0QH MxzA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:arc-authentication-results; bh=dk9paQV5usbH6NLvvr8u672mg7oPg9A0a4BPOjPa8Ek=; b=D+3EFCpg81nEmHCOAbjGLWxN1S6hSroruqs3jJZiv6KjKojCaO4fxScP8XJYVpDRJe Tx2Ev/d695mX4IcsAqRd1vqBhWMyB5DgYmbjRT5UvPdeslg8mzQ4JSiG8Q/74WmZBf0y QMEdqq8JFBbshDksBX/v8ZLUKrhjUEKS50mCkHZ/qmPj0GjEmvH098YwwrjalJdSmI0y ME554NufYT06EWHmqINcwgRpfKTGlw2/HsgerIxnXR5WjjpIQ+q0wspAhjlIuyTriKf/ aDAueMk+W47xFCbrqQBYI2kWsobDI7u5SDbONTh6H3ogEb+NKYFp48dfbm3iV3s+uWCu ZVrg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id r17-v6si29370515pfg.305.2018.06.08.04.15.32; Fri, 08 Jun 2018 04:15:47 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751486AbeFHLPI (ORCPT + 99 others); Fri, 8 Jun 2018 07:15:08 -0400 Received: from outbound-smtp25.blacknight.com ([81.17.249.193]:58502 "EHLO outbound-smtp25.blacknight.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751108AbeFHLPH (ORCPT ); Fri, 8 Jun 2018 07:15:07 -0400 Received: from mail.blacknight.com (pemlinmail04.blacknight.ie [81.17.254.17]) by outbound-smtp25.blacknight.com (Postfix) with ESMTPS id E2975B870E for ; Fri, 8 Jun 2018 12:15:05 +0100 (IST) Received: (qmail 1725 invoked from network); 8 Jun 2018 11:15:05 -0000 Received: from unknown (HELO techsingularity.net) (mgorman@techsingularity.net@[37.228.237.171]) by 81.17.254.9 with ESMTPSA (DHE-RSA-AES256-SHA encrypted, authenticated); 8 Jun 2018 11:15:05 -0000 Date: Fri, 8 Jun 2018 12:15:05 +0100 From: Mel Gorman To: Jirka Hladky Cc: Jakub Racek , linux-kernel , "Rafael J. Wysocki" , Len Brown , linux-acpi@vger.kernel.org Subject: Re: [4.17 regression] Performance drop on kernel-4.17 visible on Stream, Linpack and NAS parallel benchmarks Message-ID: <20180608111505.qlru36ycih6eltqc@techsingularity.net> References: <20180606122731.GB27707@jra-laptop.brq.redhat.com> <20180607123915.avrqbpp4adgj7ck4@techsingularity.net> <20180608074057.jtxczsw3jwx6boti@techsingularity.net> <20180608092451.mwzr6pvxh2cprzju@techsingularity.net> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-15 Content-Disposition: inline In-Reply-To: User-Agent: NeoMutt/20170912 (1.9.0) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Jun 08, 2018 at 01:02:54PM +0200, Jirka Hladky wrote: > > > > Unknown and unknowable. It depends entirely on the reference pattern of > > the different threads. If they are fully parallelised with private buffers > > that are page-aligned then I expect it to be quick (to pass the 2-reference > > filter). > > > I'm running 20 parallel processes. There is no connection between them. If > I read it correctly the migration should happen fast in this case, right? > > I have checked the source code and variables are global and static (and > thus allocated in the data segment). They are NOT 4k aligned: > > variable a is at address: 0x9e999e0 > variable b is at address: 0x524e5e0 > variable c is at address: 0x6031e0 > > static double a[N], > b[N], > c[N]; > If these are 20 completely indepent processes (and not sharing data via MPI if you're using that version of STREAM) then the migration should be relatively quick. Migrations should start within 3 seconds of the process starting. How long it takes depends on the size of the STREAM processes as it's only scanned in chunks and migrations won't start until there are two full passes of the address space. You can partially monitor the progress using /proc/pid/numa_maps. More detailed monitoring needs ftrace for some activity and the use of probes on specific functions to get detailed information. It may also be worth examining /proc/pid/sched and seeing if a task sets numa_preferred_nid to node 0 and keeps it there even after migrating to node 1 but that's doubtful. -- Mel Gorman SUSE Labs