Received: by 2002:a05:6a10:8c0a:0:0:0:0 with SMTP id go10csp3586496pxb; Mon, 25 Jan 2021 22:26:54 -0800 (PST) X-Google-Smtp-Source: ABdhPJyXDGF5tO9hiIev4Rs+VK0JSx0tDWcb5FNv89vzEbsz22ELkjP0K5Y26jqLtvwmSGZCPuPK X-Received: by 2002:a17:907:1b06:: with SMTP id mp6mr2551699ejc.408.1611642414641; Mon, 25 Jan 2021 22:26:54 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1611642414; cv=none; d=google.com; s=arc-20160816; b=WltQlPOygl5yl3sP9dxTp77nAPZYi3TRtV88mpG5HdPMWq3HP0uxsxBQ5ZkbizvWiE kEY6mdJGJisvjVW7iolFu+cUrhKrug1DYRo/4PcWDRqoUDqpyK5rdBCovEjNRcEceVeO WtpiERE4wqCKEbx9IQTu89+YqH8rompzmvaYM78lZYHdZCTMnuXO/zWy4k+GNKoPFee3 TlD4OHOvsSgStKvYSjWzNZ1CwpjE9eAqmSp0LMCWvoyiMfphwxdkU+Z/0nWmDmDdWyfA GC1etY130BuWZQnpYXd3hvsOoW7KdOf+he17CX1zSgN//1u0J3gE4Ip2RKAmHBWy9wtf WYRg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:mime-version:user-agent:references:message-id :in-reply-to:subject:cc:to:from:date; bh=3FZ+q3OiMg9NoEfCa+hsSuedqhN7pVZJIbCAcuWQUzo=; b=d3rteWQKoXOg5zHLM+bpwEiB33qbwbiqsss0SLIQclHxI3rEXbPlUcXmL+9etV3nPS WXbHN53CXWz1FGZu23rf3JCM1X1j29+OZVFUXk7wyGsXZTwU6KbGfogH8HNwd4IrEn8G G1jiRs7ZI+IKFb8aeH2lKyR29eP4kyHG0vdXJypFxs53ajVxC9h7mVa9SvDBDu6EetiK WIAr3gMdPJ2iuxrzu/bzIoM1s5QXa7KVegTzgVJNLlOGS6KRFNPDRb0R4AMVYBUk4svc 5MegenUzZwkbGK4tzXQzssjV0wSNVl3eBd44FdeNdo6ed1yZGpxFyC7jL43b9pe9RBpF QR0A== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id lc20si6722844ejc.253.2021.01.25.22.26.31; Mon, 25 Jan 2021 22:26:54 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2387702AbhAZGXq (ORCPT + 99 others); Tue, 26 Jan 2021 01:23:46 -0500 Received: from mail2-relais-roc.national.inria.fr ([192.134.164.83]:11568 "EHLO mail2-relais-roc.national.inria.fr" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1728987AbhAYNjg (ORCPT ); Mon, 25 Jan 2021 08:39:36 -0500 X-IronPort-AV: E=Sophos;i="5.79,373,1602540000"; d="scan'208";a="488960900" Received: from clt-128-93-178-244.vpn.inria.fr ([128.93.178.244]) by mail2-relais-roc.national.inria.fr with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 25 Jan 2021 14:38:36 +0100 Date: Mon, 25 Jan 2021 14:38:35 +0100 (CET) From: Julia Lawall X-X-Sender: jll@hadrien To: Vincent Guittot cc: Mel Gorman , Ingo Molnar , Peter Zijlstra , Juri Lelli , Dietmar Eggemann , Steven Rostedt , Ben Segall , Daniel Bristot de Oliveira , linux-kernel Subject: Re: [PATCH v2] sched/fair: check for idle core In-Reply-To: Message-ID: References: <1603372550-14680-1-git-send-email-Julia.Lawall@inria.fr> <20201027091936.GS32041@suse.de> <20210125091238.GE20777@suse.de> User-Agent: Alpine 2.22 (DEB 394 2020-01-19) MIME-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, 25 Jan 2021, Vincent Guittot wrote: > On Mon, 25 Jan 2021 at 10:20, Julia Lawall wrote: > > > > > > > > On Mon, 25 Jan 2021, Mel Gorman wrote: > > > > > On Sun, Jan 24, 2021 at 09:38:14PM +0100, Julia Lawall wrote: > > > > > > > > > > > > On Tue, 27 Oct 2020, Mel Gorman wrote: > > > > > > > > > On Thu, Oct 22, 2020 at 03:15:50PM +0200, Julia Lawall wrote: > > > > > > Fixes: 11f10e5420f6 ("sched/fair: Use load instead of runnable load in wakeup path") > > > > > > Signed-off-by: Julia Lawall > > > > > > Reviewed-by Vincent Guittot > > > > > > > > > > > > > > > > While not a universal win, it was mostly a win or neutral. In few cases > > > > > where there was a problem, one benchmark I'm a bit suspicious of generally > > > > > as occasionally it generates bad results for unknown and unpredictable > > > > > reasons. In another, it was very machine specific and the differences > > > > > were small in absolte time rather than relative time. Other tests on the > > > > > same machine were fine so overall; > > > > > > > > > > Acked-by: Mel Gorman > > > > > > > > Recently, we have been testing the phoronix multicore benchmarks. On v5.9 > > > > with this patch, the preparation time of phoronix slows down, from ~23 > > > > seconds to ~28 seconds. In v5.11-rc4, we see 29 seconds. It's not yet > > > > clear what causes the problem. But perhaps the patch should be removed > > > > from v5.11, until the problem is understood. > > > > > > > > commit d8fcb81f1acf651a0e50eacecca43d0524984f87 > > > > > > > > > > I'm not 100% convinved given that it was a mix of wins and losses. In > > > the wakup path in general, universal wins almost never happen. It's not > > > 100% clear from your mail what happens during the preparation patch. If > > > it included time to download the benchmarks and install then it would be > > > inherently variable due to network time (if download) or cache hotness > > > (if installing/compiling). While preparation time can be interesting -- > > > for example, if preparation involves reading a lot of files from disk, > > > it's not universally interesting when it's not the critical phase of a > > > benchmark. > > > > The benchmark is completely downloaded prior to the runs. There seems to > > be some perturbation to the activation of containerd. Normally it is > > even: * * * * > > Does it impact the benchmark results too or only the preparation prior > to running the benchmark ? Looking at a few of the benchmarks, there is no clear pattern which is better. But there is not a big degradation, like from 23 to 28 seconds for the preparation time. I will report back when we figure out more. julia > > > > > and with the patch it becomes more like: * ** ** > > > > That is every other one is on time, and every other one is late. > > > > But I don't know why this happens. > > > > julia > > > > > > > > I think it would be better to wait until the problem is fully understood > > > to see if it's a timing artifact (e.g. a race between when prev_cpu is > > > observed to be idle and when it is busy). > > I agree that a better understanding of what is happening is necessary > before any changes > > > > > > > -- > > > Mel Gorman > > > SUSE Labs > > > >