Received: by 2002:a05:6a10:f347:0:0:0:0 with SMTP id d7csp3293160pxu; Tue, 8 Dec 2020 08:19:59 -0800 (PST) X-Google-Smtp-Source: ABdhPJzCzwkYe86m9Ryrx7jKuW+kpXX19GOEOcCDIip4cWMIlEhKoewIw4weFGMJski/8EWmMH/t X-Received: by 2002:a17:906:1294:: with SMTP id k20mr23433955ejb.404.1607444398848; Tue, 08 Dec 2020 08:19:58 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1607444398; cv=none; d=google.com; s=arc-20160816; b=c3ruMO/WE4O5L4oiin20PPRMRvhwEv+zdEKbohax8nN1wMfdycN8bk58XPn8QHNJmY 4ytKvCeLcHCeKWzSVQ/z70MSikDqR9sOxo1Ybq2kutx8z4YQ7q6IAWNvUh7OGe4r/2BB Zw0Uow8E+AmCKNWGO8AxYL1D6TvJntj/y3US0Etp45vaT9J+zS9nmEUkH36k7vvGvLUI tGsQpS7DJFyvB2UbB9hKC9SoJ2b8tEJPs/jW9TBk9QGyIndvZi2XrysWP7xvGe4HbrI/ n5WUjV2eMD55HwPP96lWadPB7sJtBXeYO9S49KXoxwZn65TA+V4/fyQww4jWCPKbUbhT hq0g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=jvK/rcBZ+ncoq0z4m0U1M9RN+o8FAfMKBetL37kSxrw=; b=L66R5Gk3BiCRSo/fbDkoZABwNPEgxUFyfUmdRGhwy4hhJyjRf8BzvtzJW6NDeG6bNt wfg8JYwXAxCNf1ROf/yCQKuSoiplfE2F5L/2vcQbr40eXdfBdiJdLncHq1Rxpf/qhV2/ e4FVjvpNG/HKnpyJU3W8G6mZVD49F+iyclqrs67Hx3XCNv21b0xpTkrjcLAsNLbD1BM/ GTSWjm5Zk3S5wdhoHCeBMRQVkPUrZUKp9JSYKfccl9u/WzQkJIg3Sd1nI1tej5UQRCne /0R09plJIXHUC2Upkhe7egZWZgTEeJq9enHmP/OVXH+ke12d7MSDh+bstZ5jHa7g2rTz +Y2g== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=kyRy8HCA; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id u9si10074598edb.338.2020.12.08.08.19.32; Tue, 08 Dec 2020 08:19:58 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=kyRy8HCA; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728679AbgLHQPb (ORCPT + 99 others); Tue, 8 Dec 2020 11:15:31 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:53942 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726080AbgLHQPa (ORCPT ); Tue, 8 Dec 2020 11:15:30 -0500 Received: from mail-lj1-x243.google.com (mail-lj1-x243.google.com [IPv6:2a00:1450:4864:20::243]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 6EA27C061749 for ; Tue, 8 Dec 2020 08:14:50 -0800 (PST) Received: by mail-lj1-x243.google.com with SMTP id x23so13336879lji.7 for ; Tue, 08 Dec 2020 08:14:50 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=jvK/rcBZ+ncoq0z4m0U1M9RN+o8FAfMKBetL37kSxrw=; b=kyRy8HCAMatgEuN8e9NTUBjdlAHxTLViLrlRLxAKVmGBr6I/Zo46rojBGVq6WliZ9D I2/Vqfh+vQczS0j7bPP4nibN18cOx4iDO3r+6dnDbGfp1wxGKXWZX8rjLtPX6mL7woJE Cn3TS4FWsE24BabWAvSKfki4qVOL15vsHDIc7IWzhimo+EiiD+Sw4o1XxlTd4l23rFFm UJi85xP7eHWTU7v12jct1B0wVSenqTaoIfhr86+GqRwa3uG1x6rTH/+3zqdS0rSFL7JY 4u4QK8H/lFs+78UaxsEDSP5YmUQ1doO/I0VX0dsp7sOHD9cwu8x0gQ0zi/UWFpU7Bgqq rpOA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=jvK/rcBZ+ncoq0z4m0U1M9RN+o8FAfMKBetL37kSxrw=; b=uPDOPXzrw58J3nYDyuMQ2nHnwNe73/GGfiuxYUoxkmArT81U8PBdBiAj7Z7SwZNa/3 aG8i7A3sRrAtJNzsLZCE+104e5CQ8LfuWMwqUsU+ZHgaMmP+t2kH5pNhA0V6lZzVt2PE K39eHCrj4KTrlMZ5gr9gcJaoE2dTp28ZJUK9skLDtrTTPsOO2kNFBWoafx7tWRsKkK7a 28QMTXq769B82PnZoVqSuLVmTgMEgpgh/qoh2r8EKqUy/JI5I4MLuz8A9BxBlv6VttAX ziL7WDmjdlos+I5+essGKkatOjtSsuEuOupRN4ee0o6dt6M97xfyaqYjDzFW0Y/PGYZV u9rA== X-Gm-Message-State: AOAM531AMmgT3wjxFN9aNWml9O5W4U2UrIivweFGSlwFqiHl4+wqj4tG SvbR2l2eHznF6CslftSitmUrMn1OQzG+T7JxXVaeMg== X-Received: by 2002:a2e:7d0f:: with SMTP id y15mr10991904ljc.401.1607444088783; Tue, 08 Dec 2020 08:14:48 -0800 (PST) MIME-Version: 1.0 References: <20201208153501.1467-1-mgorman@techsingularity.net> <20201208153501.1467-4-mgorman@techsingularity.net> In-Reply-To: <20201208153501.1467-4-mgorman@techsingularity.net> From: Vincent Guittot Date: Tue, 8 Dec 2020 17:14:37 +0100 Message-ID: Subject: Re: [PATCH 3/4] sched/fair: Do not replace recent_used_cpu with the new target To: Mel Gorman Cc: Peter Ziljstra , Ingo Molnar , LKML , Aubrey Li , Barry Song , Juri Lelli , Valentin Schneider , Linux-ARM Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, 8 Dec 2020 at 16:35, Mel Gorman wrote: > > After select_idle_sibling, p->recent_used_cpu is set to the > new target. However on the next wakeup, prev will be the same as > recent_used_cpu unless the load balancer has moved the task since the last > wakeup. It still works, but is less efficient than it can be after all > the changes that went in since that reduce unnecessary migrations, load > balancer changes etc. This patch preserves recent_used_cpu for longer. > > With tbench on a 2-socket CascadeLake machine, 80 logical CPUs, HT enabled > > 5.10.0-rc6 5.10.0-rc6 > baseline-v2 altrecent-v2 > Hmean 1 508.39 ( 0.00%) 502.05 * -1.25%* > Hmean 2 986.70 ( 0.00%) 983.65 * -0.31%* > Hmean 4 1914.55 ( 0.00%) 1920.24 * 0.30%* > Hmean 8 3702.37 ( 0.00%) 3663.96 * -1.04%* > Hmean 16 6573.11 ( 0.00%) 6545.58 * -0.42%* > Hmean 32 10142.57 ( 0.00%) 10253.73 * 1.10%* > Hmean 64 14348.40 ( 0.00%) 12506.31 * -12.84%* > Hmean 128 21842.59 ( 0.00%) 21967.13 * 0.57%* > Hmean 256 20813.75 ( 0.00%) 21534.52 * 3.46%* > Hmean 320 20684.33 ( 0.00%) 21070.14 * 1.87%* > > The different was marginal except for 64 threads which showed in the > baseline that the result was very unstable where as the patch was much > more stable. This is somewhat machine specific as on a separate 80-cpu > Broadwell machine the same test reported. > > 5.10.0-rc6 5.10.0-rc6 > baseline-v2 altrecent-v2 > Hmean 1 310.36 ( 0.00%) 291.81 * -5.98%* > Hmean 2 340.86 ( 0.00%) 547.22 * 60.54%* > Hmean 4 912.29 ( 0.00%) 1063.21 * 16.54%* > Hmean 8 2116.40 ( 0.00%) 2103.60 * -0.60%* > Hmean 16 4232.90 ( 0.00%) 4362.92 * 3.07%* > Hmean 32 8442.03 ( 0.00%) 8642.10 * 2.37%* > Hmean 64 11733.91 ( 0.00%) 11473.66 * -2.22%* > Hmean 128 17727.24 ( 0.00%) 16784.23 * -5.32%* > Hmean 256 16089.23 ( 0.00%) 16110.79 * 0.13%* > Hmean 320 15992.60 ( 0.00%) 16071.64 * 0.49%* > > schedstats were not used in this series but from an earlier debugging > effort, the schedstats after the test run were as follows; > > Ops SIS Search 5653107942.00 5726545742.00 > Ops SIS Domain Search 3365067916.00 3319768543.00 > Ops SIS Scanned 112173512543.00 99194352541.00 > Ops SIS Domain Scanned 109885472517.00 96787575342.00 > Ops SIS Failures 2923185114.00 2950166441.00 > Ops SIS Recent Used Hit 56547.00 118064916.00 > Ops SIS Recent Used Miss 1590899250.00 354942791.00 > Ops SIS Recent Attempts 1590955797.00 473007707.00 > Ops SIS Search Efficiency 5.04 5.77 > Ops SIS Domain Search Eff 3.06 3.43 > Ops SIS Fast Success Rate 40.47 42.03 > Ops SIS Success Rate 48.29 48.48 > Ops SIS Recent Success Rate 0.00 24.96 > > First interesting point is the ridiculous number of times runqueues are > enabled -- almost 97 billion times over the course of 40 minutes > > With the patch, "Recent Used Hit" is over 2000 times more likely to > succeed. The failure rate also increases by quite a lot but the cost is > marginal even if the "Fast Success Rate" only increases by 2% overall. What > cannot be observed from these stats is where the biggest impact as these > stats cover low utilisation to over saturation. > > If graphed over time, the graphs show that the sched domain is only > scanned at negligible rates until the machine is fully busy. With > low utilisation, the "Fast Success Rate" is almost 100% until the > machine is fully busy. For 320 clients, the success rate is close to > 0% which is unsurprising. > > Signed-off-by: Mel Gorman Reviewed-by: Vincent Guittot > --- > kernel/sched/fair.c | 9 +-------- > 1 file changed, 1 insertion(+), 8 deletions(-) > > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > index 5c41875aec23..413d895bbbf8 100644 > --- a/kernel/sched/fair.c > +++ b/kernel/sched/fair.c > @@ -6277,17 +6277,13 @@ static int select_idle_sibling(struct task_struct *p, int prev, int target) > > /* Check a recently used CPU as a potential idle candidate: */ > recent_used_cpu = p->recent_used_cpu; > + p->recent_used_cpu = prev; > if (recent_used_cpu != prev && > recent_used_cpu != target && > cpus_share_cache(recent_used_cpu, target) && > (available_idle_cpu(recent_used_cpu) || sched_idle_cpu(recent_used_cpu)) && > cpumask_test_cpu(p->recent_used_cpu, p->cpus_ptr) && > asym_fits_capacity(task_util, recent_used_cpu)) { > - /* > - * Replace recent_used_cpu with prev as it is a potential > - * candidate for the next wake: > - */ > - p->recent_used_cpu = prev; > return recent_used_cpu; > } > > @@ -6768,9 +6764,6 @@ select_task_rq_fair(struct task_struct *p, int prev_cpu, int wake_flags) > } else if (wake_flags & WF_TTWU) { /* XXX always ? */ > /* Fast path */ > new_cpu = select_idle_sibling(p, prev_cpu, new_cpu); > - > - if (want_affine) > - current->recent_used_cpu = cpu; > } > rcu_read_unlock(); > > -- > 2.26.2 >