Received: by 2002:ad5:474a:0:0:0:0:0 with SMTP id i10csp3839143imu; Mon, 10 Dec 2018 08:36:37 -0800 (PST) X-Google-Smtp-Source: AFSGD/UIa0z4l4IFPSFiiBjqqMNlcioMnWq8EkcpjzuWqvjTi5SvkaPFCM3EEtil40gY0pt0FDtW X-Received: by 2002:a17:902:780a:: with SMTP id p10mr13076240pll.54.1544459797767; Mon, 10 Dec 2018 08:36:37 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1544459797; cv=none; d=google.com; s=arc-20160816; b=G6JNzjDI/jtErce35yiTdJawgQBAX/OgkiQ0WJsYtn8FYSnjFnNUPO0vpbjKG4mC2Q IhWF3rf0bvg2kR7KDFwV3nAQJlsaenTG/1A4M5sazkRp9lMZ2Xr5o3dLSK6P+Yj/aT9B /k5C5v5mnPZDGjF6arIqsNHIuiMOJl4utJt+UcQy5QtMPP9XO4cR96Xr+v/avKlzBuPX m6tNKrng9DF1Blq7tK3Gr9O98i5Nc5kPLMisHaHUxZsvIIfj7NJckdV5ErpXjgVCeupw LtlLawFMaKTDkASPXIeNy97byfA6IVIlxvWKhbUeTslfVenEzoMCAUI6qJDwWGwAQ3gp oyRg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=nW5mHAMxRJK905Cqwk3WaIQNzhJUtTOgPn1bRUQHJOA=; b=Zw8Jo7fd20yOQEV2pP9yFyW4WBmQOiTWt9uxgggLxOgd11jS5u2apgldQj26BIhfom Xnik8VYE568BLp9ZpFCbR5wtlu+m6K9r5V8u7x9e1dRzqxZgkVf6pFlKK/YNZuH1+Wq5 mrOC79HFIhglJxbswT+jgLXkggv5I4W1Si4t9f3XSggwfhU/ugkpC+AONmPlgV/RG6TC vHRSXpuzkDN8c7cegq3qHzqyROQ784lx2i4h4kaoZKqOQQ+6+GE3mX+VsbgnlI8auoyl ZL56F265bDjgZzmv9TElFT5u39SmSIjr5Nruf6b1uG/5r/5YR9VfppFZQbAyqz12c21B /BSQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b="WaLTaE/3"; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id l24si9942500pgb.489.2018.12.10.08.36.22; Mon, 10 Dec 2018 08:36:37 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b="WaLTaE/3"; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727405AbeLJQdg (ORCPT + 99 others); Mon, 10 Dec 2018 11:33:36 -0500 Received: from mail-io1-f68.google.com ([209.85.166.68]:39116 "EHLO mail-io1-f68.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726324AbeLJQdg (ORCPT ); Mon, 10 Dec 2018 11:33:36 -0500 Received: by mail-io1-f68.google.com with SMTP id k7so9234735iob.6 for ; Mon, 10 Dec 2018 08:33:35 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=nW5mHAMxRJK905Cqwk3WaIQNzhJUtTOgPn1bRUQHJOA=; b=WaLTaE/3/4gIj67A5h1Medap+EitqCFbgtxy8NxZcc14jgm8rplN9IJwFe7r9Lhg4g /4Rt4DM9dRF700tQEJRg6hFOmWezaSh3VU6UPTmFo0r+Vrr+GdvYkC6WoK4/tyYpOuS8 9/Ys/XIgMJCv0X1EiGmTCDFaiFyyYWZ15MRY8= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=nW5mHAMxRJK905Cqwk3WaIQNzhJUtTOgPn1bRUQHJOA=; b=HZOs7II2XDWW9McWwdhXfCo1dgJeuhA/6TqZ9jVv3gdRSAuxMeq0f/EoYB+POCQx++ HJb+s5PPt2/Wx+he2Y20hDa6B5+BbOgF8z7MNt5gzJqxwR1TIYNq712jMy3T+Q5ogIk7 pcDbZrbjGP34LU7Yl2+Chg/maLKVUHFHMeEdPVLZxpOf24j8ryjLF7HL+mLaozwoDYWM Ksbif/0yFXAcjCt8/7ftkPq9HbMWAHhPqXyRMXI2ohFF3KzOxyUL0zZ0siwx5rA9QFhl +v17eC783KAk2E/CQW90cixVxKIK7rEQRLFUF40uhJ4PbCQMt7QyBdkNo5SDR7oakXNr D3uQ== X-Gm-Message-State: AA+aEWYq3FYiy1p6UZPXks14hzOlwZjbt8UBqmAERVvOmO24ByNHCLnn Lcy2+M1Y3Rs3Hu4XZ1fu3r4LRPl00h98CyqPS+GBEQ== X-Received: by 2002:a6b:fe13:: with SMTP id x19mr9916537ioh.294.1544459614470; Mon, 10 Dec 2018 08:33:34 -0800 (PST) MIME-Version: 1.0 References: <1544131696-2888-1-git-send-email-steven.sistare@oracle.com> In-Reply-To: From: Vincent Guittot Date: Mon, 10 Dec 2018 17:33:22 +0100 Message-ID: Subject: Re: [PATCH v4 00/10] steal tasks to improve CPU utilization To: steven.sistare@oracle.com Cc: Ingo Molnar , Peter Zijlstra , subhra.mazumdar@oracle.com, Dhaval Giani , daniel.m.jordan@oracle.com, pavel.tatashin@microsoft.com, Matt Fleming , Mike Galbraith , Rik van Riel , Josef Bacik , Juri Lelli , Valentin Schneider , Quentin Perret , linux-kernel Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, 10 Dec 2018 at 17:29, Steven Sistare wrote: > > On 12/10/2018 11:10 AM, Vincent Guittot wrote: > > Hi Steven, > > > > On Thu, 6 Dec 2018 at 22:38, Steve Sistare wrote: > >> > >> When a CPU has no more CFS tasks to run, and idle_balance() fails to > >> find a task, then attempt to steal a task from an overloaded CPU in the > >> same LLC. Maintain and use a bitmap of overloaded CPUs to efficiently > >> identify candidates. To minimize search time, steal the first migratable > >> task that is found when the bitmap is traversed. For fairness, search > >> for migratable tasks on an overloaded CPU in order of next to run. > >> > >> This simple stealing yields a higher CPU utilization than idle_balance() > >> alone, because the search is cheap, so it may be called every time the CPU > >> is about to go idle. idle_balance() does more work because it searches > >> widely for the busiest queue, so to limit its CPU consumption, it declines > >> to search if the system is too busy. Simple stealing does not offload the > >> globally busiest queue, but it is much better than running nothing at all. > >> > >> The bitmap of overloaded CPUs is a new type of sparse bitmap, designed to > >> reduce cache contention vs the usual bitmap when many threads concurrently > >> set, clear, and visit elements. > >> > >> Patch 1 defines the sparsemask type and its operations. > >> > >> Patches 2, 3, and 4 implement the bitmap of overloaded CPUs. > >> > >> Patches 5 and 6 refactor existing code for a cleaner merge of later > >> patches. > >> > >> Patches 7 and 8 implement task stealing using the overloaded CPUs bitmap. > >> > >> Patch 9 disables stealing on systems with more than 2 NUMA nodes for the > >> time being because of performance regressions that are not due to stealing > >> per-se. See the patch description for details. > >> > >> Patch 10 adds schedstats for comparing the new behavior to the old, and > >> provided as a convenience for developers only, not for integration. > >> > >> The patch series is based on kernel 4.20.0-rc1. It compiles, boots, and > >> runs with/without each of CONFIG_SCHED_SMT, CONFIG_SMP, CONFIG_SCHED_DEBUG, > >> and CONFIG_PREEMPT. It runs without error with CONFIG_DEBUG_PREEMPT + > >> CONFIG_SLUB_DEBUG + CONFIG_DEBUG_PAGEALLOC + CONFIG_DEBUG_MUTEXES + > >> CONFIG_DEBUG_SPINLOCK + CONFIG_DEBUG_ATOMIC_SLEEP. CPU hot plug and CPU > >> bandwidth control were tested. > >> > >> Stealing improves utilization with only a modest CPU overhead in scheduler > >> code. In the following experiment, hackbench is run with varying numbers > >> of groups (40 tasks per group), and the delta in /proc/schedstat is shown > >> for each run, averaged per CPU, augmented with these non-standard stats: > >> > >> %find - percent of time spent in old and new functions that search for > >> idle CPUs and tasks to steal and set the overloaded CPUs bitmap. > >> > >> steal - number of times a task is stolen from another CPU. > >> > >> X6-2: 1 socket * 10 cores * 2 hyperthreads = 20 CPUs > >> Intel(R) Xeon(R) CPU E5-2630 v4 @ 2.20GHz > >> hackbench process 100000 > >> sched_wakeup_granularity_ns=15000000 > >> > >> baseline > >> grps time %busy slice sched idle wake %find steal > >> 1 8.084 75.02 0.10 105476 46291 59183 0.31 0 > >> 2 13.892 85.33 0.10 190225 70958 119264 0.45 0 > >> 3 19.668 89.04 0.10 263896 87047 176850 0.49 0 > >> 4 25.279 91.28 0.10 322171 94691 227474 0.51 0 > >> 8 47.832 94.86 0.09 630636 144141 486322 0.56 0 > >> > >> new > >> grps time %busy slice sched idle wake %find steal %speedup > >> 1 5.938 96.80 0.24 31255 7190 24061 0.63 7433 36.1 > >> 2 11.491 99.23 0.16 74097 4578 69512 0.84 19463 20.9 > >> 3 16.987 99.66 0.15 115824 1985 113826 0.77 24707 15.8 > >> 4 22.504 99.80 0.14 167188 2385 164786 0.75 29353 12.3 > >> 8 44.441 99.86 0.11 389153 1616 387401 0.67 38190 7.6 > >> > >> Elapsed time improves by 8 to 36%, and CPU busy utilization is up > >> by 5 to 22% hitting 99% for 2 or more groups (80 or more tasks). > >> The cost is at most 0.4% more find time. > > > > I have run some hackbench tests on my hikey arm64 octo cores with your > > patchset. My original intent was to send a tested-by but I have some > > performances regressions. > > This hikey is the smp one and not the asymetric hikey960 that Valentin > > used for his tests > > The sched domain topology is > > domain-0: span=0-3 level=MC and domain-0: span=4-7 level=MC > > domain-1: span=0-7 level=DIE > > > > I have run 12 times hackbench -g $j -P -l 2000 with j equals to 1 2 3 4 8 > > > > grps time > > 1 1.396 > > 2 2.699 > > 3 3.617 > > 4 4.498 > > 8 7.721 > > > > Then after disabling STEAL in sched_feature with echo NO_STEAL > > > /sys/kernel/debug/sched_features , the results become: > > grps time > > 1 1.217 > > 2 1.973 > > 3 2.855 > > 4 3.932 > > 8 7.674 > > > > I haven't looked in details about some possible reasons of such > > difference yet and haven't collected the stats that you added with > > patch 10. > > Have you got a script to collect and post process them ? > > > > Regards, > > Vincent > > Thanks Vincent. What is the value of /proc/sys/kernel/sched_wakeup_granularity_ns? it's 4000000 > Try 15000000. Your 8-core system is heavily overloaded with 40 * groups tasks, > and I suspect preemptions are killing performance. ok. I'm going to run the tests with the proposed value > > I have a python script to post-process schedstat files, but it does many things > and is large and I am not ready to share it. I can write a short bash script if > that would help. It was mainly in case you wanted the figures of these statistics Vincent > > - Steve