Received: by 2002:ac0:aa62:0:0:0:0:0 with SMTP id w31-v6csp1560105ima; Thu, 25 Oct 2018 00:52:04 -0700 (PDT) X-Google-Smtp-Source: AJdET5cQRB568DlJhLHAXpaHKxiQwfNuba+pGSnYgf3KcQZRU7dlK63N5e1oFdC3204ueN6MOBbo X-Received: by 2002:a63:d30c:: with SMTP id b12-v6mr485809pgg.61.1540453924026; Thu, 25 Oct 2018 00:52:04 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1540453923; cv=none; d=google.com; s=arc-20160816; b=I+qiKI99mKja3KQVjvQoFBhzhLEKnzdqRotvrV7hFZLf1Agd1ynvoJDRnvKqzWo5TN dk8ZUkDQ5RYnnD1zhz6kEh32j/qjadvwfWg1iw5ADQGXoafnbODR887ilTjU8JjjeX0r CJ6pSrj0g0iH9Zr2r5MVhdueAgKu4ZTqwFp8R1msnEYTpaPqraedhkkJXidzEXQsoeA/ 6b8tbHT14gX73Kvk9k6c+Se6lBlsPFyMLd/5xZRc9JuMQ2//Qibc0oSsHR2Nv6uzK21L 8KfiRgMPP/5StPDPYY1u6128GfHHE4dWP+yM4mx6dl6IO5vlT6ypKg8GMreudTtmDwi3 gWdw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=UrlZGQq4GlR0A2hA4fe5sarOzBOD7SvL2j15qdg+QKA=; b=npB5yn+NDtmhxGcc7GSfLfw7+9BvsI0ZiQc9/3CZ/a8zcKrAC3pnQsERbktTzIej7N v1wszEoRCOByjIEa/0B/PvumzEZccJSvOrQHPay7LX9CnJXJ006TDORE4goQXKIjHsdb H0v2+NYq3XIL4LCxMpqtqvQO1cGb9w6IYc0D5AuPX+8Fx6xnEgQLlV2KbN45TXIJFkgB 0sJpR4+pm9L2DeGdKLQb1vqTo6BRPxnTrzkdUJcpW8lxZUmuVduUDrJb+Nrhxtjq0Z4X oW/Nfo6aksRQeaaWZHvZPXxBNTqcMS6/IrGJEaJKpX7k4jGxT930W7JVOSifn8ZxYsNA 05Ow== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=NqMCZHf4; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id q25-v6si2268455pgd.357.2018.10.25.00.51.48; Thu, 25 Oct 2018 00:52:03 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=NqMCZHf4; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727134AbeJYQWo (ORCPT + 99 others); Thu, 25 Oct 2018 12:22:44 -0400 Received: from mail-it1-f193.google.com ([209.85.166.193]:36459 "EHLO mail-it1-f193.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726497AbeJYQWo (ORCPT ); Thu, 25 Oct 2018 12:22:44 -0400 Received: by mail-it1-f193.google.com with SMTP id h14-v6so583010itf.1 for ; Thu, 25 Oct 2018 00:51:08 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=UrlZGQq4GlR0A2hA4fe5sarOzBOD7SvL2j15qdg+QKA=; b=NqMCZHf4Uffrmnf/QBfly+Je86Vp16yTUo4i8hGcQp5eoJfJrFfG2hDfBsgSqcqHQs KHBsIHqhQfBA5yT1OcECuEiG9Nbq42lO+LFF873l6RZpqrtD2Qh5mCw5iurRfjk58HiF YtYwbG9XKr2wuf/8Lf1GC/sDocaVZpR/Wr64A= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=UrlZGQq4GlR0A2hA4fe5sarOzBOD7SvL2j15qdg+QKA=; b=X3XvPNUEKuU4zzPeBO8F0xa/4K77TVuhasPT7PvlM/jHyr9CPlMsvsJ1SMPKAcvOuc pGJiFXC1J8kkqFV/Apo5mnNqBebWQU8kQBUubhOR6MiE0tNeXNArLnYWlcM6BNZAS4VI JcVpkTqm0QnLECNrKUGT3/zz4AXATmh3wY2tfsSleIu3fIziJnWc4iBCHwmjp0lwA16t tEenp2Anjj7JeeRyu1dgjEV1eiUfdXBQ2hhNROL68Q8moRHrN+TZrBiWPBfXL5WH+QxD LfoWlp2jSpdq6KPcQ5cWl0PME/yUFBwaCBTeeschepM/WaIDiJE/qMTEGclUOt2fiwHs mlRw== X-Gm-Message-State: AGRZ1gI5IjWwmc6EpG9lwrOffzRZDZUIZ1MIT6V3FQXAQm0ekhLJMLh8 O18SAo0dxtYSlucoari0rH1Tnb0RG+2rg0qB5QP1YQ== X-Received: by 2002:a24:670a:: with SMTP id u10-v6mr307184itc.114.1540453868207; Thu, 25 Oct 2018 00:51:08 -0700 (PDT) MIME-Version: 1.0 References: <1540220381-424433-1-git-send-email-steven.sistare@oracle.com> In-Reply-To: <1540220381-424433-1-git-send-email-steven.sistare@oracle.com> From: Vincent Guittot Date: Thu, 25 Oct 2018 09:50:57 +0200 Message-ID: Subject: Re: [PATCH 00/10] steal tasks to improve CPU utilization To: steven.sistare@oracle.com Cc: Ingo Molnar , Peter Zijlstra , subhra.mazumdar@oracle.com, Dhaval Giani , Rohit Jain , daniel.m.jordan@oracle.com, pavel.tatashin@microsoft.com, Matt Fleming , Mike Galbraith , Rik van Riel , Josef Bacik , Juri Lelli , linux-kernel Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Steve, On Mon, 22 Oct 2018 at 17:10, Steve Sistare wrote: > > When a CPU has no more CFS tasks to run, and idle_balance() fails to > find a task, then attempt to steal a task from an overloaded CPU in the > same LLC. Maintain and use a bitmap of overloaded CPUs to efficiently > identify candidates. To minimize search time, steal the first migratable > task that is found when the bitmap is traversed. For fairness, search > for migratable tasks on an overloaded CPU in order of next to run. > > This simple stealing yields a higher CPU utilization than idle_balance() > alone, because the search is cheap, so it may be called every time the CPU > is about to go idle. idle_balance() does more work because it searches > widely for the busiest queue, so to limit its CPU consumption, it declines > to search if the system is too busy. Simple stealing does not offload the > globally busiest queue, but it is much better than running nothing at all. > > The bitmap of overloaded CPUs is a new type of sparse bitmap, designed to > reduce cache contention vs the usual bitmap when many threads concurrently > set, clear, and visit elements. > > Patch 1 defines the sparsemask type and its operations. > > Patches 2, 3, and 4 implement the bitmap of overloaded CPUs. > > Patches 5 and 6 refactor existing code for a cleaner merge of later > patches. > > Patches 7 and 8 implement task stealing using the overloaded CPUs bitmap. > > Patch 9 disables stealing on systems with more than 2 NUMA nodes for the > time being because of performance regressions that are not due to stealing > per-se. See the patch description for details. > > Patch 10 adds schedstats for comparing the new behavior to the old, and > provided as a convenience for developers only, not for integration. > > The patch series is based on kernel 4.19.0-rc7. It compiles, boots, and > runs with/without each of CONFIG_SCHED_SMT, CONFIG_SMP, CONFIG_SCHED_DEBUG, > and CONFIG_PREEMPT. It runs without error with CONFIG_DEBUG_PREEMPT + > CONFIG_SLUB_DEBUG + CONFIG_DEBUG_PAGEALLOC + CONFIG_DEBUG_MUTEXES + > CONFIG_DEBUG_SPINLOCK + CONFIG_DEBUG_ATOMIC_SLEEP. CPU hot plug and CPU > bandwidth control were tested. > > Stealing imprroves utilization with only a modest CPU overhead in scheduler > code. In the following experiment, hackbench is run with varying numbers > of groups (40 tasks per group), and the delta in /proc/schedstat is shown > for each run, averaged per CPU, augmented with these non-standard stats: > > %find - percent of time spent in old and new functions that search for > idle CPUs and tasks to steal and set the overloaded CPUs bitmap. > > steal - number of times a task is stolen from another CPU. > > X6-2: 1 socket * 10 cores * 2 hyperthreads = 20 CPUs > Intel(R) Xeon(R) CPU E5-2630 v4 @ 2.20GHz > hackbench process 100000 > sched_wakeup_granularity_ns=15000000 Why do you mention this sched_wakeup_granularity_ns value ? It is something that you changed for you tests ? The comment for this tunable says that default value is 1ms * ilog(ncpus) = 4ms for 20CPUs > > baseline > grps time %busy slice sched idle wake %find steal > 1 8.084 75.02 0.10 105476 46291 59183 0.31 0 > 2 13.892 85.33 0.10 190225 70958 119264 0.45 0 > 3 19.668 89.04 0.10 263896 87047 176850 0.49 0 > 4 25.279 91.28 0.10 322171 94691 227474 0.51 0 > 8 47.832 94.86 0.09 630636 144141 486322 0.56 0 > > new > grps time %busy slice sched idle wake %find steal %speedup > 1 5.938 96.80 0.24 31255 7190 24061 0.63 7433 36.1 > 2 11.491 99.23 0.16 74097 4578 69512 0.84 19463 20.9 > 3 16.987 99.66 0.15 115824 1985 113826 0.77 24707 15.8 > 4 22.504 99.80 0.14 167188 2385 164786 0.75 29353 12.3 > 8 44.441 99.86 0.11 389153 1616 387401 0.67 38190 7.6 > > Elapsed time improves by 8 to 36%, and CPU busy utilization is up > by 5 to 22% hitting 99% for 2 or more groups (80 or more tasks). > The cost is at most 0.4% more find time. >