Received: by 2002:a05:6a10:1d13:0:0:0:0 with SMTP id pp19csp1725611pxb; Mon, 23 Aug 2021 03:09:50 -0700 (PDT) X-Google-Smtp-Source: ABdhPJz8AIF5pa6CdHwglpMnOM9dioo1gyUhDK1QiLiP9ek5u2zREkELgPF6oxnuqGctokLNwW2Z X-Received: by 2002:a92:1944:: with SMTP id e4mr22383735ilm.186.1629713390606; Mon, 23 Aug 2021 03:09:50 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1629713390; cv=none; d=google.com; s=arc-20160816; b=k2esLiJX2c0eydda16lWb8EHPsm8rkJWGfpXLpjjBkUaM/4PoY76SLMPO48g3lNkBn fYxS/i3kL+/gKaV3qG1b8PDp+H68M/jaA/z/X6db0gdD3QnWRVZ+/QyiibXWeME69LQv dTalpEeXRWwYl2EGFes7pEtZRpDe1Xslc0OXYBI6D4lk66LgH35kfRRmGuJMWM7E2uX9 A3CLR0NzGOAv5YU4uymftx0F7i0ZBeK+5hmTGsZaez88nh3RpEr6exTIYsUe4dc0L18U zKYXEE7LcTAO304axK2Q2W4NxJ552Q8mulyGwVP285jgZ3e3TVRJ8nYmlyvBWEfCDmwN CjEg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=9wBfR6vpwVhRysUb5PTjYhMXJlLWyxkvdJddCxy4f30=; b=SjtCO8mzaDRhNUvlY86zGb8bL9WmLeNSZQvaOGNWr+UOAUJzY/9fddXRgCJ+zOAHNQ TWJPQmO8OFQBy1lyw8mB3fIFxuO1YRspUfWVS+Y3ujvTqpsCZ3aGNMnRyInQAMlRdS1D 5L8eZm09KaAVPPqNuLgCNykb0przE5/D1zXu9wQQdF+8qkThfJaE2pUAZYzoS8H29z3u V692Ff3vS7OlY2KFkdwttwIg+sbSYQMNW4JsEhP95mdsWCFc8+1eE9ewbdgaC2LRO3Dc GCH1P1ufaqHBbK2eNVb3Fep46oQV0bfE9+54HS6V6WnjF7rR7QnPhfQVmeNqe0b75gxd g0Aw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=vyAWfXnR; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id h14si13693143ile.129.2021.08.23.03.09.38; Mon, 23 Aug 2021 03:09:50 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=vyAWfXnR; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236125AbhHWKJR (ORCPT + 99 others); Mon, 23 Aug 2021 06:09:17 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:53836 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S236083AbhHWKJQ (ORCPT ); Mon, 23 Aug 2021 06:09:16 -0400 Received: from mail-lj1-x22b.google.com (mail-lj1-x22b.google.com [IPv6:2a00:1450:4864:20::22b]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 8574AC061575 for ; Mon, 23 Aug 2021 03:08:33 -0700 (PDT) Received: by mail-lj1-x22b.google.com with SMTP id i28so30493266ljm.7 for ; Mon, 23 Aug 2021 03:08:33 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=9wBfR6vpwVhRysUb5PTjYhMXJlLWyxkvdJddCxy4f30=; b=vyAWfXnRN6CV6ZsiL3jSXeNixNL5kMPnrZja5/z1aHdmG8CsCspXQzvOI/hxHkH8K8 JWqAS40XSEeWcAlcdbnSMg3uCKot3KXr0Mg1bV3E010wbPAVTgYZsB7sOLFm0McMG+8D zjE+4niupdKyB2O+weAx/Nnvf+IrVf1QtdcFXeJ3sfUsKuEAkMdZlDUzXX7Dsc+lCIIJ g2wu+aAlneocrahquB2SWsGD/AhJlHJxkFnOv5JysFqhZkaoAzCDwBbZ0F5+aUEGluLX lN9bEbrVejw61pvGeoGgLAmM6ngTuXSjXy9wWGoZ9X8+0FaZl4eddYSjTpj3LHOpJrwJ s+vA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=9wBfR6vpwVhRysUb5PTjYhMXJlLWyxkvdJddCxy4f30=; b=FZFzyTqx/isVP4QzqDYzcpCMzm8srOvYIM7I8LtYq+RdJ5nWhKhb8iw+xnjgUDfenj 8Exnt/0Akpc6x9Tw7+9QrF+AoGeok2B72camSDSF7dlrwrLZYtMo6sjR1PpzX8yBP16j ni/4x7J0IAzxQy7LFH6TqgGfGdfdi+w1m4wOYWvwC9/WCTK/BFdfn6uA1DYhKy0x/Uth 424vRHSFqedCkRKb8mgWb4VrFNkXUNv7U/xvCArKUjrjommbAwtk5Qk4X0c2FN6ul3bc xkdHr3b9m/YyZ/ahvAD8NgSlvKOBDOH/vCAjxQEm2mpt6brgzL7sAjfMqMVATf5NZoe3 Srgw== X-Gm-Message-State: AOAM533vseEsywlzLVdhJp8p6c9iwZ2bpl0+K1XpEMGctOeBTPc9Hf9M 4vB8kyBSTDUVUmheg/vHyryXWcjdx2yEjqK8kQHD7A== X-Received: by 2002:a2e:730b:: with SMTP id o11mr11565941ljc.176.1629713311851; Mon, 23 Aug 2021 03:08:31 -0700 (PDT) MIME-Version: 1.0 References: <20210820010403.946838-1-joshdon@google.com> <20210820010403.946838-4-joshdon@google.com> In-Reply-To: <20210820010403.946838-4-joshdon@google.com> From: Vincent Guittot Date: Mon, 23 Aug 2021 12:08:20 +0200 Message-ID: Subject: Re: [PATCH v3 3/4] sched: reduce sched slice for SCHED_IDLE entities To: Josh Don Cc: Ingo Molnar , Peter Zijlstra , Juri Lelli , Dietmar Eggemann , Steven Rostedt , Ben Segall , Mel Gorman , Daniel Bristot de Oliveira , Paul Turner , Oleg Rombakh , Viresh Kumar , Steve Sistare , Tejun Heo , Rik van Riel , linux-kernel Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, 20 Aug 2021 at 03:04, Josh Don wrote: > > Use a small, non-scaled min granularity for SCHED_IDLE entities, when > competing with normal entities. This reduces the latency of getting > a normal entity back on cpu, at the expense of increased context > switch frequency of SCHED_IDLE entities. > > The benefit of this change is to reduce the round-robin latency for > normal entities when competing with a SCHED_IDLE entity. > > Example: on a machine with HZ=1000, spawned two threads, one of which is > SCHED_IDLE, and affined to one cpu. Without this patch, the SCHED_IDLE > thread runs for 4ms then waits for 1.4s. With this patch, it runs for > 1ms and waits 340ms (as it round-robins with the other thread). > > Signed-off-by: Josh Don > --- > kernel/sched/debug.c | 2 ++ > kernel/sched/fair.c | 29 ++++++++++++++++++++++++----- > kernel/sched/sched.h | 1 + > 3 files changed, 27 insertions(+), 5 deletions(-) > > diff --git a/kernel/sched/debug.c b/kernel/sched/debug.c > index 33538579db9a..317ef560aa63 100644 > --- a/kernel/sched/debug.c > +++ b/kernel/sched/debug.c > @@ -305,6 +305,7 @@ static __init int sched_init_debug(void) > > debugfs_create_u32("latency_ns", 0644, debugfs_sched, &sysctl_sched_latency); > debugfs_create_u32("min_granularity_ns", 0644, debugfs_sched, &sysctl_sched_min_granularity); > + debugfs_create_u32("idle_min_granularity_ns", 0644, debugfs_sched, &sysctl_sched_idle_min_granularity); > debugfs_create_u32("wakeup_granularity_ns", 0644, debugfs_sched, &sysctl_sched_wakeup_granularity); > > debugfs_create_u32("latency_warn_ms", 0644, debugfs_sched, &sysctl_resched_latency_warn_ms); > @@ -806,6 +807,7 @@ static void sched_debug_header(struct seq_file *m) > SEQ_printf(m, " .%-40s: %Ld.%06ld\n", #x, SPLIT_NS(x)) > PN(sysctl_sched_latency); > PN(sysctl_sched_min_granularity); > + PN(sysctl_sched_idle_min_granularity); > PN(sysctl_sched_wakeup_granularity); > P(sysctl_sched_child_runs_first); > P(sysctl_sched_features); > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > index 19a9244c140f..31f40aa005b9 100644 > --- a/kernel/sched/fair.c > +++ b/kernel/sched/fair.c > @@ -59,6 +59,14 @@ unsigned int sysctl_sched_tunable_scaling = SCHED_TUNABLESCALING_LOG; > unsigned int sysctl_sched_min_granularity = 750000ULL; > static unsigned int normalized_sysctl_sched_min_granularity = 750000ULL; > > +/* > + * Minimal preemption granularity for CPU-bound SCHED_IDLE tasks. > + * Applies only when SCHED_IDLE tasks compete with normal tasks. > + * > + * (default: 0.75 msec) > + */ > +unsigned int sysctl_sched_idle_min_granularity = 750000ULL; > + > /* > * This value is kept at sysctl_sched_latency/sysctl_sched_min_granularity > */ > @@ -665,6 +673,8 @@ static u64 __sched_period(unsigned long nr_running) > return sysctl_sched_latency; > } > > +static bool sched_idle_cfs_rq(struct cfs_rq *cfs_rq); > + > /* > * We calculate the wall-time slice from the period by taking a part > * proportional to the weight. > @@ -674,6 +684,8 @@ static u64 __sched_period(unsigned long nr_running) > static u64 sched_slice(struct cfs_rq *cfs_rq, struct sched_entity *se) > { > unsigned int nr_running = cfs_rq->nr_running; > + struct sched_entity *init_se = se; > + unsigned int min_gran; > u64 slice; > > if (sched_feat(ALT_PERIOD)) > @@ -684,12 +696,13 @@ static u64 sched_slice(struct cfs_rq *cfs_rq, struct sched_entity *se) > for_each_sched_entity(se) { > struct load_weight *load; > struct load_weight lw; > + struct cfs_rq *qcfs_rq; > > - cfs_rq = cfs_rq_of(se); > - load = &cfs_rq->load; > + qcfs_rq = cfs_rq_of(se); > + load = &qcfs_rq->load; > > if (unlikely(!se->on_rq)) { > - lw = cfs_rq->load; > + lw = qcfs_rq->load; > > update_load_add(&lw, se->load.weight); > load = &lw; > @@ -697,8 +710,14 @@ static u64 sched_slice(struct cfs_rq *cfs_rq, struct sched_entity *se) > slice = __calc_delta(slice, se->load.weight, load); > } > > - if (sched_feat(BASE_SLICE)) > - slice = max(slice, (u64)sysctl_sched_min_granularity); > + if (sched_feat(BASE_SLICE)) { > + if (se_is_idle(init_se) && !sched_idle_cfs_rq(cfs_rq)) Like for place_entity, we should probably not dynamically switch between the 2 values below depending on the presence or not of non sched idle tasks and always use sysctl_sched_idle_min_granularity > + min_gran = sysctl_sched_idle_min_granularity; > + else > + min_gran = sysctl_sched_min_granularity; > + > + slice = max_t(u64, slice, min_gran); > + } > > return slice; > } > diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h > index 6af039e433fb..29846da35861 100644 > --- a/kernel/sched/sched.h > +++ b/kernel/sched/sched.h > @@ -2399,6 +2399,7 @@ extern const_debug unsigned int sysctl_sched_migration_cost; > #ifdef CONFIG_SCHED_DEBUG > extern unsigned int sysctl_sched_latency; > extern unsigned int sysctl_sched_min_granularity; > +extern unsigned int sysctl_sched_idle_min_granularity; > extern unsigned int sysctl_sched_wakeup_granularity; > extern int sysctl_resched_latency_warn_ms; > extern int sysctl_resched_latency_warn_once; > -- > 2.33.0.rc2.250.ged5fa647cd-goog >