Received: by 2002:a05:6358:489b:b0:bb:da1:e618 with SMTP id x27csp1827559rwn; Fri, 16 Sep 2022 01:06:14 -0700 (PDT) X-Google-Smtp-Source: AMsMyM5yKjDdmyRGJ4tvhRw9igWSsRcV5UWNSPPFUm895JolY0qeuRaeNUG8GmdKDXytP0W4AxKi X-Received: by 2002:a17:90b:1bcd:b0:203:27a3:6d0f with SMTP id oa13-20020a17090b1bcd00b0020327a36d0fmr4234276pjb.234.1663315574340; Fri, 16 Sep 2022 01:06:14 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1663315574; cv=none; d=google.com; s=arc-20160816; b=r/0hUlFRtFV1e3BNlP7O6iUUZQoDThcW45liKKkGEu/J3soZQJLCW60VoAzGZuDcEq Z7m0vDvsPcM+ARG23kVfKYBgyD8ROBtxP61LddgLckMvPxNjfRHRfNQBMEAWrwX6Qsls xymLdreeHKVakRnG+rZjNdHg69VojFauHMin6rlSBbW3Z8uyPyEag5Lo6+VFzV4zCqCX fo6LEpm/ZS5xo2kWxdCz1bvI8u1TVGbbY8upoRwEPKhTD0gE2G5KPpzaDU/JmJr9FId+ 9mr7dH1okP4CBpP9sdkONNQHUfdYdF/EJLKZnnOZQSi8yIK1QbLwSVITvm2B8biOOv3v UX2Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:references:in-reply-to:message-id:date:subject :cc:to:from:dkim-signature; bh=7xCFBgFTOWFPGz0fkdQnzhLrl99Sda3Rg9hYgdSgxLY=; b=REQZzuTio9kTwzyZiOp2vJlz91KxLCfdiGJ2/IM2xPQehPKaS3dGvQ3ycE+e6yTtxn lI/XfVb5c1wjyZwNiITG/9Za0jA4G5OG5Ecet/BsVQXIWN7aEyzxSsA2sX3N2G7sdGAj /zsdhaKuTvf6kGfAE2QizyxMcsbXGosqO0L8MMZ2g77ejwTXWrrVkSWMpZ0BLbkAKfqo mLHCH79KzS7ZRcl2rLZ9NURoNBtWeurnpJ++U4t2O+3m/p98zN4PrvRaPwoJW8cxPmnD rLkinKzYTrzDCoibRCbkur4CHkEgMjyw1gy/kRKAteFnV++NEPgZTCrLO2CC3qWEsHkW 5Gfg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=HYA3RPLb; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id r13-20020a65508d000000b00429d2838aa5si9914687pgp.635.2022.09.16.01.06.03; Fri, 16 Sep 2022 01:06:14 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=HYA3RPLb; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229831AbiIPIEb (ORCPT + 99 others); Fri, 16 Sep 2022 04:04:31 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:43706 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230109AbiIPIDc (ORCPT ); Fri, 16 Sep 2022 04:03:32 -0400 Received: from mail-wr1-x42b.google.com (mail-wr1-x42b.google.com [IPv6:2a00:1450:4864:20::42b]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 5CD42A4B30 for ; Fri, 16 Sep 2022 01:03:24 -0700 (PDT) Received: by mail-wr1-x42b.google.com with SMTP id e16so34687370wrx.7 for ; Fri, 16 Sep 2022 01:03:24 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=references:in-reply-to:message-id:date:subject:cc:to:from:from:to :cc:subject:date; bh=7xCFBgFTOWFPGz0fkdQnzhLrl99Sda3Rg9hYgdSgxLY=; b=HYA3RPLbfEEeZaybbiWEKmkJ5Xcgdtv42Zw4t1+8X33NoNQND7AUVIgz9jCW0YneDc dG3Q41i+xRf0niRuL/vZ0g/7tkAFXHuuRXyegQWPaIFFKKjIV+bebeGewhsmaUXHeDKR wFEBAZMDMQl2dS2q4rv2n3bewrqcMUnorYn8Pbv0ffu+gDImFVHnqi4X1zdfnXqQ7wMn cFp0XweHgZ/zlidM0dIrTmYz1JFeBEmEBaNXqLZba91MFLQTCVIHhGZejPhQC6BpS7Mi 1U2LCShIdLD+0AAQFm7xi/7dpWp6mKP068g5cyjhpQPLYA886QdtrMzQU5zhjETlZ1xf y3qg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=references:in-reply-to:message-id:date:subject:cc:to:from :x-gm-message-state:from:to:cc:subject:date; bh=7xCFBgFTOWFPGz0fkdQnzhLrl99Sda3Rg9hYgdSgxLY=; b=6rLzPhT0g6FhyVHLbPTGzWvFi0K7iJuUL0iiMYMk7eEQ0m4WD5NAjsMUibMLQ9/93q OyCP7Pa6s7MhPHvb9/aSQ7CZjtDdSrlnrnY5cxZUM4vF91j1l1OfD3rwmfc42R7scJCn MAiaPzTZhR8vNWWO8yd4iNECIDpsbKmOmTlpjrlB/4YXPz76beHy2ahDmtLtjHrPOG9C gKlsCOqXo8DmXhNwNouKUN+M8I6q6UvXorPeHUV0To3DBmok5Qmt1Rv/UANR8ZIopiSx 0v+LItTGhqaZOhHe7YJq2mJPdiCgOdX7w1lu1HvSBzGxaD273BcY3vTSkqA51DmaCWvh 2pwQ== X-Gm-Message-State: ACrzQf2BvL52Xl7PPBCYXesXScTiHK90NjZbvKhXoKZ87jBzonlPrYpo XvYvRd0dmKZJjX8IInDb6MD//g== X-Received: by 2002:a5d:5289:0:b0:225:4852:4248 with SMTP id c9-20020a5d5289000000b0022548524248mr1918568wrv.228.1663315402738; Fri, 16 Sep 2022 01:03:22 -0700 (PDT) Received: from localhost.localdomain ([2a01:e0a:f:6020:9ab:1164:827a:1548]) by smtp.gmail.com with ESMTPSA id c9-20020adffb49000000b00228dbf15072sm4418116wrs.62.2022.09.16.01.03.20 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 16 Sep 2022 01:03:21 -0700 (PDT) From: Vincent Guittot To: mingo@redhat.com, peterz@infradead.org, juri.lelli@redhat.com, dietmar.eggemann@arm.com, rostedt@goodmis.org, bsegall@google.com, mgorman@suse.de, bristot@redhat.com, vschneid@redhat.com, linux-kernel@vger.kernel.org, parth@linux.ibm.com Cc: qais.yousef@arm.com, chris.hyser@oracle.com, valentin.schneider@arm.com, patrick.bellasi@matbug.net, David.Laight@aculab.com, pjt@google.com, pavel@ucw.cz, tj@kernel.org, qperret@google.com, tim.c.chen@linux.intel.com, joshdon@google.com, Vincent Guittot Subject: [PATCH v4 5/8] sched/fair: Take into account latency priority at wakeup Date: Fri, 16 Sep 2022 10:03:02 +0200 Message-Id: <20220916080305.29574-6-vincent.guittot@linaro.org> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20220916080305.29574-1-vincent.guittot@linaro.org> References: <20220916080305.29574-1-vincent.guittot@linaro.org> X-Spam-Status: No, score=-1.6 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,PDS_BTC_ID,RCVD_IN_DNSWL_NONE, SPF_HELO_NONE,SPF_PASS autolearn=no autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Take into account the latency priority of a thread when deciding to preempt the current running thread. We don't want to provide more CPU bandwidth to a thread but reorder the scheduling to run latency sensitive task first whenever possible. As long as a thread didn't use its bandwidth, it will be able to preempt the current thread. At the opposite, a thread with a low latency priority will preempt current thread at wakeup only to keep fair CPU bandwidth sharing. Otherwise it will wait for the tick to get its sched slice. curr vruntime | sysctl_sched_wakeup_granularity <--> ----------------------------------|----|-----------------------|--------------- | |<---------------------> | . sysctl_sched_latency | . default/current latency entity | . | . 1111111111111111111111111111111111|0000|-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1- se preempts curr at wakeup ------>|<- se doesn't preempt curr ----------------- | . | . | . low latency entity | . ---------------------->| % of sysctl_sched_latency | 1111111111111111111111111111111111111111111111111111111111|0000|-1-1-1-1-1-1-1- preempt ------------------------------------------------->|<- do not preempt -- | . | . | . high latency entity | . |<-----------------------| . | % of sysctl_sched_latency . 111111111|0000|-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1-1 preempt->|<- se doesn't preempt curr ------------------------------------------ Tests results of nice latency impact on heavy load like hackbench: hackbench -l (2560 / group) -g group group latency 0 latency 19 1 1.350(+/- 1.10%) 1.287(+/- 1.60%) + 5% 4 1.392(+/- 1.67%) 1.248(+/- 1.51%) +11% 8 1.294(+/- 1.56%) 1.254(+/- 1.96%) + 3% 16 1.315(+/- 1.02%) 1.288(+/- 1.53%) + 2% hackbench -p -l (2560 / group) -g group group 1 1.768(+/- 13%) 0.805(+/- 2%) +54% 4 1.634(+/- 13%) 0.765(+/- 1%) +53% 8 1.305(+/- 4%) 0.745(+/- 2%) +43% 16 0.786(+/- 4%) 0.705(+/- 2%) +10% By deacreasing the latency prio, we reduce the number of preemption at wakeup and help hackbench making progress. Test results of nice latency impact on short live load like cyclictest while competing with heavy load like hackbench: hackbench -l 10000 -g $group & cyclictest --policy other -D 5 -q -n latency 0 latency -20 group min avg max min avg max 0 17 18 28 16 18 28 1 67 319 7369 63 76 2479 4 64 453 51699 45 95 8788 8 65 814 51699 63 116 19535 16 64 1275 37744 63 159 51134 group = 0 means that hackbench is not running. The avg is significantly improved with nice latency -20 especially with large number of groups but min and max remain quite similar. If we add the histogram parameter to get details of latency, we have : hackbench -l 10000 -g 16 & cyclictest --policy other -D 5 -q -n -H 20000 --histfile data.txt latency 0 latency -20 Min Latencies: 63 63 Avg Latencies: 1459 214 Max Latencies: 58274 82358 50% latencies: 164 87 75% latencies: 848 91 85% latencies: 1246 94 90% latencies: 2149 96 95% latencies: 8463 99 99% latencies:>20000 120 With percentile details, we see the benefit of nice latency -20 as only 1% of the latencies stays above 120us whereas the default latency has got 25% are above 848us, 10% over the 2ms and 1% above 20ms. Signed-off-by: Vincent Guittot --- include/linux/sched.h | 4 +++- init/init_task.c | 2 +- kernel/sched/core.c | 39 ++++++++++++++++++++++++------ kernel/sched/debug.c | 2 +- kernel/sched/fair.c | 56 +++++++++++++++++++++++++++++++++++++++++-- kernel/sched/sched.h | 14 ++++++++++- 6 files changed, 104 insertions(+), 13 deletions(-) diff --git a/include/linux/sched.h b/include/linux/sched.h index a0adb55efa1c..2d2a361d65ee 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h @@ -557,6 +557,8 @@ struct sched_entity { /* cached value of my_q->h_nr_running */ unsigned long runnable_weight; #endif + /* preemption offset in ns */ + long latency_offset; #ifdef CONFIG_SMP /* @@ -773,7 +775,7 @@ struct task_struct { int static_prio; int normal_prio; unsigned int rt_priority; - int latency_nice; + int latency_prio; struct sched_entity se; struct sched_rt_entity rt; diff --git a/init/init_task.c b/init/init_task.c index 225d11a39bc9..e98c71f24981 100644 --- a/init/init_task.c +++ b/init/init_task.c @@ -78,7 +78,7 @@ struct task_struct init_task .prio = MAX_PRIO - 20, .static_prio = MAX_PRIO - 20, .normal_prio = MAX_PRIO - 20, - .latency_nice = DEFAULT_LATENCY_NICE, + .latency_prio = NICE_WIDTH - 20, .policy = SCHED_NORMAL, .cpus_ptr = &init_task.cpus_mask, .user_cpus_ptr = NULL, diff --git a/kernel/sched/core.c b/kernel/sched/core.c index 3c79c5419d1b..13cf794708ee 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -1288,6 +1288,13 @@ static void set_load_weight(struct task_struct *p, bool update_load) } } +static void set_latency_offset(struct task_struct *p) +{ + long weight = sched_latency_to_weight[p->latency_prio]; + + p->se.latency_offset = (sysctl_sched_latency * weight) >> NICE_LATENCY_SHIFT; +} + #ifdef CONFIG_UCLAMP_TASK /* * Serializes updates of utilization clamp values @@ -4512,7 +4519,7 @@ int sched_fork(unsigned long clone_flags, struct task_struct *p) p->prio = current->normal_prio; /* Propagate the parent's latency requirements to the child as well */ - p->latency_nice = current->latency_nice; + p->latency_prio = current->latency_prio; uclamp_fork(p); @@ -4530,7 +4537,9 @@ int sched_fork(unsigned long clone_flags, struct task_struct *p) p->prio = p->normal_prio = p->static_prio; set_load_weight(p, false); - p->latency_nice = DEFAULT_LATENCY_NICE; + p->latency_prio = NICE_TO_LATENCY(0); + set_latency_offset(p); + /* * We don't need the reset flag anymore after the fork. It has * fulfilled its duty: @@ -7295,8 +7304,10 @@ static void __setscheduler_params(struct task_struct *p, static void __setscheduler_latency(struct task_struct *p, const struct sched_attr *attr) { - if (attr->sched_flags & SCHED_FLAG_LATENCY_NICE) - p->latency_nice = attr->sched_latency_nice; + if (attr->sched_flags & SCHED_FLAG_LATENCY_NICE) { + p->latency_prio = NICE_TO_LATENCY(attr->sched_latency_nice); + set_latency_offset(p); + } } /* @@ -7445,7 +7456,7 @@ static int __sched_setscheduler(struct task_struct *p, if (attr->sched_latency_nice < MIN_LATENCY_NICE) return -EINVAL; /* Use the same security checks as NICE */ - if (attr->sched_latency_nice < p->latency_nice && + if (attr->sched_latency_nice < LATENCY_TO_NICE(p->latency_prio) && !capable(CAP_SYS_NICE)) return -EPERM; } @@ -7485,7 +7496,7 @@ static int __sched_setscheduler(struct task_struct *p, if (attr->sched_flags & SCHED_FLAG_UTIL_CLAMP) goto change; if (attr->sched_flags & SCHED_FLAG_LATENCY_NICE && - attr->sched_latency_nice != p->latency_nice) + attr->sched_latency_nice != LATENCY_TO_NICE(p->latency_prio)) goto change; p->sched_reset_on_fork = reset_on_fork; @@ -8026,7 +8037,7 @@ SYSCALL_DEFINE4(sched_getattr, pid_t, pid, struct sched_attr __user *, uattr, get_params(p, &kattr); kattr.sched_flags &= SCHED_FLAG_ALL; - kattr.sched_latency_nice = p->latency_nice; + kattr.sched_latency_nice = LATENCY_TO_NICE(p->latency_prio); #ifdef CONFIG_UCLAMP_TASK /* @@ -11177,6 +11188,20 @@ const u32 sched_prio_to_wmult[40] = { /* 15 */ 119304647, 148102320, 186737708, 238609294, 286331153, }; +/* + * latency weight for wakeup preemption + */ +const int sched_latency_to_weight[40] = { + /* -20 */ -1024, -973, -922, -870, -819, + /* -15 */ -768, -717, -666, -614, -563, + /* -10 */ -512, -461, -410, -358, -307, + /* -5 */ -256, -205, -154, -102, -51, + /* 0 */ 0, 51, 102, 154, 205, + /* 5 */ 256, 307, 358, 410, 461, + /* 10 */ 512, 563, 614, 666, 717, + /* 15 */ 768, 819, 870, 922, 973, +}; + void call_trace_sched_update_nr_running(struct rq *rq, int count) { trace_sched_update_nr_running_tp(rq, count); diff --git a/kernel/sched/debug.c b/kernel/sched/debug.c index a3f7876217a6..06aaa0c81d4b 100644 --- a/kernel/sched/debug.c +++ b/kernel/sched/debug.c @@ -1042,7 +1042,7 @@ void proc_sched_show_task(struct task_struct *p, struct pid_namespace *ns, #endif P(policy); P(prio); - P(latency_nice); + P(latency_prio); if (task_has_dl_policy(p)) { P(dl.runtime); P(dl.deadline); diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index e8c1b889dcbb..a20eadb0af97 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -4574,6 +4574,8 @@ dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) update_idle_cfs_rq_clock_pelt(cfs_rq); } +static long wakeup_latency_gran(struct sched_entity *curr, struct sched_entity *se); + /* * Preempt the current task with a newly woken task if needed: */ @@ -4606,6 +4608,7 @@ check_preempt_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr) se = __pick_first_entity(cfs_rq); delta = curr->vruntime - se->vruntime; + delta -= wakeup_latency_gran(curr, se); if (delta < 0) return; @@ -5732,6 +5735,35 @@ static int sched_idle_cpu(int cpu) } #endif +static void set_next_buddy(struct sched_entity *se); + +static void check_preempt_from_others(struct cfs_rq *cfs, struct sched_entity *se) +{ + struct sched_entity *next; + + if (se->latency_offset >= 0) + return; + + if (cfs->nr_running <= 1) + return; + /* + * When waking from idle, we don't need to check to preempt at wakeup + * the idle thread and don't set next buddy as a candidate for being + * picked in priority. + * In case of simultaneous wakeup from idle, the latency sensitive tasks + * lost opportunity to preempt non sensitive tasks which woke up + * simultaneously. + */ + + if (cfs->next) + next = cfs->next; + else + next = __pick_first_entity(cfs); + + if (next && wakeup_preempt_entity(next, se) == 1) + set_next_buddy(se); +} + /* * The enqueue_task method is called before nr_running is * increased. Here we update the fair scheduling stats and @@ -5818,14 +5850,15 @@ enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags) if (!task_new) update_overutilized_status(rq); + if (rq->curr->sched_class != &fair_sched_class) + check_preempt_from_others(cfs_rq_of(&p->se), &p->se); + enqueue_throttle: assert_list_leaf_cfs_rq(rq); hrtick_update(rq); } -static void set_next_buddy(struct sched_entity *se); - /* * The dequeue_task method is called before nr_running is * decreased. We remove the task from the rbtree and @@ -7148,6 +7181,22 @@ balance_fair(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) } #endif /* CONFIG_SMP */ +static long wakeup_latency_gran(struct sched_entity *curr, struct sched_entity *se) +{ + long latency_offset = se->latency_offset; + + /* + * A negative latency weigth means that the sched_entity has latency + * requirement that needs to be evaluated versus other entity. + * Otherwise, use the latency weight to evaluate how much scheduling + * delay is acceptable by se. + */ + if ((se->latency_offset < 0) || (curr->latency_offset < 0)) + latency_offset -= curr->latency_offset; + + return latency_offset; +} + static unsigned long wakeup_gran(struct sched_entity *se) { unsigned long gran = sysctl_sched_wakeup_granularity; @@ -7187,6 +7236,9 @@ wakeup_preempt_entity(struct sched_entity *curr, struct sched_entity *se) { s64 gran, vdiff = curr->vruntime - se->vruntime; + /* Take into account latency priority */ + vdiff -= wakeup_latency_gran(curr, se); + if (vdiff <= 0) return -1; diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h index b927269b84f9..a4cb8058f1b2 100644 --- a/kernel/sched/sched.h +++ b/kernel/sched/sched.h @@ -141,6 +141,17 @@ extern int sched_rr_timeslice; * Default tasks should be treated as a task with latency_nice = 0. */ #define DEFAULT_LATENCY_NICE 0 +#define DEFAULT_LATENCY_PRIO (DEFAULT_LATENCY_NICE + LATENCY_NICE_WIDTH/2) + +/* + * Convert user-nice values [ -20 ... 0 ... 19 ] + * to static latency [ 0..39 ], + * and back. + */ +#define NICE_TO_LATENCY(nice) ((nice) + DEFAULT_LATENCY_PRIO) +#define LATENCY_TO_NICE(prio) ((prio) - DEFAULT_LATENCY_PRIO) +#define NICE_LATENCY_SHIFT (SCHED_FIXEDPOINT_SHIFT) +#define NICE_LATENCY_WEIGHT_MAX (1L << NICE_LATENCY_SHIFT) /* * Increase resolution of nice-level calculations for 64-bit architectures. @@ -2114,6 +2125,7 @@ static_assert(WF_TTWU == SD_BALANCE_WAKE); extern const int sched_prio_to_weight[40]; extern const u32 sched_prio_to_wmult[40]; +extern const int sched_latency_to_weight[40]; /* * {de,en}queue flags: @@ -2442,8 +2454,8 @@ extern void check_preempt_curr(struct rq *rq, struct task_struct *p, int flags); extern const_debug unsigned int sysctl_sched_nr_migrate; extern const_debug unsigned int sysctl_sched_migration_cost; -#ifdef CONFIG_SCHED_DEBUG extern unsigned int sysctl_sched_latency; +#ifdef CONFIG_SCHED_DEBUG extern unsigned int sysctl_sched_min_granularity; extern unsigned int sysctl_sched_idle_min_granularity; extern unsigned int sysctl_sched_wakeup_granularity; -- 2.17.1