Received: by 10.223.185.116 with SMTP id b49csp2161148wrg; Thu, 22 Feb 2018 09:03:35 -0800 (PST) X-Google-Smtp-Source: AH8x227Kr6pKwScYQC3fqVcpP1Fs3sO5U03anh1xArEcgqwRyRzMeP7oyPFyUyatocC+uVNJtN7u X-Received: by 10.99.140.85 with SMTP id q21mr1525407pgn.51.1519319015180; Thu, 22 Feb 2018 09:03:35 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1519319015; cv=none; d=google.com; s=arc-20160816; b=R3AKbxI2NxxQXFQXAjMpvlN0IcE7iYiPUWG+Ff3bKMNQ0bRp3tJijR8s8QZtcr+EoZ ephzt0TDCt/ALoqIqkyQIsEsmbIPcphYa0JaW2H/L9uq3gkwIs52AqTycEaiNX1JoFMH 8YxHJHXCqny+Sn7xVOMBCI+K0gH8r985XhGCaaSL+FQMjynBNBOM37DvujLpaaeFSvcd nZ1PtejraTfFYNpvtIiLiDEeqaStP59AvCLlGNtgkS7VepINkXKy9h+YjNJGmZj2kZaS sah6Yo09XQYAXpr2Cw/C9E+x1vlJESNCb4ILKk3AATiHKlG/x2PUL4eB5VDRp4onWj0T Hilw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:arc-authentication-results; bh=2B4ywUK/FLjK9NPdZN6/ReiC9BklQ8y/Da9dty3aBTw=; b=EjRVeGy2FDK1jbYdINadJ7Tu5dY3NsYv1ELiGivTBc/OTiruOWNbiGGJqD4K9qusVc AucS0zEPrD90kKd7AN1W4p4BKsg7GCShZci0bxBiQVC1kR1h2GxfEn8tCAfii5XhQkMg 17qO0P3Hc72b7d2OHn6T0YjYk6C5sLq6V2FW/8fgW45ejdbQtnzZ7BZOwtWoPW4u2Y3w X7FJSOqjqMrVPwWhzJxvS68NMayPNF5dM+75TBMZij6dRZxzv9atC7Nu85doMyww/Nni YVt6QCe7nqDZ0okMiZSnqKaRxTKKZ6iz0FH3PPReprE5MdNnPV3m/kfkD5Ok5SGh/dbe AWew== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id k9si237605pgc.628.2018.02.22.09.03.19; Thu, 22 Feb 2018 09:03:35 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S933417AbeBVRCc (ORCPT + 99 others); Thu, 22 Feb 2018 12:02:32 -0500 Received: from usa-sjc-mx-foss1.foss.arm.com ([217.140.101.70]:45302 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S933193AbeBVRC3 (ORCPT ); Thu, 22 Feb 2018 12:02:29 -0500 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.72.51.249]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 5341780D; Thu, 22 Feb 2018 09:02:29 -0800 (PST) Received: from e110439-lin.cambridge.arm.com (e110439-lin.cambridge.arm.com [10.1.210.68]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPA id CA2353F318; Thu, 22 Feb 2018 09:02:26 -0800 (PST) From: Patrick Bellasi To: linux-kernel@vger.kernel.org, linux-pm@vger.kernel.org Cc: Ingo Molnar , Peter Zijlstra , "Rafael J . Wysocki" , Viresh Kumar , Vincent Guittot , Paul Turner , Dietmar Eggemann , Morten Rasmussen , Juri Lelli , Todd Kjos , Joel Fernandes , Steve Muckle Subject: [PATCH v5 4/4] sched/fair: update util_est only on util_avg updates Date: Thu, 22 Feb 2018 17:01:53 +0000 Message-Id: <20180222170153.673-5-patrick.bellasi@arm.com> X-Mailer: git-send-email 2.15.1 In-Reply-To: <20180222170153.673-1-patrick.bellasi@arm.com> References: <20180222170153.673-1-patrick.bellasi@arm.com> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org The estimated utilization of a task is currently updated every time the task is dequeued. However, to keep overheads under control, PELT signals are effectively updated at maximum once every 1ms. Thus, for really short running tasks, it can happen that their util_avg value has not been updates since their last enqueue. If such tasks are also frequently running tasks (e.g. the kind of workload generated by hackbench) it can also happen that their util_avg is updated only every few activations. This means that updating util_est at every dequeue potentially introduces not necessary overheads and it's also conceptually wrong if the util_avg signal has never been updated during a task activation. Let's introduce a throttling mechanism on task's util_est updates to sync them with util_avg updates. To make the solution memory efficient, both in terms of space and load/store operations, we encode a synchronization flag into the LSB of util_est.enqueued. This makes util_est an even values only metric, which is still considered good enough for its purpose. The synchronization bit is (re)set by __update_load_avg_se() once the PELT signal of a task has been updated during its last activation. Such a throttling mechanism allows to keep under control util_est overheads in the wakeup hot path, thus making it a suitable mechanism which can be enabled also on high-intensity workload systems. Thus, this now switches on by default the estimation utilization scheduler feature. Suggested-by: Chris Redpath Signed-off-by: Patrick Bellasi --- Changes in v5: - set SCHED_FEAT(UTIL_EST, true) as default (Peter) --- kernel/sched/fair.c | 39 +++++++++++++++++++++++++++++++++++---- kernel/sched/features.h | 2 +- 2 files changed, 36 insertions(+), 5 deletions(-) diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 8364771f7301..1bf9a86ebc39 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -3047,6 +3047,29 @@ static inline void cfs_rq_util_change(struct cfs_rq *cfs_rq) } } +/* + * When a task is dequeued, its estimated utilization should not be update if + * its util_avg has not been updated at least once. + * This flag is used to synchronize util_avg updates with util_est updates. + * We map this information into the LSB bit of the utilization saved at + * dequeue time (i.e. util_est.dequeued). + */ +#define UTIL_EST_NEED_UPDATE_FLAG 0x1 + +static inline void cfs_se_util_change(struct sched_avg *avg) +{ + if (sched_feat(UTIL_EST)) { + struct util_est ue = READ_ONCE(avg->util_est); + + if (!(ue.enqueued & UTIL_EST_NEED_UPDATE_FLAG)) + return; + + /* Reset flag to report util_avg has been updated */ + ue.enqueued &= ~UTIL_EST_NEED_UPDATE_FLAG; + WRITE_ONCE(avg->util_est, ue); + } +} + #ifdef CONFIG_SMP /* * Approximate: @@ -3308,6 +3331,7 @@ __update_load_avg_se(u64 now, int cpu, struct cfs_rq *cfs_rq, struct sched_entit cfs_rq->curr == se)) { ___update_load_avg(&se->avg, se_weight(se), se_runnable(se)); + cfs_se_util_change(&se->avg); return 1; } @@ -5218,7 +5242,7 @@ static inline void util_est_enqueue(struct cfs_rq *cfs_rq, /* Update root cfs_rq's estimated utilization */ enqueued = READ_ONCE(cfs_rq->avg.util_est.enqueued); - enqueued += _task_util_est(p); + enqueued += (_task_util_est(p) | 0x1); WRITE_ONCE(cfs_rq->avg.util_est.enqueued, enqueued); } @@ -5310,7 +5334,7 @@ static inline void util_est_dequeue(struct cfs_rq *cfs_rq, if (cfs_rq->nr_running) { ue.enqueued = READ_ONCE(cfs_rq->avg.util_est.enqueued); ue.enqueued -= min_t(unsigned int, ue.enqueued, - _task_util_est(p)); + (_task_util_est(p) | UTIL_EST_NEED_UPDATE_FLAG)); } WRITE_ONCE(cfs_rq->avg.util_est.enqueued, ue.enqueued); @@ -5321,12 +5345,19 @@ static inline void util_est_dequeue(struct cfs_rq *cfs_rq, if (!task_sleep) return; + /* + * Skip update of task's estimated utilization if the PELT signal has + * never been updated (at least once) since last enqueue time. + */ + ue = READ_ONCE(p->se.avg.util_est); + if (ue.enqueued & UTIL_EST_NEED_UPDATE_FLAG) + return; + /* * Skip update of task's estimated utilization when its EWMA is * already ~1% close to its last activation value. */ - ue = READ_ONCE(p->se.avg.util_est); - ue.enqueued = task_util(p); + ue.enqueued = (task_util(p) | UTIL_EST_NEED_UPDATE_FLAG); last_ewma_diff = ue.enqueued - ue.ewma; if (within_margin(last_ewma_diff, (SCHED_CAPACITY_SCALE / 100))) return; diff --git a/kernel/sched/features.h b/kernel/sched/features.h index c459a4b61544..85ae8488039c 100644 --- a/kernel/sched/features.h +++ b/kernel/sched/features.h @@ -89,4 +89,4 @@ SCHED_FEAT(WA_BIAS, true) /* * UtilEstimation. Use estimated CPU utilization. */ -SCHED_FEAT(UTIL_EST, false) +SCHED_FEAT(UTIL_EST, true) -- 2.15.1