Received: by 10.192.165.156 with SMTP id m28csp961006imm; Wed, 18 Apr 2018 01:15:07 -0700 (PDT) X-Google-Smtp-Source: AIpwx4+WxO6pPz0RhKj1OfGjHccjEuaiY78OdW79s7hz71RXK6n7MHp/IY3ZSHMTfZDsPxXbH7Ci X-Received: by 10.98.137.218 with SMTP id n87mr1121881pfk.48.1524039307732; Wed, 18 Apr 2018 01:15:07 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1524039307; cv=none; d=google.com; s=arc-20160816; b=Znwij1so1Ot4Js9QHRgW52dYr+/rddQoF1SrBFsR9hFs0lQDf8yhuIkdzSxohGsNPR rj0HU49Sx+mzMTT7EuDYuikONk5GMAwcmcGHlNo5OPbbC06RaYUFRTBBg0umb66x5EHn yI2ESr7ED949f4hN52mgd8juG8pk5uf/H3RONq4u4TR2CLe/gjUSGEhvqeobmzGP6/y1 +Dmq39CchkKsu+fTwFBY7PTxvXG6f6WHvq+ADqLbPe+KwJmf6ORkr4S7/eCitrA3HrO1 Xoj6gXOwBPcuMWfloTaZrxr9AsDxYobJBzYccg8nBvp7PYMWYbvV0q40caqMPsbrhJir LW2g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:arc-authentication-results; bh=JBun0DtGqYp2TwoNxX6+JkeWkcrl3EOhFmwHzkA+an4=; b=BU+PWVhE8x3rtwX9Ofsa4Mr5gr4ylgG7fogtU989CxWJWXDwqi6kCaUUZeqUGg/oN6 BkX1YEcpf/IUI7j+gPT6EHVeHpPI3FWkPGJWxWbJZXjfBEv9fCP2i+var32ynwi8t2Xh nK1blBrW1EbVRVUWziddqJRvndQKD/kMkv6NfBmsQA16Jq6BKlIEZ3C0c2cPvtuEjOJi 4dX23KTCOl3G6CT25ucgXCZzce7R0T1l+6YV1+mqDgvN6ltRMONv9W85glwV75fCWLoB kibRuv5DUQjYWVwTin93PIjCsxjxjNKl94xBXLlSqsCUrilkgEaGbmzPwa7Kru6cSf0c sDuw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id f68si748582pfb.30.2018.04.18.01.14.53; Wed, 18 Apr 2018 01:15:07 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753560AbeDRINr (ORCPT + 99 others); Wed, 18 Apr 2018 04:13:47 -0400 Received: from foss.arm.com ([217.140.101.70]:51444 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1753251AbeDRINo (ORCPT ); Wed, 18 Apr 2018 04:13:44 -0400 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.72.51.249]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 880511435; Wed, 18 Apr 2018 01:13:44 -0700 (PDT) Received: from e108498-lin.cambridge.arm.com (e108498-lin.cambridge.arm.com [10.1.210.84]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 7404D3F487; Wed, 18 Apr 2018 01:13:41 -0700 (PDT) Date: Wed, 18 Apr 2018 09:13:39 +0100 From: Quentin Perret To: Leo Yan Cc: Dietmar Eggemann , linux-kernel@vger.kernel.org, Peter Zijlstra , Thara Gopinath , linux-pm@vger.kernel.org, Morten Rasmussen , Chris Redpath , Patrick Bellasi , Valentin Schneider , "Rafael J . Wysocki" , Greg Kroah-Hartman , Vincent Guittot , Viresh Kumar , Todd Kjos , Joel Fernandes , Juri Lelli , Steve Muckle , Eduardo Valentin Subject: Re: [RFC PATCH v2 4/6] sched/fair: Introduce an energy estimation helper function Message-ID: <20180418081339.GB3943@e108498-lin.cambridge.arm.com> References: <20180406153607.17815-1-dietmar.eggemann@arm.com> <20180406153607.17815-5-dietmar.eggemann@arm.com> <20180417152213.GC18509@leoy-ThinkPad-X240s> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20180417152213.GC18509@leoy-ThinkPad-X240s> User-Agent: Mutt/1.8.3 (2017-05-23) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tuesday 17 Apr 2018 at 23:22:13 (+0800), Leo Yan wrote: > On Fri, Apr 06, 2018 at 04:36:05PM +0100, Dietmar Eggemann wrote: > > From: Quentin Perret > > > > In preparation for the definition of an energy-aware wakeup path, a > > helper function is provided to estimate the consequence on system energy > > when a specific task wakes-up on a specific CPU. compute_energy() > > estimates the OPPs to be reached by all frequency domains and estimates > > the consumption of each online CPU according to its energy model and its > > percentage of busy time. > > > > Cc: Ingo Molnar > > Cc: Peter Zijlstra > > Signed-off-by: Quentin Perret > > Signed-off-by: Dietmar Eggemann > > --- > > include/linux/sched/energy.h | 20 +++++++++++++ > > kernel/sched/fair.c | 68 ++++++++++++++++++++++++++++++++++++++++++++ > > kernel/sched/sched.h | 2 +- > > 3 files changed, 89 insertions(+), 1 deletion(-) > > > > diff --git a/include/linux/sched/energy.h b/include/linux/sched/energy.h > > index 941071eec013..b4110b145228 100644 > > --- a/include/linux/sched/energy.h > > +++ b/include/linux/sched/energy.h > > @@ -27,6 +27,24 @@ static inline bool sched_energy_enabled(void) > > return static_branch_unlikely(&sched_energy_present); > > } > > > > +static inline > > +struct capacity_state *find_cap_state(int cpu, unsigned long util) > > +{ > > + struct sched_energy_model *em = *per_cpu_ptr(energy_model, cpu); > > + struct capacity_state *cs = NULL; > > + int i; > > + > > + util += util >> 2; > > + > > + for (i = 0; i < em->nr_cap_states; i++) { > > + cs = &em->cap_states[i]; > > + if (cs->cap >= util) > > + break; > > + } > > + > > + return cs; > > 'cs' is possible to return NULL. Only if em-nr_cap_states==0, and that shouldn't be possible if sched_energy_present==True, so this code should be safe :-) > > > +} > > + > > static inline struct cpumask *freq_domain_span(struct freq_domain *fd) > > { > > return &fd->span; > > @@ -42,6 +60,8 @@ struct freq_domain; > > static inline bool sched_energy_enabled(void) { return false; } > > static inline struct cpumask > > *freq_domain_span(struct freq_domain *fd) { return NULL; } > > +static inline struct capacity_state > > +*find_cap_state(int cpu, unsigned long util) { return NULL; } > > static inline void init_sched_energy(void) { } > > #define for_each_freq_domain(fdom) for (; fdom; fdom = NULL) > > #endif > > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > > index 6960e5ef3c14..8cb9fb04fff2 100644 > > --- a/kernel/sched/fair.c > > +++ b/kernel/sched/fair.c > > @@ -6633,6 +6633,74 @@ static int wake_cap(struct task_struct *p, int cpu, int prev_cpu) > > } > > > > /* > > + * Returns the util of "cpu" if "p" wakes up on "dst_cpu". > > + */ > > +static unsigned long cpu_util_next(int cpu, struct task_struct *p, int dst_cpu) > > +{ > > + unsigned long util, util_est; > > + struct cfs_rq *cfs_rq; > > + > > + /* Task is where it should be, or has no impact on cpu */ > > + if ((task_cpu(p) == dst_cpu) || (cpu != task_cpu(p) && cpu != dst_cpu)) > > + return cpu_util(cpu); > > + > > + cfs_rq = &cpu_rq(cpu)->cfs; > > + util = READ_ONCE(cfs_rq->avg.util_avg); > > + > > + if (dst_cpu == cpu) > > + util += task_util(p); > > + else > > + util = max_t(long, util - task_util(p), 0); > > I tried to understand the logic at here, below code is more clear for > myself: > > int prev_cpu = task_cpu(p); > > cfs_rq = &cpu_rq(cpu)->cfs; > util = READ_ONCE(cfs_rq->avg.util_avg); > > /* Bail out if src and dst CPUs are the same one */ > if (prev_cpu == cpu && dst_cpu == cpu) > return util; > > /* Remove task utilization for src CPU */ > if (cpu == prev_cpu) > util = max_t(long, util - task_util(p), 0); > > /* Add task utilization for dst CPU */ > if (dst_cpu == cpu) > util += task_util(p); > > BTW, CPU utilization is decayed value and task_util() is not decayed > value, so 'util - task_util(p)' calculates a smaller value than the > prev CPU pure utilization, right? task_util() is the raw PELT signal, without UTIL_EST, so I think it's fine to do `util - task_util()`. > > Another question is can we reuse the function cpu_util_wake() and > just compenstate task util for dst cpu? Well it's not that simple. cpu_util_wake() will give you the max between the util_avg and the util_est value, so which task_util() should you add to it ? The util_avg or the uti_est value ? Here we are trying to predict what will be the cpu_util signal in the future, so the only always-correct implementation of this function has to predict what will be the CPU util_avg and util_est signals in parallel and take the max of the two. > > > + if (sched_feat(UTIL_EST)) { > > + util_est = READ_ONCE(cfs_rq->avg.util_est.enqueued); > > + if (dst_cpu == cpu) > > + util_est += _task_util_est(p); > > + else > > + util_est = max_t(long, util_est - _task_util_est(p), 0); > > + util = max(util, util_est); > > + } > > + > > + return min_t(unsigned long, util, capacity_orig_of(cpu)); > > +} > > + > > +/* > > + * Estimates the system level energy assuming that p wakes-up on dst_cpu. > > + * > > + * compute_energy() is safe to call only if an energy model is available for > > + * the platform, which is when sched_energy_enabled() is true. > > + */ > > +static unsigned long compute_energy(struct task_struct *p, int dst_cpu) > > +{ > > + unsigned long util, max_util, sum_util; > > + struct capacity_state *cs; > > + unsigned long energy = 0; > > + struct freq_domain *fd; > > + int cpu; > > + > > + for_each_freq_domain(fd) { > > + max_util = sum_util = 0; > > + for_each_cpu_and(cpu, freq_domain_span(fd), cpu_online_mask) { > > + util = cpu_util_next(cpu, p, dst_cpu); > > + util += cpu_util_dl(cpu_rq(cpu)); > > + max_util = max(util, max_util); > > + sum_util += util; > > + } > > + > > + /* > > + * Here we assume that the capacity states of CPUs belonging to > > + * the same frequency domains are shared. Hence, we look at the > > + * capacity state of the first CPU and re-use it for all. > > + */ > > + cpu = cpumask_first(freq_domain_span(fd)); > > + cs = find_cap_state(cpu, max_util); > > + energy += cs->power * sum_util / cs->cap; > > + } > > This means all CPUs will be iterated for calculation, the complexity is > O(n)... > > > + return energy; > > +} > > + > > +/* > > * select_task_rq_fair: Select target runqueue for the waking task in domains > > * that have the 'sd_flag' flag set. In practice, this is SD_BALANCE_WAKE, > > * SD_BALANCE_FORK, or SD_BALANCE_EXEC. > > diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h > > index 5d552c0d7109..6eb38f41d5d9 100644 > > --- a/kernel/sched/sched.h > > +++ b/kernel/sched/sched.h > > @@ -2156,7 +2156,7 @@ static inline void cpufreq_update_util(struct rq *rq, unsigned int flags) {} > > # define arch_scale_freq_invariant() false > > #endif > > > > -#ifdef CONFIG_CPU_FREQ_GOV_SCHEDUTIL > > +#ifdef CONFIG_SMP > > static inline unsigned long cpu_util_dl(struct rq *rq) > > { > > return (rq->dl.running_bw * SCHED_CAPACITY_SCALE) >> BW_SHIFT; > > -- > > 2.11.0 > >