Received: by 2002:ac0:a5a7:0:0:0:0:0 with SMTP id m36-v6csp2573706imm; Thu, 16 Aug 2018 11:44:15 -0700 (PDT) X-Google-Smtp-Source: AA+uWPz2K1o43HuvHeYrZ8uESdHsaOb1t3F/dR6YbL7EYlFugrHfQRzSMTSO8KzAwRC7HPydRnHa X-Received: by 2002:a63:f751:: with SMTP id f17-v6mr30652470pgk.410.1534445055259; Thu, 16 Aug 2018 11:44:15 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1534445055; cv=none; d=google.com; s=arc-20160816; b=bg9MCQ9PbhLt1OjBzg3PgV3ZG/Yx0tGvTpxuOOInZenRB0GfcG47Vk5oIM//BObXr9 aMaytmPF/ieBtguDeWaEXwBDu7ohQ1sSdw95ZvE7Kp80/GrrEUMIuUvCb/sp6sXspCxf /1dRnR54YP++Sc9d6lN1v64KZ7kZM9sRxSqKAWSUm551Rq3sos0QLb0qqIQWDJ9kGDTY +w3cN2mFwycwbLb7P4KtViRzoaQ7CH9ZdCsm86OyooTsbxH9keLzHtjQf6sIL0PkkZiX GsvdXW04kp9M952LOTF/GI3rs6Fw1esywTC2XTrbPvpWJhzz6IrZMASrxAQZUq8geztq h/4A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject:arc-authentication-results; bh=rNjuk9qeuFFG7E0oMn2PO+z//Yij2tPl73mqYvV5Ejo=; b=WvyWpM7kf4bt7OtlEdYp4JQDBvXSydeA2GjhTNqNLIy74mr06N9rC5/2nybxLWk5O+ U0VNrppTvZUTcew9Y5xEmF1Ekro+JRj4gIgGiQHNfAZuXhDl9Pp8R06HUylf25rIp35b wQh73VKPDhxK9oUlcNHR44HgIkQzNBkpzCJ2bShoQnOERn6gRIqdLizstOMXVyoF/C4D Nh3B1eZdFA95whUd+1q/qS9wwOtlaXKX/dpSuZjQxA2KX0yXroHt1SxSUSfHjdN7ppBm HxdUH1UyIfiyjZtPL5rXZuCfkwzdOdvxg1EU9J8zT9KOChLn9YnY0Ihdmg047KlzIpEq mguw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id u134-v6si34399pfc.244.2018.08.16.11.43.59; Thu, 16 Aug 2018 11:44:15 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2404064AbeHPSmr (ORCPT + 99 others); Thu, 16 Aug 2018 14:42:47 -0400 Received: from usa-sjc-mx-foss1.foss.arm.com ([217.140.101.70]:37936 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S2392098AbeHPSmp (ORCPT ); Thu, 16 Aug 2018 14:42:45 -0400 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.72.51.249]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id F28D17A9; Thu, 16 Aug 2018 08:43:29 -0700 (PDT) Received: from [0.0.0.0] (e107985-lin.emea.arm.com [10.4.12.239]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id B6A743F5BD; Thu, 16 Aug 2018 08:43:21 -0700 (PDT) Subject: Re: [PATCH v3 07/14] sched/core: uclamp: enforce last task UCLAMP_MAX To: Patrick Bellasi , linux-kernel@vger.kernel.org, linux-pm@vger.kernel.org Cc: Ingo Molnar , Peter Zijlstra , Tejun Heo , "Rafael J . Wysocki" , Viresh Kumar , Vincent Guittot , Paul Turner , Morten Rasmussen , Juri Lelli , Todd Kjos , Joel Fernandes , Steve Muckle , Suren Baghdasaryan References: <20180806163946.28380-1-patrick.bellasi@arm.com> <20180806163946.28380-8-patrick.bellasi@arm.com> From: Dietmar Eggemann Message-ID: <2366fe11-db1f-4f39-df03-960535611319@arm.com> Date: Thu, 16 Aug 2018 17:43:18 +0200 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:52.0) Gecko/20100101 Thunderbird/52.9.1 MIME-Version: 1.0 In-Reply-To: <20180806163946.28380-8-patrick.bellasi@arm.com> Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-GB Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 08/06/2018 06:39 PM, Patrick Bellasi wrote: > When a util_max clamped task sleeps, its clamp constraints are removed > from the CPU. However, the blocked utilization on that CPU can still be > higher than the max clamp value enforced while that task was running. > This max clamp removal when a CPU is going to be idle could thus allow > unwanted CPU frequency increases, right while the task is not running. So 'rq->uclamp.flags == UCLAMP_FLAG_IDLE' means CPU is IDLE because non-clamped tasks are tracked as well ((group_id = 0)). Maybe this is worth mentioning here? > This can happen, for example, where there is another (smaller) task > running on a different CPU of the same frequency domain. > In this case, when we aggregate the utilization of all the CPUs in a > shared frequency domain, schedutil can still see the full non clamped > blocked utilization of all the CPUs and thus eventually increase the > frequency. > > Let's fix this by using: > > uclamp_cpu_put_id(UCLAMP_MAX) > uclamp_cpu_update(last_clamp_value) > > to detect when a CPU has no more RUNNABLE clamped tasks and to flag this > condition. Thus, while a CPU is idle, we can still enforce the last used > clamp value for it. > > To the contrary, we do not track any UCLAMP_MIN since, while a CPU is > idle, we don't want to enforce any minimum frequency > Indeed, we rely just on blocked load decay to smoothly reduce the > frequency. [...] > diff --git a/kernel/sched/core.c b/kernel/sched/core.c > index bc2beedec7bf..ff76b000bbe8 100644 > --- a/kernel/sched/core.c > +++ b/kernel/sched/core.c > @@ -906,7 +906,8 @@ uclamp_group_find(int clamp_id, unsigned int clamp_value) > * For the specified clamp index, this method computes the new CPU utilization > * clamp to use until the next change on the set of RUNNABLE tasks on that CPU. > */ > -static inline void uclamp_cpu_update(struct rq *rq, int clamp_id) > +static inline void uclamp_cpu_update(struct rq *rq, int clamp_id, > + unsigned int last_clamp_value) > { > struct uclamp_group *uc_grp = &rq->uclamp.group[clamp_id][0]; > int max_value = UCLAMP_NOT_VALID; > @@ -924,6 +925,19 @@ static inline void uclamp_cpu_update(struct rq *rq, int clamp_id) The condition: if (!uclamp_group_active(uc_grp, group_id)) continue; in 'for (group_id = 0; group_id <= CONFIG_UCLAMP_GROUPS_COUNT; ++group_id) {}' makes sure that 'max_value == UCLAMP_NOT_VALID' is true for the if condition (*): > if (max_value >= SCHED_CAPACITY_SCALE) > break; > } > + > + /* > + * Just for the UCLAMP_MAX value, in case there are no RUNNABLE > + * task, we keep the CPU clamped to the last task's clamp value. > + * This avoids frequency spikes to MAX when one CPU, with an high > + * blocked utilization, sleeps and another CPU, in the same frequency > + * domain, do not see anymore the clamp on the first CPU. > + */ > + if (clamp_id == UCLAMP_MAX && max_value == UCLAMP_NOT_VALID) { > + rq->uclamp.flags |= UCLAMP_FLAG_IDLE; > + max_value = last_clamp_value; > + } > + (*): So the uc_grp[group_id].value stays last_clamp_value? What do you do when the blocked utilization decays below this enforced last_clamp_value on that CPU? I assume there are plenty of this kind of corner cases because we have blocked signals (including all tasks) and clamping (including runnable tasks).