Received: by 10.192.165.148 with SMTP id m20csp5253009imm; Wed, 9 May 2018 01:51:57 -0700 (PDT) X-Google-Smtp-Source: AB8JxZrU0hOAzJQd2hLY633FR95b0ZkuU0p1ParTrOItStzT8V9xC+QaGel1jqb9wN1RjmSBwsWE X-Received: by 2002:a63:7351:: with SMTP id d17-v6mr34280752pgn.297.1525855917831; Wed, 09 May 2018 01:51:57 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1525855917; cv=none; d=google.com; s=arc-20160816; b=WvDnXB/HunGwtUSdWEFH8aUyWdwD8LUjKIxWVH6n1l8uZkOg+SBYmgVq//niXUd8XR Q5a9GIvf9N/lPuDFoJ4TdvQZTBUrJ0GpjA/C/0H1nGsWJLJZuSUHByHzDQry6u9hsuaa 6kMdgW7Y6JaMlODB0YW+RG4/42XfBCYq1fz2zI7E4N2ClNllq7RUiTwd1Z8rmC2atQip LMD0Ofbvx/B5/fpvLVB2BDfzVUc+8g+sMMXADVyogu9LiW5Befo/XpSWVMqqUy5je50s NrEkAlhxqB+/4rrIR0LeADcgltQQV5vN3iYr7zKlvQCeWo1WSqx9dqATR1rrZt8oin4e C6eA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dkim-signature:arc-authentication-results; bh=ugXfcS9jnh1h+BfWBCPdnoyH2JdMbr19WhFpc/bo2pc=; b=uEbO1s22/2fBilJoq2qKP8RJ/HmashAgv9h9FwnCCDo6Jpt26eMG8NxYXEKEkqNKAs DYOfNM16zF3vbb4hcn93gMao8ARPwHkvilb1TIdE5CZ0et4WfW7pfRdDdrZlFoHv5CWr 2op8V/7bENZEpFK7m5JlP1/I4M+Uc7Z4TahtzELnrS1piKgeGefCgJkRVh4yuOQk3SKa B5pbwnZyveYpVobsgw7ocDAK9k+bgP6oegcAresTySWMblvOZA4CTTBP+2LtV2ixsQEK C8UZHJntx0COjw7Y2nj7r2ub8Phd6j4xFRM/vmCEZ/TYW7lgN6a/4Y07fHWAOYLWoDxs /7aw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@joelfernandes-org.20150623.gappssmtp.com header.s=20150623 header.b=UFLCnsNT; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id q4-v6si25677459plb.251.2018.05.09.01.51.43; Wed, 09 May 2018 01:51:57 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@joelfernandes-org.20150623.gappssmtp.com header.s=20150623 header.b=UFLCnsNT; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S934187AbeEIIvZ (ORCPT + 99 others); Wed, 9 May 2018 04:51:25 -0400 Received: from mail-pf0-f193.google.com ([209.85.192.193]:43199 "EHLO mail-pf0-f193.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S934168AbeEIIvS (ORCPT ); Wed, 9 May 2018 04:51:18 -0400 Received: by mail-pf0-f193.google.com with SMTP id j20so10983408pff.10 for ; Wed, 09 May 2018 01:51:17 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=joelfernandes-org.20150623.gappssmtp.com; s=20150623; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to:user-agent; bh=ugXfcS9jnh1h+BfWBCPdnoyH2JdMbr19WhFpc/bo2pc=; b=UFLCnsNTvkvXeXqbDZzjtnh6ZtCSpkeBEhX+QeJfFhQGRfOElm79lCxG3mMIVOIqVB 1WQ7qiNEr74GazMU0ljeGEKyCzfXEbg4f9HhFpsAwY3Xgid+RjdnPNSSd/4Vt/gYnT9C jOHpVCzZvtJjX9Noa1ILkWasgolZXHW3OzmlBE15+Z5FBdlV53veYkv+CF2jNJvRaLEr WwJUXM7T84G7FF+RTuwiERMa7rwCQ3kO+KbEnSv+9GwsQWwnmEqEd/+LiRSoBqlBMuR3 0cDeIcu3CqH4OihQF/HJw0qsuqTLXD8xuseFEGowVu+RIUYujTWHg397NDiKVLv6Z1Ct 7Oiw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=ugXfcS9jnh1h+BfWBCPdnoyH2JdMbr19WhFpc/bo2pc=; b=b5p8oL5jgqgtEmMz+56eutSbrVeCmdmZKsdlh/XncC5wdkKUOAuk6jJJUzQmmRcaw7 1JS+ZwtZDZORoUo3Vit4QgwpROOhg/CvUI33ze7DceNigBy+qLdki3Lw5LHkwTWbv8H1 IprFwv5ABcnyk/ymg7BxjUB85m9kBMjuC5UTRfg3oxskJlcbnZ3iH32W5KvMnnnnWcIA d30u1dbSgh992wIddbzynHJ28WY4QtKj5/MxveVxiFgIrvPFYCOpK68T8nombfvGGYd/ cPiJPg7DlMYZkbv60RL5K88yzoSFZgKmwZ2SanGikd/6T6EauO4WvLLeN+dKOulOL32d HF1Q== X-Gm-Message-State: ALQs6tC10LxVP3j4vPruEV0xRDaTCt/5QaEbNmRwbJ9Ku04Q5DbW53VO Bwr+iGgox02T5DlSu6hKVt/J/w== X-Received: by 10.98.60.16 with SMTP id j16mr43240962pfa.7.1525855877363; Wed, 09 May 2018 01:51:17 -0700 (PDT) Received: from localhost ([2620:0:1000:1600:3122:ea9c:d178:eb]) by smtp.gmail.com with ESMTPSA id o5-v6sm21240880pgv.47.2018.05.09.01.51.16 (version=TLS1_2 cipher=ECDHE-RSA-CHACHA20-POLY1305 bits=256/256); Wed, 09 May 2018 01:51:16 -0700 (PDT) Date: Wed, 9 May 2018 01:51:16 -0700 From: Joel Fernandes To: "Rafael J. Wysocki" Cc: Juri Lelli , Viresh Kumar , Claudio Scordino , Linux Kernel Mailing List , "Rafael J . Wysocki" , Peter Zijlstra , Ingo Molnar , Patrick Bellasi , Luca Abeni , Joel Fernandes , Linux PM Subject: Re: [RFC PATCH] sched/cpufreq/schedutil: handling urgent frequency requests Message-ID: <20180509085116.GC76874@joelaf.mtv.corp.google.com> References: <1525704215-8683-1-git-send-email-claudio@evidence.eu.com> <20180508065435.bcht6dyb3rpp6gk5@vireshk-i7> <20180509045425.GA158882@joelaf.mtv.corp.google.com> <20180509064530.GA1681@localhost.localdomain> <20180509080644.GA76874@joelaf.mtv.corp.google.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.9.2 (2017-12-15) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, May 09, 2018 at 10:30:37AM +0200, Rafael J. Wysocki wrote: > On Wed, May 9, 2018 at 10:06 AM, Joel Fernandes wrote: > > On Wed, May 09, 2018 at 08:45:30AM +0200, Juri Lelli wrote: > >> On 08/05/18 21:54, Joel Fernandes wrote: > >> > >> [...] > >> > >> > Just for discussion sake, is there any need for work_in_progress? If we can > >> > queue multiple work say kthread_queue_work can handle it, then just queuing > >> > works whenever they are available should be Ok and the kthread loop can > >> > handle them. __cpufreq_driver_target is also protected by the work lock if > >> > there is any concern that can have races... only thing is rate-limiting of > >> > the requests, but we are doing a rate limiting, just not for the "DL > >> > increased utilization" type requests (which I don't think we are doing at the > >> > moment for urgent DL requests anyway). > >> > > >> > Following is an untested diff to show the idea. What do you think? > >> > > >> > thanks, > >> > > >> > - Joel > >> > > >> > ----8<--- > >> > diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c > >> > index d2c6083304b4..862634ff4bf3 100644 > >> > --- a/kernel/sched/cpufreq_schedutil.c > >> > +++ b/kernel/sched/cpufreq_schedutil.c > >> > @@ -38,7 +38,6 @@ struct sugov_policy { > >> > struct mutex work_lock; > >> > struct kthread_worker worker; > >> > struct task_struct *thread; > >> > - bool work_in_progress; > >> > > >> > bool need_freq_update; > >> > }; > >> > @@ -92,16 +91,8 @@ static bool sugov_should_update_freq(struct sugov_policy *sg_policy, u64 time) > >> > !cpufreq_can_do_remote_dvfs(sg_policy->policy)) > >> > return false; > >> > > >> > - if (sg_policy->work_in_progress) > >> > - return false; > >> > - > >> > if (unlikely(sg_policy->need_freq_update)) { > >> > sg_policy->need_freq_update = false; > >> > - /* > >> > - * This happens when limits change, so forget the previous > >> > - * next_freq value and force an update. > >> > - */ > >> > - sg_policy->next_freq = UINT_MAX; > >> > return true; > >> > } > >> > > >> > @@ -129,7 +120,6 @@ static void sugov_update_commit(struct sugov_policy *sg_policy, u64 time, > >> > policy->cur = next_freq; > >> > trace_cpu_frequency(next_freq, smp_processor_id()); > >> > } else { > >> > - sg_policy->work_in_progress = true; > >> > irq_work_queue(&sg_policy->irq_work); > >> > >> Isn't this potentially introducing unneeded irq pressure (and doing the > >> whole wakeup the kthread thing), while the already active kthread could > >> simply handle multiple back-to-back requests before going to sleep? > > > > How about this? Will use the latest request, and also doesn't do unnecessary > > irq_work_queue: > > > > (untested) > > -----8<-------- > > diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c > > index d2c6083304b4..6a3e42b01f52 100644 > > --- a/kernel/sched/cpufreq_schedutil.c > > +++ b/kernel/sched/cpufreq_schedutil.c > > @@ -38,7 +38,7 @@ struct sugov_policy { > > struct mutex work_lock; > > struct kthread_worker worker; > > struct task_struct *thread; > > - bool work_in_progress; > > + bool work_in_progress; /* Has kthread been kicked */ > > > > bool need_freq_update; > > }; > > @@ -92,9 +92,6 @@ static bool sugov_should_update_freq(struct sugov_policy *sg_policy, u64 time) > > !cpufreq_can_do_remote_dvfs(sg_policy->policy)) > > return false; > > > > - if (sg_policy->work_in_progress) > > - return false; > > - > > Why this change? > > Doing the below is rather pointless if work_in_progress is set, isn't it? The issue being discussed is that if a work was already in progress, then new frequency updates will be dropped. So say even if DL increased in utilization, nothing will happen because if work_in_progress = true and need_freq_update = true, we would skip an update. In this diff, I am allowing the frequency request to be possible while work_in_progress is true. In the end the latest update will be picked. > > You'll drop the results of it on the floor going forward anyway then AFAICS. Why? If sg_policy->need_freq_update = true, sugov_should_update_freq() will return true. thanks, - Joel > > > if (unlikely(sg_policy->need_freq_update)) { > > sg_policy->need_freq_update = false; > > /* > > @@ -129,8 +126,11 @@ static void sugov_update_commit(struct sugov_policy *sg_policy, u64 time, > > policy->cur = next_freq; > > trace_cpu_frequency(next_freq, smp_processor_id()); > > } else { > > - sg_policy->work_in_progress = true; > > - irq_work_queue(&sg_policy->irq_work); > > + /* work_in_progress helps us not queue unnecessarily */ > > + if (!sg_policy->work_in_progress) { > > + sg_policy->work_in_progress = true; > > + irq_work_queue(&sg_policy->irq_work); > > + } > > } > > } > > > > @@ -381,13 +381,23 @@ sugov_update_shared(struct update_util_data *hook, u64 time, unsigned int flags) > > static void sugov_work(struct kthread_work *work) > > { > > struct sugov_policy *sg_policy = container_of(work, struct sugov_policy, work); > > + unsigned int freq; > > + > > + /* > > + * Hold sg_policy->update_lock just enough to handle the case where: > > + * if sg_policy->next_freq is updated before work_in_progress is set to > > + * false, we may miss queueing the new update request since > > + * work_in_progress would appear to be true. > > + */ > > + raw_spin_lock(&sg_policy->update_lock); > > + freq = sg_policy->next_freq; > > + sg_policy->work_in_progress = false; > > + raw_spin_unlock(&sg_policy->update_lock); > > > > mutex_lock(&sg_policy->work_lock); > > - __cpufreq_driver_target(sg_policy->policy, sg_policy->next_freq, > > + __cpufreq_driver_target(sg_policy->policy, freq, > > CPUFREQ_RELATION_L); > > mutex_unlock(&sg_policy->work_lock); > > - > > - sg_policy->work_in_progress = false; > > } > > > > static void sugov_irq_work(struct irq_work *irq_work)