Received: by 10.192.165.148 with SMTP id m20csp5217745imm; Wed, 9 May 2018 01:07:22 -0700 (PDT) X-Google-Smtp-Source: AB8JxZrAsXCV/Iwea6njxbslxMN6GIhTGKgqt3qfjLpF2n5nzgT9OqCn1ATtDGJU4GLMozbu+AX1 X-Received: by 2002:a63:6ec6:: with SMTP id j189-v6mr34388743pgc.86.1525853242761; Wed, 09 May 2018 01:07:22 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1525853242; cv=none; d=google.com; s=arc-20160816; b=Aw/ruvtyOokpCpB9xZ9dXUvTW5TLqDjwCTnMFLgmC+DPCo6VVg5nA40+XOsrNPnFkH 8y9kA5gLRuqIQB4m4kAWe/U+m8Um6ckSU5onZoUrzXgAcUfemD19PPGTQB0azScO3bUF VxjaiiR8WRtPoH41O+Y8mDEWGJZkqT2gpSZM72fiNXPSO94Rv0k/2iioJg8UDzsGeshU Z7jI1GXjDQLfymY2PhEpD5LPGoFXEJnMVLLcSp5u2/EMiRR/OHaKJQ1uYfGfxCY9Ihso lEF9c51XOjM2t4Yoqofvac5n3OGT/bqRsBXd2eKctmPOFfehBVdQMWuldShe1FweB/Uz TKIw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dkim-signature:arc-authentication-results; bh=nSELlwoXbrYW1QsDwi1PZayrCJ2THC4ht6oV9vAkz2Q=; b=O8LIbcBT/GD3fabycoNPnMnCGM5ph3RBzFniErDURTl10h/EO3pqB/SXUyGktuWL+w Pff2mzIA6FY/HsUHZVp9ME7hVjqa2ltGrJu9KPThVv2bayL24mamSQb3kwQWFKoz6juX YM1BATt2P2hJJPl/XO15Bt+CtYGwOLcWyu40hUUPlc6EtO2etPDvUmDOmmEHGKfuK2Em YqJiPknKCpXCvkbdxe+gt/jNYN3JK2y0hkNxEHtJKWs7GiPBV71UYW5Hgw3ZYiWOdjav RHcHPhrkDttdtQUjNrDdi5VZPBdgpFiWcH3QCN0Gd4PWtLuBrgrcoUUf5PPGZruWRdkr zY9g== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@joelfernandes-org.20150623.gappssmtp.com header.s=20150623 header.b=1dftfK+v; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id e9-v6si26159693pln.72.2018.05.09.01.07.08; Wed, 09 May 2018 01:07:22 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@joelfernandes-org.20150623.gappssmtp.com header.s=20150623 header.b=1dftfK+v; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1756185AbeEIIGw (ORCPT + 99 others); Wed, 9 May 2018 04:06:52 -0400 Received: from mail-pl0-f65.google.com ([209.85.160.65]:44300 "EHLO mail-pl0-f65.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1756118AbeEIIGr (ORCPT ); Wed, 9 May 2018 04:06:47 -0400 Received: by mail-pl0-f65.google.com with SMTP id e6-v6so3835784plt.11 for ; Wed, 09 May 2018 01:06:46 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=joelfernandes-org.20150623.gappssmtp.com; s=20150623; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to:user-agent; bh=nSELlwoXbrYW1QsDwi1PZayrCJ2THC4ht6oV9vAkz2Q=; b=1dftfK+vA6fZ8VpZWDu1IxiBVbtPxu8rU0/tLtCC6uk7tfPs4g8rX3PAs3ZSGb2RGa xEeq6hCmpSiZqltefopwGAhjrQURxl/aEwDr/dsyi/UCJ84sWYWARzrIJDCmA6nhxeI2 gFKEch1nkbMZD1XpQ6vu0C8pTa3vtOweCXQ9CcIvxn3qV1FBAMJRQKy8HXgpvU/xebqN m5/y+d1IG/brRDc5yGUjQ+ibj4C9jP+WF4i8z2eBUiKtSFe6LyGsQwH8OGroYN8LLqU7 7Lz4ugbkEYQ4gDGz7IF3pzrZ/WiJ5ppZX0w5Q3J0ngf+Uw7kD+33Q/L8QByc2kDMVoqS YE5g== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=nSELlwoXbrYW1QsDwi1PZayrCJ2THC4ht6oV9vAkz2Q=; b=M9ws/jDz2oUptxwfoRqCd2bDXvJ9juM2BGaWFQ67Ctwy/zFdIiMVdhD/y0RJdn4obw sD3rU4cTxcPZk6rBLLnAHgNDL9kDW1Tp8LzM9WJ/C05fI0upAAzf3ocpiocYv/juY79s IELNRUW7BdIZkx+inr0gQnqFyhkxJFElbYypyoiWDikHOH5HnqXdUSWyZWlIRNnaSwTn Ud/Sp5uBtSpX9JGnKqjrcX8/v7At4GwuEgAt4lplEkI/ia59mRWrPYl9UD5DWx95x3Cy HOfdMWfMEGt6VznBnC3fTAWG1eTs97wFG6DIRjQCzALHR3TrxXBWLmatiSNvKKuo364E pfEQ== X-Gm-Message-State: ALQs6tCZbxDbHsBlb+EA5MKVcGdszCDpQz+P6dkgmJbcyIEnM8Q/+nts ffIH1NdyjyWzP+omAz0vEkOK9Q== X-Received: by 2002:a17:902:d909:: with SMTP id c9-v6mr18040466plz.293.1525853205908; Wed, 09 May 2018 01:06:45 -0700 (PDT) Received: from localhost ([2620:0:1000:1600:3122:ea9c:d178:eb]) by smtp.gmail.com with ESMTPSA id i1sm34891436pfi.133.2018.05.09.01.06.45 (version=TLS1_2 cipher=ECDHE-RSA-CHACHA20-POLY1305 bits=256/256); Wed, 09 May 2018 01:06:45 -0700 (PDT) Date: Wed, 9 May 2018 01:06:44 -0700 From: Joel Fernandes To: Juri Lelli Cc: Viresh Kumar , Claudio Scordino , linux-kernel@vger.kernel.org, "Rafael J . Wysocki" , Peter Zijlstra , Ingo Molnar , Patrick Bellasi , Luca Abeni , Joel Fernandes , linux-pm@vger.kernel.org Subject: Re: [RFC PATCH] sched/cpufreq/schedutil: handling urgent frequency requests Message-ID: <20180509080644.GA76874@joelaf.mtv.corp.google.com> References: <1525704215-8683-1-git-send-email-claudio@evidence.eu.com> <20180508065435.bcht6dyb3rpp6gk5@vireshk-i7> <20180509045425.GA158882@joelaf.mtv.corp.google.com> <20180509064530.GA1681@localhost.localdomain> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20180509064530.GA1681@localhost.localdomain> User-Agent: Mutt/1.9.2 (2017-12-15) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, May 09, 2018 at 08:45:30AM +0200, Juri Lelli wrote: > On 08/05/18 21:54, Joel Fernandes wrote: > > [...] > > > Just for discussion sake, is there any need for work_in_progress? If we can > > queue multiple work say kthread_queue_work can handle it, then just queuing > > works whenever they are available should be Ok and the kthread loop can > > handle them. __cpufreq_driver_target is also protected by the work lock if > > there is any concern that can have races... only thing is rate-limiting of > > the requests, but we are doing a rate limiting, just not for the "DL > > increased utilization" type requests (which I don't think we are doing at the > > moment for urgent DL requests anyway). > > > > Following is an untested diff to show the idea. What do you think? > > > > thanks, > > > > - Joel > > > > ----8<--- > > diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c > > index d2c6083304b4..862634ff4bf3 100644 > > --- a/kernel/sched/cpufreq_schedutil.c > > +++ b/kernel/sched/cpufreq_schedutil.c > > @@ -38,7 +38,6 @@ struct sugov_policy { > > struct mutex work_lock; > > struct kthread_worker worker; > > struct task_struct *thread; > > - bool work_in_progress; > > > > bool need_freq_update; > > }; > > @@ -92,16 +91,8 @@ static bool sugov_should_update_freq(struct sugov_policy *sg_policy, u64 time) > > !cpufreq_can_do_remote_dvfs(sg_policy->policy)) > > return false; > > > > - if (sg_policy->work_in_progress) > > - return false; > > - > > if (unlikely(sg_policy->need_freq_update)) { > > sg_policy->need_freq_update = false; > > - /* > > - * This happens when limits change, so forget the previous > > - * next_freq value and force an update. > > - */ > > - sg_policy->next_freq = UINT_MAX; > > return true; > > } > > > > @@ -129,7 +120,6 @@ static void sugov_update_commit(struct sugov_policy *sg_policy, u64 time, > > policy->cur = next_freq; > > trace_cpu_frequency(next_freq, smp_processor_id()); > > } else { > > - sg_policy->work_in_progress = true; > > irq_work_queue(&sg_policy->irq_work); > > Isn't this potentially introducing unneeded irq pressure (and doing the > whole wakeup the kthread thing), while the already active kthread could > simply handle multiple back-to-back requests before going to sleep? How about this? Will use the latest request, and also doesn't do unnecessary irq_work_queue: (untested) -----8<-------- diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c index d2c6083304b4..6a3e42b01f52 100644 --- a/kernel/sched/cpufreq_schedutil.c +++ b/kernel/sched/cpufreq_schedutil.c @@ -38,7 +38,7 @@ struct sugov_policy { struct mutex work_lock; struct kthread_worker worker; struct task_struct *thread; - bool work_in_progress; + bool work_in_progress; /* Has kthread been kicked */ bool need_freq_update; }; @@ -92,9 +92,6 @@ static bool sugov_should_update_freq(struct sugov_policy *sg_policy, u64 time) !cpufreq_can_do_remote_dvfs(sg_policy->policy)) return false; - if (sg_policy->work_in_progress) - return false; - if (unlikely(sg_policy->need_freq_update)) { sg_policy->need_freq_update = false; /* @@ -129,8 +126,11 @@ static void sugov_update_commit(struct sugov_policy *sg_policy, u64 time, policy->cur = next_freq; trace_cpu_frequency(next_freq, smp_processor_id()); } else { - sg_policy->work_in_progress = true; - irq_work_queue(&sg_policy->irq_work); + /* work_in_progress helps us not queue unnecessarily */ + if (!sg_policy->work_in_progress) { + sg_policy->work_in_progress = true; + irq_work_queue(&sg_policy->irq_work); + } } } @@ -381,13 +381,23 @@ sugov_update_shared(struct update_util_data *hook, u64 time, unsigned int flags) static void sugov_work(struct kthread_work *work) { struct sugov_policy *sg_policy = container_of(work, struct sugov_policy, work); + unsigned int freq; + + /* + * Hold sg_policy->update_lock just enough to handle the case where: + * if sg_policy->next_freq is updated before work_in_progress is set to + * false, we may miss queueing the new update request since + * work_in_progress would appear to be true. + */ + raw_spin_lock(&sg_policy->update_lock); + freq = sg_policy->next_freq; + sg_policy->work_in_progress = false; + raw_spin_unlock(&sg_policy->update_lock); mutex_lock(&sg_policy->work_lock); - __cpufreq_driver_target(sg_policy->policy, sg_policy->next_freq, + __cpufreq_driver_target(sg_policy->policy, freq, CPUFREQ_RELATION_L); mutex_unlock(&sg_policy->work_lock); - - sg_policy->work_in_progress = false; } static void sugov_irq_work(struct irq_work *irq_work)