Received: by 10.192.165.148 with SMTP id m20csp5247785imm; Wed, 9 May 2018 01:45:57 -0700 (PDT) X-Google-Smtp-Source: AB8JxZrytzmcQNMGqZAtHetnIsc/gzjvGhg6+Yo4D3et3zyjJa+IIah0s08yQITPs443WNnJHhbc X-Received: by 2002:a17:902:b589:: with SMTP id a9-v6mr6273637pls.161.1525855557304; Wed, 09 May 2018 01:45:57 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1525855557; cv=none; d=google.com; s=arc-20160816; b=J1qvx3a55k7P6JAeHdXThnSNu4ukMF/mINhvwPGeClulMATwlXLxnfmKxzKa5q6fU5 VtZN7b9u684YWJX8EEYZmM7lCRYQRmbhi6ITrYH1e5jBLwfcGScLoOAwIXRnx5ARCqmu nOdyueS7mzzsYEREk4vrj31S3NKLDn6Wj6HhqPIvIIsvb3IGr5VRxzIO6mOp/dSKYSPo hYxu+Xr7gXby9UWpkFVyjX1Z4oefe+DIC2SOCCxnW0KaPXHKwdN7uIe6X4c+R0umieQg 3ZkmbT5nU1qaszENkA9VtmztiMCN7UQvLYp3tR4ULPwUD+N2rs9zvSKjHUGOX4I8+YOT JBDw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dkim-signature:arc-authentication-results; bh=cR7wvCbeV5U9V/XDZ1AJncwFEJy3kGNmmpgB5fTBoLA=; b=CAtXeryTDFGkT5ZOcxMRGRqdGouV/RKMXhqISzUSwK4ll+8SX8thfalreCVvSwChdu ZVqQLBQg5nY0WtcZdRYmJ14h+WAJwUfWA5hXRCj3vJd/Wk5hmdse437CI9GSPKRvewD8 IyNQk2h6A3mheAmd1+Ql5xiXCQFjm9IBgabaaTsx1aU/hOTBxsLhQPBixtn/HtOhYHxZ /hvS+aFUN7UZQMkRVvaXrJIdZqn5xNnk5T1QJg70nIm0p7YueJNI2n7DaGU0E5MXRf9V WIE3sS3Jg6iHBVbWLqI9HTo9LnEsTPR8h5TO07Kbnr23StWER7txhkPPWrv0QLrooKLQ M4uA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=ABpq3Tlk; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id g34-v6si26196528pld.411.2018.05.09.01.45.42; Wed, 09 May 2018 01:45:57 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=ABpq3Tlk; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S934303AbeEIIkN (ORCPT + 99 others); Wed, 9 May 2018 04:40:13 -0400 Received: from mail-pg0-f66.google.com ([74.125.83.66]:43856 "EHLO mail-pg0-f66.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S934278AbeEIIkE (ORCPT ); Wed, 9 May 2018 04:40:04 -0400 Received: by mail-pg0-f66.google.com with SMTP id k11-v6so21242361pgo.10 for ; Wed, 09 May 2018 01:40:04 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to:user-agent; bh=cR7wvCbeV5U9V/XDZ1AJncwFEJy3kGNmmpgB5fTBoLA=; b=ABpq3TlkDNlAk8oOM5k9EwmJCV2Urm4M+BSxUiWlfRI1e01e3MSkjpqr0Z6raXdGG9 uVb6T36Knpkuroy7CgULc6zcB/ZvdHR6ll9Q0TV5wsI5/KiHwgIxxvA9yjwmgZDJvPNm N/5SLbjTtCXNYTxjjNtU1/Yw+4nV9vqLEqV5U= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=cR7wvCbeV5U9V/XDZ1AJncwFEJy3kGNmmpgB5fTBoLA=; b=izc1oIOQY1ewUpFIoR+3askbErTbKyisBi7Yz9L9mFWpm/nTyYtYO2Kk7mmWFQFQ24 yWT+lfi3J3+X20S04Njp/RrkPdhQtj63aTPECJV6YDtW7h93wjxroq8TFTeMRjiBx4GN IBezATfNGAS93s397h6HuJy9v4Sfrr6xa3ru0GSherNmdXZTOYzJ2sT3N4zvDsQEMbhk p2vdvflN5jlzUUw/xLCaAbYRzmp3hFAF4T9wjwYKQ3E3fIUucUUycC6OwVf/LXGVPrqV G0lrByA6O09ZBP3977iII2EXU9zeXQtAxOaYVsQL79AuCfUSHuj89ROX04CUvFxvNtis ORxA== X-Gm-Message-State: ALQs6tBNs/K3ZKBcyQLTwXUinJshuGTYbACfqDLoD8AE1/CdYv4ny71N bkoa0v1ve6//RnU0rBYNy99ENA== X-Received: by 2002:a65:4dc7:: with SMTP id q7-v6mr25893611pgt.48.1525855203987; Wed, 09 May 2018 01:40:03 -0700 (PDT) Received: from localhost ([122.167.163.112]) by smtp.gmail.com with ESMTPSA id g11-v6sm15451869pgs.21.2018.05.09.01.40.02 (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 09 May 2018 01:40:03 -0700 (PDT) Date: Wed, 9 May 2018 14:10:01 +0530 From: Viresh Kumar To: "Rafael J. Wysocki" Cc: Joel Fernandes , Juri Lelli , Claudio Scordino , Linux Kernel Mailing List , "Rafael J . Wysocki" , Peter Zijlstra , Ingo Molnar , Patrick Bellasi , Luca Abeni , Joel Fernandes , Linux PM Subject: Re: [RFC PATCH] sched/cpufreq/schedutil: handling urgent frequency requests Message-ID: <20180509084001.bghnwpv3a3xnuxce@vireshk-i7> References: <1525704215-8683-1-git-send-email-claudio@evidence.eu.com> <20180508065435.bcht6dyb3rpp6gk5@vireshk-i7> <20180509045425.GA158882@joelaf.mtv.corp.google.com> <20180509064530.GA1681@localhost.localdomain> <20180509080644.GA76874@joelaf.mtv.corp.google.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: User-Agent: NeoMutt/20180323-120-3dd1ac Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 09-05-18, 10:30, Rafael J. Wysocki wrote: > On Wed, May 9, 2018 at 10:06 AM, Joel Fernandes wrote: > > How about this? Will use the latest request, and also doesn't do unnecessary > > irq_work_queue: I almost wrote the same stuff before I went for lunch :) > > (untested) > > -----8<-------- > > diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c > > index d2c6083304b4..6a3e42b01f52 100644 > > --- a/kernel/sched/cpufreq_schedutil.c > > +++ b/kernel/sched/cpufreq_schedutil.c > > @@ -38,7 +38,7 @@ struct sugov_policy { > > struct mutex work_lock; > > struct kthread_worker worker; > > struct task_struct *thread; > > - bool work_in_progress; > > + bool work_in_progress; /* Has kthread been kicked */ > > > > bool need_freq_update; > > }; > > @@ -92,9 +92,6 @@ static bool sugov_should_update_freq(struct sugov_policy *sg_policy, u64 time) > > !cpufreq_can_do_remote_dvfs(sg_policy->policy)) > > return false; > > > > - if (sg_policy->work_in_progress) > > - return false; > > - > > Why this change? > > Doing the below is rather pointless if work_in_progress is set, isn't it? > > You'll drop the results of it on the floor going forward anyway then AFAICS. > > > if (unlikely(sg_policy->need_freq_update)) { > > sg_policy->need_freq_update = false; > > /* > > @@ -129,8 +126,11 @@ static void sugov_update_commit(struct sugov_policy *sg_policy, u64 time, > > policy->cur = next_freq; > > trace_cpu_frequency(next_freq, smp_processor_id()); > > } else { > > - sg_policy->work_in_progress = true; > > - irq_work_queue(&sg_policy->irq_work); > > + /* work_in_progress helps us not queue unnecessarily */ > > + if (!sg_policy->work_in_progress) { > > + sg_policy->work_in_progress = true; > > + irq_work_queue(&sg_policy->irq_work); > > + } > > } > > } Right, none of the above changes are required now. > > @@ -381,13 +381,23 @@ sugov_update_shared(struct update_util_data *hook, u64 time, unsigned int flags) > > static void sugov_work(struct kthread_work *work) > > { > > struct sugov_policy *sg_policy = container_of(work, struct sugov_policy, work); > > + unsigned int freq; > > + > > + /* > > + * Hold sg_policy->update_lock just enough to handle the case where: > > + * if sg_policy->next_freq is updated before work_in_progress is set to > > + * false, we may miss queueing the new update request since > > + * work_in_progress would appear to be true. > > + */ > > + raw_spin_lock(&sg_policy->update_lock); > > + freq = sg_policy->next_freq; > > + sg_policy->work_in_progress = false; > > + raw_spin_unlock(&sg_policy->update_lock); One problem we still have is that sg_policy->update_lock is only used in the shared policy case and not in the single CPU per policy case, so the race isn't solved there yet. > > mutex_lock(&sg_policy->work_lock); > > - __cpufreq_driver_target(sg_policy->policy, sg_policy->next_freq, > > + __cpufreq_driver_target(sg_policy->policy, freq, > > CPUFREQ_RELATION_L); > > mutex_unlock(&sg_policy->work_lock); > > - > > - sg_policy->work_in_progress = false; > > } > > > > static void sugov_irq_work(struct irq_work *irq_work) -- viresh