Received: by 2002:ac0:a594:0:0:0:0:0 with SMTP id m20-v6csp3353325imm; Thu, 17 May 2018 07:29:15 -0700 (PDT) X-Google-Smtp-Source: AB8JxZoMQ1Ktbg1I3XmRrcF605yd6fnNx+lE83QWqJ7vvZAs4W+PeX+6eQwQub7y3W4bNuorA++O X-Received: by 2002:a62:cf43:: with SMTP id b64-v6mr5363158pfg.248.1526567355583; Thu, 17 May 2018 07:29:15 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1526567355; cv=none; d=google.com; s=arc-20160816; b=gx3YHt4dz53ylBl8Zc0LjrAn4ejixynx6Rtbqw9qaRIQ1Q1Z7hyAeTKbQ/xscfNTNv B/Bv//Uzyq2hkZECQ1BiYmI0L+rq+DnDkuecqF6OcreD6HQbzjSvw4PAG9RKqVPvPiE9 jI2FDhDvc9B747YfGDcIambc3RM0QC0zFrenFddnl9xqQKs+9RpxLmH5Msx2sknNddgG 41xvQHoWlpqG+IRe9r+HFPrJORYqEzHF8l2BWfe24XE0kfMIM0uPuXaSsSb6uvDVEaFa l68upR+xlBYTFZIq54Q/Jc3EceveggKBcqRt87SuLLX41BYHjUqDUkBioHauUW7aBXt9 hmVA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:arc-authentication-results; bh=8yr0/xzf+0twzrvALF4H5y1mdXZ/WvQwPd70kF/LLr0=; b=ZLn5CyV1IFjda+lJvjcfwwv1dWDFJtRpmmQm8zuXuauMS2aIiGehIAKhb1SfiDTslf 8RqiNzLwCzK6PUY4G4/IJb1ydDtQid0j4B02sqUtNcAt8Ic4I17uw/5TqTzSwl2jTSRL CWLFwZ4SJQOTev+TaV/8l6QuHDeYELCYHdZjNUxiVrgUF3FJ9z+3a+KVk5AgJO5i7+OY YrUP28LM52dBWvh6gZJrHWrzGjRYqhM4cWPUYzdh4qSgV/4tdpTOFvr1Z3s1t+apT1sY RPMUK74wbRY8cAzE14ybrm3eEYIwmRIx6tBZfGKPT11E8++WRxQ2KNlquw1g/YtcAwcU KOVw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id o2-v6si4883812plk.527.2018.05.17.07.29.01; Thu, 17 May 2018 07:29:15 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752245AbeEQO23 (ORCPT + 99 others); Thu, 17 May 2018 10:28:29 -0400 Received: from mail-wr0-f194.google.com ([209.85.128.194]:46962 "EHLO mail-wr0-f194.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751798AbeEQO21 (ORCPT ); Thu, 17 May 2018 10:28:27 -0400 Received: by mail-wr0-f194.google.com with SMTP id x9-v6so3051120wrl.13 for ; Thu, 17 May 2018 07:28:27 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=8yr0/xzf+0twzrvALF4H5y1mdXZ/WvQwPd70kF/LLr0=; b=h4hCILdctJZlzqmiXDGH37mPGbnLxTOE9NHd26HYS2x7+o9At5svrhcT2XPhkrOdf1 IOP9/dQX/jxmN/aeBtzUJhpZwveM8uaeEEDB5iwP4iIZbfXfpd+njH2h/in2mqe8QC6D uTg/xRV90d8w5xgoWz1aHf2RUwaYRYLpVePxgTY/4mZSk2tbvo6uV/ASpauQe3uDEb/B H2CbOH2H81NaJWAffVgThujHFt0T52gIUt0xsozNDkqKICYb64BROGQOdXAIofPGNiE+ O4vGjSrtsgkgyxwqyiaZ2DyO9Rj/2G2HvWeirHg5xftNohdhJ49lwF7mG7bmkeRNdV02 o/2Q== X-Gm-Message-State: ALKqPwdRyzODBkzRFGphBG+yxm+B0LIYbGNbVcjW6MaBHPvYqOZzFyEs u/t/zXpBUO64t6yS3DEZO9rmWw== X-Received: by 2002:adf:8486:: with SMTP id 6-v6mr4441875wrg.148.1526567306422; Thu, 17 May 2018 07:28:26 -0700 (PDT) Received: from localhost.localdomain ([151.15.207.48]) by smtp.gmail.com with ESMTPSA id v196-v6sm4375888wmf.36.2018.05.17.07.28.25 (version=TLS1_2 cipher=ECDHE-RSA-CHACHA20-POLY1305 bits=256/256); Thu, 17 May 2018 07:28:25 -0700 (PDT) Date: Thu, 17 May 2018 16:28:23 +0200 From: Juri Lelli To: Joel Fernandes Cc: Viresh Kumar , linux-kernel@vger.kernel.org, "Rafael J . Wysocki" , Peter Zijlstra , Ingo Molnar , Patrick Bellasi , Luca Abeni , Joel Fernandes , linux-pm@vger.kernel.org, kernel-team@android.com Subject: Re: [PATCH RFC] schedutil: Allow cpufreq requests to be made even when kthread kicked Message-ID: <20180517142823.GD22493@localhost.localdomain> References: <20180516224518.109891-1-joel@joelfernandes.org> <20180517070026.GA22493@localhost.localdomain> <20180517102024.s3dxo4uepujh5f65@vireshk-i7> <20180517105358.GB22493@localhost.localdomain> <20180517130704.GA139147@joelaf.mtv.corp.google.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20180517130704.GA139147@joelaf.mtv.corp.google.com> User-Agent: Mutt/1.9.2 (2017-12-15) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 17/05/18 06:07, Joel Fernandes wrote: > On Thu, May 17, 2018 at 12:53:58PM +0200, Juri Lelli wrote: > > On 17/05/18 15:50, Viresh Kumar wrote: > > > On 17-05-18, 09:00, Juri Lelli wrote: > > > > Hi Joel, > > > > > > > > On 16/05/18 15:45, Joel Fernandes (Google) wrote: > > > > > > > > [...] > > > > > > > > > @@ -382,13 +391,24 @@ sugov_update_shared(struct update_util_data *hook, u64 time, unsigned int flags) > > > > > static void sugov_work(struct kthread_work *work) > > > > > { > > > > > struct sugov_policy *sg_policy = container_of(work, struct sugov_policy, work); > > > > > + unsigned int freq; > > > > > + unsigned long flags; > > > > > + > > > > > + /* > > > > > + * Hold sg_policy->update_lock shortly to handle the case where: > > > > > + * incase sg_policy->next_freq is read here, and then updated by > > > > > + * sugov_update_shared just before work_in_progress is set to false > > > > > + * here, we may miss queueing the new update. > > > > > + */ > > > > > + raw_spin_lock_irqsave(&sg_policy->update_lock, flags); > > > > > + freq = sg_policy->next_freq; > > > > > + sg_policy->work_in_progress = false; > > > > > + raw_spin_unlock_irqrestore(&sg_policy->update_lock, flags); > > > > > > > > OK, we queue the new request up, but still we need to let this kthread > > > > activation complete and then wake it up again to service the request > > > > already queued, right? Wasn't what Claudio proposed (service back to > > > > back requests all in the same kthread activation) better from an > > > > overhead pow? > > Hmm, from that perspective, yeah. But note that my patch doesn't increase the > overhead from what it already is.. because we don't queue the irq_work again > unless work_in_progress is cleared, which wouldn't be if the kthread didn't > run yet. > > > > > > > We would need more locking stuff in the work handler in that case and > > > I think there maybe a chance of missing the request in that solution > > > if the request happens right at the end of when sugov_work returns. > > > > Mmm, true. Ideally we might want to use some sort of queue where to > > atomically insert requests and then consume until queue is empty from > > sugov kthread. > > IMO we don't really need a queue or anything, we should need the kthread to > process the *latest* request it sees since that's the only one that matters. Yep, makes sense. > > But, I guess that's going to be too much complexity for an (hopefully) > > corner case. > > I thought of this corner case too, I'd argue its still an improvement over > not doing anything, but we could tighten this up a bit more if you wanted by Indeed! :) > doing something like this on top of my patch. Thoughts? > > ---8<----------------------- > > diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c > index a87fc281893d..e45ec24b810b 100644 > --- a/kernel/sched/cpufreq_schedutil.c > +++ b/kernel/sched/cpufreq_schedutil.c > @@ -394,6 +394,7 @@ static void sugov_work(struct kthread_work *work) > unsigned int freq; > unsigned long flags; > > +redo_work: > /* > * Hold sg_policy->update_lock shortly to handle the case where: > * incase sg_policy->next_freq is read here, and then updated by > @@ -409,6 +410,9 @@ static void sugov_work(struct kthread_work *work) > __cpufreq_driver_target(sg_policy->policy, freq, > CPUFREQ_RELATION_L); > mutex_unlock(&sg_policy->work_lock); > + > + if (sg_policy->work_in_progress) > + goto redo_work; Didn't we already queue up another irq_work at this point?