Received: by 2002:ac0:a594:0:0:0:0:0 with SMTP id m20-v6csp3459486imm; Thu, 17 May 2018 09:05:13 -0700 (PDT) X-Google-Smtp-Source: AB8JxZqKc9zuzcQP6Dyua5jvCxQ8+ev+gtfw+n4+dE7plLkyAlBYKqpJIF4izWNRq9MYbt85LikT X-Received: by 2002:a63:77c9:: with SMTP id s192-v6mr4586073pgc.364.1526573113896; Thu, 17 May 2018 09:05:13 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1526573113; cv=none; d=google.com; s=arc-20160816; b=OWSc06TLELKIBpAI0pV8/COq38RpL5GOffSrVdmFFKHhOUZCetwxDnOcblWPqGZZPL 4cunWkiYDCkqt/M6/F/LlR3PIIUmdvEijRUa2pmPZsZDC19U3fpw9GMw5eP0rVtBHmt9 iK/tk8ZObY3/niPUse+tm9Lj2gQ5YSbDacdKcdIBzxtoHgPV+SMpoZJK1z82FHreOwjg bwhh3wneq+hzBuQPSF3ytWcVaPXXIUsbFn/pvo5E03cYEU3cgQj2gWuGS7irBIU1I5Ip kImODGgrTZMOA5MDNCArR/LbQG1TD8UFs9MomewMFK0Gk2s/UqP5eEJMNJIn5Ct7Z3p2 xKaQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dkim-signature:arc-authentication-results; bh=vQXnMRFyk+oKPooUbxs5xTRVMmRnPhj3gUT0iiHlrcg=; b=Wi8BSIsvN9MuKspqaEi6Ih++rIky/0E0riRsRfw0VQ6+m2Ibx5o+j3jOZGsp+ro4Qs LwuGio2ry71N2YFZsaZAzLm1HmEbMDEnhN2TKYvE6sBgQukph2jS+0eQZ9nTKg/923fA rYua53Y9Lzqvq0XskHY/ZQEUhmpo3RfH7LZbZrs5WPdVZzlEXAp/rZCc7OFK3AQbfBQD CUhrcB9qhh9ZCTGhWRWaDWpxjcs0kPVLZRQTz4i8gAt6gxK7rV0zfJ1Rw5ePMPCzDTev ss4CUDa90J4wwaTrjNRqsVLMTSLWyIbqWqCruSEXG1U+3qZ060ILfCVKU/2XF7vFeDJr 6ztA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@joelfernandes.org header.s=google header.b=BFJ8+vkv; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id h190-v6si4269830pgc.663.2018.05.17.09.04.51; Thu, 17 May 2018 09:05:13 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@joelfernandes.org header.s=google header.b=BFJ8+vkv; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752089AbeEQQEh (ORCPT + 99 others); Thu, 17 May 2018 12:04:37 -0400 Received: from mail-pf0-f196.google.com ([209.85.192.196]:35811 "EHLO mail-pf0-f196.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751481AbeEQQEg (ORCPT ); Thu, 17 May 2018 12:04:36 -0400 Received: by mail-pf0-f196.google.com with SMTP id x9-v6so2324058pfm.2 for ; Thu, 17 May 2018 09:04:35 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=joelfernandes.org; s=google; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to:user-agent; bh=vQXnMRFyk+oKPooUbxs5xTRVMmRnPhj3gUT0iiHlrcg=; b=BFJ8+vkvE8ajO1I39TPi58YZxoW5W77/8Lf8EnpuoO994S0FNrfvV/W6B0yodFkFgz 9Mvau7MdlCuT3/WWyzuTUqzUVFoQJS89vzvAjQ5y/iRppPqkG/AXPcdZMqVFdChzb4tg qJA61srFJMSU5PpU6yk2cpzXI0ao8tZsrCJCg= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=vQXnMRFyk+oKPooUbxs5xTRVMmRnPhj3gUT0iiHlrcg=; b=j9sLhNQCza+ClwfXft2IUBIHvymXaiIZjmC6Xrjxwj6XoB8+5+uT/j4QDHCn38ZMOB DrlhoUsLhhuFzdEewA/RYnxwp8nPvhnRw7Zkefd9lMxaVv2pBTjpAT3WjJ2BhVZv/K5w jPwOfq3Kd54wV+og3PgC/u5OJvHz3M5pNutXc9FkuMZPoOUaAArqdlABhSTsZeiAERk9 d70a2/dLu9SzGKERDoeiKk/do29Bsa2u1bFhXKXNv+tL+aeoRQf4VjXljDQT6iklk1NK 7aoIMB25i7HZizjvIHnoLtUJoB9snjzZW/HlgFUd8I+f0hE12sDUT7wEwA+AodafnR7i hRGA== X-Gm-Message-State: ALKqPwfue1Zr7D/4BuyjaiLgA0GFdKdgp9jzxQZ8JSBzr//BiEITE1KX vXN+eYlzTUzOEaKMGLOLT41drw== X-Received: by 2002:a62:14c3:: with SMTP id 186-v6mr5726506pfu.92.1526573075568; Thu, 17 May 2018 09:04:35 -0700 (PDT) Received: from localhost ([2620:0:1000:1600:3122:ea9c:d178:eb]) by smtp.gmail.com with ESMTPSA id u19-v6sm6985133pgv.25.2018.05.17.09.04.34 (version=TLS1_2 cipher=ECDHE-RSA-CHACHA20-POLY1305 bits=256/256); Thu, 17 May 2018 09:04:34 -0700 (PDT) Date: Thu, 17 May 2018 09:04:33 -0700 From: Joel Fernandes To: Juri Lelli Cc: Viresh Kumar , linux-kernel@vger.kernel.org, "Rafael J . Wysocki" , Peter Zijlstra , Ingo Molnar , Patrick Bellasi , Luca Abeni , linux-pm@vger.kernel.org, kernel-team@android.com Subject: Re: [PATCH RFC] schedutil: Allow cpufreq requests to be made even when kthread kicked Message-ID: <20180517160433.GD162290@joelaf.mtv.corp.google.com> References: <20180516224518.109891-1-joel@joelfernandes.org> <20180517070026.GA22493@localhost.localdomain> <20180517102024.s3dxo4uepujh5f65@vireshk-i7> <20180517105358.GB22493@localhost.localdomain> <20180517130704.GA139147@joelaf.mtv.corp.google.com> <20180517142823.GD22493@localhost.localdomain> <20180517144359.GA162290@joelaf.mtv.corp.google.com> <20180517152312.GG22493@localhost.localdomain> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20180517152312.GG22493@localhost.localdomain> User-Agent: Mutt/1.9.2 (2017-12-15) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, May 17, 2018 at 05:23:12PM +0200, Juri Lelli wrote: > On 17/05/18 07:43, Joel Fernandes wrote: > > On Thu, May 17, 2018 at 04:28:23PM +0200, Juri Lelli wrote: > > [...] > > > > > > We would need more locking stuff in the work handler in that case and > > > > > > I think there maybe a chance of missing the request in that solution > > > > > > if the request happens right at the end of when sugov_work returns. > > > > > > > > > > Mmm, true. Ideally we might want to use some sort of queue where to > > > > > atomically insert requests and then consume until queue is empty from > > > > > sugov kthread. > > > > > > > > IMO we don't really need a queue or anything, we should need the kthread to > > > > process the *latest* request it sees since that's the only one that matters. > > > > > > Yep, makes sense. > > > > > > > > But, I guess that's going to be too much complexity for an (hopefully) > > > > > corner case. > > > > > > > > I thought of this corner case too, I'd argue its still an improvement over > > > > not doing anything, but we could tighten this up a bit more if you wanted by > > > > > > Indeed! :) > > > > > > > doing something like this on top of my patch. Thoughts? > > > > > > > > ---8<----------------------- > > > > > > > > diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c > > > > index a87fc281893d..e45ec24b810b 100644 > > > > --- a/kernel/sched/cpufreq_schedutil.c > > > > +++ b/kernel/sched/cpufreq_schedutil.c > > > > @@ -394,6 +394,7 @@ static void sugov_work(struct kthread_work *work) > > > > unsigned int freq; > > > > unsigned long flags; > > > > > > > > +redo_work: > > > > /* > > > > * Hold sg_policy->update_lock shortly to handle the case where: > > > > * incase sg_policy->next_freq is read here, and then updated by > > > > @@ -409,6 +410,9 @@ static void sugov_work(struct kthread_work *work) > > > > __cpufreq_driver_target(sg_policy->policy, freq, > > > > CPUFREQ_RELATION_L); > > > > mutex_unlock(&sg_policy->work_lock); > > > > + > > > > + if (sg_policy->work_in_progress) > > > > + goto redo_work; > > > > > > Didn't we already queue up another irq_work at this point? > > > > Oh yeah, so the case I was thinking was if the kthread was active, while the > > new irq_work raced and finished. > > > > Since that would just mean a new kthread_work for the worker, the loop I > > mentioned above isn't needed. Infact there's already a higher level loop > > taking care of it in kthread_worker_fn as below. So the governor thread > > will not sleep and we'll keep servicing all pending requests till > > they're done. So I think we're good with my original patch. > > > > repeat: > > [...] > > if (!list_empty(&worker->work_list)) { > > work = list_first_entry(&worker->work_list, > > struct kthread_work, node); > > list_del_init(&work->node); > > } > > worker->current_work = work; > > spin_unlock_irq(&worker->lock); > > > > if (work) { > > __set_current_state(TASK_RUNNING); > > work->func(work); > > } else if (!freezing(current)) > > schedule(); > > > > try_to_freeze(); > > cond_resched(); > > goto repeat; > > Ah, right. Your original patch LGTM then. :) Cool, thanks. :) > Maybe add a comment about this higher level loop? Sure, will do. thanks, - Joel