Received: by 10.192.165.148 with SMTP id m20csp5344030imm; Wed, 9 May 2018 03:35:15 -0700 (PDT) X-Google-Smtp-Source: AB8JxZov0+MqF6Up2zhVwYTh3FO7kblcWn4nlo0jkiW07oMk/RDVOIJNC44B/OqUyc82OZxnE45/ X-Received: by 2002:a17:902:8d8c:: with SMTP id v12-v6mr44918031plo.366.1525862115044; Wed, 09 May 2018 03:35:15 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1525862115; cv=none; d=google.com; s=arc-20160816; b=Gfm2Tta8RQcqRu2wbSlFxc/2bogw3LtOiCb8PIR/YePpvSQUi9Eo9rjeyeX4E9BPSR N61VYDm1wbZZMCat0MtyWGtkY9bdT109ewV1lXIqaqsurmyCKPLfCGPBG5ihEhU9unGN +xCVVKdyU8fcfYpPuPIDKg+dZ8VR/UBIlwPEY40pqfthNNhKKwmo7sqB3ef1V5MTD/8p 2AwACWNg+B3W3M+Ok4tR3fI3XCH4rzizscQOVAvCli42MAjsMtozu66aMWdb1GaFyITk t7t2g14leu1m+Wnw/6S9q5BLRpFKzjVH/00gZ1rZRkg+q8oZ+qHO/cJ8rksyOUKOfx5k mBHA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dkim-signature:arc-authentication-results; bh=HyBpY5nMJDYwsU9GmJ11VSAEut8/QmfVBCHtwNats70=; b=urwS2Gl7cIDEAZEfDkGwnOpmOkySFioRaUZ53XdWpg5Bo6eX/NJyIyh5v/jBiGqK1Q hiSuiUYSHpZF1rZ0UAExPXHqxqjhyQDoAREfosBAAmzys7h1mZztHsta67PZR5zcrR3S G0qp2ew2MYH3brAVuCQ4yw7HLzetGk1px0GgHQ2IDgS00zIO+1Yo+qVVU1z3+XD3nc7M 4mUIM0S/t9Tzofx7TKh6conYX9d7hZWpFlFgvqPSGT/k90Pwre2UU3ubC5IxZB9u8vfx K24fwhByu4/XmNeOXDfMEwrHPIAAi3JQki+14VOPiMm1xa2dbm6RjUAyxUTRdlljh2fJ rT/A== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@joelfernandes-org.20150623.gappssmtp.com header.s=20150623 header.b=pe3bFUWH; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id q7-v6si20882093pgs.193.2018.05.09.03.35.00; Wed, 09 May 2018 03:35:14 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@joelfernandes-org.20150623.gappssmtp.com header.s=20150623 header.b=pe3bFUWH; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S934563AbeEIKei (ORCPT + 99 others); Wed, 9 May 2018 06:34:38 -0400 Received: from mail-pf0-f194.google.com ([209.85.192.194]:46521 "EHLO mail-pf0-f194.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S933864AbeEIKeg (ORCPT ); Wed, 9 May 2018 06:34:36 -0400 Received: by mail-pf0-f194.google.com with SMTP id p12so25402719pff.13 for ; Wed, 09 May 2018 03:34:36 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=joelfernandes-org.20150623.gappssmtp.com; s=20150623; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to:user-agent; bh=HyBpY5nMJDYwsU9GmJ11VSAEut8/QmfVBCHtwNats70=; b=pe3bFUWHQ6jNsYCfv+o0gjZXYnoPq07EVCHGz679IMwFj8mx9Wod30Kw4dxBgbaPm8 53VQmC0S5pXOFPJtrkcvcyHh3flzOQbfesgqMMrI1+QuOaIIPDiGDmYGB+l5HlWvhDPN keivJ7byO33okPVpSWcJjAdDVomHWiPF2YNiu7aC4twAKUj1mcT8GBP+42SLHRp7xOdw auYcvOerJ7wUHuHXzKC9nR9TzFT2ld8kajc2hhIWwcdRQwWwww1/4hwXgbuGkva4nKZr 340kgX+dWwnz7yZNa9PEejmY2Sykgurkps3AeMhS2o+9zxbepqRjvzCKSv6rChfORj7p 5oUQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=HyBpY5nMJDYwsU9GmJ11VSAEut8/QmfVBCHtwNats70=; b=QdZ46p/spcezsDAdjChublOegMUJGJ2qtmzo6y4bkmhZkocqxL48GhIcq/5vckKDa2 15w8wpwgos7CgmtBlE6DAP8zgch8qtjMDPrXULEAb4cUwyzmJOt3Sg71stxuE0rDXsXR x78FHBdynB8NsS3owA64wmF6HoSN+sxcg3jK49WlhlsMFagWvwLia+EyQOEOHCbMOsZT EREsrqqThM9itxkTwvaeaAD/YtvCUwoJST2b0wBNzvh8IsQJ5wM7PqpQWD0Dvh79Lese GhExqFwsa5vM34aN0bPS6R1Rb3nnd4iUxhMhPWxhcK6p4dOIUzxQf/KOepO8Z9svXWRO I/pA== X-Gm-Message-State: ALKqPwf18D6xuk5uYXHT7wDjXXoqcl5BHqaRBqOSDL7guzQz0nw4oyLy 7Xd/10WJmekarTeCW/g39tlZGIgBAfY= X-Received: by 2002:a65:6414:: with SMTP id a20-v6mr4765795pgv.226.1525862076307; Wed, 09 May 2018 03:34:36 -0700 (PDT) Received: from localhost ([2620:0:1000:1600:3122:ea9c:d178:eb]) by smtp.gmail.com with ESMTPSA id s88sm76404960pfe.43.2018.05.09.03.34.35 (version=TLS1_2 cipher=ECDHE-RSA-CHACHA20-POLY1305 bits=256/256); Wed, 09 May 2018 03:34:35 -0700 (PDT) Date: Wed, 9 May 2018 03:34:34 -0700 From: Joel Fernandes To: Viresh Kumar Cc: "Rafael J. Wysocki" , Juri Lelli , Claudio Scordino , Linux Kernel Mailing List , "Rafael J . Wysocki" , Peter Zijlstra , Ingo Molnar , Patrick Bellasi , Luca Abeni , Joel Fernandes , Linux PM Subject: Re: [RFC PATCH] sched/cpufreq/schedutil: handling urgent frequency requests Message-ID: <20180509103434.GF76874@joelaf.mtv.corp.google.com> References: <1525704215-8683-1-git-send-email-claudio@evidence.eu.com> <20180508065435.bcht6dyb3rpp6gk5@vireshk-i7> <20180509045425.GA158882@joelaf.mtv.corp.google.com> <20180509064530.GA1681@localhost.localdomain> <20180509080644.GA76874@joelaf.mtv.corp.google.com> <20180509084001.bghnwpv3a3xnuxce@vireshk-i7> <20180509090259.GD76874@joelaf.mtv.corp.google.com> <20180509092823.sfph4gomnblb7jgr@vireshk-i7> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20180509092823.sfph4gomnblb7jgr@vireshk-i7> User-Agent: Mutt/1.9.2 (2017-12-15) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, May 09, 2018 at 02:58:23PM +0530, Viresh Kumar wrote: > On 09-05-18, 02:02, Joel Fernandes wrote: > > On Wed, May 09, 2018 at 02:10:01PM +0530, Viresh Kumar wrote: > > > Right, none of the above changes are required now. > > > > I didn't follow what you mean the changes are not required? I was developing > > against Linus mainline. Also I replied to Rafael's comment in the other > > thread. > > At least for the shared policy case the entire sequence of > sugov_should_update_freq() followed by sugov_update_commit() is > executed from within spinlock protected region and you are using the > same lock below. And so either the above two routines or the kthread > routine below will execute at a given point of time. > > So in case kthread has started doing the update and acquired the lock, > the util update handler will wait until the time work_in_progress is > set to false, that's not a problem we are trying to solve here. > > And if kthread hasn't acquired the lock yet and util handler has > started executing sugov_should_update_freq() .... > > And ^^^ this is where I understood that your earlier change is > actually required, so that we accumulate the latest updated next_freq > value. > > And with all that we wouldn't require a while loop in the kthread > code. Oh yeah, totally. So I think we are on the same page now about that. > > > > > @@ -381,13 +381,23 @@ sugov_update_shared(struct update_util_data *hook, u64 time, unsigned int flags) > > > > > static void sugov_work(struct kthread_work *work) > > > > > { > > > > > struct sugov_policy *sg_policy = container_of(work, struct sugov_policy, work); > > > > > + unsigned int freq; > > > > > + > > > > > + /* > > > > > + * Hold sg_policy->update_lock just enough to handle the case where: > > > > > + * if sg_policy->next_freq is updated before work_in_progress is set to > > > > > + * false, we may miss queueing the new update request since > > > > > + * work_in_progress would appear to be true. > > > > > + */ > > > > > + raw_spin_lock(&sg_policy->update_lock); > > > > > + freq = sg_policy->next_freq; > > > > > + sg_policy->work_in_progress = false; > > > > > + raw_spin_unlock(&sg_policy->update_lock); > > > > > > One problem we still have is that sg_policy->update_lock is only used > > > in the shared policy case and not in the single CPU per policy case, > > > so the race isn't solved there yet. > > > > True.. I can make the single CPU case acquire the update_lock very briefly > > around sugov_update_commit call in sugov_update_single. > > Rafael was very clear from the beginning that he wouldn't allow a spin > lock in the un-shared policy case :) That's fair. Probably we can just not do this trickery at all for the single case for now, incase work_in_progress is set. That way we still get the benefit for the shared case, and the single case isn't changed from what it is today. thanks, - Joel