Received: by 2002:ac0:a594:0:0:0:0:0 with SMTP id m20-v6csp1416302imm; Tue, 22 May 2018 03:51:46 -0700 (PDT) X-Google-Smtp-Source: AB8JxZrEfdJfmEjatGsstZNr3aNv8owskJufrKVl9t4WkSFmGBLNDoke0BPaouX4eLGP2hYZBO04 X-Received: by 2002:a63:a341:: with SMTP id v1-v6mr2062334pgn.442.1526986306432; Tue, 22 May 2018 03:51:46 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1526986306; cv=none; d=google.com; s=arc-20160816; b=OIx4fzBHHGJ0yw9IACWF5GwdTWFcExtB3WWAUkCtEdkksGX6RG7QLAgK6zGlo5rUT6 entGh7lXk2/45COA2apcbjBEuf19NGJUPWnnCUzcM9dFhYw4WtQ8Z1HI6kxiItzxhQXN /WrPH29XlNMDEvhLzNmcCrW+5WGmVmnDivF+QTMqYYGnxNVkCZk47k/MTZXMJVaVDpKH MgjNjxIBlR/hEAeARsx/8Vgqyp7/sbxasnIS2Ee/K3MqGf1aQFSPKd4ZTk5HO0YRSkKe uGztOn8aOjsN7iMamUQKC1ABQKAT6n1z2a9arugygiZBVxXwy19UGiIatRq/oKUBVlpZ HW2g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :arc-authentication-results; bh=Mp7Lk3berIqJu/eAPCHVDnFjG3u6CwhQyMZ6R4HZ194=; b=iH3Op879ouejxeT7GoC2SVDpMKpXW1QflXuaho6xm+2hliHbhm7x92ShXolXTrDcAO S4ipbgD+aUP2FiO9ENvOFa0g74pjKkHCnycE4Q3OZeX7lHYKbQkAxjtvjNcvwht5RS6O KSNE9mw4ze0jQgmRSjDm5suHNFtQybJqwX2WCQtxyBjQXnEjiyO6vkn85KuyqOd4lPKU A9Et0TiZ7x/BH+uwzHZFTMn6WzhAVI+P91/sEEBtpeEoDC0GfTrT+eR5nNElbYZenAl9 MKZuU6yGwGdf4G7cBNeawQWmsYmxgenGHsvJPFEz3kb0g4je4Vp3toY6Lhp1jIT2be0c aabA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id i64-v6si16263552pli.274.2018.05.22.03.51.31; Tue, 22 May 2018 03:51:46 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751361AbeEVKv2 (ORCPT + 99 others); Tue, 22 May 2018 06:51:28 -0400 Received: from cloudserver094114.home.pl ([79.96.170.134]:63024 "EHLO cloudserver094114.home.pl" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751199AbeEVKv1 (ORCPT ); Tue, 22 May 2018 06:51:27 -0400 Received: from 79.184.253.39.ipv4.supernova.orange.pl (79.184.253.39) (HELO aspire.rjw.lan) by serwer1319399.home.pl (79.96.170.134) with SMTP (IdeaSmtpServer 0.83) id 9f3492a0ca45975a; Tue, 22 May 2018 12:51:24 +0200 From: "Rafael J. Wysocki" To: Viresh Kumar Cc: "Joel Fernandes (Google.)" , linux-kernel@vger.kernel.org, "Joel Fernandes (Google)" , "Rafael J . Wysocki" , Peter Zijlstra , Ingo Molnar , Patrick Bellasi , Juri Lelli , Luca Abeni , Todd Kjos , claudio@evidence.eu.com, kernel-team@android.com, linux-pm@vger.kernel.org Subject: Re: [PATCH v2] schedutil: Allow cpufreq requests to be made even when kthread kicked Date: Tue, 22 May 2018 12:50:46 +0200 Message-ID: <4393838.cQWhzXzUcj@aspire.rjw.lan> In-Reply-To: <20180522105006.hvntwnhzzvhznij2@vireshk-i7> References: <20180518185501.173552-1-joel@joelfernandes.org> <20180522103415.cuutobi5kbhj4gcw@vireshk-i7> <20180522105006.hvntwnhzzvhznij2@vireshk-i7> MIME-Version: 1.0 Content-Transfer-Encoding: 7Bit Content-Type: text/plain; charset="us-ascii" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tuesday, May 22, 2018 12:50:06 PM CEST Viresh Kumar wrote: > On 22-05-18, 16:04, Viresh Kumar wrote: > > Okay, me and Rafael were discussing this patch, locking and races around this. > > > > On 18-05-18, 11:55, Joel Fernandes (Google.) wrote: > > > diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c > > > index e13df951aca7..5c482ec38610 100644 > > > --- a/kernel/sched/cpufreq_schedutil.c > > > +++ b/kernel/sched/cpufreq_schedutil.c > > > @@ -92,9 +92,6 @@ static bool sugov_should_update_freq(struct sugov_policy *sg_policy, u64 time) > > > !cpufreq_can_do_remote_dvfs(sg_policy->policy)) > > > return false; > > > > > > - if (sg_policy->work_in_progress) > > > - return false; > > > - > > > if (unlikely(sg_policy->need_freq_update)) { > > > sg_policy->need_freq_update = false; > > > /* > > > @@ -128,7 +125,7 @@ static void sugov_update_commit(struct sugov_policy *sg_policy, u64 time, > > > > > > policy->cur = next_freq; > > > trace_cpu_frequency(next_freq, smp_processor_id()); > > > - } else { > > > + } else if (!sg_policy->work_in_progress) { > > > sg_policy->work_in_progress = true; > > > irq_work_queue(&sg_policy->irq_work); > > > } > > > @@ -291,6 +288,13 @@ static void sugov_update_single(struct update_util_data *hook, u64 time, > > > > > > ignore_dl_rate_limit(sg_cpu, sg_policy); > > > > > > + /* > > > + * For slow-switch systems, single policy requests can't run at the > > > + * moment if update is in progress, unless we acquire update_lock. > > > + */ > > > + if (sg_policy->work_in_progress) > > > + return; > > > + > > > if (!sugov_should_update_freq(sg_policy, time)) > > > return; > > > > > > @@ -382,13 +386,27 @@ sugov_update_shared(struct update_util_data *hook, u64 time, unsigned int flags) > > > static void sugov_work(struct kthread_work *work) > > > { > > > struct sugov_policy *sg_policy = container_of(work, struct sugov_policy, work); > > > + unsigned int freq; > > > + unsigned long flags; > > > + > > > + /* > > > + * Hold sg_policy->update_lock shortly to handle the case where: > > > + * incase sg_policy->next_freq is read here, and then updated by > > > + * sugov_update_shared just before work_in_progress is set to false > > > + * here, we may miss queueing the new update. > > > + * > > > + * Note: If a work was queued after the update_lock is released, > > > + * sugov_work will just be called again by kthread_work code; and the > > > + * request will be proceed before the sugov thread sleeps. > > > + */ > > > + raw_spin_lock_irqsave(&sg_policy->update_lock, flags); > > > + freq = sg_policy->next_freq; > > > + sg_policy->work_in_progress = false; > > > + raw_spin_unlock_irqrestore(&sg_policy->update_lock, flags); > > > > > > mutex_lock(&sg_policy->work_lock); > > > - __cpufreq_driver_target(sg_policy->policy, sg_policy->next_freq, > > > - CPUFREQ_RELATION_L); > > > + __cpufreq_driver_target(sg_policy->policy, freq, CPUFREQ_RELATION_L); > > > mutex_unlock(&sg_policy->work_lock); > > > - > > > - sg_policy->work_in_progress = false; > > > } > > > > And I do see a race here for single policy systems doing slow switching. > > > > Kthread Sched update > > > > sugov_work() sugov_update_single() > > > > lock(); > > // The CPU is free to rearrange below > > // two in any order, so it may clear > > // the flag first and then read next > > // freq. Lets assume it does. > > work_in_progress = false > > > > if (work_in_progress) > > return; > > > > sg_policy->next_freq = 0; > > freq = sg_policy->next_freq; > > sg_policy->next_freq = real-next-freq; > > unlock(); > > > > > > > > Is the above theory right or am I day dreaming ? :) > > And here comes the ugly fix: > > diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c > index 35826f4ec43c..1665da31862e 100644 > --- a/kernel/sched/cpufreq_schedutil.c > +++ b/kernel/sched/cpufreq_schedutil.c > @@ -283,6 +283,9 @@ static void sugov_update_single(struct update_util_data *hook, u64 time, > > ignore_dl_rate_limit(sg_cpu, sg_policy); > > + if (!policy->fast_switch_enabled) > + raw_spin_lock(&sg_policy->update_lock); > + > /* > * For slow-switch systems, single policy requests can't run at the > * moment if update is in progress, unless we acquire update_lock. > @@ -312,6 +315,9 @@ static void sugov_update_single(struct update_util_data *hook, u64 time, > } > > sugov_update_commit(sg_policy, time, next_f); > + > + if (!policy->fast_switch_enabled) > + raw_spin_unlock(&sg_policy->update_lock); > } > > static unsigned int sugov_next_freq_shared(struct sugov_cpu *sg_cpu, u64 time) > > > Ugly indeed.