Received: by 2002:a25:ad19:0:0:0:0:0 with SMTP id y25csp3936391ybi; Mon, 29 Jul 2019 15:43:33 -0700 (PDT) X-Google-Smtp-Source: APXvYqw9O2xmZGHAxB8dEXGnRXiyDtG9T8Sj6Ev5jf1yRd3F/npYFD10eNxvtadq14FKc+VKIjYw X-Received: by 2002:a17:90a:710c:: with SMTP id h12mr111959068pjk.36.1564440213589; Mon, 29 Jul 2019 15:43:33 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1564440213; cv=none; d=google.com; s=arc-20160816; b=Agv2dZLj5zSIYh0LSG1K0KeThkTVKSbD5EFSQ3QbLNKlqG3x1eN8z0Ot98M82booCm hbdGcM7f2mV3/yFhUsgR7WFEVMitEclOUuW6NIkJad3YGtV6ns/Jg7VPzcxh6nDwFCGx rKdZYYGknuJYBvlvRYAVykO/9xoZqKJKQdc6bAvHvaflJzL57PiCqsdCacb6DkF8IXzW 3ike/48k7LaMYpKB4qukfAlnFWFqvOPnl/YFrrjhGo2GpFokOqLDmvoqVC3YdPAX+Vut BeVSPrLi4L7BM9KQ5Ajz9YsZSg95prfnSkQmZj0jjc6HjCw6kLocD+rq/HJpEgFBtWqZ 8Ohg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dmarc-filter:dkim-signature:dkim-signature; bh=jIZa/La/kVnzDamsVgNwizAADmu4McS0WibCE+Ppuhk=; b=hLsMtFStEHfHFvKm24tW84gflTgV7B+HvEtdxXQgAY68qyLuHtZbvzqw6XJUGNvWBJ PZ9kpcRlB09Z8NHhFrNl/DLgj+PSBICETT/eyne0JXGLQn7M7DIt7N9JcAkpKNEQQMoO YPrjblC4+YutqSZYbWfFiy6IhKWCkQVZZAujnB75chveppZT+BCV/FFUKxqR6dQZitTQ 4YhyVm0LyalwNG5JtV+tGL5DhdPvzBb91p81pC6Qhe+X+lYs2B7XdOnO5ozenUx0YP67 bpRhkdRfWihPJMnfhgSYBnbPvpMi9zik0xkL4nQmpCgC6yqXROvxgYvBAn9ppYW8k7VB 3zxg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@codeaurora.org header.s=default header.b=C0h6AhOp; dkim=pass header.i=@codeaurora.org header.s=default header.b=W5H6a+XL; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id q2si27535779pll.230.2019.07.29.15.43.18; Mon, 29 Jul 2019 15:43:33 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@codeaurora.org header.s=default header.b=C0h6AhOp; dkim=pass header.i=@codeaurora.org header.s=default header.b=W5H6a+XL; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728794AbfG2TBn (ORCPT + 99 others); Mon, 29 Jul 2019 15:01:43 -0400 Received: from smtp.codeaurora.org ([198.145.29.96]:42386 "EHLO smtp.codeaurora.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726680AbfG2TBn (ORCPT ); Mon, 29 Jul 2019 15:01:43 -0400 Received: by smtp.codeaurora.org (Postfix, from userid 1000) id B1B566079C; Mon, 29 Jul 2019 19:01:41 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=codeaurora.org; s=default; t=1564426901; bh=k6DT1Yj8P17N2Ic6TgQ6KjGznFDxpl3EwDYW1Vfrrmw=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=C0h6AhOpWRjvFo2Gn0mz4BQAHulOLWu1QUkhWSwJiDJU5GBFDRbWJmoGQaj8I8oO6 34Fc95jJpY4ZwCUhRRhAWEjKtV9R5icgqF69c3AdvzjHIHGfhfrcIF0ROd5+3JKtel 75ybH2suGS3/YvQWKXlKqYwnvIuZyyG9w6VwQ+rw= X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on pdx-caf-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-2.7 required=2.0 tests=ALL_TRUSTED,BAYES_00, DKIM_INVALID,DKIM_SIGNED,SPF_NONE autolearn=no autolearn_force=no version=3.4.0 Received: from localhost (i-global254.qualcomm.com [199.106.103.254]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) (Authenticated sender: ilina@smtp.codeaurora.org) by smtp.codeaurora.org (Postfix) with ESMTPSA id 8FDED60256; Mon, 29 Jul 2019 19:01:40 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=codeaurora.org; s=default; t=1564426900; bh=k6DT1Yj8P17N2Ic6TgQ6KjGznFDxpl3EwDYW1Vfrrmw=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=W5H6a+XL8y5Q8GQReJ5K2IdZBTNoNFzNnhDdMvwJCbS/Qqy+OOvFaF55FTfGuzcLH jFHUKUcqARDjiMDuiLW/sC01COo7qx8cvWoG2PsOKicnDNxgX2KHZRS9yWt/9NGE13 XNLl+NYoKwu6uI3wveLKzx07bjLHhrowOdW7Htrk= DMARC-Filter: OpenDMARC Filter v1.3.2 smtp.codeaurora.org 8FDED60256 Authentication-Results: pdx-caf-mail.web.codeaurora.org; dmarc=none (p=none dis=none) header.from=codeaurora.org Authentication-Results: pdx-caf-mail.web.codeaurora.org; spf=none smtp.mailfrom=ilina@codeaurora.org Date: Mon, 29 Jul 2019 13:01:39 -0600 From: Lina Iyer To: Doug Anderson Cc: Stephen Boyd , Andy Gross , Bjorn Andersson , linux-arm-msm , "open list:ARM/QUALCOMM SUPPORT" , Rajendra Nayak , LKML , Linux PM , mkshah@codeaurora.org Subject: Re: [PATCH V2 2/4] drivers: qcom: rpmh-rsc: avoid locking in the interrupt handler Message-ID: <20190729190139.GH18620@codeaurora.org> References: <20190722215340.3071-1-ilina@codeaurora.org> <20190722215340.3071-2-ilina@codeaurora.org> <5d3769df.1c69fb81.55d03.aa33@mx.google.com> <20190724145251.GB18620@codeaurora.org> <5d38b38e.1c69fb81.e8e5d.035b@mx.google.com> <20190724203610.GE18620@codeaurora.org> <20190725151851.GG18620@codeaurora.org> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii; format=flowed Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.11.3 (2019-02-01) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Jul 25 2019 at 09:44 -0600, Doug Anderson wrote: >Hi, > >On Thu, Jul 25, 2019 at 8:18 AM Lina Iyer wrote: >> >> On Wed, Jul 24 2019 at 17:28 -0600, Doug Anderson wrote: >> >Hi, >> > >> >On Wed, Jul 24, 2019 at 1:36 PM Lina Iyer wrote: >> >> >> >> On Wed, Jul 24 2019 at 13:38 -0600, Stephen Boyd wrote: >> >> >Quoting Lina Iyer (2019-07-24 07:52:51) >> >> >> On Tue, Jul 23 2019 at 14:11 -0600, Stephen Boyd wrote: >> >> >> >Quoting Lina Iyer (2019-07-22 14:53:38) >> >> >> >> Avoid locking in the interrupt context to improve latency. Since we >> >> >> >> don't lock in the interrupt context, it is possible that we now could >> >> >> >> race with the DRV_CONTROL register that writes the enable register and >> >> >> >> cleared by the interrupt handler. For fire-n-forget requests, the >> >> >> >> interrupt may be raised as soon as the TCS is triggered and the IRQ >> >> >> >> handler may clear the enable bit before the DRV_CONTROL is read back. >> >> >> >> >> >> >> >> Use the non-sync variant when enabling the TCS register to avoid reading >> >> >> >> back a value that may been cleared because the interrupt handler ran >> >> >> >> immediately after triggering the TCS. >> >> >> >> >> >> >> >> Signed-off-by: Lina Iyer >> >> >> >> --- >> >> >> > >> >> >> >I have to read this patch carefully. The commit text isn't convincing me >> >> >> >that it is actually safe to make this change. It mostly talks about the >> >> >> >performance improvements and how we need to fix __tcs_trigger(), which >> >> >> >is good, but I was hoping to be convinced that not grabbing the lock >> >> >> >here is safe. >> >> >> > >> >> >> >How do we ensure that drv->tcs_in_use is cleared before we call >> >> >> >tcs_write() and try to look for a free bit? Isn't it possible that we'll >> >> >> >get into a situation where the bitmap is all used up but the hardware >> >> >> >has just received an interrupt and is going to clear out a bit and then >> >> >> >an rpmh write fails with -EBUSY? >> >> >> > >> >> >> If we have a situation where there are no available free bits, we retry >> >> >> and that is part of the function. Since we have only 2 TCSes avaialble >> >> >> to write to the hardware and there could be multiple requests coming in, >> >> >> it is a very common situation. We try and acquire the drv->lock and if >> >> >> there are free TCS available and if available mark them busy and send >> >> >> our requests. If there are none available, we keep retrying. >> >> >> >> >> > >> >> >Ok. I wonder if we need some sort of barriers here too, like an >> >> >smp_mb__after_atomic()? That way we can make sure that the write to >> >> >clear the bit is seen by another CPU that could be spinning forever >> >> >waiting for that bit to be cleared? Before this change the spinlock >> >> >would be guaranteed to make these barriers for us, but now that doesn't >> >> >seem to be the case. I really hope that this whole thing can be changed >> >> >to be a mutex though, in which case we can use the bit_wait() API, etc. >> >> >to put tasks to sleep while RPMh is processing things. >> >> > >> >> We have drivers that want to send requests in atomic contexts and >> >> therefore mutex locks would not work. >> > >> >Jumping in without reading all the context, but I saw this fly by and >> >it seemed odd. If I'm way off base then please ignore... >> > >> >Can you give more details? Why are these drivers in atomic contexts? >> >If they are in atomic contexts because they are running in the context >> >of an interrupt then your next patch in the series isn't so correct. >> > >> >Also: when people submit requests in atomic context are they always >> >submitting an asynchronous request? In that case we could >> >(presumably) just use a spinlock to protect the queue of async >> >requests and a mutex for everything else? >> Yes, drivers only make async requests in interrupt contexts. > >So correct me if I'm off base, but you're saying that drivers make >requests in interrupt contexts even after your whole series and that's >why you're using spinlocks instead of mutexes. ...but then in patch >#3 in your series you say: > >> Switch over from using _irqsave/_irqrestore variants since we no longer >> race with a lock from the interrupt handler. > >Those seem like contradictions. What happens if someone is holding >the lock, then an interrupt fires, then the interrupt routine wants to >do an async request. Boom, right? > The interrupt routine is handled by the driver and only completes the waiting object (for sync requests). No other requests can be made from our interrupt handler. >> They cannot >> use the sync variants. The async and sync variants are streamlined into >> the same code path. Hence the use of spinlocks instead of mutexes >> through the critical path. > >I will perhaps defer to Stephen who was the one thinking that a mutex >would be a big win here. ...but if a mutex truly is a big win then it >doesn't seem like it'd be that hard to have a linked list (protected >by a spinlock) and then some type of async worker that: > >1. Grab the spinlock, pops one element off the linked list, release the spinlock >2. Grab the mutex, send the one element, release the mutex This would be a problem when the request is made from an irq handler. We want to keep things simple and quick. >3. Go back to step #1. > >This will keep the spinlock held for as little time as possible.