Received: by 2002:a25:824b:0:0:0:0:0 with SMTP id d11csp1801808ybn; Thu, 26 Sep 2019 02:24:37 -0700 (PDT) X-Google-Smtp-Source: APXvYqyKUGUcxIn0IePj2LDiXcDsrHI+38kEL4hS518xamugOuLOp9nYmEJfVEb1KJyb7TP9zJ58 X-Received: by 2002:a05:6402:2054:: with SMTP id bc20mr2510504edb.55.1569489877746; Thu, 26 Sep 2019 02:24:37 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1569489877; cv=none; d=google.com; s=arc-20160816; b=VEb4BB+6Fjeawzv+nhKlDBu4t5zbUw3Ifv+iLyqt1c+lxJKqvirGuaxg5dZBvUlan0 yaeHs+dCYnxIHqyd4lusLVlCJ8nhq9NAf1jPft57s4dIBjeABvgWY4jrRfsCuOfS+3s8 QLTsFeEjR2W/WnTpYdiS9hlQfETKO6CApaMKVfUVCrWMvovWRDaXs6T0w68oqu85RpbR QYzfrwM9ZPMa2tz7L8GpKPrqq+QJQsLXzUAOe1lwHYAcseAjfBr/9dbcgVLsI71zL2n7 rskVuXILxkP1WglEQz/PLT3I9zsK/0WphhhBoOYOWWqMu6d4u/ptxh/1KKUavXAuma8p Q8OQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :organization:references:in-reply-to:message-id:subject:cc:to:from :date; bh=I5gTzuXDy/8qwwX89IA2df+C/IaAop2IZlhumCV2HZ4=; b=tlzMUYFLrMSDkYG24vK41yAuanhZhI6PPebOQsto8Ynohrd9ml4n8ISLujKr21yh5O IDFWcLv3rB+0++1nErGIHwJ6XcXEUen4XZI0QYZM7WrvVw2mKLcQP/JD9GLqbogKD4/t /xtPpceJQp8ftVU4sDDD9x/ry78ensu42HWDyqvah5HeL5PNqhmjmDDhVhyvclunDn91 E1H9mk/q07HTmBMozrOZuJtkIy0pA7dN73oUno7p+HXko8k0P84V/a8fYouKEFrNXkXV bQozaXIgDtOqSyawrGZk4m5hry+tz6Xapnurm0D532lJJb0AmWC9i6G2xSc7FI6zGKx5 Wjfw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id h34si826838ede.247.2019.09.26.02.24.14; Thu, 26 Sep 2019 02:24:37 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1732833AbfIYJeR (ORCPT + 99 others); Wed, 25 Sep 2019 05:34:17 -0400 Received: from relay7-d.mail.gandi.net ([217.70.183.200]:57543 "EHLO relay7-d.mail.gandi.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1732789AbfIYJdz (ORCPT ); Wed, 25 Sep 2019 05:33:55 -0400 X-Originating-IP: 86.250.200.211 Received: from windsurf (lfbn-1-17395-211.w86-250.abo.wanadoo.fr [86.250.200.211]) (Authenticated sender: thomas.petazzoni@bootlin.com) by relay7-d.mail.gandi.net (Postfix) with ESMTPSA id 30D7820005; Wed, 25 Sep 2019 09:33:52 +0000 (UTC) Date: Wed, 25 Sep 2019 11:33:51 +0200 From: Thomas Petazzoni To: Remi Pommarel Cc: Lorenzo Pieralisi , Bjorn Helgaas , linux-pci@vger.kernel.org, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org Subject: Re: [PATCH] PCI: aardvark: Don't rely on jiffies while holding spinlock Message-ID: <20190925113351.0b53d2e9@windsurf> In-Reply-To: <20190901142303.27815-1-repk@triplefau.lt> References: <20190901142303.27815-1-repk@triplefau.lt> Organization: Bootlin X-Mailer: Claws Mail 3.17.3 (GTK+ 2.24.32; x86_64-redhat-linux-gnu) MIME-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hello Remi, Thanks for the patch, I have a few comments/questions below. On Sun, 1 Sep 2019 16:23:03 +0200 Remi Pommarel wrote: > diff --git a/drivers/pci/controller/pci-aardvark.c b/drivers/pci/controller/pci-aardvark.c > index fc0fe4d4de49..1fa6d04ad7aa 100644 > --- a/drivers/pci/controller/pci-aardvark.c > +++ b/drivers/pci/controller/pci-aardvark.c > @@ -175,7 +175,8 @@ > (PCIE_CONF_BUS(bus) | PCIE_CONF_DEV(PCI_SLOT(devfn)) | \ > PCIE_CONF_FUNC(PCI_FUNC(devfn)) | PCIE_CONF_REG(where)) > > -#define PIO_TIMEOUT_MS 1 > +#define PIO_RETRY_CNT 10 > +#define PIO_RETRY_DELAY 100 /* 100 us*/ > > #define LINK_WAIT_MAX_RETRIES 10 > #define LINK_WAIT_USLEEP_MIN 90000 > @@ -383,17 +384,16 @@ static void advk_pcie_check_pio_status(struct advk_pcie *pcie) > static int advk_pcie_wait_pio(struct advk_pcie *pcie) > { > struct device *dev = &pcie->pdev->dev; > - unsigned long timeout; > + size_t i; Is it common to use a size_t for a loop counter ? > > - timeout = jiffies + msecs_to_jiffies(PIO_TIMEOUT_MS); > - > - while (time_before(jiffies, timeout)) { > + for (i = 0; i < PIO_RETRY_CNT; ++i) { I find it more common to use post-increment for loop counters rather than pre-increment, but that's a really nitpick and I don't care much. > u32 start, isr; > > start = advk_readl(pcie, PIO_START); > isr = advk_readl(pcie, PIO_ISR); > if (!start && isr) > return 0; > + udelay(PIO_RETRY_DELAY); But the bigger issue is that this change causes a 100us delay at *every* single PIO read or write operation. Indeed, at the first iteration of the loop, the PIO operation has not completed, so you will always hit the udelay(100) a first time, and it's only at the second iteration of the loop that the PIO operation has completed (for successful PIO operations of course, which don't hit the timeout). I took a measurement around wait_pio() with sched_clock before and after the patch. Before the patch, I have measurements like this (in nanoseconds): [ 1.562801] time = 6000 [ 1.565310] time = 6000 [ 1.567809] time = 6080 [ 1.570327] time = 6080 [ 1.572836] time = 6080 [ 1.575339] time = 6080 [ 1.577858] time = 2720 [ 1.580366] time = 2720 [ 1.582862] time = 6000 [ 1.585377] time = 2720 [ 1.587890] time = 2720 [ 1.590393] time = 2720 So it takes a few microseconds for each PIO operation. With your patch applied: [ 2.267291] time = 101680 [ 2.270002] time = 100880 [ 2.272852] time = 100800 [ 2.275573] time = 100880 [ 2.278285] time = 100800 [ 2.281005] time = 100880 [ 2.283722] time = 100800 [ 2.286444] time = 100880 [ 2.289264] time = 100880 [ 2.291981] time = 100800 [ 2.294690] time = 100800 [ 2.297405] time = 100800 We're jumping to 100us for every PIO read/write operation. To be honest, I don't know if this is very important, there are not that many PIO operations, and they are not used in any performance hot path. But I thought it was worth pointing out the additional delay caused by this implementation change. Best regards, Thomas -- Thomas Petazzoni, CTO, Bootlin Embedded Linux and Kernel engineering https://bootlin.com