Received: by 10.213.65.68 with SMTP id h4csp2724897imn; Mon, 2 Apr 2018 12:49:06 -0700 (PDT) X-Google-Smtp-Source: AIpwx4/cIRFTZ4ZftImsgEoIUMaMt+fRT++8faprWlfgzHSWv9BeqE3tpoECsC4HMhgFrcPjJmqa X-Received: by 10.99.125.87 with SMTP id m23mr6932923pgn.297.1522698546426; Mon, 02 Apr 2018 12:49:06 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1522698546; cv=none; d=google.com; s=arc-20160816; b=SQPYxPsawcj6zrAOQUbR4qMFsoHpkCoLr9eBBy0OLWL/m8o/JQ2Kxla4mD0agoO2oZ U2UwU3BhVvWEsyYkUltJVsqGjb6rxQ/gZgmJYy6jQ3G20lFFrezTjDk5tNAdAPxPA/O8 YzkxeSCmgx8EJxMvFU3W+5qTo9juwLK75IvDpapJ7MryV+JXtplT2k30h3oBdc4Zyd/i Xj4o59aDlXePoETePBHyinBwvSFXIxkh6T7hkb90mHeiHaOaVNOsGvF0sC4OkgDZgxqw MUTJL0+OQkeH/kEmU0MG6IBN+CiRZrH1OpkOUAy+KtLJCSfGkXEev4tAb0ZAZ5AhZnyQ GpgA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dmarc-filter:arc-authentication-results; bh=pqGlMctHr+le9xgjUaLUEET47K/JqVreTEz5YKaYYDM=; b=O8KfYp0hPF63H4062wecWS6GkgSaX5xdI/4VDO+9rHdthZKIj0DqlRfHbn42ij+2xv H9W2PJcZa3FTaMBKeJyG3hWleJNCKVHvbQFRhaBmjPZ04W6EKg8DzubfSF/4AGOE1NkK 48neeeV3FC+26qQBwdWmL+MSv+FcKzU+B7Ysru4ovuZdSChU39c1WcA8Ic8kL7n3BG6i CojDTe1p6b+vwE695ILdzclE7un80Ek8p5MsiUCIX/dlLYexqjMnesyhczP84Ddw9n5Z 4wgZUvIZagQhzSu+EZ4KIQEwivVaek17EPw5Tx1QmhHOLyrUpwPVQJZrSE1ud9J6NET9 DU+w== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id u2-v6si997638plm.671.2018.04.02.12.48.52; Mon, 02 Apr 2018 12:49:06 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1757126AbeDBTh6 (ORCPT + 99 others); Mon, 2 Apr 2018 15:37:58 -0400 Received: from mail.kernel.org ([198.145.29.99]:33228 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754655AbeDBTh4 (ORCPT ); Mon, 2 Apr 2018 15:37:56 -0400 Received: from localhost (unknown [69.71.5.252]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id DC668217A2; Mon, 2 Apr 2018 19:37:55 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org DC668217A2 Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=kernel.org Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=helgaas@kernel.org Date: Mon, 2 Apr 2018 14:37:54 -0500 From: Bjorn Helgaas To: "Keller, Jacob E" Cc: Tal Gilboa , Tariq Toukan , Ariel Elior , Ganesh Goudar , "Kirsher, Jeffrey T" , "everest-linux-l2@cavium.com" , "intel-wired-lan@lists.osuosl.org" , "netdev@vger.kernel.org" , "linux-kernel@vger.kernel.org" , "linux-pci@vger.kernel.org" Subject: Re: [PATCH v5 03/14] PCI: Add pcie_bandwidth_capable() to compute max supported link bandwidth Message-ID: <20180402193754.GK9322@bhelgaas-glaptop.roam.corp.google.com> References: <152244269202.135666.3064353823697623332.stgit@bhelgaas-glaptop.roam.corp.google.com> <152244390359.135666.14890735614456271032.stgit@bhelgaas-glaptop.roam.corp.google.com> <31e66048-e8b8-47ba-baf5-023560b4c124@mellanox.com> <20180402004049.GA131023@bhelgaas-glaptop.roam.corp.google.com> <50346f44-de3f-b226-69ad-6de45e94e261@mellanox.com> <20180402140501.GA244675@bhelgaas-glaptop.roam.corp.google.com> <3decbba0-74a2-c906-b5b4-a04f261860eb@mellanox.com> <02874ECE860811409154E81DA85FBB5882D4980A@ORSMSX115.amr.corp.intel.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <02874ECE860811409154E81DA85FBB5882D4980A@ORSMSX115.amr.corp.intel.com> User-Agent: Mutt/1.9.2 (2017-12-15) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Apr 02, 2018 at 04:00:16PM +0000, Keller, Jacob E wrote: > > -----Original Message----- > > From: Tal Gilboa [mailto:talgi@mellanox.com] > > Sent: Monday, April 02, 2018 7:34 AM > > To: Bjorn Helgaas > > Cc: Tariq Toukan ; Keller, Jacob E > > ; Ariel Elior ; Ganesh > > Goudar ; Kirsher, Jeffrey T > > ; everest-linux-l2@cavium.com; intel-wired- > > lan@lists.osuosl.org; netdev@vger.kernel.org; linux-kernel@vger.kernel.org; > > linux-pci@vger.kernel.org > > Subject: Re: [PATCH v5 03/14] PCI: Add pcie_bandwidth_capable() to compute > > max supported link bandwidth > > > > On 4/2/2018 5:05 PM, Bjorn Helgaas wrote: > > > On Mon, Apr 02, 2018 at 10:34:58AM +0300, Tal Gilboa wrote: > > >> On 4/2/2018 3:40 AM, Bjorn Helgaas wrote: > > >>> On Sun, Apr 01, 2018 at 11:38:53PM +0300, Tal Gilboa wrote: > > >>>> On 3/31/2018 12:05 AM, Bjorn Helgaas wrote: > > >>>>> From: Tal Gilboa > > >>>>> > > >>>>> Add pcie_bandwidth_capable() to compute the max link bandwidth > > supported by > > >>>>> a device, based on the max link speed and width, adjusted by the > > encoding > > >>>>> overhead. > > >>>>> > > >>>>> The maximum bandwidth of the link is computed as: > > >>>>> > > >>>>> max_link_speed * max_link_width * (1 - encoding_overhead) > > >>>>> > > >>>>> The encoding overhead is about 20% for 2.5 and 5.0 GT/s links using > > 8b/10b > > >>>>> encoding, and about 1.5% for 8 GT/s or higher speed links using 128b/130b > > >>>>> encoding. > > >>>>> > > >>>>> Signed-off-by: Tal Gilboa > > >>>>> [bhelgaas: adjust for pcie_get_speed_cap() and pcie_get_width_cap() > > >>>>> signatures, don't export outside drivers/pci] > > >>>>> Signed-off-by: Bjorn Helgaas > > >>>>> Reviewed-by: Tariq Toukan > > >>>>> --- > > >>>>> drivers/pci/pci.c | 21 +++++++++++++++++++++ > > >>>>> drivers/pci/pci.h | 9 +++++++++ > > >>>>> 2 files changed, 30 insertions(+) > > >>>>> > > >>>>> diff --git a/drivers/pci/pci.c b/drivers/pci/pci.c > > >>>>> index 43075be79388..9ce89e254197 100644 > > >>>>> --- a/drivers/pci/pci.c > > >>>>> +++ b/drivers/pci/pci.c > > >>>>> @@ -5208,6 +5208,27 @@ enum pcie_link_width > > pcie_get_width_cap(struct pci_dev *dev) > > >>>>> return PCIE_LNK_WIDTH_UNKNOWN; > > >>>>> } > > >>>>> +/** > > >>>>> + * pcie_bandwidth_capable - calculates a PCI device's link bandwidth > > capability > > >>>>> + * @dev: PCI device > > >>>>> + * @speed: storage for link speed > > >>>>> + * @width: storage for link width > > >>>>> + * > > >>>>> + * Calculate a PCI device's link bandwidth by querying for its link speed > > >>>>> + * and width, multiplying them, and applying encoding overhead. > > >>>>> + */ > > >>>>> +u32 pcie_bandwidth_capable(struct pci_dev *dev, enum pci_bus_speed > > *speed, > > >>>>> + enum pcie_link_width *width) > > >>>>> +{ > > >>>>> + *speed = pcie_get_speed_cap(dev); > > >>>>> + *width = pcie_get_width_cap(dev); > > >>>>> + > > >>>>> + if (*speed == PCI_SPEED_UNKNOWN || *width == > > PCIE_LNK_WIDTH_UNKNOWN) > > >>>>> + return 0; > > >>>>> + > > >>>>> + return *width * PCIE_SPEED2MBS_ENC(*speed); > > >>>>> +} > > >>>>> + > > >>>>> /** > > >>>>> * pci_select_bars - Make BAR mask from the type of resource > > >>>>> * @dev: the PCI device for which BAR mask is made > > >>>>> diff --git a/drivers/pci/pci.h b/drivers/pci/pci.h > > >>>>> index 66738f1050c0..2a50172b9803 100644 > > >>>>> --- a/drivers/pci/pci.h > > >>>>> +++ b/drivers/pci/pci.h > > >>>>> @@ -261,8 +261,17 @@ void pci_disable_bridge_window(struct pci_dev > > *dev); > > >>>>> (speed) == PCIE_SPEED_2_5GT ? "2.5 GT/s" : \ > > >>>>> "Unknown speed") > > >>>>> +/* PCIe speed to Mb/s with encoding overhead: 20% for gen2, ~1.5% for > > gen3 */ > > >>>>> +#define PCIE_SPEED2MBS_ENC(speed) \ > > >>>> > > >>>> Missing gen4. > > >>> > > >>> I made it "gen3+". I think that's accurate, isn't it? The spec > > >>> doesn't seem to actually use "gen3" as a specific term, but sec 4.2.2 > > >>> says rates of 8 GT/s or higher (which I think includes gen3 and gen4) > > >>> use 128b/130b encoding. > > >>> > > >> > > >> I meant that PCIE_SPEED_16_0GT will return 0 from this macro since it wasn't > > >> added. Need to return 15754. > > > > > > Oh, duh, of course! Sorry for being dense. What about the following? > > > I included the calculation as opposed to just the magic numbers to try > > > to make it clear how they're derived. This has the disadvantage of > > > truncating the result instead of rounding, but I doubt that's > > > significant in this context. If it is, we could use the magic numbers > > > and put the computation in a comment. > > > > We can always use DIV_ROUND_UP((speed * enc_nominator), > > enc_denominator). I think this is confusing and since this introduces a > > bandwidth limit I would prefer to give a wider limit than a wrong one, > > even it is by less than 1Mb/s. My vote is for leaving it as you wrote below. > > > > > Another question: we currently deal in Mb/s, not MB/s. Mb/s has the > > > advantage of sort of corresponding to the GT/s numbers, but using MB/s > > > would have the advantage of smaller numbers that match the table here: > > > https://en.wikipedia.org/wiki/PCI_Express#History_and_revisions, > > > but I don't know what's most typical in user-facing situations. > > > What's better? > > > > I don't know what's better but for network devices we measure bandwidth > > in Gb/s, so presenting bandwidth in MB/s would mean additional > > calculations. The truth is I would have prefer to use Gb/s instead of > > Mb/s, but again, don't want to loss up to 1Gb/s. > > I prefer this version with the calculation in line since it makes > the derivation clear. Keeping them in Mb/s makes it easier to > convert to Gb/s, which is what most people would expect. OK, let's keep this patch as-is since returning Mb/s means we don't have to worry about floating point, and it sounds like we agree the truncation isn't a big deal. I'll post a proposal to convert to Gb/s when printing. > > > commit 946435491b35b7782157e9a4d1bd73071fba7709 > > > Author: Tal Gilboa > > > Date: Fri Mar 30 08:32:03 2018 -0500 > > > > > > PCI: Add pcie_bandwidth_capable() to compute max supported link > > bandwidth > > > > > > Add pcie_bandwidth_capable() to compute the max link bandwidth > > supported by > > > a device, based on the max link speed and width, adjusted by the encoding > > > overhead. > > > > > > The maximum bandwidth of the link is computed as: > > > > > > max_link_width * max_link_speed * (1 - encoding_overhead) > > > > > > 2.5 and 5.0 GT/s links use 8b/10b encoding, which reduces the raw > > bandwidth > > > available by 20%; 8.0 GT/s and faster links use 128b/130b encoding, which > > > reduces it by about 1.5%. > > > > > > The result is in Mb/s, i.e., megabits/second, of raw bandwidth. > > > > > > Signed-off-by: Tal Gilboa > > > [bhelgaas: add 16 GT/s, adjust for pcie_get_speed_cap() and > > > pcie_get_width_cap() signatures, don't export outside drivers/pci] > > > Signed-off-by: Bjorn Helgaas > > > Reviewed-by: Tariq Toukan > > > > > > diff --git a/drivers/pci/pci.c b/drivers/pci/pci.c > > > index 43075be79388..ff1e72060952 100644 > > > --- a/drivers/pci/pci.c > > > +++ b/drivers/pci/pci.c > > > @@ -5208,6 +5208,28 @@ enum pcie_link_width pcie_get_width_cap(struct > > pci_dev *dev) > > > return PCIE_LNK_WIDTH_UNKNOWN; > > > } > > > > > > +/** > > > + * pcie_bandwidth_capable - calculate a PCI device's link bandwidth capability > > > + * @dev: PCI device > > > + * @speed: storage for link speed > > > + * @width: storage for link width > > > + * > > > + * Calculate a PCI device's link bandwidth by querying for its link speed > > > + * and width, multiplying them, and applying encoding overhead. The result > > > + * is in Mb/s, i.e., megabits/second of raw bandwidth. > > > + */ > > > +u32 pcie_bandwidth_capable(struct pci_dev *dev, enum pci_bus_speed > > *speed, > > > + enum pcie_link_width *width) > > > +{ > > > + *speed = pcie_get_speed_cap(dev); > > > + *width = pcie_get_width_cap(dev); > > > + > > > + if (*speed == PCI_SPEED_UNKNOWN || *width == > > PCIE_LNK_WIDTH_UNKNOWN) > > > + return 0; > > > + > > > + return *width * PCIE_SPEED2MBS_ENC(*speed); > > > +} > > > + > > > /** > > > * pci_select_bars - Make BAR mask from the type of resource > > > * @dev: the PCI device for which BAR mask is made > > > diff --git a/drivers/pci/pci.h b/drivers/pci/pci.h > > > index 66738f1050c0..37f9299ed623 100644 > > > --- a/drivers/pci/pci.h > > > +++ b/drivers/pci/pci.h > > > @@ -261,8 +261,18 @@ void pci_disable_bridge_window(struct pci_dev *dev); > > > (speed) == PCIE_SPEED_2_5GT ? "2.5 GT/s" : \ > > > "Unknown speed") > > > > > > +/* PCIe speed to Mb/s reduced by encoding overhead */ > > > +#define PCIE_SPEED2MBS_ENC(speed) \ > > > + ((speed) == PCIE_SPEED_16_0GT ? (16000*(128/130)) : \ > > > + (speed) == PCIE_SPEED_8_0GT ? (8000*(128/130)) : \ > > > + (speed) == PCIE_SPEED_5_0GT ? (5000*(8/10)) : \ > > > + (speed) == PCIE_SPEED_2_5GT ? (2500*(8/10)) : \ > > > + 0) > > > + > > > enum pci_bus_speed pcie_get_speed_cap(struct pci_dev *dev); > > > enum pcie_link_width pcie_get_width_cap(struct pci_dev *dev); > > > +u32 pcie_bandwidth_capable(struct pci_dev *dev, enum pci_bus_speed > > *speed, > > > + enum pcie_link_width *width); > > > > > > /* Single Root I/O Virtualization */ > > > struct pci_sriov { > > >