Received: by 2002:a25:868d:0:0:0:0:0 with SMTP id z13csp3675023ybk; Tue, 19 May 2020 10:14:47 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzoGZROTMTmNSoTSt10yWIJWvNioheg/M3ifwjKmBlIxyWvep03DsbDqKUJgWp+Ei/fIhPj X-Received: by 2002:a17:906:6d10:: with SMTP id m16mr276204ejr.84.1589908487739; Tue, 19 May 2020 10:14:47 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1589908487; cv=none; d=google.com; s=arc-20160816; b=ZLqHJJDV82dxgZGohYZcDbmgsia6qya8s8OuWDl4Bwi8vnlK8dUrGnfc+VeVLouuM9 j2R5OwDs/nf+0EOq7Y0thNOyw3whEH3dRd5q81lv4EmzON3sT71dWhAGKmbJkvpHFyb5 JG+qCsmnU9+35Pr9jI4LR5QxzmzHbBUTvGq5/LEu1b4rSdIGe1S/mvBE3U99kKxdpQuo aM8iBHZwWmERqjZ4SwfSieu9SYI/6Te6FAzFUoWAQaGrfkZbStHxZX2QGRrPHueSeT5F TuqW/X0aWJ9FCz/WhoIA7ynhTvvWw3fXUx6hg7g4D5DS4A3hwQ4sUKVZ8eKUCWlTWeUe BmcA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=JkSx/LQjXlAf56xSCOPqQjuwfHhEaq28wKsH0guS880=; b=eyCdSl7jw4eykWTJsQYFrS49VunHnnGcuhNPr7h/bFoHQv1bK150tgrjpn6MjR7BYD sAUtYp/tl494FE6nWmoL2ejs7my29ACybJz1mOqPaUM07e9IceMX3QB+LinjgZInHFfb RHwfYaY3XI1Vy2vn7Gz2sxRSd1ImxjOH0FiELWAppiF/02JERCtIE5LLG29JLuESbUez FDcowKGp3jyv46Ws78sdzvnYh7swXP50iqw0hfyX78GaozqNXiQ46FIg59Mevxif+2TC mqUlV05kUbtxAqVDLF+AFfALEyhdmjR+2wOVrTl7h6GQPBQfvuyXvHxrvZ/ikZa8n+Wc oriQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=default header.b=TLbCPhb4; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id qh14si266578ejb.171.2020.05.19.10.14.23; Tue, 19 May 2020 10:14:47 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=default header.b=TLbCPhb4; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1729442AbgESRL4 (ORCPT + 99 others); Tue, 19 May 2020 13:11:56 -0400 Received: from mail.kernel.org ([198.145.29.99]:41238 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1729001AbgESRL4 (ORCPT ); Tue, 19 May 2020 13:11:56 -0400 Received: from mail-oi1-f176.google.com (mail-oi1-f176.google.com [209.85.167.176]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 78AA7207D8; Tue, 19 May 2020 17:11:55 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=default; t=1589908315; bh=J1Y30cmsXnTK+NpTbd/xD2zpvVMzNGnZNpopIBUAdOI=; h=References:In-Reply-To:From:Date:Subject:To:Cc:From; b=TLbCPhb4n0z4OOZth5bBxZ6VeNK677WoS6fBfxN7NTf8lXpCGiYv5YiDEsXfmiT0t XrwIh0sAqkggc1hAxQbSeB6Zi6l+JIgCmwyPx0GHCi9VI9aDsjf/KyS7UgG0Ip79XR lQUV4pbrxKCykGqXmxuD72w2Bid7Uv7Ret2WcZj4= Received: by mail-oi1-f176.google.com with SMTP id y85so364629oie.11; Tue, 19 May 2020 10:11:55 -0700 (PDT) X-Gm-Message-State: AOAM531FmMnGSLLvn+AcEpondCGO1+N3L3egPZWUs9xrCinSo2sqq+tf Lhi52ZJDwgf0AX5LfVHy2+Om2/PrleGyToBdrQ== X-Received: by 2002:a05:6808:24f:: with SMTP id m15mr347220oie.152.1589908314771; Tue, 19 May 2020 10:11:54 -0700 (PDT) MIME-Version: 1.0 References: <20200508130646.23939-1-kishon@ti.com> <20200508130646.23939-5-kishon@ti.com> In-Reply-To: <20200508130646.23939-5-kishon@ti.com> From: Rob Herring Date: Tue, 19 May 2020 11:11:43 -0600 X-Gmail-Original-Message-ID: Message-ID: Subject: Re: [PATCH v3 4/4] PCI: cadence: Use "dma-ranges" instead of "cdns,no-bar-match-nbits" property To: Kishon Vijay Abraham I Cc: Bjorn Helgaas , Lorenzo Pieralisi , Robin Murphy , Tom Joseph , PCI , devicetree@vger.kernel.org, "linux-kernel@vger.kernel.org" , "moderated list:ARM/FREESCALE IMX / MXC ARM ARCHITECTURE" Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, May 8, 2020 at 7:07 AM Kishon Vijay Abraham I wrote: > > Cadence PCIe core driver (host mode) uses "cdns,no-bar-match-nbits" > property to configure the number of bits passed through from PCIe > address to internal address in Inbound Address Translation register. > This only used the NO MATCH BAR. > > However standard PCI dt-binding already defines "dma-ranges" to > describe the address ranges accessible by PCIe controller. Add support > in Cadence PCIe host driver to parse dma-ranges and configure the > inbound regions for BAR0, BAR1 and NO MATCH BAR. Cadence IP specifies > maximum size for BAR0 as 256GB, maximum size for BAR1 as 2 GB, so if > the dma-ranges specifies a size larger than the maximum allowed, the > driver will split and configure the BARs. Would be useful to know what your dma-ranges contains now. > Legacy device tree binding compatibility is maintained by retaining > support for "cdns,no-bar-match-nbits". > > Signed-off-by: Kishon Vijay Abraham I > --- > .../controller/cadence/pcie-cadence-host.c | 141 ++++++++++++++++-- > drivers/pci/controller/cadence/pcie-cadence.h | 17 ++- > 2 files changed, 141 insertions(+), 17 deletions(-) > > diff --git a/drivers/pci/controller/cadence/pcie-cadence-host.c b/drivers/pci/controller/cadence/pcie-cadence-host.c > index 6ecebb79057a..2485ecd8434d 100644 > --- a/drivers/pci/controller/cadence/pcie-cadence-host.c > +++ b/drivers/pci/controller/cadence/pcie-cadence-host.c > @@ -11,6 +11,12 @@ > > #include "pcie-cadence.h" > > +static u64 cdns_rp_bar_max_size[] = { > + [RP_BAR0] = _ULL(128 * SZ_2G), > + [RP_BAR1] = SZ_2G, > + [RP_NO_BAR] = SZ_64T, > +}; > + > void __iomem *cdns_pci_map_bus(struct pci_bus *bus, unsigned int devfn, > int where) > { > @@ -106,6 +112,117 @@ static int cdns_pcie_host_init_root_port(struct cdns_pcie_rc *rc) > return 0; > } > > +static void cdns_pcie_host_bar_ib_config(struct cdns_pcie_rc *rc, > + enum cdns_pcie_rp_bar bar, > + u64 cpu_addr, u32 aperture) > +{ > + struct cdns_pcie *pcie = &rc->pcie; > + u32 addr0, addr1; > + > + addr0 = CDNS_PCIE_AT_IB_RP_BAR_ADDR0_NBITS(aperture) | > + (lower_32_bits(cpu_addr) & GENMASK(31, 8)); > + addr1 = upper_32_bits(cpu_addr); > + cdns_pcie_writel(pcie, CDNS_PCIE_AT_IB_RP_BAR_ADDR0(bar), addr0); > + cdns_pcie_writel(pcie, CDNS_PCIE_AT_IB_RP_BAR_ADDR1(bar), addr1); > +} > + > +static int cdns_pcie_host_bar_config(struct cdns_pcie_rc *rc, > + struct resource_entry *entry, > + enum cdns_pcie_rp_bar *index) > +{ > + u64 cpu_addr, pci_addr, size, winsize; > + struct cdns_pcie *pcie = &rc->pcie; > + struct device *dev = pcie->dev; > + enum cdns_pcie_rp_bar bar; > + unsigned long flags; > + u32 aperture; > + u32 value; > + > + cpu_addr = entry->res->start; > + flags = entry->res->flags; > + pci_addr = entry->res->start - entry->offset; > + size = resource_size(entry->res); > + bar = *index; > + > + if (entry->offset) { > + dev_err(dev, "Cannot map PCI addr: %llx to CPU addr: %llx\n", > + pci_addr, cpu_addr); Would be a bit more clear to say PCI addr must equal CPU addr. > + return -EINVAL; > + } > + > + value = cdns_pcie_readl(pcie, CDNS_PCIE_LM_RC_BAR_CFG); > + while (size > 0) { > + if (bar > RP_NO_BAR) { > + dev_err(dev, "Failed to map inbound regions!\n"); > + return -EINVAL; > + } > + > + winsize = size; > + if (size > cdns_rp_bar_max_size[bar]) > + winsize = cdns_rp_bar_max_size[bar]; > + > + aperture = ilog2(winsize); > + > + cdns_pcie_host_bar_ib_config(rc, bar, cpu_addr, aperture); > + > + if (bar == RP_NO_BAR) > + break; > + > + if (winsize + cpu_addr >= SZ_4G) { > + if (!(flags & IORESOURCE_PREFETCH)) > + value |= LM_RC_BAR_CFG_CTRL_MEM_64BITS(bar); > + value |= LM_RC_BAR_CFG_CTRL_PREF_MEM_64BITS(bar); > + } else { > + if (!(flags & IORESOURCE_PREFETCH)) > + value |= LM_RC_BAR_CFG_CTRL_MEM_32BITS(bar); > + value |= LM_RC_BAR_CFG_CTRL_PREF_MEM_32BITS(bar); > + } > + > + value |= LM_RC_BAR_CFG_APERTURE(bar, aperture); > + > + size -= winsize; > + cpu_addr += winsize; > + bar++; > + } > + cdns_pcie_writel(pcie, CDNS_PCIE_LM_RC_BAR_CFG, value); > + *index = bar; > + > + return 0; > +} > + > +static int cdns_pcie_host_map_dma_ranges(struct cdns_pcie_rc *rc) > +{ > + enum cdns_pcie_rp_bar bar = RP_BAR0; > + struct cdns_pcie *pcie = &rc->pcie; > + struct device *dev = pcie->dev; > + struct device_node *np = dev->of_node; > + struct pci_host_bridge *bridge; > + struct resource_entry *entry; > + u32 no_bar_nbits = 32; > + int err; > + > + bridge = pci_host_bridge_from_priv(rc); > + if (!bridge) > + return -ENOMEM; > + > + if (list_empty(&bridge->dma_ranges)) { > + of_property_read_u32(np, "cdns,no-bar-match-nbits", > + &no_bar_nbits); > + cdns_pcie_host_bar_ib_config(rc, RP_NO_BAR, 0x0, no_bar_nbits); > + return 0; > + } > + > + resource_list_for_each_entry(entry, &bridge->dma_ranges) { > + err = cdns_pcie_host_bar_config(rc, entry, &bar); Seems like this should have some better logic to pick which BAR to use. Something like find the biggest region and then find the smallest BAR that it fits in. Then get the next biggest... > + if (err) { > + dev_err(dev, "Fail to configure IB using dma-ranges\n"); > + return err; > + } > + } > + > + return 0; > +} > +