Received: by 2002:a05:6602:18e:0:0:0:0 with SMTP id m14csp5815928ioo; Wed, 1 Jun 2022 13:11:11 -0700 (PDT) X-Google-Smtp-Source: ABdhPJyJ+VTwiNo+WD6cbbwhPXS+z4mCTF3CcVotCh5wMyrToEM4oJ6BtcNBD5WG8mhvax0wi/nt X-Received: by 2002:a17:90b:1e06:b0:1e3:7d5:b542 with SMTP id pg6-20020a17090b1e0600b001e307d5b542mr17000481pjb.230.1654114271062; Wed, 01 Jun 2022 13:11:11 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1654114271; cv=none; d=google.com; s=arc-20160816; b=RFfT5ldO3QREN22Si74fYhw6mYm45LvJwDykLfQ/HVV29sPoekSX4KbGPx2u82V7X/ DDT5fzaOnI9YNHnn+cnzW2vhKFIu2+Uain9e1qv7MAPZp3s6NSttCQnCQiXkOS2MOPaR uqdmdZr/lYuCgY1bMEp4tLz4iu1MU2KzACDW04abkGZVvmAZeB5FeIFMsXpw5s/dTewl Q6zU8U0JD40I/EHDAhZLp6RubARvm9eNlebmzKHfbjmVC5FBfWWK4WtScBlLbvUZ6v2O YAy7DIS0TLAymAEZMVFB0TjCBvbu3v0uLi4r08O1CxX+eno+RbUFXhqQxq3n0DrJJOjf pvjg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date; bh=vjc960V+Fsr5l6vnOihKqexHn0oeA8wZ4TyG7Y6hAq0=; b=rLzMEWLG7taGl+qWdwr/09XaJRlgFtzvma60rWnZh2BrcfXh2S5BJtC/cMUlWrmjT0 4kqm/HwE3Z9ubZppmq7oIaWrJRn4dzHkaeTO0kuqHJNsEmTU0bqOVzcmqvIjFuDqZPvt g1ERlcIdyVCaXEMQeHuJqcCZSgRHyfVNwkSdRkaZc+aSbYErzK4oOJZR7uWMayhJ7GdY rK8n01LV1wm4irwz+JBnTvDYD0VswHF5F5nE2tD4zOivjiES5HL7z6/imp9FiCmuG8A/ hYdVZaKWa7zcFpl0eMaqF+erUPcGGPfoWvK51UU2vlAwQsbwCIG/lzfl2OjOIUhBSwJy BOoQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=bwidawsk.net Return-Path: Received: from lindbergh.monkeyblade.net (lindbergh.monkeyblade.net. [2620:137:e000::1:18]) by mx.google.com with ESMTPS id lp4-20020a17090b4a8400b001cb8bfcc721si3919946pjb.7.2022.06.01.13.11.10 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 01 Jun 2022 13:11:11 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:18 as permitted sender) client-ip=2620:137:e000::1:18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=bwidawsk.net Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by lindbergh.monkeyblade.net (Postfix) with ESMTP id 7CBEF163F51; Wed, 1 Jun 2022 12:24:43 -0700 (PDT) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1346743AbiEaR6h (ORCPT + 99 others); Tue, 31 May 2022 13:58:37 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:44262 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S239219AbiEaR6f (ORCPT ); Tue, 31 May 2022 13:58:35 -0400 Received: from mail.bwidawsk.net (mail.bwidawsk.net [107.170.211.233]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 47E8B62CF2; Tue, 31 May 2022 10:58:33 -0700 (PDT) Received: by mail.bwidawsk.net (Postfix, from userid 5001) id A159112328F; Tue, 31 May 2022 10:50:27 -0700 (PDT) X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net X-Spam-Level: X-Spam-ASN: X-Spam-Status: No, score=-1.9 required=5.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,RDNS_NONE, SPF_HELO_NONE,T_SCC_BODY_TEXT_LINE autolearn=no autolearn_force=no version=3.4.6 Received: from mail.bwidawsk.net (c-24-20-163-23.hsd1.or.comcast.net [24.20.163.23]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange ECDHE (P-256) server-signature RSA-PSS (2048 bits) server-digest SHA256) (Client did not present a certificate) by mail.bwidawsk.net (Postfix) with ESMTPSA id 4D309120013; Tue, 31 May 2022 10:50:22 -0700 (PDT) Date: Tue, 31 May 2022 10:50:20 -0700 From: Ben Widawsky To: ira.weiny@intel.com Cc: Dan Williams , Bjorn Helgaas , Jonathan Cameron , Alison Schofield , Vishal Verma , Dave Jiang , linux-kernel@vger.kernel.org, linux-cxl@vger.kernel.org, linux-pci@vger.kernel.org Subject: Re: [PATCH V9 4/9] cxl/pci: Create PCI DOE mailbox's for memory devices Message-ID: <20220531175020.efqfth7ubbyhoubp@mail.bwidawsk.net> References: <20220531152632.1397976-1-ira.weiny@intel.com> <20220531152632.1397976-5-ira.weiny@intel.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20220531152632.1397976-5-ira.weiny@intel.com> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 22-05-31 08:26:27, ira.weiny@intel.com wrote: > From: Ira Weiny > > DOE mailbox objects will be needed for various mailbox communications > with each memory device. > > Iterate each DOE mailbox capability and create PCI DOE mailbox objects > as found. > > It is not anticipated that this is the final resting place for the > iteration of the DOE devices. The support of ports may drive this code > into the pcie side. In this imagined architecture the CXL port driver > would then query into the PCI device for the DOE mailbox array. Not sure if direction has changed, but initially it would have been the cxl_pci driver who would query this and pass it along when the port driver probes. Personally, I've never had an issue with non cxl_pci drivers using PCI interfaces and semantics, but it is something we've taken specific care to avoid. > > For now this is good enough for the endpoints and the split is similar > to the envisioned architecture where getting the mailbox array is > separated from the various protocol needs. For example, it is not > anticipated that the CDAT code will need to move because it is only > needed by the cxl_ports. > > Likewise irq's are separated out in a similar design pattern to the > PCIe port driver. But a much simpler irq enabling flag is used and only > DOE interrupts are supported. > > Signed-off-by: Ira Weiny > > --- > Changes from V8: > Move PCI_DOE selection to CXL_BUS to support future patches > which move queries into the port code. > Remove Auxiliary device arch > Squash the functionality of the auxiliary driver into this > patch. > Split out the irq handling a bit. > > Changes from V7: > Minor code clean ups > Rebased on cxl-pending > > Changes from V6: > Move all the auxiliary device stuff to the CXL layer > > Changes from V5: > Split the CXL specific stuff off from the PCI DOE create > auxiliary device code. > --- > drivers/cxl/Kconfig | 1 + > drivers/cxl/cxlmem.h | 6 +++ > drivers/cxl/pci.c | 111 +++++++++++++++++++++++++++++++++++++++++++ > 3 files changed, 118 insertions(+) > > diff --git a/drivers/cxl/Kconfig b/drivers/cxl/Kconfig > index f64e3984689f..7adaaf80b302 100644 > --- a/drivers/cxl/Kconfig > +++ b/drivers/cxl/Kconfig > @@ -2,6 +2,7 @@ > menuconfig CXL_BUS > tristate "CXL (Compute Express Link) Devices Support" > depends on PCI > + select PCI_DOE > help > CXL is a bus that is electrically compatible with PCI Express, but > layers three protocols on that signalling (CXL.io, CXL.cache, and > diff --git a/drivers/cxl/cxlmem.h b/drivers/cxl/cxlmem.h > index 60d10ee1e7fc..4d2764b865ab 100644 > --- a/drivers/cxl/cxlmem.h > +++ b/drivers/cxl/cxlmem.h > @@ -191,6 +191,8 @@ struct cxl_endpoint_dvsec_info { > * @component_reg_phys: register base of component registers > * @info: Cached DVSEC information about the device. > * @serial: PCIe Device Serial Number > + * @doe_mbs: PCI DOE mailbox array > + * @num_mbs: Number of DOE mailboxes > * @mbox_send: @dev specific transport for transmitting mailbox commands > * > * See section 8.2.9.5.2 Capacity Configuration and Label Storage for > @@ -224,6 +226,10 @@ struct cxl_dev_state { > resource_size_t component_reg_phys; > u64 serial; > > + bool doe_use_irq; > + struct pci_doe_mb **doe_mbs; > + int num_mbs; > + > int (*mbox_send)(struct cxl_dev_state *cxlds, struct cxl_mbox_cmd *cmd); > }; > > diff --git a/drivers/cxl/pci.c b/drivers/cxl/pci.c > index 5a0ae46d4989..131f89dec8e7 100644 > --- a/drivers/cxl/pci.c > +++ b/drivers/cxl/pci.c > @@ -8,6 +8,7 @@ > #include > #include > #include > +#include > #include > #include "cxlmem.h" > #include "cxlpci.h" > @@ -386,6 +387,113 @@ static int cxl_setup_regs(struct pci_dev *pdev, enum cxl_regloc_type type, > return rc; > } > > +static void cxl_pci_free_irq_vectors(void *data) > +{ > + pci_free_irq_vectors(data); > +} > + > +static void cxl_doe_destroy_mb(void *ds) > +{ > + struct cxl_dev_state *cxlds = ds; > + int i; > + > + for (i = 0; i < cxlds->num_mbs; i++) { > + if (cxlds->doe_mbs[i]) > + pci_doe_destroy_mb(cxlds->doe_mbs[i]); > + } > +} > + > +static void cxl_alloc_irq_vectors(struct cxl_dev_state *cxlds) > +{ > + struct device *dev = cxlds->dev; > + struct pci_dev *pdev = to_pci_dev(dev); > + int num_irqs = 0; > + int off = 0; > + int rc; > + > + /* Account for all the DOE vectors needed */ > + pci_doe_for_each_off(pdev, off) { > + int irq = pci_doe_get_irq_num(pdev, off); > + > + if (irq < 0) > + continue; > + num_irqs = max(num_irqs, irq + 1); This seems overly complicated. Isn't it just num_irqs++? > + } > + > + /* > + * Allocate enough vectors for the DOE's > + */ > + rc = pci_alloc_irq_vectors(pdev, num_irqs, num_irqs, PCI_IRQ_MSI | > + PCI_IRQ_MSIX); > + if (rc != num_irqs) { > + pci_err(pdev, "Not enough interrupts; use polling\n"); > + /* Some got allocated; clean them up */ > + if (rc > 0) > + cxl_pci_free_irq_vectors(pdev); > + cxlds->doe_use_irq = false; > + return; > + } > + > + rc = devm_add_action_or_reset(dev, cxl_pci_free_irq_vectors, pdev); > + if (rc) { > + cxlds->doe_use_irq = false; > + return; > + } > + > + cxlds->doe_use_irq = true; If you named it doe_poll, you could avoid having to do anything at the end of the function... If you felt like it. if (failure) return; if (other_failure) return; cxld->do_use_poll = false; > +} > + > +/** > + * devm_cxl_pci_create_doe - Scan and set up DOE mailboxes > + * > + * @cxlds: The CXL device state > + * > + * RETURNS: 0 on success -ERRNO on failure. > + */ > +static int devm_cxl_pci_create_doe(struct cxl_dev_state *cxlds) > +{ > + struct device *dev = cxlds->dev; > + struct pci_dev *pdev = to_pci_dev(dev); > + u16 off = 0; > + int num_mbs = 0; > + int rc; > + > + pci_doe_for_each_off(pdev, off) > + num_mbs++; > + Do you want to bail here if num_mbs == 0? > + cxlds->doe_mbs = devm_kcalloc(dev, num_mbs, sizeof(*cxlds->doe_mbs), > + GFP_KERNEL); > + if (!cxlds->doe_mbs) > + return -ENOMEM; > + > + pci_doe_for_each_off(pdev, off) { > + struct pci_doe_mb *doe_mb; > + int irq = -1; > + > + if (cxlds->doe_use_irq) > + irq = pci_doe_get_irq_num(pdev, off); > + > + doe_mb = pci_doe_create_mb(pdev, off, irq); > + if (IS_ERR(doe_mb)) { > + pci_err(pdev, > + "Failed to create MB object for MB @ %x\n", > + off); > + doe_mb = NULL; > + } > + > + cxlds->doe_mbs[cxlds->num_mbs] = doe_mb; > + cxlds->num_mbs++; > + } > + > + rc = devm_add_action_or_reset(dev, cxl_doe_destroy_mb, cxlds); > + if (rc) > + return rc; > + > + pci_info(pdev, "Configured %d DOE mailbox's\n", cxlds->num_mbs); > + > + return 0; > +} > + > static int cxl_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id) > { > struct cxl_register_map map; > @@ -454,6 +562,9 @@ static int cxl_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id) > if (IS_ERR(cxlmd)) > return PTR_ERR(cxlmd); > > + cxl_alloc_irq_vectors(cxlds); > + devm_cxl_pci_create_doe(cxlds); If you're not going to check the return value, just make the functions void. > + > if (range_len(&cxlds->pmem_range) && IS_ENABLED(CONFIG_CXL_PMEM)) > rc = devm_cxl_add_nvdimm(&pdev->dev, cxlmd); > > -- > 2.35.1 >