Received: by 2002:a05:6358:d09b:b0:dc:cd0c:909e with SMTP id jc27csp883817rwb; Wed, 16 Nov 2022 08:55:37 -0800 (PST) X-Google-Smtp-Source: AA0mqf5dLuWhghd3UE5BVIU5WOmbLmNt32wa52lFdMTbk+qqv0GPp5Z+/iRAL1Urm26yVHKR/u2P X-Received: by 2002:a17:90a:62c2:b0:20d:bc7f:4fee with SMTP id k2-20020a17090a62c200b0020dbc7f4feemr4648434pjs.168.1668617737047; Wed, 16 Nov 2022 08:55:37 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1668617737; cv=none; d=google.com; s=arc-20160816; b=wc2un+uNg62zspcdTxaksOsTC8AyVd+KbXtn+59Z0hf4cyTrWkKwA5ZfghSn5pDgfP Nwbo6ZzZhB5vacpOC78mEtlQnlTS24iIIYa/gKitNGxBfDiGLGjKweu78a6b8Ct+NeRi ypdO7DB+9lWpGYj3C1hbd+Tj4Q7R6AEtPGQGfrGO3NwO6CVjkXENsfXt0AjNZuu02rwa 8AyTWJwBXd0h/ghE+MvKHY4LlF+zoz+WFETVHNdeR9Z1ymvIku4zzlSK4G0SNtXr8c4w o9dpNa7rGBbInuMD2uNOB/FyUKiQlT8kFW87Db+8XWxsKydtS/97j1ARy3e3zpmx1Xh7 voug== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :message-id:subject:cc:to:from:date:dkim-signature; bh=17KbU+atjdZ/UYEewLP1ijyQj2ZSJ/lzAAOfTEViOpg=; b=tmd2n+5EVJ83K0Ebh6KEMUfJ6fWCa+VyZg4qVcdWjCFeu10wMPQzh+SXPw99kRq28T bPqvV+ijpIV16d2+N7hQN6ZudkRNmKjiEb7/PXy5b8rvXXPXAsvPxoW2CRGCPbWV0/ot SAE97UIP9Dto6/CyBqyalKY1hwhr3G80npWLmRXY9wTXehdkeKkB47J6zLIwIHCAhrFv WyIiAQPG+sMJ5CoDAMdxPi7w9XajOZCxYpN89yedORkntAm8tWehG/5NMt4rnlBrAbFq xxcjPqao+xlYzwuVc1iLed1lsbSlbQs5EkkFtZISYNu68Z5dmsYjblS9qYg0UG86pMpW +Phg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=pzSMBZDV; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id t33-20020a632261000000b0046f55fec417si15511556pgm.650.2022.11.16.08.55.25; Wed, 16 Nov 2022 08:55:37 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=pzSMBZDV; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S233950AbiKPQg2 (ORCPT + 91 others); Wed, 16 Nov 2022 11:36:28 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:47920 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S238805AbiKPQgJ (ORCPT ); Wed, 16 Nov 2022 11:36:09 -0500 Received: from ams.source.kernel.org (ams.source.kernel.org [145.40.68.75]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 993FD4D5F3; Wed, 16 Nov 2022 08:28:47 -0800 (PST) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ams.source.kernel.org (Postfix) with ESMTPS id 14D08B81DDB; Wed, 16 Nov 2022 16:28:46 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id 62109C433C1; Wed, 16 Nov 2022 16:28:44 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1668616124; bh=za45kr7NXviT1L5zMoLfqwPhtRcnhnUs9DCgfu4JINo=; h=Date:From:To:Cc:Subject:In-Reply-To:From; b=pzSMBZDVV7SU+2TvgvbA3WDHR6gDaj0p23v+E9GqDrn84u7cKpKRNn8iVVASn613o nc29Fp/X04wT5d0KClsXJNtt6BKMVAQv/C56RWZxMRr7dK0rzSdAyRpdTsPQoMgiRC SbObANlTDSbAnFfKT4lU0o8ItUBfXtIzhM94qjURnCPnddMDrp0VPETEejEhbLl9/m xsfQzl1d4Yya4bT9y2kv7KPzGw49ok5WYsGcIp3KwYLVTBdVUYCXDvhGFth8Uk97OR xrEKTZc1ZuoQsD3w00s7U6o2xxfEdi1j6Ipyi3apQjobSPn7lnK1uHmUL6PSXBAyYW RxUeegbK+b2kA== Date: Wed, 16 Nov 2022 10:28:43 -0600 From: Bjorn Helgaas To: Thomas Gleixner Cc: LKML , x86@kernel.org, Joerg Roedel , Will Deacon , linux-pci@vger.kernel.org, Bjorn Helgaas , Lorenzo Pieralisi , Marc Zyngier , Greg Kroah-Hartman , Jason Gunthorpe , Dave Jiang , Alex Williamson , Kevin Tian , Dan Williams , Logan Gunthorpe , Ashok Raj , Jon Mason , Allen Hubbe , "Ahmed S. Darwish" , Reinette Chatre , Michael Ellerman , Christophe Leroy , linuxppc-dev@lists.ozlabs.org Subject: Re: [patch 32/39] PCI/MSI: Reorder functions in msi.c Message-ID: <20221116162843.GA1116077@bhelgaas> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20221111122015.459089736@linutronix.de> X-Spam-Status: No, score=-7.1 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_HI, SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Nov 11, 2022 at 02:55:06PM +0100, Thomas Gleixner wrote: > From: Ahmed S. Darwish > > There is no way to navigate msi.c without banging the head against the wall > every now and then because MSI and MSI-X specific functions are > intermingled and the code flow is completely non-obvious. > > Reorder everthing so common helpers, MSI and MSI-X specific functions are > grouped together. s/everthing/everything/ > Suggested-by: Thomas Gleixner > Signed-off-by: Ahmed S. Darwish > Signed-off-by: Thomas Gleixner Acked-by: Bjorn Helgaas I assume this is pure code movement, so I didn't even look at the text below. > --- > drivers/pci/msi/msi.c | 577 +++++++++++++++++++++++++------------------------- > 1 file changed, 295 insertions(+), 282 deletions(-) > > --- a/drivers/pci/msi/msi.c > +++ b/drivers/pci/msi/msi.c > @@ -16,6 +16,97 @@ > int pci_msi_enable = 1; > int pci_msi_ignore_mask; > > +/** > + * pci_msi_supported - check whether MSI may be enabled on a device > + * @dev: pointer to the pci_dev data structure of MSI device function > + * @nvec: how many MSIs have been requested? > + * > + * Look at global flags, the device itself, and its parent buses > + * to determine if MSI/-X are supported for the device. If MSI/-X is > + * supported return 1, else return 0. > + **/ > +static int pci_msi_supported(struct pci_dev *dev, int nvec) > +{ > + struct pci_bus *bus; > + > + /* MSI must be globally enabled and supported by the device */ > + if (!pci_msi_enable) > + return 0; > + > + if (!dev || dev->no_msi) > + return 0; > + > + /* > + * You can't ask to have 0 or less MSIs configured. > + * a) it's stupid .. > + * b) the list manipulation code assumes nvec >= 1. > + */ > + if (nvec < 1) > + return 0; > + > + /* > + * Any bridge which does NOT route MSI transactions from its > + * secondary bus to its primary bus must set NO_MSI flag on > + * the secondary pci_bus. > + * > + * The NO_MSI flag can either be set directly by: > + * - arch-specific PCI host bus controller drivers (deprecated) > + * - quirks for specific PCI bridges > + * > + * or indirectly by platform-specific PCI host bridge drivers by > + * advertising the 'msi_domain' property, which results in > + * the NO_MSI flag when no MSI domain is found for this bridge > + * at probe time. > + */ > + for (bus = dev->bus; bus; bus = bus->parent) > + if (bus->bus_flags & PCI_BUS_FLAGS_NO_MSI) > + return 0; > + > + return 1; > +} > + > +static void pcim_msi_release(void *pcidev) > +{ > + struct pci_dev *dev = pcidev; > + > + dev->is_msi_managed = false; > + pci_free_irq_vectors(dev); > +} > + > +/* > + * Needs to be separate from pcim_release to prevent an ordering problem > + * vs. msi_device_data_release() in the MSI core code. > + */ > +static int pcim_setup_msi_release(struct pci_dev *dev) > +{ > + int ret; > + > + if (!pci_is_managed(dev) || dev->is_msi_managed) > + return 0; > + > + ret = devm_add_action(&dev->dev, pcim_msi_release, dev); > + if (!ret) > + dev->is_msi_managed = true; > + return ret; > +} > + > +/* > + * Ordering vs. devres: msi device data has to be installed first so that > + * pcim_msi_release() is invoked before it on device release. > + */ > +static int pci_setup_msi_context(struct pci_dev *dev) > +{ > + int ret = msi_setup_device_data(&dev->dev); > + > + if (!ret) > + ret = pcim_setup_msi_release(dev); > + return ret; > +} > + > +/* > + * Helper functions for mask/unmask and MSI message handling > + */ > + > void pci_msi_update_mask(struct msi_desc *desc, u32 clear, u32 set) > { > raw_spinlock_t *lock = &to_pci_dev(desc->dev)->msi_lock; > @@ -163,15 +254,8 @@ void pci_write_msi_msg(unsigned int irq, > } > EXPORT_SYMBOL_GPL(pci_write_msi_msg); > > -void pci_free_msi_irqs(struct pci_dev *dev) > -{ > - pci_msi_teardown_msi_irqs(dev); > > - if (dev->msix_base) { > - iounmap(dev->msix_base); > - dev->msix_base = NULL; > - } > -} > +/* PCI/MSI specific functionality */ > > static void pci_intx_for_msi(struct pci_dev *dev, int enable) > { > @@ -190,111 +274,6 @@ static void pci_msi_set_enable(struct pc > pci_write_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, control); > } > > -/* > - * Architecture override returns true when the PCI MSI message should be > - * written by the generic restore function. > - */ > -bool __weak arch_restore_msi_irqs(struct pci_dev *dev) > -{ > - return true; > -} > - > -void __pci_restore_msi_state(struct pci_dev *dev) > -{ > - struct msi_desc *entry; > - u16 control; > - > - if (!dev->msi_enabled) > - return; > - > - entry = irq_get_msi_desc(dev->irq); > - > - pci_intx_for_msi(dev, 0); > - pci_msi_set_enable(dev, 0); > - if (arch_restore_msi_irqs(dev)) > - __pci_write_msi_msg(entry, &entry->msg); > - > - pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &control); > - pci_msi_update_mask(entry, 0, 0); > - control &= ~PCI_MSI_FLAGS_QSIZE; > - control |= (entry->pci.msi_attrib.multiple << 4) | PCI_MSI_FLAGS_ENABLE; > - pci_write_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, control); > -} > - > -static void pci_msix_clear_and_set_ctrl(struct pci_dev *dev, u16 clear, u16 set) > -{ > - u16 ctrl; > - > - pci_read_config_word(dev, dev->msix_cap + PCI_MSIX_FLAGS, &ctrl); > - ctrl &= ~clear; > - ctrl |= set; > - pci_write_config_word(dev, dev->msix_cap + PCI_MSIX_FLAGS, ctrl); > -} > - > -void __pci_restore_msix_state(struct pci_dev *dev) > -{ > - struct msi_desc *entry; > - bool write_msg; > - > - if (!dev->msix_enabled) > - return; > - > - /* route the table */ > - pci_intx_for_msi(dev, 0); > - pci_msix_clear_and_set_ctrl(dev, 0, > - PCI_MSIX_FLAGS_ENABLE | PCI_MSIX_FLAGS_MASKALL); > - > - write_msg = arch_restore_msi_irqs(dev); > - > - msi_lock_descs(&dev->dev); > - msi_for_each_desc(entry, &dev->dev, MSI_DESC_ALL) { > - if (write_msg) > - __pci_write_msi_msg(entry, &entry->msg); > - pci_msix_write_vector_ctrl(entry, entry->pci.msix_ctrl); > - } > - msi_unlock_descs(&dev->dev); > - > - pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_MASKALL, 0); > -} > - > -static void pcim_msi_release(void *pcidev) > -{ > - struct pci_dev *dev = pcidev; > - > - dev->is_msi_managed = false; > - pci_free_irq_vectors(dev); > -} > - > -/* > - * Needs to be separate from pcim_release to prevent an ordering problem > - * vs. msi_device_data_release() in the MSI core code. > - */ > -static int pcim_setup_msi_release(struct pci_dev *dev) > -{ > - int ret; > - > - if (!pci_is_managed(dev) || dev->is_msi_managed) > - return 0; > - > - ret = devm_add_action(&dev->dev, pcim_msi_release, dev); > - if (!ret) > - dev->is_msi_managed = true; > - return ret; > -} > - > -/* > - * Ordering vs. devres: msi device data has to be installed first so that > - * pcim_msi_release() is invoked before it on device release. > - */ > -static int pci_setup_msi_context(struct pci_dev *dev) > -{ > - int ret = msi_setup_device_data(&dev->dev); > - > - if (!ret) > - ret = pcim_setup_msi_release(dev); > - return ret; > -} > - > static int msi_setup_msi_desc(struct pci_dev *dev, int nvec, > struct irq_affinity_desc *masks) > { > @@ -415,6 +394,149 @@ static int msi_capability_init(struct pc > return ret; > } > > +int __pci_enable_msi_range(struct pci_dev *dev, int minvec, int maxvec, > + struct irq_affinity *affd) > +{ > + int nvec; > + int rc; > + > + if (!pci_msi_supported(dev, minvec) || dev->current_state != PCI_D0) > + return -EINVAL; > + > + /* Check whether driver already requested MSI-X IRQs */ > + if (dev->msix_enabled) { > + pci_info(dev, "can't enable MSI (MSI-X already enabled)\n"); > + return -EINVAL; > + } > + > + if (maxvec < minvec) > + return -ERANGE; > + > + if (WARN_ON_ONCE(dev->msi_enabled)) > + return -EINVAL; > + > + nvec = pci_msi_vec_count(dev); > + if (nvec < 0) > + return nvec; > + if (nvec < minvec) > + return -ENOSPC; > + > + if (nvec > maxvec) > + nvec = maxvec; > + > + rc = pci_setup_msi_context(dev); > + if (rc) > + return rc; > + > + for (;;) { > + if (affd) { > + nvec = irq_calc_affinity_vectors(minvec, nvec, affd); > + if (nvec < minvec) > + return -ENOSPC; > + } > + > + rc = msi_capability_init(dev, nvec, affd); > + if (rc == 0) > + return nvec; > + > + if (rc < 0) > + return rc; > + if (rc < minvec) > + return -ENOSPC; > + > + nvec = rc; > + } > +} > + > +/** > + * pci_msi_vec_count - Return the number of MSI vectors a device can send > + * @dev: device to report about > + * > + * This function returns the number of MSI vectors a device requested via > + * Multiple Message Capable register. It returns a negative errno if the > + * device is not capable sending MSI interrupts. Otherwise, the call succeeds > + * and returns a power of two, up to a maximum of 2^5 (32), according to the > + * MSI specification. > + **/ > +int pci_msi_vec_count(struct pci_dev *dev) > +{ > + int ret; > + u16 msgctl; > + > + if (!dev->msi_cap) > + return -EINVAL; > + > + pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &msgctl); > + ret = 1 << ((msgctl & PCI_MSI_FLAGS_QMASK) >> 1); > + > + return ret; > +} > +EXPORT_SYMBOL(pci_msi_vec_count); > + > +/* > + * Architecture override returns true when the PCI MSI message should be > + * written by the generic restore function. > + */ > +bool __weak arch_restore_msi_irqs(struct pci_dev *dev) > +{ > + return true; > +} > + > +void __pci_restore_msi_state(struct pci_dev *dev) > +{ > + struct msi_desc *entry; > + u16 control; > + > + if (!dev->msi_enabled) > + return; > + > + entry = irq_get_msi_desc(dev->irq); > + > + pci_intx_for_msi(dev, 0); > + pci_msi_set_enable(dev, 0); > + if (arch_restore_msi_irqs(dev)) > + __pci_write_msi_msg(entry, &entry->msg); > + > + pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &control); > + pci_msi_update_mask(entry, 0, 0); > + control &= ~PCI_MSI_FLAGS_QSIZE; > + control |= (entry->pci.msi_attrib.multiple << 4) | PCI_MSI_FLAGS_ENABLE; > + pci_write_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, control); > +} > + > +void pci_msi_shutdown(struct pci_dev *dev) > +{ > + struct msi_desc *desc; > + > + if (!pci_msi_enable || !dev || !dev->msi_enabled) > + return; > + > + pci_msi_set_enable(dev, 0); > + pci_intx_for_msi(dev, 1); > + dev->msi_enabled = 0; > + > + /* Return the device with MSI unmasked as initial states */ > + desc = msi_first_desc(&dev->dev, MSI_DESC_ALL); > + if (!WARN_ON_ONCE(!desc)) > + pci_msi_unmask(desc, msi_multi_mask(desc)); > + > + /* Restore dev->irq to its default pin-assertion IRQ */ > + dev->irq = desc->pci.msi_attrib.default_irq; > + pcibios_alloc_irq(dev); > +} > + > +/* PCI/MSI-X specific functionality */ > + > +static void pci_msix_clear_and_set_ctrl(struct pci_dev *dev, u16 clear, u16 set) > +{ > + u16 ctrl; > + > + pci_read_config_word(dev, dev->msix_cap + PCI_MSIX_FLAGS, &ctrl); > + ctrl &= ~clear; > + ctrl |= set; > + pci_write_config_word(dev, dev->msix_cap + PCI_MSIX_FLAGS, ctrl); > +} > + > static void __iomem *msix_map_region(struct pci_dev *dev, > unsigned int nr_entries) > { > @@ -599,101 +721,6 @@ static int msix_capability_init(struct p > return ret; > } > > -/** > - * pci_msi_supported - check whether MSI may be enabled on a device > - * @dev: pointer to the pci_dev data structure of MSI device function > - * @nvec: how many MSIs have been requested? > - * > - * Look at global flags, the device itself, and its parent buses > - * to determine if MSI/-X are supported for the device. If MSI/-X is > - * supported return 1, else return 0. > - **/ > -static int pci_msi_supported(struct pci_dev *dev, int nvec) > -{ > - struct pci_bus *bus; > - > - /* MSI must be globally enabled and supported by the device */ > - if (!pci_msi_enable) > - return 0; > - > - if (!dev || dev->no_msi) > - return 0; > - > - /* > - * You can't ask to have 0 or less MSIs configured. > - * a) it's stupid .. > - * b) the list manipulation code assumes nvec >= 1. > - */ > - if (nvec < 1) > - return 0; > - > - /* > - * Any bridge which does NOT route MSI transactions from its > - * secondary bus to its primary bus must set NO_MSI flag on > - * the secondary pci_bus. > - * > - * The NO_MSI flag can either be set directly by: > - * - arch-specific PCI host bus controller drivers (deprecated) > - * - quirks for specific PCI bridges > - * > - * or indirectly by platform-specific PCI host bridge drivers by > - * advertising the 'msi_domain' property, which results in > - * the NO_MSI flag when no MSI domain is found for this bridge > - * at probe time. > - */ > - for (bus = dev->bus; bus; bus = bus->parent) > - if (bus->bus_flags & PCI_BUS_FLAGS_NO_MSI) > - return 0; > - > - return 1; > -} > - > -/** > - * pci_msi_vec_count - Return the number of MSI vectors a device can send > - * @dev: device to report about > - * > - * This function returns the number of MSI vectors a device requested via > - * Multiple Message Capable register. It returns a negative errno if the > - * device is not capable sending MSI interrupts. Otherwise, the call succeeds > - * and returns a power of two, up to a maximum of 2^5 (32), according to the > - * MSI specification. > - **/ > -int pci_msi_vec_count(struct pci_dev *dev) > -{ > - int ret; > - u16 msgctl; > - > - if (!dev->msi_cap) > - return -EINVAL; > - > - pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &msgctl); > - ret = 1 << ((msgctl & PCI_MSI_FLAGS_QMASK) >> 1); > - > - return ret; > -} > -EXPORT_SYMBOL(pci_msi_vec_count); > - > -void pci_msi_shutdown(struct pci_dev *dev) > -{ > - struct msi_desc *desc; > - > - if (!pci_msi_enable || !dev || !dev->msi_enabled) > - return; > - > - pci_msi_set_enable(dev, 0); > - pci_intx_for_msi(dev, 1); > - dev->msi_enabled = 0; > - > - /* Return the device with MSI unmasked as initial states */ > - desc = msi_first_desc(&dev->dev, MSI_DESC_ALL); > - if (!WARN_ON_ONCE(!desc)) > - pci_msi_unmask(desc, msi_multi_mask(desc)); > - > - /* Restore dev->irq to its default pin-assertion IRQ */ > - dev->irq = desc->pci.msi_attrib.default_irq; > - pcibios_alloc_irq(dev); > -} > - > static int __pci_enable_msix(struct pci_dev *dev, struct msix_entry *entries, > int nvec, struct irq_affinity *affd, int flags) > { > @@ -729,57 +756,23 @@ static int __pci_enable_msix(struct pci_ > return msix_capability_init(dev, entries, nvec, affd); > } > > -void pci_msix_shutdown(struct pci_dev *dev) > -{ > - struct msi_desc *desc; > - > - if (!pci_msi_enable || !dev || !dev->msix_enabled) > - return; > - > - if (pci_dev_is_disconnected(dev)) { > - dev->msix_enabled = 0; > - return; > - } > - > - /* Return the device with MSI-X masked as initial states */ > - msi_for_each_desc(desc, &dev->dev, MSI_DESC_ALL) > - pci_msix_mask(desc); > - > - pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_ENABLE, 0); > - pci_intx_for_msi(dev, 1); > - dev->msix_enabled = 0; > - pcibios_alloc_irq(dev); > -} > - > -int __pci_enable_msi_range(struct pci_dev *dev, int minvec, int maxvec, > - struct irq_affinity *affd) > +int __pci_enable_msix_range(struct pci_dev *dev, > + struct msix_entry *entries, int minvec, > + int maxvec, struct irq_affinity *affd, > + int flags) > { > - int nvec; > - int rc; > - > - if (!pci_msi_supported(dev, minvec) || dev->current_state != PCI_D0) > - return -EINVAL; > - > - /* Check whether driver already requested MSI-X IRQs */ > - if (dev->msix_enabled) { > - pci_info(dev, "can't enable MSI (MSI-X already enabled)\n"); > - return -EINVAL; > - } > + int rc, nvec = maxvec; > > if (maxvec < minvec) > return -ERANGE; > > - if (WARN_ON_ONCE(dev->msi_enabled)) > + if (dev->msi_enabled) { > + pci_info(dev, "can't enable MSI-X (MSI already enabled)\n"); > return -EINVAL; > + } > > - nvec = pci_msi_vec_count(dev); > - if (nvec < 0) > - return nvec; > - if (nvec < minvec) > - return -ENOSPC; > - > - if (nvec > maxvec) > - nvec = maxvec; > + if (WARN_ON_ONCE(dev->msix_enabled)) > + return -EINVAL; > > rc = pci_setup_msi_context(dev); > if (rc) > @@ -792,7 +785,7 @@ int __pci_enable_msi_range(struct pci_de > return -ENOSPC; > } > > - rc = msi_capability_init(dev, nvec, affd); > + rc = __pci_enable_msix(dev, entries, nvec, affd, flags); > if (rc == 0) > return nvec; > > @@ -805,48 +798,68 @@ int __pci_enable_msi_range(struct pci_de > } > } > > -int __pci_enable_msix_range(struct pci_dev *dev, > - struct msix_entry *entries, int minvec, > - int maxvec, struct irq_affinity *affd, > - int flags) > +void __pci_restore_msix_state(struct pci_dev *dev) > { > - int rc, nvec = maxvec; > + struct msi_desc *entry; > + bool write_msg; > > - if (maxvec < minvec) > - return -ERANGE; > + if (!dev->msix_enabled) > + return; > > - if (dev->msi_enabled) { > - pci_info(dev, "can't enable MSI-X (MSI already enabled)\n"); > - return -EINVAL; > + /* route the table */ > + pci_intx_for_msi(dev, 0); > + pci_msix_clear_and_set_ctrl(dev, 0, > + PCI_MSIX_FLAGS_ENABLE | PCI_MSIX_FLAGS_MASKALL); > + > + write_msg = arch_restore_msi_irqs(dev); > + > + msi_lock_descs(&dev->dev); > + msi_for_each_desc(entry, &dev->dev, MSI_DESC_ALL) { > + if (write_msg) > + __pci_write_msi_msg(entry, &entry->msg); > + pci_msix_write_vector_ctrl(entry, entry->pci.msix_ctrl); > } > + msi_unlock_descs(&dev->dev); > > - if (WARN_ON_ONCE(dev->msix_enabled)) > - return -EINVAL; > + pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_MASKALL, 0); > +} > > - rc = pci_setup_msi_context(dev); > - if (rc) > - return rc; > +void pci_msix_shutdown(struct pci_dev *dev) > +{ > + struct msi_desc *desc; > > - for (;;) { > - if (affd) { > - nvec = irq_calc_affinity_vectors(minvec, nvec, affd); > - if (nvec < minvec) > - return -ENOSPC; > - } > + if (!pci_msi_enable || !dev || !dev->msix_enabled) > + return; > > - rc = __pci_enable_msix(dev, entries, nvec, affd, flags); > - if (rc == 0) > - return nvec; > + if (pci_dev_is_disconnected(dev)) { > + dev->msix_enabled = 0; > + return; > + } > > - if (rc < 0) > - return rc; > - if (rc < minvec) > - return -ENOSPC; > + /* Return the device with MSI-X masked as initial states */ > + msi_for_each_desc(desc, &dev->dev, MSI_DESC_ALL) > + pci_msix_mask(desc); > > - nvec = rc; > + pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_ENABLE, 0); > + pci_intx_for_msi(dev, 1); > + dev->msix_enabled = 0; > + pcibios_alloc_irq(dev); > +} > + > +/* Common interfaces */ > + > +void pci_free_msi_irqs(struct pci_dev *dev) > +{ > + pci_msi_teardown_msi_irqs(dev); > + > + if (dev->msix_base) { > + iounmap(dev->msix_base); > + dev->msix_base = NULL; > } > } > > +/* Misc. infrastructure */ > + > struct pci_dev *msi_desc_to_pci_dev(struct msi_desc *desc) > { > return to_pci_dev(desc->dev); >