Received: by 2002:a05:6358:700f:b0:131:369:b2a3 with SMTP id 15csp2301731rwo; Thu, 3 Aug 2023 07:35:46 -0700 (PDT) X-Google-Smtp-Source: APBJJlGMvVciFIqO65oBtMFXa/VEvPGfgh4y2qB7K+YoMiizmU4MDqKuwIoEVBrVqOisLuxpHvhI X-Received: by 2002:a05:6a21:9998:b0:13f:52f1:fe66 with SMTP id ve24-20020a056a21999800b0013f52f1fe66mr2937515pzb.19.1691073346563; Thu, 03 Aug 2023 07:35:46 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1691073346; cv=none; d=google.com; s=arc-20160816; b=qNxFBfGhGJNSuNinFH5hgF6Rk1I2kcUNyoA/UrQffrrIMWXzl4bUAyvHS9tLSkHjYm YVAaUo6HcdTdVa56kXMXpSOGdtGe+hObJRtUnaKyqtjQ5HXczFWlgSs3qJTHdI2vggiA 1I4nSDM2tLVisMcmI7nPP5DBQb5hwZPwzg+4t+ByidivqvmxpaGVDkZvRzy7DHlXxhYv gi59eBVuXxL3Rab4ZMYqTyijY/Ww5Sosp0c7z4JezcakGYowzYhWfeMPVB23epYvfPcC hdZDlwemv90y//J4MXOoaifcl+HFeCezQxK5SCutgC+uDOYAlf3/12c3eUy4PN/y14TN EYyA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:mime-version:references:message-id:in-reply-to :subject:cc:to:from:date:dkim-signature; bh=wM2R/Nrj1RerExCmvUoaGBkma1+Ni59g2BIAlnxzgV8=; fh=N8SjsVivOHLbAuBPCAmGLNzmMXWpzplG3/3HXmmex50=; b=ZvOxXT2DzzoSlYcelREhO7cJ6MEtxTt940bAfBrYekMrwFqxp5Pyg6k3JMWZCxts7f QsFsIVD9l+AzES+nah+N3bUOw0UddEIVDPc4hgdh8aXy9BgBGGQ4NhVzos5fIpDW+sS8 elWE2PCCQQQ7ATANSmcQx4lS/OwE3cNQVE5hdMd4Rde208mrNxNoz9q3l4+peydhAjka 8Atv0w3AHl4aIi0FuOFTiMPzCeGh/0GrO7ilhVt3MSfDnHYAUb/Egkb8Bs1q+08eIIfN 1zIA9lAohwd+wWsDOvsIrrgM8a/BPNgTNVbFSfsp2Ou5Y7BWFp9wTSHfhlDuXa5LO1RB gPmg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@intel.com header.s=Intel header.b=AngfhjRb; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id g8-20020a636b08000000b005649a48340esi1567389pgc.797.2023.08.03.07.35.34; Thu, 03 Aug 2023 07:35:46 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@intel.com header.s=Intel header.b=AngfhjRb; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230356AbjHCONo (ORCPT + 99 others); Thu, 3 Aug 2023 10:13:44 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:58980 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234772AbjHCOM4 (ORCPT ); Thu, 3 Aug 2023 10:12:56 -0400 Received: from mgamail.intel.com (unknown [192.55.52.151]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id E7B6549C0; Thu, 3 Aug 2023 07:12:10 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1691071930; x=1722607930; h=date:from:to:cc:subject:in-reply-to:message-id: references:mime-version; bh=NsxrA53UgZ7A/dLW9sZ979+1bM+DyTg3FTRgoUO9Eik=; b=AngfhjRbPf0qegcp1dSS1yk7U/MdB1CncMsIG+trNmJAYBoz5K/p8mWk 7yMFFPYBOaz50w7Fkw1KqhbJbH1zv734wKDrNnD16H4t6xUeQmDw80R94 0OVpLi4SoDPny5ySwDdyV6NNaZU9fMbMIQ5N6/Q/tdEkdV6S+e9x2+aJ2 AOlwC86wjtQW4FgJiyrX8xfnG+MTOizEcaimcNSgnb13UIC1wCpeypHgi i4qE9VIG1d2rIDGddAanqFlS2ZczvpL7LVMkI+SDCqDIPDac4wlwyXRHQ z5hLSHEFs4w15gxYgYqlcaGKXlgJNYpVn6z1F+ZUt4quxJ51KbJUbpqwv Q==; X-IronPort-AV: E=McAfee;i="6600,9927,10791"; a="350178089" X-IronPort-AV: E=Sophos;i="6.01,252,1684825200"; d="scan'208";a="350178089" Received: from fmsmga007.fm.intel.com ([10.253.24.52]) by fmsmga107.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 03 Aug 2023 07:12:09 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6600,9927,10791"; a="732784452" X-IronPort-AV: E=Sophos;i="6.01,252,1684825200"; d="scan'208";a="732784452" Received: from eozturk-mobl.ger.corp.intel.com ([10.249.38.219]) by fmsmga007-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 03 Aug 2023 07:12:04 -0700 Date: Thu, 3 Aug 2023 17:12:02 +0300 (EEST) From: =?ISO-8859-15?Q?Ilpo_J=E4rvinen?= To: Alex Deucher cc: Bjorn Helgaas , =?ISO-8859-2?Q?Krzysztof_Wilczy=F1ski?= , Emmanuel Grumbach , Jammy Zhou , linux-pci@vger.kernel.org, dri-devel@lists.freedesktop.org, Lorenzo Pieralisi , =?ISO-8859-15?Q?Jonas_Dre=DFler?= , "Rafael J . Wysocki" , amd-gfx@lists.freedesktop.org, Dean Luick , Bjorn Helgaas , Andy Shevchenko , "Pan, Xinhui" , LKML , stable@vger.kernel.org, Alex Deucher , Ken Wang , =?ISO-8859-15?Q?Christian_K=F6nig?= , Heiner Kallweit Subject: Re: [PATCH v5 05/11] drm/amdgpu: Use RMW accessors for changing LNKCTL In-Reply-To: Message-ID: <5aebcbcf-9e76-e9f5-ae76-74b8d4f8f6e8@linux.intel.com> References: <20230720215550.GA554900@bhelgaas> MIME-Version: 1.0 Content-Type: multipart/mixed; boundary="8323329-1571245239-1691071929=:1954" X-Spam-Status: No, score=-2.0 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_EF,RCVD_IN_DNSWL_BLOCKED, SPF_HELO_NONE,SPF_NONE,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org This message is in MIME format. The first part should be readable text, while the remaining parts are likely unreadable without MIME-aware tools. --8323329-1571245239-1691071929=:1954 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8BIT On Fri, 21 Jul 2023, Alex Deucher wrote: > On Fri, Jul 21, 2023 at 4:18 AM Ilpo Järvinen > wrote: > > > > On Thu, 20 Jul 2023, Bjorn Helgaas wrote: > > > > > On Mon, Jul 17, 2023 at 03:04:57PM +0300, Ilpo Järvinen wrote: > > > > Don't assume that only the driver would be accessing LNKCTL. ASPM > > > > policy changes can trigger write to LNKCTL outside of driver's control. > > > > And in the case of upstream bridge, the driver does not even own the > > > > device it's changing the registers for. > > > > > > > > Use RMW capability accessors which do proper locking to avoid losing > > > > concurrent updates to the register value. > > > > > > > > Fixes: a2e73f56fa62 ("drm/amdgpu: Add support for CIK parts") > > > > Fixes: 62a37553414a ("drm/amdgpu: add si implementation v10") > > > > Suggested-by: Lukas Wunner > > > > Signed-off-by: Ilpo Järvinen > > > > Cc: stable@vger.kernel.org > > > > > > Do we have any reports of problems that are fixed by this patch (or by > > > others in the series)? If not, I'm not sure it really fits the usual > > > stable kernel criteria: > > > > > > https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/Documentation/process/stable-kernel-rules.rst?id=v6.4 > > > > I was on the edge with this. The answer to your direct question is no, > > there are no such reports so it would be okay to leave stable out I think. > > This applies to all patches in this series. > > > > Basically, this series came to be after Lukas noted the potential > > concurrency issues with how LNKCTL is unprotected when reviewing > > (internally) my bandwidth controller series. Then I went to look around > > all LNKCTL usage and realized existing things might alreary have similar > > issues. > > > > Do you want me to send another version w/o cc stable or you'll take care > > of that? > > > > > > --- > > > > drivers/gpu/drm/amd/amdgpu/cik.c | 36 +++++++++----------------------- > > > > drivers/gpu/drm/amd/amdgpu/si.c | 36 +++++++++----------------------- > > > > 2 files changed, 20 insertions(+), 52 deletions(-) > > > > > > > > diff --git a/drivers/gpu/drm/amd/amdgpu/cik.c b/drivers/gpu/drm/amd/amdgpu/cik.c > > > > index 5641cf05d856..e63abdf52b6c 100644 > > > > --- a/drivers/gpu/drm/amd/amdgpu/cik.c > > > > +++ b/drivers/gpu/drm/amd/amdgpu/cik.c > > > > @@ -1574,17 +1574,8 @@ static void cik_pcie_gen3_enable(struct amdgpu_device *adev) > > > > u16 bridge_cfg2, gpu_cfg2; > > > > u32 max_lw, current_lw, tmp; > > > > > > > > - pcie_capability_read_word(root, PCI_EXP_LNKCTL, > > > > - &bridge_cfg); > > > > - pcie_capability_read_word(adev->pdev, PCI_EXP_LNKCTL, > > > > - &gpu_cfg); > > > > - > > > > - tmp16 = bridge_cfg | PCI_EXP_LNKCTL_HAWD; > > > > - pcie_capability_write_word(root, PCI_EXP_LNKCTL, tmp16); > > > > - > > > > - tmp16 = gpu_cfg | PCI_EXP_LNKCTL_HAWD; > > > > - pcie_capability_write_word(adev->pdev, PCI_EXP_LNKCTL, > > > > - tmp16); > > > > + pcie_capability_set_word(root, PCI_EXP_LNKCTL, PCI_EXP_LNKCTL_HAWD); > > > > + pcie_capability_set_word(adev->pdev, PCI_EXP_LNKCTL, PCI_EXP_LNKCTL_HAWD); > > > > > > > > tmp = RREG32_PCIE(ixPCIE_LC_STATUS1); > > > > max_lw = (tmp & PCIE_LC_STATUS1__LC_DETECTED_LINK_WIDTH_MASK) >> > > > > @@ -1637,21 +1628,14 @@ static void cik_pcie_gen3_enable(struct amdgpu_device *adev) > > > > msleep(100); > > > > > > > > /* linkctl */ > > > > - pcie_capability_read_word(root, PCI_EXP_LNKCTL, > > > > - &tmp16); > > > > - tmp16 &= ~PCI_EXP_LNKCTL_HAWD; > > > > - tmp16 |= (bridge_cfg & PCI_EXP_LNKCTL_HAWD); > > > > - pcie_capability_write_word(root, PCI_EXP_LNKCTL, > > > > - tmp16); > > > > - > > > > - pcie_capability_read_word(adev->pdev, > > > > - PCI_EXP_LNKCTL, > > > > - &tmp16); > > > > - tmp16 &= ~PCI_EXP_LNKCTL_HAWD; > > > > - tmp16 |= (gpu_cfg & PCI_EXP_LNKCTL_HAWD); > > > > - pcie_capability_write_word(adev->pdev, > > > > - PCI_EXP_LNKCTL, > > > > - tmp16); > > > > + pcie_capability_clear_and_set_word(root, PCI_EXP_LNKCTL, > > > > + PCI_EXP_LNKCTL_HAWD, > > > > + bridge_cfg & > > > > + PCI_EXP_LNKCTL_HAWD); > > > > + pcie_capability_clear_and_set_word(adev->pdev, PCI_EXP_LNKCTL, > > > > + PCI_EXP_LNKCTL_HAWD, > > > > + gpu_cfg & > > > > + PCI_EXP_LNKCTL_HAWD); > > > > > > Wow, there's a lot of pointless-looking work going on here: > > > > > > set root PCI_EXP_LNKCTL_HAWD > > > set GPU PCI_EXP_LNKCTL_HAWD > > > > > > for (i = 0; i < 10; i++) { > > > read root PCI_EXP_LNKCTL > > > read GPU PCI_EXP_LNKCTL > > > > > > clear root PCI_EXP_LNKCTL_HAWD > > > if (root PCI_EXP_LNKCTL_HAWD was set) > > > set root PCI_EXP_LNKCTL_HAWD > > > > > > clear GPU PCI_EXP_LNKCTL_HAWD > > > if (GPU PCI_EXP_LNKCTL_HAWD was set) > > > set GPU PCI_EXP_LNKCTL_HAWD > > > } > > > > > > If it really *is* pointless, it would be nice to clean it up, but that > > > wouldn't be material for this patch, so what you have looks good. > > > > I really don't know if it's needed or not. There's stuff which looks hw > > specific going on besides those things you point out and I've not really > > understood what all that does. > > > > One annoying thing is that this code has been copy-pasted to appear in > > almost identical form in 4 files. > > > > I agree it certainly looks there might be room for cleaning things up here > > but such cleanups look a bit too scary to me w/o hw to test them. > > > > > > /* linkctl2 */ > > > > pcie_capability_read_word(root, PCI_EXP_LNKCTL2, > > > > > > The PCI_EXP_LNKCTL2 stuff also includes RMW updates. I don't see any > > > uses of PCI_EXP_LNKCTL2 outside this driver that look relevant, so I > > > guess we don't care about making the PCI_EXP_LNKCTL2 updates atomic? > > > > Currently no, which is why I left it out from this patchset. > > > > It is going to change soon though as I intend to submit bandwidth > > controller series after this series which will add RMW ops for LNKCTL2. > > The LNKCTL2 RMW parts are now in that series rather than in this one. > > > > After adding the bandwidth controller, this driver might be able to use > > it instead of tweaking LNKCTL2 directly to alter PCIe link speed (but I > > don't expect myself to be able to test these drivers and it feels too > > risky to make such a change without testing it, unfortunately). > > Thanks for the background. It was not clear what the point of this > patch set was. Thanks for chimming in! There are two reasons, one is to fix the existing RMW races and the other is the addition of BW controller. The RMW race issue was discovered while Lukas was reviewing BW controller's code and it was disjoint/generic enough from the BW controller to go into its own patchset. > This code and the similar code in radeon is just to > change the link speed of the GPU. Some older platforms used default > to slower link on boot so we added this code to renegotiate the link > to a faster speed when the driver loaded. If you are adding core > infrastructure to do that, we can switch to that. This was just the > programming sequence I got from the hardware team back when this code > was written. Most platforms I've seen these days come up at the max > supported speed of the platform and endpoint so I don't think the code > actually gets used much anymore. If I understood the code correctly, it's the device side with these GPUs which triggers the actual Link Speed change based on those HW specific writes? With the BW controller, the speed change would be initiated by setting the Root Port to do Link retraining. > Taking a step back, what is the end goal of the bandwidth controller > changes? The reason I ask is that today, we look at the currently > negotiated speed of the link and use that for the baseline in the > driver. The driver then enables PCIe dynamic power management where > the system management unit on the GPU dynamically adjusts the link > speed, width, and clock on demand based on the PCIe bandwidth > requirements of the currently executing GPU jobs to save power. This > might conflict with software if the goal is for some software > component to do something similar. BW controller is mainly done for thermal reasons (a thermal side cooling device is provided per root port for thermald/userspace to tap into) but it is not to say there couldn't be other usecases. If there's another actor besides BW controller altering Link Speeds, I think what would happen is that BW controller setting Root Port's Target Speed would just upper-bound the Link Speed requests from GPU (which makes sense at least from thermal point-of-view). -- i. --8323329-1571245239-1691071929=:1954--