2022-08-26 00:44:33

by William McVicker

[permalink] [raw]
Subject: [PATCH v6 0/2] PCI: dwc: Add support for 64-bit MSI target addresses

Hi All,

I've update patch 2/2 to address Robin's suggestions. This includes:

* Dropping the while-loop for retrying with a 64-bit mask in favor of
retrying within the error if-statement.
* Using an int for the DMA mask instead of a bool and ternary operation.

Thanks again for the reviews and sorry for the extra revision today!
Hopefully this is the last one :) If not, I'd be fine to submit patch 1/2
without 2/2 to avoid resending patch 1/2 for future revisions of patch 2/2
(unless I don't need to do that anyway).

Thanks,
Will

Will McVicker (2):
PCI: dwc: Drop dependency on ZONE_DMA32

v6:
* Retrying DMA allocation with 64-bit mask within the error if-statement.
* Use an int for the DMA mask instead of a bool and ternary operation.

v5:
* Updated patch 2/2 to first try with a 32-bit DMA mask. On failure,
retry with a 64-bit mask if supported.

v4:
* Updated commit descriptions.
* Renamed msi_64b -> msi_64bit.
* Dropped msi_64bit ternary use.
* Dropped export of dw_pcie_msi_capabilities.

v3:
* Switched to a managed DMA allocation.
* Simplified the DMA allocation cleanup.
* Dropped msi_page from struct dw_pcie_rp.
* Allocating a u64 instead of a full page.

v2:
* Fixed build error caught by kernel test robot
* Fixed error handling reported by Isaac Manjarres
PCI: dwc: Add support for 64-bit MSI target address

.../pci/controller/dwc/pcie-designware-host.c | 43 +++++++++----------
drivers/pci/controller/dwc/pcie-designware.c | 8 ++++
drivers/pci/controller/dwc/pcie-designware.h | 2 +-
3 files changed, 30 insertions(+), 23 deletions(-)


base-commit: 568035b01cfb107af8d2e4bd2fb9aea22cf5b868
--
2.37.2.672.g94769d06f0-goog


2022-08-29 08:31:26

by Lorenzo Pieralisi

[permalink] [raw]
Subject: Re: [PATCH v6 0/2] PCI: dwc: Add support for 64-bit MSI target addresses

On Thu, 25 Aug 2022 23:54:01 +0000, Will McVicker wrote:
> I've update patch 2/2 to address Robin's suggestions. This includes:
>
> * Dropping the while-loop for retrying with a 64-bit mask in favor of
> retrying within the error if-statement.
> * Using an int for the DMA mask instead of a bool and ternary operation.
>
> Thanks again for the reviews and sorry for the extra revision today!
> Hopefully this is the last one :) If not, I'd be fine to submit patch 1/2
> without 2/2 to avoid resending patch 1/2 for future revisions of patch 2/2
> (unless I don't need to do that anyway).
>
> [...]

Applied to pci/dwc, thanks!

[1/2] PCI: dwc: Drop dependency on ZONE_DMA32
https://git.kernel.org/lpieralisi/pci/c/423511ec23e2
[2/2] PCI: dwc: Add support for 64-bit MSI target address
https://git.kernel.org/lpieralisi/pci/c/e99d8c5e803b

Thanks,
Lorenzo

2023-01-31 12:30:09

by Evgenii Shatokhin

[permalink] [raw]
Subject: Re: [PATCH v6 0/2] PCI: dwc: Add support for 64-bit MSI target addresses

Hi,

On 26.08.2022 02:54, Will McVicker wrote:
> Hi All,
>
> I've update patch 2/2 to address Robin's suggestions. This includes:
>
> * Dropping the while-loop for retrying with a 64-bit mask in favor of
> retrying within the error if-statement.
> * Using an int for the DMA mask instead of a bool and ternary operation.
>
> Thanks again for the reviews and sorry for the extra revision today!
> Hopefully this is the last one :) If not, I'd be fine to submit patch 1/2
> without 2/2 to avoid resending patch 1/2 for future revisions of patch 2/2
> (unless I don't need to do that anyway).

The first patch of the series made it into the mainline kernel, but, it
seems, the second one ("PCI: dwc: Add support for 64-bit MSI target
address") did not. As of 6.2-rc6, it is still missing.

Was it intentionally dropped because of some issues or, perhaps, just by
accident? If it was by accident, could you please queue it for inclusion
into mainline again?

Support for 64-bit MSI target addresses is needed for some of our SoCs.
I ran into a situation when there was no available RAM in ZONE_DMA32
during initialization of PCIe host. Hence, dmam_alloc_coherent() failed
in dw_pcie_msi_host_init() and initialization failed with -ENOMEM:

[ 0.374834] dw-pcie 4000000.pcie0: host bridge /soc/pcie0@4000000 ranges:
[ 0.375813] dw-pcie 4000000.pcie0: MEM
0x0041000000..0x004fffffff -> 0x0041000000
[ 0.376171] dw-pcie 4000000.pcie0: IB MEM
0x0400000000..0x07ffffffff -> 0x0400000000
[ 0.377914] dw-pcie 4000000.pcie0: Failed to alloc and map MSI data
[ 0.378191] dw-pcie 4000000.pcie0: Failed to initialize host
[ 0.378255] dw-pcie: probe of 4000000.pcie0 failed with error -12

Mainline kernel 6.2-rc6 was used in that test.

The hardware supports 64-bit target addresses, so the patch "PCI: dwc:
Add support for 64-bit MSI target address" should help with this
particular failure.


>
> Thanks,
> Will
>
> Will McVicker (2):
> PCI: dwc: Drop dependency on ZONE_DMA32
>
> v6:
> * Retrying DMA allocation with 64-bit mask within the error if-statement.
> * Use an int for the DMA mask instead of a bool and ternary operation.
>
> v5:
> * Updated patch 2/2 to first try with a 32-bit DMA mask. On failure,
> retry with a 64-bit mask if supported.
>
> v4:
> * Updated commit descriptions.
> * Renamed msi_64b -> msi_64bit.
> * Dropped msi_64bit ternary use.
> * Dropped export of dw_pcie_msi_capabilities.
>
> v3:
> * Switched to a managed DMA allocation.
> * Simplified the DMA allocation cleanup.
> * Dropped msi_page from struct dw_pcie_rp.
> * Allocating a u64 instead of a full page.
>
> v2:
> * Fixed build error caught by kernel test robot
> * Fixed error handling reported by Isaac Manjarres
> PCI: dwc: Add support for 64-bit MSI target address
>
> .../pci/controller/dwc/pcie-designware-host.c | 43 +++++++++----------
> drivers/pci/controller/dwc/pcie-designware.c | 8 ++++
> drivers/pci/controller/dwc/pcie-designware.h | 2 +-
> 3 files changed, 30 insertions(+), 23 deletions(-)
>
>
> base-commit: 568035b01cfb107af8d2e4bd2fb9aea22cf5b868

Thank you in advance.

Regards,
Evgenii




2023-01-31 12:43:12

by Robin Murphy

[permalink] [raw]
Subject: Re: [PATCH v6 0/2] PCI: dwc: Add support for 64-bit MSI target addresses

On 2023-01-31 12:29, Evgenii Shatokhin wrote:
> Hi,
>
> On 26.08.2022 02:54, Will McVicker wrote:
>> Hi All,
>>
>> I've update patch 2/2 to address Robin's suggestions. This includes:
>>
>>   * Dropping the while-loop for retrying with a 64-bit mask in favor of
>>     retrying within the error if-statement.
>>   * Using an int for the DMA mask instead of a bool and ternary
>> operation.
>>
>> Thanks again for the reviews and sorry for the extra revision today!
>> Hopefully this is the last one :) If not, I'd be fine to submit patch 1/2
>> without 2/2 to avoid resending patch 1/2 for future revisions of patch
>> 2/2
>> (unless I don't need to do that anyway).
>
> The first patch of the series made it into the mainline kernel, but, it
> seems, the second one ("PCI: dwc: Add support for 64-bit MSI target
> address") did not. As of 6.2-rc6, it is still missing.
>
> Was it intentionally dropped because of some issues or, perhaps, just by
> accident? If it was by accident, could you please queue it for inclusion
> into mainline again?

Yes, it was dropped due to the PCI_MSI_FLAGS_64BIT usage apparently
being incorrect, and some other open debate (which all happened on the
v5 thread):

https://lore.kernel.org/linux-pci/YzVTmy9MWh+AjshC@lpieralisi/

The DMA mask issues have now been sorted out, so you, or Will, or anyone
else interested should be free to rework this on top of linux-next
(although at this point, more realistically on top of 6.3-rc1 in a few
weeks).

Thanks,
Robin.

> Support for 64-bit MSI target addresses is needed for some of our SoCs.
> I ran into a situation when there was no available RAM in ZONE_DMA32
> during initialization of PCIe host. Hence, dmam_alloc_coherent() failed
> in dw_pcie_msi_host_init() and initialization failed with -ENOMEM:
>
> [    0.374834] dw-pcie 4000000.pcie0: host bridge /soc/pcie0@4000000
> ranges:
> [    0.375813] dw-pcie 4000000.pcie0:      MEM
> 0x0041000000..0x004fffffff -> 0x0041000000
> [    0.376171] dw-pcie 4000000.pcie0:   IB MEM
> 0x0400000000..0x07ffffffff -> 0x0400000000
> [    0.377914] dw-pcie 4000000.pcie0: Failed to alloc and map MSI data
> [    0.378191] dw-pcie 4000000.pcie0: Failed to initialize host
> [    0.378255] dw-pcie: probe of 4000000.pcie0 failed with error -12
>
> Mainline kernel 6.2-rc6 was used in that test.
>
> The hardware supports 64-bit target addresses, so the patch "PCI: dwc:
> Add support for 64-bit MSI target address" should help with this
> particular failure.
>
>
>>
>> Thanks,
>> Will
>>
>> Will McVicker (2):
>>    PCI: dwc: Drop dependency on ZONE_DMA32
>>
>> v6:
>>   * Retrying DMA allocation with 64-bit mask within the error
>> if-statement.
>>   * Use an int for the DMA mask instead of a bool and ternary operation.
>>
>> v5:
>>   * Updated patch 2/2 to first try with a 32-bit DMA mask. On failure,
>>     retry with a 64-bit mask if supported.
>>
>> v4:
>>   * Updated commit descriptions.
>>   * Renamed msi_64b -> msi_64bit.
>>   * Dropped msi_64bit ternary use.
>>   * Dropped export of dw_pcie_msi_capabilities.
>>
>> v3:
>>    * Switched to a managed DMA allocation.
>>    * Simplified the DMA allocation cleanup.
>>    * Dropped msi_page from struct dw_pcie_rp.
>>    * Allocating a u64 instead of a full page.
>>
>> v2:
>>    * Fixed build error caught by kernel test robot
>>    * Fixed error handling reported by Isaac Manjarres
>>   PCI: dwc: Add support for 64-bit MSI target address
>>
>>   .../pci/controller/dwc/pcie-designware-host.c | 43 +++++++++----------
>>   drivers/pci/controller/dwc/pcie-designware.c  |  8 ++++
>>   drivers/pci/controller/dwc/pcie-designware.h  |  2 +-
>>   3 files changed, 30 insertions(+), 23 deletions(-)
>>
>>
>> base-commit: 568035b01cfb107af8d2e4bd2fb9aea22cf5b868
>
> Thank you in advance.
>
> Regards,
> Evgenii
>
>
>

2023-02-01 13:56:28

by Evgenii Shatokhin

[permalink] [raw]
Subject: Re: [PATCH v6 0/2] PCI: dwc: Add support for 64-bit MSI target addresses

On 31.01.2023 15:42, Robin Murphy wrote:
>
> On 2023-01-31 12:29, Evgenii Shatokhin wrote:
>> Hi,
>>
>> On 26.08.2022 02:54, Will McVicker wrote:
>>> Hi All,
>>>
>>> I've update patch 2/2 to address Robin's suggestions. This includes:
>>>
>>>   * Dropping the while-loop for retrying with a 64-bit mask in favor of
>>>     retrying within the error if-statement.
>>>   * Using an int for the DMA mask instead of a bool and ternary
>>> operation.
>>>
>>> Thanks again for the reviews and sorry for the extra revision today!
>>> Hopefully this is the last one :) If not, I'd be fine to submit patch
>>> 1/2
>>> without 2/2 to avoid resending patch 1/2 for future revisions of patch
>>> 2/2
>>> (unless I don't need to do that anyway).
>>
>> The first patch of the series made it into the mainline kernel, but, it
>> seems, the second one ("PCI: dwc: Add support for 64-bit MSI target
>> address") did not. As of 6.2-rc6, it is still missing.
>>
>> Was it intentionally dropped because of some issues or, perhaps, just by
>> accident? If it was by accident, could you please queue it for inclusion
>> into mainline again?
>
> Yes, it was dropped due to the PCI_MSI_FLAGS_64BIT usage apparently
> being incorrect, and some other open debate (which all happened on the
> v5 thread):
>
> https://lore.kernel.org/linux-pci/YzVTmy9MWh+AjshC@lpieralisi/

I see. If I understand it correctly, the problem was that
PCI_MSI_FLAGS_64BIT flag did not guarantee that 64-bit mask could be
used for that particular allocation. Right?

>
> The DMA mask issues have now been sorted out,

I suppose, you mean
https://lore.kernel.org/all/[email protected]/?

It still breaks our particular case when the SoC has no
32-bit-addressable RAM. We'd set DMA masks to DMA_BIT_MASK(36) in the
platform-specific driver before calling dw_pcie_host_init(). However,
dw_pcie_msi_host_init() resets it to 32-bit, tries dmam_alloc_coherent()
and fails.

With 36-bit masks, the kernel seems to play well with the devices in our
case.

I saw your comment in
https://lore.kernel.org/linux-pci/[email protected]/
that drivers should always explicitly set their masks.

Is it a really bad idea to check the current coherent mask's bits in
dw_pcie_msi_host_init() and if it is more than 32 - just issue a warning
rather than reset it to 32-bit unconditionally? That would help in our
case. Or, perhaps, there is a better workaround.

Looking forward to your comments.


> so you, or Will, or anyone
> else interested should be free to rework this on top of linux-next
> (although at this point, more realistically on top of 6.3-rc1 in a few
> weeks).
>
> Thanks,
> Robin.
>
>> Support for 64-bit MSI target addresses is needed for some of our SoCs.
>> I ran into a situation when there was no available RAM in ZONE_DMA32
>> during initialization of PCIe host. Hence, dmam_alloc_coherent() failed
>> in dw_pcie_msi_host_init() and initialization failed with -ENOMEM:
>>
>> [    0.374834] dw-pcie 4000000.pcie0: host bridge /soc/pcie0@4000000
>> ranges:
>> [    0.375813] dw-pcie 4000000.pcie0:      MEM
>> 0x0041000000..0x004fffffff -> 0x0041000000
>> [    0.376171] dw-pcie 4000000.pcie0:   IB MEM
>> 0x0400000000..0x07ffffffff -> 0x0400000000
>> [    0.377914] dw-pcie 4000000.pcie0: Failed to alloc and map MSI data
>> [    0.378191] dw-pcie 4000000.pcie0: Failed to initialize host
>> [    0.378255] dw-pcie: probe of 4000000.pcie0 failed with error -12
>>
>> Mainline kernel 6.2-rc6 was used in that test.
>>
>> The hardware supports 64-bit target addresses, so the patch "PCI: dwc:
>> Add support for 64-bit MSI target address" should help with this
>> particular failure.
>>
>>
>>>
>>> Thanks,
>>> Will
>>>
>>> Will McVicker (2):
>>>    PCI: dwc: Drop dependency on ZONE_DMA32
>>>
>>> v6:
>>>   * Retrying DMA allocation with 64-bit mask within the error
>>> if-statement.
>>>   * Use an int for the DMA mask instead of a bool and ternary operation.
>>>
>>> v5:
>>>   * Updated patch 2/2 to first try with a 32-bit DMA mask. On failure,
>>>     retry with a 64-bit mask if supported.
>>>
>>> v4:
>>>   * Updated commit descriptions.
>>>   * Renamed msi_64b -> msi_64bit.
>>>   * Dropped msi_64bit ternary use.
>>>   * Dropped export of dw_pcie_msi_capabilities.
>>>
>>> v3:
>>>    * Switched to a managed DMA allocation.
>>>    * Simplified the DMA allocation cleanup.
>>>    * Dropped msi_page from struct dw_pcie_rp.
>>>    * Allocating a u64 instead of a full page.
>>>
>>> v2:
>>>    * Fixed build error caught by kernel test robot
>>>    * Fixed error handling reported by Isaac Manjarres
>>>   PCI: dwc: Add support for 64-bit MSI target address
>>>
>>>   .../pci/controller/dwc/pcie-designware-host.c | 43 +++++++++----------
>>>   drivers/pci/controller/dwc/pcie-designware.c  |  8 ++++
>>>   drivers/pci/controller/dwc/pcie-designware.h  |  2 +-
>>>   3 files changed, 30 insertions(+), 23 deletions(-)
>>>
>>>
>>> base-commit: 568035b01cfb107af8d2e4bd2fb9aea22cf5b868
>>
>> Thank you in advance.
>>
>> Regards,
>> Evgenii
>>
>>
>>
>



2023-02-03 22:14:06

by Serge Semin

[permalink] [raw]
Subject: Re: [PATCH v6 0/2] PCI: dwc: Add support for 64-bit MSI target addresses

Hi Evgenii

On Wed, Feb 01, 2023 at 04:54:55PM +0300, Evgenii Shatokhin wrote:
> On 31.01.2023 15:42, Robin Murphy wrote:
> >
> > On 2023-01-31 12:29, Evgenii Shatokhin wrote:
> > > Hi,
> > >
> > > On 26.08.2022 02:54, Will McVicker wrote:
> > > > Hi All,
> > > >
> > > > I've update patch 2/2 to address Robin's suggestions. This includes:
> > > >
> > > > ? * Dropping the while-loop for retrying with a 64-bit mask in favor of
> > > > ??? retrying within the error if-statement.
> > > > ? * Using an int for the DMA mask instead of a bool and ternary
> > > > operation.
> > > >
> > > > Thanks again for the reviews and sorry for the extra revision today!
> > > > Hopefully this is the last one :) If not, I'd be fine to submit
> > > > patch 1/2
> > > > without 2/2 to avoid resending patch 1/2 for future revisions of patch
> > > > 2/2
> > > > (unless I don't need to do that anyway).
> > >
> > > The first patch of the series made it into the mainline kernel, but, it
> > > seems, the second one ("PCI: dwc: Add support for 64-bit MSI target
> > > address") did not. As of 6.2-rc6, it is still missing.
> > >
> > > Was it intentionally dropped because of some issues or, perhaps, just by
> > > accident? If it was by accident, could you please queue it for inclusion
> > > into mainline again?
> >
> > Yes, it was dropped due to the PCI_MSI_FLAGS_64BIT usage apparently
> > being incorrect, and some other open debate (which all happened on the
> > v5 thread):
> >
> > https://lore.kernel.org/linux-pci/YzVTmy9MWh+AjshC@lpieralisi/
>

> I see. If I understand it correctly, the problem was that
> PCI_MSI_FLAGS_64BIT flag did not guarantee that 64-bit mask could be used
> for that particular allocation. Right?
>

William was trying to utilize for only software cause. Setting
PCI_MSI_FLAGS_64BIT didn't actually change the hardware behavior.
He could have as well provided just a driver private capability
flag. (see below for a more detailed problem description)

> >
> > The DMA mask issues have now been sorted out,
>
> I suppose, you mean https://lore.kernel.org/all/[email protected]/?

Well, the way the DMA-mask issue has been solved was a bit of the
hacky. I wouldn't call it a fully proper solution. The problem with
pointlessly allocating physical memory for the iMSI-RX engine (it
doesn't perform any DMA) and artificially restricting the coherent-DMA
mask is still there. The patch in the subject was a compromise in
order to at least permit unrestricted streaming DMAs but limiting the
coherent DMAs for the MSI setup to work properly for all peripheral
devices.

>
> It still breaks our particular case when the SoC has no 32-bit-addressable
> RAM. We'd set DMA masks to DMA_BIT_MASK(36) in the platform-specific driver
> before calling dw_pcie_host_init(). However, dw_pcie_msi_host_init() resets
> it to 32-bit, tries dmam_alloc_coherent() and fails.

Yeah. That's another problem with the implemented approach. But are
your sure the driver had worked even before this patch? AFAICS the
driver allocated the MSI-targeted page from DMA32 zone before this
modification. So the allocation must have failed on your platform too.

>
> With 36-bit masks, the kernel seems to play well with the devices in our
> case.
>
> I saw your comment in https://lore.kernel.org/linux-pci/[email protected]/
> that drivers should always explicitly set their masks.
>

> Is it a really bad idea to check the current coherent mask's bits in
> dw_pcie_msi_host_init() and if it is more than 32 - just issue a warning
> rather than reset it to 32-bit unconditionally? That would help in our case.
> Or, perhaps, there is a better workaround.

The problem isn't in the value the mask is set to. The problem is
two-leveled, but is mainly connected with the PCIe device detected on
the PCIe bus. There are some of them which can't send MSI TLPs to the
64-bit addresses. Since we can't predict whether such devices exist on
the bus beforehand the LLDD probe is performed together with the
MSI-engine initialization, the solution was to just restrict the MSIs
base address to be allocated within the lowest 4GB. Moreover as I said
above the iMSI-RX engine doesn't actually cause any DMA thus there is
no need in any memory allocation. Instead reserving some PCIe-bus
space/DWORD for MSIs would be enough. Alas the PCIe-subsystem doesn't
provide a way to do so. That's why we have what you see in the driver:
DMA mask restriction and coherent DMA memory allocation.

If only we had a way to auto-detected the PCIe-bus space with no
physical memory behind it and take out a DWORD from it to initialize
the iMSI-RX engine we could have immediately got rid from the mask
setting operation and the memory allocation. It would have solved your
problem too.

-Serge(y)

>
> Looking forward to your comments.



>
>
> > so you, or Will, or anyone
> > else interested should be free to rework this on top of linux-next
> > (although at this point, more realistically on top of 6.3-rc1 in a few
> > weeks).
> >
> > Thanks,
> > Robin.
> >
> > > Support for 64-bit MSI target addresses is needed for some of our SoCs.
> > > I ran into a situation when there was no available RAM in ZONE_DMA32
> > > during initialization of PCIe host. Hence, dmam_alloc_coherent() failed
> > > in dw_pcie_msi_host_init() and initialization failed with -ENOMEM:
> > >
> > > [??? 0.374834] dw-pcie 4000000.pcie0: host bridge /soc/pcie0@4000000
> > > ranges:
> > > [??? 0.375813] dw-pcie 4000000.pcie0:????? MEM
> > > 0x0041000000..0x004fffffff -> 0x0041000000
> > > [??? 0.376171] dw-pcie 4000000.pcie0:?? IB MEM
> > > 0x0400000000..0x07ffffffff -> 0x0400000000
> > > [??? 0.377914] dw-pcie 4000000.pcie0: Failed to alloc and map MSI data
> > > [??? 0.378191] dw-pcie 4000000.pcie0: Failed to initialize host
> > > [??? 0.378255] dw-pcie: probe of 4000000.pcie0 failed with error -12
> > >
> > > Mainline kernel 6.2-rc6 was used in that test.
> > >
> > > The hardware supports 64-bit target addresses, so the patch "PCI: dwc:
> > > Add support for 64-bit MSI target address" should help with this
> > > particular failure.
> > >
> > >
> > > >
> > > > Thanks,
> > > > Will
> > > >
> > > > Will McVicker (2):
> > > > ?? PCI: dwc: Drop dependency on ZONE_DMA32
> > > >
> > > > v6:
> > > > ? * Retrying DMA allocation with 64-bit mask within the error
> > > > if-statement.
> > > > ? * Use an int for the DMA mask instead of a bool and ternary operation.
> > > >
> > > > v5:
> > > > ? * Updated patch 2/2 to first try with a 32-bit DMA mask. On failure,
> > > > ??? retry with a 64-bit mask if supported.
> > > >
> > > > v4:
> > > > ? * Updated commit descriptions.
> > > > ? * Renamed msi_64b -> msi_64bit.
> > > > ? * Dropped msi_64bit ternary use.
> > > > ? * Dropped export of dw_pcie_msi_capabilities.
> > > >
> > > > v3:
> > > > ?? * Switched to a managed DMA allocation.
> > > > ?? * Simplified the DMA allocation cleanup.
> > > > ?? * Dropped msi_page from struct dw_pcie_rp.
> > > > ?? * Allocating a u64 instead of a full page.
> > > >
> > > > v2:
> > > > ?? * Fixed build error caught by kernel test robot
> > > > ?? * Fixed error handling reported by Isaac Manjarres
> > > > ? PCI: dwc: Add support for 64-bit MSI target address
> > > >
> > > > ? .../pci/controller/dwc/pcie-designware-host.c | 43 +++++++++----------
> > > > ? drivers/pci/controller/dwc/pcie-designware.c? |? 8 ++++
> > > > ? drivers/pci/controller/dwc/pcie-designware.h? |? 2 +-
> > > > ? 3 files changed, 30 insertions(+), 23 deletions(-)
> > > >
> > > >
> > > > base-commit: 568035b01cfb107af8d2e4bd2fb9aea22cf5b868
> > >
> > > Thank you in advance.
> > >
> > > Regards,
> > > Evgenii
> > >
> > >
> > >
> >
>
>
>


2023-02-06 11:27:59

by Evgenii Shatokhin

[permalink] [raw]
Subject: Re: [PATCH v6 0/2] PCI: dwc: Add support for 64-bit MSI target addresses

Hi Sergey,

First of all, thank you for the detailed explanation. It is clearer now
what is going on and why it is that way.

On 04.02.2023 01:12, Serge Semin wrote:
> Hi Evgenii
>
> On Wed, Feb 01, 2023 at 04:54:55PM +0300, Evgenii Shatokhin wrote:
>> On 31.01.2023 15:42, Robin Murphy wrote:
>>>
>>> On 2023-01-31 12:29, Evgenii Shatokhin wrote:
>>>> Hi,
>>>>
>>>> On 26.08.2022 02:54, Will McVicker wrote:
>>>>> Hi All,
>>>>>
>>>>> I've update patch 2/2 to address Robin's suggestions. This includes:
>>>>>
>>>>> * Dropping the while-loop for retrying with a 64-bit mask in favor of
>>>>> retrying within the error if-statement.
>>>>> * Using an int for the DMA mask instead of a bool and ternary
>>>>> operation.
>>>>>
>>>>> Thanks again for the reviews and sorry for the extra revision today!
>>>>> Hopefully this is the last one :) If not, I'd be fine to submit
>>>>> patch 1/2
>>>>> without 2/2 to avoid resending patch 1/2 for future revisions of patch
>>>>> 2/2
>>>>> (unless I don't need to do that anyway).
>>>>
>>>> The first patch of the series made it into the mainline kernel, but, it
>>>> seems, the second one ("PCI: dwc: Add support for 64-bit MSI target
>>>> address") did not. As of 6.2-rc6, it is still missing.
>>>>
>>>> Was it intentionally dropped because of some issues or, perhaps, just by
>>>> accident? If it was by accident, could you please queue it for inclusion
>>>> into mainline again?
>>>
>>> Yes, it was dropped due to the PCI_MSI_FLAGS_64BIT usage apparently
>>> being incorrect, and some other open debate (which all happened on the
>>> v5 thread):
>>>
>>> https://lore.kernel.org/linux-pci/YzVTmy9MWh+AjshC@lpieralisi/
>>
>
>> I see. If I understand it correctly, the problem was that
>> PCI_MSI_FLAGS_64BIT flag did not guarantee that 64-bit mask could be used
>> for that particular allocation. Right?
>>
>
> William was trying to utilize for only software cause. Setting
> PCI_MSI_FLAGS_64BIT didn't actually change the hardware behavior.
> He could have as well provided just a driver private capability
> flag. (see below for a more detailed problem description)
>
>>>
>>> The DMA mask issues have now been sorted out,
>>
>> I suppose, you mean https://lore.kernel.org/all/[email protected]/?
>
> Well, the way the DMA-mask issue has been solved was a bit of the
> hacky. I wouldn't call it a fully proper solution. The problem with
> pointlessly allocating physical memory for the iMSI-RX engine (it
> doesn't perform any DMA) and artificially restricting the coherent-DMA
> mask is still there. The patch in the subject was a compromise in
> order to at least permit unrestricted streaming DMAs but limiting the
> coherent DMAs for the MSI setup to work properly for all peripheral
> devices.
>
>>
>> It still breaks our particular case when the SoC has no 32-bit-addressable
>> RAM. We'd set DMA masks to DMA_BIT_MASK(36) in the platform-specific driver
>> before calling dw_pcie_host_init(). However, dw_pcie_msi_host_init() resets
>> it to 32-bit, tries dmam_alloc_coherent() and fails.
>
> Yeah. That's another problem with the implemented approach. But are
> your sure the driver had worked even before this patch? AFAICS the
> driver allocated the MSI-targeted page from DMA32 zone before this
> modification. So the allocation must have failed on your platform too.

You are right. I did not notice earlier that the kernel based on
6.0-stable we used before did actually contain our SoC-specific
workaround for this. Without that custom patch, initialization of PCIe
host does not work. So, yes, the problem was present earlier too.

>
>>
>> With 36-bit masks, the kernel seems to play well with the devices in our
>> case.
>>
>> I saw your comment in https://lore.kernel.org/linux-pci/[email protected]/
>> that drivers should always explicitly set their masks.
>>
>
>> Is it a really bad idea to check the current coherent mask's bits in
>> dw_pcie_msi_host_init() and if it is more than 32 - just issue a warning
>> rather than reset it to 32-bit unconditionally? That would help in our case.
>> Or, perhaps, there is a better workaround.
>
> The problem isn't in the value the mask is set to. The problem is
> two-leveled, but is mainly connected with the PCIe device detected on
> the PCIe bus. There are some of them which can't send MSI TLPs to the
> 64-bit addresses. Since we can't predict whether such devices exist on
> the bus beforehand the LLDD probe is performed together with the
> MSI-engine initialization, the solution was to just restrict the MSIs
> base address to be allocated within the lowest 4GB. Moreover as I said
> above the iMSI-RX engine doesn't actually cause any DMA thus there is
> no need in any memory allocation. Instead reserving some PCIe-bus
> space/DWORD for MSIs would be enough. Alas the PCIe-subsystem doesn't
> provide a way to do so. That's why we have what you see in the driver:
> DMA mask restriction and coherent DMA memory allocation.

So, if I understand you correctly, what is needed here is a small area
of PCIe address space accessible to any of the connected PCIe devices.
As the kernel does not know in advance, which restrictions the devices
have, it tries to allocate 32-bit-addressable memory, suitable for DMA.
This way, it would be OK for any attached PCIe device. Right?

>
> If only we had a way to auto-detected the PCIe-bus space with no
> physical memory behind it and take out a DWORD from it to initialize
> the iMSI-RX engine we could have immediately got rid from the mask
> setting operation and the memory allocation. It would have solved your
> problem too.

Yes, it would solve our issue too. I do not know, however, if a generic
solution is possible here, but I am no expert in PCIe.

For now, we are probably better off with SoC-specific patches, when we
know which PCIe devices can possibly be used and what their restrictions
are.

>
> -Serge(y)
>
>>
>> Looking forward to your comments.
>
>
>
>>
>>
>>> so you, or Will, or anyone
>>> else interested should be free to rework this on top of linux-next
>>> (although at this point, more realistically on top of 6.3-rc1 in a few
>>> weeks).
>>>
>>> Thanks,
>>> Robin.
>>>
>>>> Support for 64-bit MSI target addresses is needed for some of our SoCs.
>>>> I ran into a situation when there was no available RAM in ZONE_DMA32
>>>> during initialization of PCIe host. Hence, dmam_alloc_coherent() failed
>>>> in dw_pcie_msi_host_init() and initialization failed with -ENOMEM:
>>>>
>>>> [    0.374834] dw-pcie 4000000.pcie0: host bridge /soc/pcie0@4000000
>>>> ranges:
>>>> [    0.375813] dw-pcie 4000000.pcie0: MEM
>>>> 0x0041000000..0x004fffffff -> 0x0041000000
>>>> [    0.376171] dw-pcie 4000000.pcie0: IB MEM
>>>> 0x0400000000..0x07ffffffff -> 0x0400000000
>>>> [    0.377914] dw-pcie 4000000.pcie0: Failed to alloc and map MSI data
>>>> [    0.378191] dw-pcie 4000000.pcie0: Failed to initialize host
>>>> [    0.378255] dw-pcie: probe of 4000000.pcie0 failed with error -12
>>>>
>>>> Mainline kernel 6.2-rc6 was used in that test.
>>>>
>>>> The hardware supports 64-bit target addresses, so the patch "PCI: dwc:
>>>> Add support for 64-bit MSI target address" should help with this
>>>> particular failure.
>>>>
>>>>
>>>>>
>>>>> Thanks,
>>>>> Will
>>>>>
>>>>> Will McVicker (2):
>>>>> PCI: dwc: Drop dependency on ZONE_DMA32
>>>>>
>>>>> v6:
>>>>> * Retrying DMA allocation with 64-bit mask within the error
>>>>> if-statement.
>>>>> * Use an int for the DMA mask instead of a bool and ternary operation.
>>>>>
>>>>> v5:
>>>>> * Updated patch 2/2 to first try with a 32-bit DMA mask. On failure,
>>>>> retry with a 64-bit mask if supported.
>>>>>
>>>>> v4:
>>>>> * Updated commit descriptions.
>>>>> * Renamed msi_64b -> msi_64bit.
>>>>> * Dropped msi_64bit ternary use.
>>>>> * Dropped export of dw_pcie_msi_capabilities.
>>>>>
>>>>> v3:
>>>>> * Switched to a managed DMA allocation.
>>>>> * Simplified the DMA allocation cleanup.
>>>>> * Dropped msi_page from struct dw_pcie_rp.
>>>>> * Allocating a u64 instead of a full page.
>>>>>
>>>>> v2:
>>>>> * Fixed build error caught by kernel test robot
>>>>> * Fixed error handling reported by Isaac Manjarres
>>>>> PCI: dwc: Add support for 64-bit MSI target address
>>>>>
>>>>> .../pci/controller/dwc/pcie-designware-host.c | 43 +++++++++----------
>>>>> drivers/pci/controller/dwc/pcie-designware.c |  8 ++++
>>>>> drivers/pci/controller/dwc/pcie-designware.h |  2 +-
>>>>> 3 files changed, 30 insertions(+), 23 deletions(-)
>>>>>
>>>>>
>>>>> base-commit: 568035b01cfb107af8d2e4bd2fb9aea22cf5b868
>>>>
>>>> Thank you in advance.

Regards,
Evgenii



2023-02-09 00:49:00

by Serge Semin

[permalink] [raw]
Subject: Re: [PATCH v6 0/2] PCI: dwc: Add support for 64-bit MSI target addresses

On Mon, Feb 06, 2023 at 02:27:41PM +0300, Evgenii Shatokhin wrote:
> Hi Sergey,
>
> First of all, thank you for the detailed explanation. It is clearer now what
> is going on and why it is that way.
>
> On 04.02.2023 01:12, Serge Semin wrote:
> > Hi Evgenii
> >
> > On Wed, Feb 01, 2023 at 04:54:55PM +0300, Evgenii Shatokhin wrote:
> > > On 31.01.2023 15:42, Robin Murphy wrote:
> > > >
> > > > On 2023-01-31 12:29, Evgenii Shatokhin wrote:
> > > > > Hi,
> > > > >
> > > > > On 26.08.2022 02:54, Will McVicker wrote:
> > > > > > Hi All,
> > > > > >
> > > > > > I've update patch 2/2 to address Robin's suggestions. This includes:
> > > > > >
> > > > > > * Dropping the while-loop for retrying with a 64-bit mask in favor of
> > > > > > retrying within the error if-statement.
> > > > > > * Using an int for the DMA mask instead of a bool and ternary
> > > > > > operation.
> > > > > >
> > > > > > Thanks again for the reviews and sorry for the extra revision today!
> > > > > > Hopefully this is the last one :) If not, I'd be fine to submit
> > > > > > patch 1/2
> > > > > > without 2/2 to avoid resending patch 1/2 for future revisions of patch
> > > > > > 2/2
> > > > > > (unless I don't need to do that anyway).
> > > > >
> > > > > The first patch of the series made it into the mainline kernel, but, it
> > > > > seems, the second one ("PCI: dwc: Add support for 64-bit MSI target
> > > > > address") did not. As of 6.2-rc6, it is still missing.
> > > > >
> > > > > Was it intentionally dropped because of some issues or, perhaps, just by
> > > > > accident? If it was by accident, could you please queue it for inclusion
> > > > > into mainline again?
> > > >
> > > > Yes, it was dropped due to the PCI_MSI_FLAGS_64BIT usage apparently
> > > > being incorrect, and some other open debate (which all happened on the
> > > > v5 thread):
> > > >
> > > > https://lore.kernel.org/linux-pci/YzVTmy9MWh+AjshC@lpieralisi/
> > >
> >
> > > I see. If I understand it correctly, the problem was that
> > > PCI_MSI_FLAGS_64BIT flag did not guarantee that 64-bit mask could be used
> > > for that particular allocation. Right?
> > >
> >
> > William was trying to utilize for only software cause. Setting
> > PCI_MSI_FLAGS_64BIT didn't actually change the hardware behavior.
> > He could have as well provided just a driver private capability
> > flag. (see below for a more detailed problem description)
> >
> > > >
> > > > The DMA mask issues have now been sorted out,
> > >
> > > I suppose, you mean https://lore.kernel.org/all/[email protected]/?
> >
> > Well, the way the DMA-mask issue has been solved was a bit of the
> > hacky. I wouldn't call it a fully proper solution. The problem with
> > pointlessly allocating physical memory for the iMSI-RX engine (it
> > doesn't perform any DMA) and artificially restricting the coherent-DMA
> > mask is still there. The patch in the subject was a compromise in
> > order to at least permit unrestricted streaming DMAs but limiting the
> > coherent DMAs for the MSI setup to work properly for all peripheral
> > devices.
> >
> > >
> > > It still breaks our particular case when the SoC has no 32-bit-addressable
> > > RAM. We'd set DMA masks to DMA_BIT_MASK(36) in the platform-specific driver
> > > before calling dw_pcie_host_init(). However, dw_pcie_msi_host_init() resets
> > > it to 32-bit, tries dmam_alloc_coherent() and fails.
> >
> > Yeah. That's another problem with the implemented approach. But are
> > your sure the driver had worked even before this patch? AFAICS the
> > driver allocated the MSI-targeted page from DMA32 zone before this
> > modification. So the allocation must have failed on your platform too.
>
> You are right. I did not notice earlier that the kernel based on 6.0-stable
> we used before did actually contain our SoC-specific workaround for this.
> Without that custom patch, initialization of PCIe host does not work. So,
> yes, the problem was present earlier too.
>
> >
> > >
> > > With 36-bit masks, the kernel seems to play well with the devices in our
> > > case.
> > >
> > > I saw your comment in https://lore.kernel.org/linux-pci/[email protected]/
> > > that drivers should always explicitly set their masks.
> > >
> >
> > > Is it a really bad idea to check the current coherent mask's bits in
> > > dw_pcie_msi_host_init() and if it is more than 32 - just issue a warning
> > > rather than reset it to 32-bit unconditionally? That would help in our case.
> > > Or, perhaps, there is a better workaround.
> >
> > The problem isn't in the value the mask is set to. The problem is
> > two-leveled, but is mainly connected with the PCIe device detected on
> > the PCIe bus. There are some of them which can't send MSI TLPs to the
> > 64-bit addresses. Since we can't predict whether such devices exist on
> > the bus beforehand the LLDD probe is performed together with the
> > MSI-engine initialization, the solution was to just restrict the MSIs
> > base address to be allocated within the lowest 4GB. Moreover as I said
> > above the iMSI-RX engine doesn't actually cause any DMA thus there is
> > no need in any memory allocation. Instead reserving some PCIe-bus
> > space/DWORD for MSIs would be enough. Alas the PCIe-subsystem doesn't
> > provide a way to do so. That's why we have what you see in the driver:
> > DMA mask restriction and coherent DMA memory allocation.
>

> So, if I understand you correctly, what is needed here is a small area of
> PCIe address space accessible to any of the connected PCIe devices. As the
> kernel does not know in advance, which restrictions the devices have, it
> tries to allocate 32-bit-addressable memory, suitable for DMA. This way, it
> would be OK for any attached PCIe device. Right?

Right. The restriction is the 64-bit MSI capability. If any PCIe
peripheral device has no PCI_MSI_64_BIT_ADDR_CAP flag set and the MSI
target address is selected from the space above 4GB then such device
MSIs won't be handled.

Note as I said above no DMA actually performed if at least one MSI
vector is enabled. The driver just needs a DWORD within the PCIe bus
space for the MSI MWr TLPs target address and EP+vector data. The MSI
TLP data is decoded by the iMSI-RX engine in order to set the
corresponding flag in the MSI IRQ status register. Such TLPs won't be
passed to the master AXI-bus.

>
> >
> > If only we had a way to auto-detected the PCIe-bus space with no
> > physical memory behind it and take out a DWORD from it to initialize
> > the iMSI-RX engine we could have immediately got rid from the mask
> > setting operation and the memory allocation. It would have solved your
> > problem too.
>

> Yes, it would solve our issue too. I do not know, however, if a generic
> solution is possible here, but I am no expert in PCIe.

Currently the kernel PCIe subsystem doesn't provide a way to reserve a
range within the PCIe bus memory with no physical RAM behind and left
unused during the BARs resource initialization. Implementing such
functionality (perhaps in the framework of the P2P module or based on
it) would give the generic solution.

>
> For now, we are probably better off with SoC-specific patches, when we know
> which PCIe devices can possibly be used and what their restrictions are.

Since you know that there is no any RAM below 4GB and you have
matching CPU and PCIe address spaces, then you can just take any
address below 4GB and use it to initialize the MSI-target address
(dw_pcie_rp.msi_data). But make sure that the peripheral PCIe-devices
don't use it for something application-specific (like accessing CPU
MMIO devices mapped to that base address). That seems like the most
universal solution for your case.

-Serge(y)

>
> >
> > -Serge(y)
> >
> > >
> > > Looking forward to your comments.
> >
> >
> >
> > >
> > >
> > > > so you, or Will, or anyone
> > > > else interested should be free to rework this on top of linux-next
> > > > (although at this point, more realistically on top of 6.3-rc1 in a few
> > > > weeks).
> > > >
> > > > Thanks,
> > > > Robin.
> > > >
> > > > > Support for 64-bit MSI target addresses is needed for some of our SoCs.
> > > > > I ran into a situation when there was no available RAM in ZONE_DMA32
> > > > > during initialization of PCIe host. Hence, dmam_alloc_coherent() failed
> > > > > in dw_pcie_msi_host_init() and initialization failed with -ENOMEM:
> > > > >
> > > > > [??? 0.374834] dw-pcie 4000000.pcie0: host bridge /soc/pcie0@4000000
> > > > > ranges:
> > > > > [??? 0.375813] dw-pcie 4000000.pcie0: MEM
> > > > > 0x0041000000..0x004fffffff -> 0x0041000000
> > > > > [??? 0.376171] dw-pcie 4000000.pcie0: IB MEM
> > > > > 0x0400000000..0x07ffffffff -> 0x0400000000
> > > > > [??? 0.377914] dw-pcie 4000000.pcie0: Failed to alloc and map MSI data
> > > > > [??? 0.378191] dw-pcie 4000000.pcie0: Failed to initialize host
> > > > > [??? 0.378255] dw-pcie: probe of 4000000.pcie0 failed with error -12
> > > > >
> > > > > Mainline kernel 6.2-rc6 was used in that test.
> > > > >
> > > > > The hardware supports 64-bit target addresses, so the patch "PCI: dwc:
> > > > > Add support for 64-bit MSI target address" should help with this
> > > > > particular failure.
> > > > >
> > > > >
> > > > > >
> > > > > > Thanks,
> > > > > > Will
> > > > > >
> > > > > > Will McVicker (2):
> > > > > > PCI: dwc: Drop dependency on ZONE_DMA32
> > > > > >
> > > > > > v6:
> > > > > > * Retrying DMA allocation with 64-bit mask within the error
> > > > > > if-statement.
> > > > > > * Use an int for the DMA mask instead of a bool and ternary operation.
> > > > > >
> > > > > > v5:
> > > > > > * Updated patch 2/2 to first try with a 32-bit DMA mask. On failure,
> > > > > > retry with a 64-bit mask if supported.
> > > > > >
> > > > > > v4:
> > > > > > * Updated commit descriptions.
> > > > > > * Renamed msi_64b -> msi_64bit.
> > > > > > * Dropped msi_64bit ternary use.
> > > > > > * Dropped export of dw_pcie_msi_capabilities.
> > > > > >
> > > > > > v3:
> > > > > > * Switched to a managed DMA allocation.
> > > > > > * Simplified the DMA allocation cleanup.
> > > > > > * Dropped msi_page from struct dw_pcie_rp.
> > > > > > * Allocating a u64 instead of a full page.
> > > > > >
> > > > > > v2:
> > > > > > * Fixed build error caught by kernel test robot
> > > > > > * Fixed error handling reported by Isaac Manjarres
> > > > > > PCI: dwc: Add support for 64-bit MSI target address
> > > > > >
> > > > > > .../pci/controller/dwc/pcie-designware-host.c | 43 +++++++++----------
> > > > > > drivers/pci/controller/dwc/pcie-designware.c |? 8 ++++
> > > > > > drivers/pci/controller/dwc/pcie-designware.h |? 2 +-
> > > > > > 3 files changed, 30 insertions(+), 23 deletions(-)
> > > > > >
> > > > > >
> > > > > > base-commit: 568035b01cfb107af8d2e4bd2fb9aea22cf5b868
> > > > >
> > > > > Thank you in advance.
>
> Regards,
> Evgenii
>
>

2023-02-09 08:03:20

by Evgenii Shatokhin

[permalink] [raw]
Subject: Re: [PATCH v6 0/2] PCI: dwc: Add support for 64-bit MSI target addresses

On 09.02.2023 03:48, Serge Semin wrote:
> On Mon, Feb 06, 2023 at 02:27:41PM +0300, Evgenii Shatokhin wrote:
>> Hi Sergey,
>>
>> First of all, thank you for the detailed explanation. It is clearer now what
>> is going on and why it is that way.
>>
>> On 04.02.2023 01:12, Serge Semin wrote:
>>> Hi Evgenii
>>>
>>> On Wed, Feb 01, 2023 at 04:54:55PM +0300, Evgenii Shatokhin wrote:
>>>> On 31.01.2023 15:42, Robin Murphy wrote:
>>>>>
>>>>> On 2023-01-31 12:29, Evgenii Shatokhin wrote:
>>>>>> Hi,
>>>>>>
>>>>>> On 26.08.2022 02:54, Will McVicker wrote:
>>>>>>> Hi All,
>>>>>>>
>>>>>>> I've update patch 2/2 to address Robin's suggestions. This includes:
>>>>>>>
>>>>>>> * Dropping the while-loop for retrying with a 64-bit mask in favor of
>>>>>>> retrying within the error if-statement.
>>>>>>> * Using an int for the DMA mask instead of a bool and ternary
>>>>>>> operation.
>>>>>>>
>>>>>>> Thanks again for the reviews and sorry for the extra revision today!
>>>>>>> Hopefully this is the last one :) If not, I'd be fine to submit
>>>>>>> patch 1/2
>>>>>>> without 2/2 to avoid resending patch 1/2 for future revisions of patch
>>>>>>> 2/2
>>>>>>> (unless I don't need to do that anyway).
>>>>>>
>>>>>> The first patch of the series made it into the mainline kernel, but, it
>>>>>> seems, the second one ("PCI: dwc: Add support for 64-bit MSI target
>>>>>> address") did not. As of 6.2-rc6, it is still missing.
>>>>>>
>>>>>> Was it intentionally dropped because of some issues or, perhaps, just by
>>>>>> accident? If it was by accident, could you please queue it for inclusion
>>>>>> into mainline again?
>>>>>
>>>>> Yes, it was dropped due to the PCI_MSI_FLAGS_64BIT usage apparently
>>>>> being incorrect, and some other open debate (which all happened on the
>>>>> v5 thread):
>>>>>
>>>>> https://lore.kernel.org/linux-pci/YzVTmy9MWh+AjshC@lpieralisi/
>>>>
>>>
>>>> I see. If I understand it correctly, the problem was that
>>>> PCI_MSI_FLAGS_64BIT flag did not guarantee that 64-bit mask could be used
>>>> for that particular allocation. Right?
>>>>
>>>
>>> William was trying to utilize for only software cause. Setting
>>> PCI_MSI_FLAGS_64BIT didn't actually change the hardware behavior.
>>> He could have as well provided just a driver private capability
>>> flag. (see below for a more detailed problem description)
>>>
>>>>>
>>>>> The DMA mask issues have now been sorted out,
>>>>
>>>> I suppose, you mean https://lore.kernel.org/all/[email protected]/?
>>>
>>> Well, the way the DMA-mask issue has been solved was a bit of the
>>> hacky. I wouldn't call it a fully proper solution. The problem with
>>> pointlessly allocating physical memory for the iMSI-RX engine (it
>>> doesn't perform any DMA) and artificially restricting the coherent-DMA
>>> mask is still there. The patch in the subject was a compromise in
>>> order to at least permit unrestricted streaming DMAs but limiting the
>>> coherent DMAs for the MSI setup to work properly for all peripheral
>>> devices.
>>>
>>>>
>>>> It still breaks our particular case when the SoC has no 32-bit-addressable
>>>> RAM. We'd set DMA masks to DMA_BIT_MASK(36) in the platform-specific driver
>>>> before calling dw_pcie_host_init(). However, dw_pcie_msi_host_init() resets
>>>> it to 32-bit, tries dmam_alloc_coherent() and fails.
>>>
>>> Yeah. That's another problem with the implemented approach. But are
>>> your sure the driver had worked even before this patch? AFAICS the
>>> driver allocated the MSI-targeted page from DMA32 zone before this
>>> modification. So the allocation must have failed on your platform too.
>>
>> You are right. I did not notice earlier that the kernel based on 6.0-stable
>> we used before did actually contain our SoC-specific workaround for this.
>> Without that custom patch, initialization of PCIe host does not work. So,
>> yes, the problem was present earlier too.
>>
>>>
>>>>
>>>> With 36-bit masks, the kernel seems to play well with the devices in our
>>>> case.
>>>>
>>>> I saw your comment in https://lore.kernel.org/linux-pci/[email protected]/
>>>> that drivers should always explicitly set their masks.
>>>>
>>>
>>>> Is it a really bad idea to check the current coherent mask's bits in
>>>> dw_pcie_msi_host_init() and if it is more than 32 - just issue a warning
>>>> rather than reset it to 32-bit unconditionally? That would help in our case.
>>>> Or, perhaps, there is a better workaround.
>>>
>>> The problem isn't in the value the mask is set to. The problem is
>>> two-leveled, but is mainly connected with the PCIe device detected on
>>> the PCIe bus. There are some of them which can't send MSI TLPs to the
>>> 64-bit addresses. Since we can't predict whether such devices exist on
>>> the bus beforehand the LLDD probe is performed together with the
>>> MSI-engine initialization, the solution was to just restrict the MSIs
>>> base address to be allocated within the lowest 4GB. Moreover as I said
>>> above the iMSI-RX engine doesn't actually cause any DMA thus there is
>>> no need in any memory allocation. Instead reserving some PCIe-bus
>>> space/DWORD for MSIs would be enough. Alas the PCIe-subsystem doesn't
>>> provide a way to do so. That's why we have what you see in the driver:
>>> DMA mask restriction and coherent DMA memory allocation.
>>
>
>> So, if I understand you correctly, what is needed here is a small area of
>> PCIe address space accessible to any of the connected PCIe devices. As the
>> kernel does not know in advance, which restrictions the devices have, it
>> tries to allocate 32-bit-addressable memory, suitable for DMA. This way, it
>> would be OK for any attached PCIe device. Right?
>
> Right. The restriction is the 64-bit MSI capability. If any PCIe
> peripheral device has no PCI_MSI_64_BIT_ADDR_CAP flag set and the MSI
> target address is selected from the space above 4GB then such device
> MSIs won't be handled.
>
> Note as I said above no DMA actually performed if at least one MSI
> vector is enabled. The driver just needs a DWORD within the PCIe bus
> space for the MSI MWr TLPs target address and EP+vector data. The MSI
> TLP data is decoded by the iMSI-RX engine in order to set the
> corresponding flag in the MSI IRQ status register. Such TLPs won't be
> passed to the master AXI-bus.
>
>>
>>>
>>> If only we had a way to auto-detected the PCIe-bus space with no
>>> physical memory behind it and take out a DWORD from it to initialize
>>> the iMSI-RX engine we could have immediately got rid from the mask
>>> setting operation and the memory allocation. It would have solved your
>>> problem too.
>>
>
>> Yes, it would solve our issue too. I do not know, however, if a generic
>> solution is possible here, but I am no expert in PCIe.
>
> Currently the kernel PCIe subsystem doesn't provide a way to reserve a
> range within the PCIe bus memory with no physical RAM behind and left
> unused during the BARs resource initialization. Implementing such
> functionality (perhaps in the framework of the P2P module or based on
> it) would give the generic solution.
>
>>
>> For now, we are probably better off with SoC-specific patches, when we know
>> which PCIe devices can possibly be used and what their restrictions are.
>
> Since you know that there is no any RAM below 4GB and you have
> matching CPU and PCIe address spaces, then you can just take any
> address below 4GB and use it to initialize the MSI-target address
> (dw_pcie_rp.msi_data). But make sure that the peripheral PCIe-devices
> don't use it for something application-specific (like accessing CPU
> MMIO devices mapped to that base address). That seems like the most
> universal solution for your case.

Interesting idea!
Thank you, Sergey.

>
> -Serge(y)
>
>>
>>>
>>> -Serge(y)
>>>
>>>>
>>>> Looking forward to your comments.
>>>
>>>
>>>
>>>>
>>>>
>>>>> so you, or Will, or anyone
>>>>> else interested should be free to rework this on top of linux-next
>>>>> (although at this point, more realistically on top of 6.3-rc1 in a few
>>>>> weeks).
>>>>>
>>>>> Thanks,
>>>>> Robin.
>>>>>
>>>>>> Support for 64-bit MSI target addresses is needed for some of our SoCs.
>>>>>> I ran into a situation when there was no available RAM in ZONE_DMA32
>>>>>> during initialization of PCIe host. Hence, dmam_alloc_coherent() failed
>>>>>> in dw_pcie_msi_host_init() and initialization failed with -ENOMEM:
>>>>>>
>>>>>> [    0.374834] dw-pcie 4000000.pcie0: host bridge /soc/pcie0@4000000
>>>>>> ranges:
>>>>>> [    0.375813] dw-pcie 4000000.pcie0: MEM
>>>>>> 0x0041000000..0x004fffffff -> 0x0041000000
>>>>>> [    0.376171] dw-pcie 4000000.pcie0: IB MEM
>>>>>> 0x0400000000..0x07ffffffff -> 0x0400000000
>>>>>> [    0.377914] dw-pcie 4000000.pcie0: Failed to alloc and map MSI data
>>>>>> [    0.378191] dw-pcie 4000000.pcie0: Failed to initialize host
>>>>>> [    0.378255] dw-pcie: probe of 4000000.pcie0 failed with error -12
>>>>>>
>>>>>> Mainline kernel 6.2-rc6 was used in that test.
>>>>>>
>>>>>> The hardware supports 64-bit target addresses, so the patch "PCI: dwc:
>>>>>> Add support for 64-bit MSI target address" should help with this
>>>>>> particular failure.
>>>>>>
>>>>>>
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Will
>>>>>>>
>>>>>>> Will McVicker (2):
>>>>>>> PCI: dwc: Drop dependency on ZONE_DMA32
>>>>>>>
>>>>>>> v6:
>>>>>>> * Retrying DMA allocation with 64-bit mask within the error
>>>>>>> if-statement.
>>>>>>> * Use an int for the DMA mask instead of a bool and ternary operation.
>>>>>>>
>>>>>>> v5:
>>>>>>> * Updated patch 2/2 to first try with a 32-bit DMA mask. On failure,
>>>>>>> retry with a 64-bit mask if supported.
>>>>>>>
>>>>>>> v4:
>>>>>>> * Updated commit descriptions.
>>>>>>> * Renamed msi_64b -> msi_64bit.
>>>>>>> * Dropped msi_64bit ternary use.
>>>>>>> * Dropped export of dw_pcie_msi_capabilities.
>>>>>>>
>>>>>>> v3:
>>>>>>> * Switched to a managed DMA allocation.
>>>>>>> * Simplified the DMA allocation cleanup.
>>>>>>> * Dropped msi_page from struct dw_pcie_rp.
>>>>>>> * Allocating a u64 instead of a full page.
>>>>>>>
>>>>>>> v2:
>>>>>>> * Fixed build error caught by kernel test robot
>>>>>>> * Fixed error handling reported by Isaac Manjarres
>>>>>>> PCI: dwc: Add support for 64-bit MSI target address
>>>>>>>
>>>>>>> .../pci/controller/dwc/pcie-designware-host.c | 43 +++++++++----------
>>>>>>> drivers/pci/controller/dwc/pcie-designware.c |  8 ++++
>>>>>>> drivers/pci/controller/dwc/pcie-designware.h |  2 +-
>>>>>>> 3 files changed, 30 insertions(+), 23 deletions(-)
>>>>>>>
>>>>>>>
>>>>>>> base-commit: 568035b01cfb107af8d2e4bd2fb9aea22cf5b868
>>>>>>
>>>>>> Thank you in advance.
>>
>> Regards,
>> Evgenii
>>
>>
>



2023-02-09 15:55:04

by Robin Murphy

[permalink] [raw]
Subject: Re: [PATCH v6 0/2] PCI: dwc: Add support for 64-bit MSI target addresses

On 2023-02-09 08:03, Evgenii Shatokhin wrote:
[...]
>> Currently the kernel PCIe subsystem doesn't provide a way to reserve a
>> range within the PCIe bus memory with no physical RAM behind and left
>> unused during the BARs resource initialization. Implementing such
>> functionality (perhaps in the framework of the P2P module or based on
>> it) would give the generic solution.
>>
>>>
>>> For now, we are probably better off with SoC-specific patches, when
>>> we know
>>> which PCIe devices can possibly be used and what their restrictions are.
>>
>> Since you know that there is no any RAM below 4GB and you have
>> matching CPU and PCIe address spaces, then you can just take any
>> address below 4GB and use it to initialize the MSI-target address
>> (dw_pcie_rp.msi_data). But make sure that the peripheral PCIe-devices
>> don't use it for something application-specific (like accessing CPU
>> MMIO devices mapped to that base address). That seems like the most
>> universal solution for your case.
>
> Interesting idea!
> Thank you, Sergey.

Yes, if the platform-specific driver knows enough to be able to pick a
suitable "safe" MSI address, then I think allowing it to pre-set
pp.msi_data, and only falling back to the dma-mapping workaround
otherwise, sounds reasonable.

Thanks,
Robin.