From: Niklas Cassel <cassel@kernel.org>
To: Hans Zhang <18255117159@163.com>
Cc: lpieralisi@kernel.org, kw@linux.com, bhelgaas@google.com,
heiko@sntech.de, thomas.petazzoni@bootlin.com,
manivannan.sadhasivam@linaro.org, yue.wang@amlogic.com,
pali@kernel.org, neil.armstrong@linaro.org, robh@kernel.org,
jingoohan1@gmail.com, khilman@baylibre.com, jbrunet@baylibre.com,
martin.blumenstingl@googlemail.com, linux-pci@vger.kernel.org,
linux-kernel@vger.kernel.org,
linux-arm-kernel@lists.infradead.org,
linux-amlogic@lists.infradead.org,
linux-rockchip@lists.infradead.org
Subject: Re: [PATCH v2 1/2] PCI: Configure root port MPS to hardware maximum during host probing
Date: Fri, 25 Apr 2025 15:47:10 +0200 [thread overview]
Message-ID: <aAuSXhmRiKQabjLO@ryzen> (raw)
In-Reply-To: <a4963173-dd9a-4341-b7f9-5fdb9485233a@163.com>
Hello Hans,
On Fri, Apr 25, 2025 at 06:56:53PM +0800, Hans Zhang wrote:
>
> But I discovered a problem:
>
> 0001:90:00.0 PCI bridge: Device 1f6c:0001 (prog-if 00 [Normal decode])
> ......
> Capabilities: [c0] Express (v2) Root Port (Slot-), MSI 00
> DevCap: MaxPayload 512 bytes, PhantFunc 0
> ExtTag- RBE+
> DevCtl: CorrErr+ NonFatalErr+ FatalErr+ UnsupReq+
> RlxdOrd+ ExtTag- PhantFunc- AuxPwr- NoSnoop+
> MaxPayload 512 bytes, MaxReadReq 1024 bytes
>
>
>
> Should the DevCtl MaxPayload be 256B?
>
> But I tested that the file reading and writing were normal. Is the display
> of 512B here what we expected?
>
> Root Port 0003:30:00.0 has the same problem. May I ask what your opinion is?
>
>
> ......
> 0001:91:00.0 Non-Volatile memory controller: Samsung Electronics Co Ltd
> NVMe SSD Controller PM9A1/PM9A3/980PRO (prog-if 02 [NVM Express])
> ......
> Capabilities: [70] Express (v2) Endpoint, MSI 00
> DevCap: MaxPayload 256 bytes, PhantFunc 0, Latency L0s
> unlimited, L1 unlimited
> ExtTag+ AttnBtn- AttnInd- PwrInd- RBE+ FLReset+
> SlotPowerLimit 0W
> DevCtl: CorrErr+ NonFatalErr+ FatalErr+ UnsupReq+
> RlxdOrd+ ExtTag+ PhantFunc- AuxPwr- NoSnoop+
> FLReset-
> MaxPayload 256 bytes, MaxReadReq 512 bytes
> ......
Here we see that the bridge has a higher DevCtl.MPS than the DevCap.MPS of
the endpoint.
Let me quote Bjorn from the previous mail thread:
"""
- I don't think it's safe to set MPS higher in all cases. If we set
the Root Port MPS=256, and an Endpoint only supports MPS=128, the
Endpoint may do a 256-byte DMA read (assuming its MRRS>=256). In
that case the RP may respond with a 256-byte payload the Endpoint
can't handle.
"""
I think the problem with this patch is that pcie_write_mps() call in
pci_host_probe() is done after the pci_scan_root_bus_bridge() call in
pci_host_probe().
So pci_configure_mps() (called by pci_configure_device()),
which does the limiting of the bus to what the endpoint supports,
is actually called before the pcie_write_mps() call added by this patch
(which increases DevCtl.MPS for the bridge).
So I think the code added in this patch needs to be executed before
pci_configure_device() is done for the EP.
It appears that pci_configure_device() is called for each device
during scan, first for the bridges and then for the EPs.
So I think something like this should work (totally untested):
--- a/drivers/pci/probe.c
+++ b/drivers/pci/probe.c
@@ -45,6 +45,8 @@ struct pci_domain_busn_res {
int domain_nr;
};
+static void pcie_write_mps(struct pci_dev *dev, int mps);
+
static struct resource *get_pci_domain_busn_res(int domain_nr)
{
struct pci_domain_busn_res *r;
@@ -2178,6 +2180,11 @@ static void pci_configure_mps(struct pci_dev *dev)
return;
}
+ if (pci_pcie_type(dev) == PCI_EXP_TYPE_ROOT_PORT &&
+ pcie_bus_config != PCIE_BUS_TUNE_OFF) {
+ pcie_write_mps(dev, 128 << dev->pcie_mpss);
+ }
+
if (!bridge || !pci_is_pcie(bridge))
return;
But we would probably need to move some code to avoid the
forward declaration.
Kind regards,
Niklas
next prev parent reply other threads:[~2025-04-25 15:33 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-04-25 9:57 [PATCH v2 0/2] PCI: Configure max payload size on pci_host_probe Hans Zhang
2025-04-25 9:57 ` [PATCH v2 1/2] PCI: Configure root port MPS to hardware maximum during host probing Hans Zhang
2025-04-25 10:23 ` Niklas Cassel
2025-04-25 10:56 ` Hans Zhang
2025-04-25 13:47 ` Niklas Cassel [this message]
2025-04-25 14:17 ` Hans Zhang
2025-04-25 9:57 ` [PATCH v2 2/2] PCI: Remove redundant MPS configuration Hans Zhang
2025-04-25 10:17 ` Niklas Cassel
2025-04-25 10:26 ` Hans Zhang
2025-04-25 11:59 ` neil.armstrong
2025-04-25 14:20 ` Hans Zhang
2025-04-25 18:13 ` Pali Rohár
2025-04-26 15:02 ` Hans Zhang
2025-04-26 15:06 ` Pali Rohár
2025-04-26 15:20 ` Hans Zhang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aAuSXhmRiKQabjLO@ryzen \
--to=cassel@kernel.org \
--cc=18255117159@163.com \
--cc=bhelgaas@google.com \
--cc=heiko@sntech.de \
--cc=jbrunet@baylibre.com \
--cc=jingoohan1@gmail.com \
--cc=khilman@baylibre.com \
--cc=kw@linux.com \
--cc=linux-amlogic@lists.infradead.org \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-pci@vger.kernel.org \
--cc=linux-rockchip@lists.infradead.org \
--cc=lpieralisi@kernel.org \
--cc=manivannan.sadhasivam@linaro.org \
--cc=martin.blumenstingl@googlemail.com \
--cc=neil.armstrong@linaro.org \
--cc=pali@kernel.org \
--cc=robh@kernel.org \
--cc=thomas.petazzoni@bootlin.com \
--cc=yue.wang@amlogic.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).