From: Gerd Bayer <gbayer@linux.ibm.com>
To: "Nikolaou Alexandros (SO/PAF1-Mb)"
<Alexandros.Nikolaou@de.bosch.com>,
"D . Wythe" <alibuda@linux.alibaba.com>,
Wen Gu <guwen@linux.alibaba.com>,
Tony Lu <tonylu@linux.alibaba.com>,
Nils Hoppmann <niho@linux.ibm.com>
Cc: "linux-s390@vger.kernel.org" <linux-s390@vger.kernel.org>,
netdev <netdev@vger.kernel.org>,
Wenjia Zhang <wenjia@linux.ibm.com>,
Jan Karcher <jaka@linux.ibm.com>
Subject: Re: SMC-R throughput drops for specific message sizes
Date: Wed, 13 Dec 2023 13:17:47 +0100 [thread overview]
Message-ID: <2c460a84c6e725187dda05fc553981ce3022bb78.camel@linux.ibm.com> (raw)
In-Reply-To: <PAWPR10MB7270731C91544AEF25E0A33CC084A@PAWPR10MB7270.EURPRD10.PROD.OUTLOOK.COM>
Hi Nikolaou,
thank you for providing more details about your setup.
On Wed, 2023-12-06 at 15:28 +0000, Nikolaou Alexandros (SO/PAF1-Mb)
wrote:
> Dear Wenjia,
while Wenjia is out, I'm writing primarily to getting some more folks'
attention to this topic. Furthermore, I'm moving the discussion to the
netdev mailing list where SMC discussions usually take place.
> Thanks for getting back to me. Some further details on the
> experiments are:
>
> - The tests had been conducted on a one-to-one connection between two
> Mellanox-powered (mlx5, ConnectX-5) PCs.
> - Attached you may find the client log of the qperf output. You may
> notice that for the majority of message size values, the bandwidth is
> around 3.2GB/s which matches the maximum throughput of the
> mellanox NICs.
> According to a periodic regular pattern though, with the first
> occurring at a message size of 473616 – 522192 (with a step of
> 12144kB), the 3.2GB/s throughput drops substantially. The
> corresponding commands for these drops are
> server: smc_run qperf
> client: smc_run qperf -v -uu -H worker1 -m 473616 tcp_bw
> - Our smc version (3E92E1460DA96BE2B2DDC2F, smc-tools-1.2.2) does not
> provide us with the smcr info, smc_rnics -a and smcr -d
> stats commands. As an alternative, you may also find attached the
> output of ibv_devinfo -v.
> - Buffer size:
> sudo sysctl -w net.ipv4.tcp_rmem="4096 1048576 6291456"
> sudo sysctl -w net.ipv4.tcp_wmem="4096 1048576 6291456"
> - MTU size: 9000
>
> Should you require further information, please let me know.
Wenjia and I belong to a group of Linux on Z developers that maintains
the SMC protocol on s390 mainframe systems. Nils Hoppmann is our expert
for performance and might be able to shed some light on his experiences
with throughput drops for particular SMC message sizes. Our experience
is heavily biased towards IBM Z systems, though - with their distinct
cache and PCI root-complex hardware designs.
Over the last few years there's a group around D. Wythe, Wen Gu and
Tony Lu who adopted and extended the SMC protocol for use-cases on x86
architectures. I address them here explicitly, soliciting feedback on
their experiences.
All in all there are several moving parts involved here, that could
play a role:
- firmware level of your Mellanox/NVidia NICs,
- platform specific hardware designs re. cache and root-complexes,
interrupt distribution, ...
- exact code level of the device drivers and the SMC protocol
This is just a heads-up, that there may be requests to try things with
newer code levels ;)
Thank you,
Gerd
--
Gerd Bayer
Linux on IBM Z Development - IBM Germany R&D
next prev parent reply other threads:[~2023-12-13 12:17 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-12-01 13:33 SMC-R throughput drops for specific message sizes Nikolaou Alexandros (SO/PAF1-Mb)
2023-12-04 16:09 ` Wenjia Zhang
2023-12-06 15:28 ` Nikolaou Alexandros (SO/PAF1-Mb)
2023-12-13 12:17 ` Gerd Bayer [this message]
2023-12-13 13:38 ` Wen Gu
-- strict thread matches above, loose matches on Subject: below --
2023-12-13 15:52 Nikolaou Alexandros (SO/PAF1-Mb)
2024-02-01 13:50 Iordache Costin (XC-AS/EAE-UK)
2024-02-05 3:50 ` Wen Gu
2024-02-19 8:44 ` Wen Gu
2024-02-27 11:28 ` Iordache Costin (XC-AS/EAE-UK)
[not found] <GV2PR10MB8037B30A9D2CE67F267D5E61BB3B2@GV2PR10MB8037.EURPRD10.PROD.OUTLOOK.COM>
[not found] ` <GV2PR10MB80376BEB9EE8E03F98CC86A1BB3B2@GV2PR10MB8037.EURPRD10.PROD.OUTLOOK.COM>
2024-03-28 12:18 ` Goerlitz Andreas (SO/PAF1-Mb)
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=2c460a84c6e725187dda05fc553981ce3022bb78.camel@linux.ibm.com \
--to=gbayer@linux.ibm.com \
--cc=Alexandros.Nikolaou@de.bosch.com \
--cc=alibuda@linux.alibaba.com \
--cc=guwen@linux.alibaba.com \
--cc=jaka@linux.ibm.com \
--cc=linux-s390@vger.kernel.org \
--cc=netdev@vger.kernel.org \
--cc=niho@linux.ibm.com \
--cc=tonylu@linux.alibaba.com \
--cc=wenjia@linux.ibm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox