From: Chris Arges <carges@cloudflare.com>
To: Jesper Dangaard Brouer <hawk@kernel.org>
Cc: Dragos Tatulea <dtatulea@nvidia.com>,
Dipayaan Roy <dipayanroy@linux.microsoft.com>,
horms@kernel.org, kuba@kernel.org, kys@microsoft.com,
haiyangz@microsoft.com, wei.liu@kernel.org, decui@microsoft.com,
andrew+netdev@lunn.ch, davem@davemloft.net, edumazet@google.com,
pabeni@redhat.com, longli@microsoft.com, kotaranov@microsoft.com,
ast@kernel.org, daniel@iogearbox.net, john.fastabend@gmail.com,
sdf@fomichev.me, lorenzo@kernel.org, michal.kubiak@intel.com,
ernis@linux.microsoft.com, shradhagupta@linux.microsoft.com,
shirazsaleem@microsoft.com, rosenp@gmail.com,
netdev@vger.kernel.org, linux-hyperv@vger.kernel.org,
linux-rdma@vger.kernel.org, bpf@vger.kernel.org,
linux-kernel@vger.kernel.org, ssengar@linux.microsoft.com,
dipayanroy@microsoft.com,
kernel-team <kernel-team@cloudflare.com>,
Tariq Toukan <tariqt@nvidia.com>,
Saeed Mahameed <saeedm@nvidia.com>,
Yunsheng Lin <linyunsheng@huawei.com>
Subject: Re: [PATCH v2] net: mana: Use page pool fragments for RX buffers instead of full pages to improve memory efficiency.
Date: Thu, 31 Jul 2025 11:49:40 -0500 [thread overview]
Message-ID: <aIuepME92Q9iR22Z@861G6M3> (raw)
In-Reply-To: <01c9284d-58c2-4a90-8833-67439a28e541@kernel.org>
On 2025-07-31 18:36:04, Jesper Dangaard Brouer wrote:
>
>
> On 30/07/2025 09.31, Dragos Tatulea wrote:
> > On Tue, Jul 29, 2025 at 01:20:07PM -0700, Dipayaan Roy wrote:
> > > On Tue, Jul 29, 2025 at 12:15:23PM +0200, Jesper Dangaard Brouer wrote:
> > > >
> > > >
> > > > On 23/07/2025 21.07, Dipayaan Roy wrote:
> > > > > This patch enhances RX buffer handling in the mana driver by allocating
> > > > > pages from a page pool and slicing them into MTU-sized fragments, rather
> > > > > than dedicating a full page per packet. This approach is especially
> > > > > beneficial on systems with large page sizes like 64KB.
> > > > >
> > > > > Key improvements:
> > > > >
> > > > > - Proper integration of page pool for RX buffer allocations.
> > > > > - MTU-sized buffer slicing to improve memory utilization.
> > > > > - Reduce overall per Rx queue memory footprint.
> > > > > - Automatic fallback to full-page buffers when:
> > > > > * Jumbo frames are enabled (MTU > PAGE_SIZE / 2).
> > > > > * The XDP path is active, to avoid complexities with fragment reuse.
> > > > > - Removal of redundant pre-allocated RX buffers used in scenarios like MTU
> > > > > changes, ensuring consistency in RX buffer allocation.
> > > > >
> > > > > Testing on VMs with 64KB pages shows around 200% throughput improvement.
> > > > > Memory efficiency is significantly improved due to reduced wastage in page
> > > > > allocations. Example: We are now able to fit 35 rx buffers in a single 64kb
> > > > > page for MTU size of 1500, instead of 1 rx buffer per page previously.
> > > > >
> > > > > Tested:
> > > > >
> > > > > - iperf3, iperf2, and nttcp benchmarks.
> > > > > - Jumbo frames with MTU 9000.
> > > > > - Native XDP programs (XDP_PASS, XDP_DROP, XDP_TX, XDP_REDIRECT) for
> > > > > testing the XDP path in driver.
> > > > > - Page leak detection (kmemleak).
> > > > > - Driver load/unload, reboot, and stress scenarios.
> > > >
> > > > Chris (Cc) discovered a crash/bug[1] with page pool fragments used
> > > > from the mlx5 driver.
> > > > He put together a BPF program that reproduces the issue here:
> > > > - [2] https://github.com/arges/xdp-redirector
> > > >
> > > > Can I ask you to test that your driver against this reproducer?
> > > >
> > > >
> > > > [1] https://lore.kernel.org/all/aIEuZy6fUj_4wtQ6@861G6M3/
> > > >
> > > > --Jesper
> > > >
> > >
> > > Hi Jesper,
> > >
> > > I was unable to reproduce this issue on mana driver.
> > >
> > Please note that I had to make a few adjustments to get reprodduction on
> > mlx5:
> >
> > - Make sure that the veth MACs are recognized by the device. Otherwise
> > traffic might be dropped by the device.
> >
> > - Enable GRO on the veth device. Otherwise packets get dropped before
> > they reach the devmap BPF program.
> >
> > Try starting the test program with one thread and see if you see packets
> > coming through veth1-ns1 end of the veth pair.
> >
>
> Hi Dipayaan,
>
> Enabling GRO on the veth device is quite important for the test to be valid.
>
> I've asked Chris to fix this in the reproducer. He can report back when
> he have done this, so you can re-run the test. It is also good advice
> from Dragos that you should check packets are coming through the veth
> pair, to make sure the test is working.
>
> The setup.sh script also need to be modified, as it is loading xdp on a
> net_device called "ext0" [0], which is specific to our systems (which
> default also have GRO enabled for veth).
>
> [0] https://github.com/arges/xdp-redirector/blob/main/setup.sh#L28
>
> --Jesper
I pushed some updates to the setup script to make it easier to use. If you have
issues running the script, please share the output.
--chris
prev parent reply other threads:[~2025-07-31 16:49 UTC|newest]
Thread overview: 9+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-07-23 19:07 [PATCH v2] net: mana: Use page pool fragments for RX buffers instead of full pages to improve memory efficiency Dipayaan Roy
2025-07-23 19:55 ` Haiyang Zhang
2025-07-26 0:54 ` Jakub Kicinski
2025-07-28 18:20 ` Dipayaan Roy
2025-07-29 10:15 ` Jesper Dangaard Brouer
2025-07-29 20:20 ` Dipayaan Roy
2025-07-30 7:31 ` Dragos Tatulea
2025-07-31 16:36 ` Jesper Dangaard Brouer
2025-07-31 16:49 ` Chris Arges [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aIuepME92Q9iR22Z@861G6M3 \
--to=carges@cloudflare.com \
--cc=andrew+netdev@lunn.ch \
--cc=ast@kernel.org \
--cc=bpf@vger.kernel.org \
--cc=daniel@iogearbox.net \
--cc=davem@davemloft.net \
--cc=decui@microsoft.com \
--cc=dipayanroy@linux.microsoft.com \
--cc=dipayanroy@microsoft.com \
--cc=dtatulea@nvidia.com \
--cc=edumazet@google.com \
--cc=ernis@linux.microsoft.com \
--cc=haiyangz@microsoft.com \
--cc=hawk@kernel.org \
--cc=horms@kernel.org \
--cc=john.fastabend@gmail.com \
--cc=kernel-team@cloudflare.com \
--cc=kotaranov@microsoft.com \
--cc=kuba@kernel.org \
--cc=kys@microsoft.com \
--cc=linux-hyperv@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-rdma@vger.kernel.org \
--cc=linyunsheng@huawei.com \
--cc=longli@microsoft.com \
--cc=lorenzo@kernel.org \
--cc=michal.kubiak@intel.com \
--cc=netdev@vger.kernel.org \
--cc=pabeni@redhat.com \
--cc=rosenp@gmail.com \
--cc=saeedm@nvidia.com \
--cc=sdf@fomichev.me \
--cc=shirazsaleem@microsoft.com \
--cc=shradhagupta@linux.microsoft.com \
--cc=ssengar@linux.microsoft.com \
--cc=tariqt@nvidia.com \
--cc=wei.liu@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox