From: Tariq Toukan <tariqt@nvidia.com>
To: Eric Dumazet <edumazet@google.com>,
Jakub Kicinski <kuba@kernel.org>, Paolo Abeni <pabeni@redhat.com>,
Andrew Lunn <andrew+netdev@lunn.ch>,
"David S. Miller" <davem@davemloft.net>
Cc: Saeed Mahameed <saeedm@nvidia.com>,
Leon Romanovsky <leon@kernel.org>,
Tariq Toukan <tariqt@nvidia.com>, Mark Bloch <mbloch@nvidia.com>,
<netdev@vger.kernel.org>, <linux-rdma@vger.kernel.org>,
<linux-kernel@vger.kernel.org>, Gal Pressman <gal@nvidia.com>,
Nimrod Oren <noren@nvidia.com>,
Dragos Tatulea <dtatulea@nvidia.com>
Subject: [PATCH net-next 2/2] net/mlx5: add debugfs stats for frag buf dma pools
Date: Thu, 14 May 2026 13:49:25 +0300 [thread overview]
Message-ID: <20260514104925.337570-3-tariqt@nvidia.com> (raw)
In-Reply-To: <20260514104925.337570-1-tariqt@nvidia.com>
From: Nimrod Oren <noren@nvidia.com>
Add a debugfs file exposing per-node DMA pool usage for mlx5_frag_buf
allocations.
# cat /sys/kernel/debug/mlx5/<dev>/frag_buf_dma_pools
node block_size used_blocks allocated_blocks
0 4096 0 0
0 8192 0 0
0 16384 0 0
0 32768 0 0
0 65536 0 0
1 4096 0 0
1 8192 0 0
1 16384 0 0
1 32768 0 0
1 65536 0 0
Signed-off-by: Nimrod Oren <noren@nvidia.com>
Signed-off-by: Tariq Toukan <tariqt@nvidia.com>
---
.../net/ethernet/mellanox/mlx5/core/alloc.c | 86 +++++++++++++++++++
include/linux/mlx5/driver.h | 1 +
2 files changed, 87 insertions(+)
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/alloc.c b/drivers/net/ethernet/mellanox/mlx5/core/alloc.c
index 16d6b126a486..4fe9d7d4f143 100644
--- a/drivers/net/ethernet/mellanox/mlx5/core/alloc.c
+++ b/drivers/net/ethernet/mellanox/mlx5/core/alloc.c
@@ -38,6 +38,8 @@
#include <linux/dma-mapping.h>
#include <linux/vmalloc.h>
#include <linux/nodemask.h>
+#include <linux/debugfs.h>
+#include <linux/seq_file.h>
#include <linux/mlx5/driver.h>
#include "mlx5_core.h"
@@ -74,6 +76,13 @@ struct mlx5_frag_buf_node_pools {
struct mlx5_dma_pool *pools[MLX5_FRAG_BUF_POOLS_NUM];
};
+struct mlx5_dma_pool_stats {
+ int node;
+ size_t block_size;
+ size_t used_blocks;
+ size_t allocated_blocks;
+};
+
/* Handling for queue buffers -- we allocate a bunch of memory and
* register it in a memory region at HCA virtual address 0.
*/
@@ -225,6 +234,43 @@ static void mlx5_dma_pool_free(struct mlx5_dma_pool *pool,
mutex_unlock(&pool->lock);
}
+static void mlx5_dma_pool_debugfs_get_stats(struct mlx5_dma_pool *pool,
+ struct mlx5_dma_pool_stats *stats)
+{
+ int blocks_per_page = BIT(PAGE_SHIFT - pool->block_shift);
+ struct mlx5_dma_pool_page *page;
+ size_t free_blocks = 0;
+ size_t pages = 0;
+
+ mutex_lock(&pool->lock);
+ list_for_each_entry(page, &pool->page_list, pool_link) {
+ pages++;
+ free_blocks += bitmap_weight(page->bitmap, blocks_per_page);
+ }
+ mutex_unlock(&pool->lock);
+
+ stats->node = pool->node;
+ stats->block_size = BIT(pool->block_shift);
+ stats->allocated_blocks = pages * blocks_per_page;
+ stats->used_blocks = stats->allocated_blocks - free_blocks;
+}
+
+static void mlx5_dma_pool_debugfs_stats_print(struct seq_file *file,
+ struct mlx5_dma_pool *pool)
+{
+ struct mlx5_dma_pool_stats stats = {};
+
+ mlx5_dma_pool_debugfs_get_stats(pool, &stats);
+ seq_printf(file, "%4d %5zu %7zu %7zu\n",
+ stats.node, stats.block_size, stats.used_blocks,
+ stats.allocated_blocks);
+}
+
+static void mlx5_dma_pools_debugfs_print_header(struct seq_file *file)
+{
+ seq_puts(file, "node block_size used_blocks allocated_blocks\n");
+}
+
static void
mlx5_frag_buf_node_pools_destroy(struct mlx5_frag_buf_node_pools *node_pools)
{
@@ -257,11 +303,46 @@ mlx5_frag_buf_node_pools_create(struct mlx5_core_dev *dev, int node)
return node_pools;
}
+static int
+mlx5_frag_buf_dma_pools_debugfs_show(struct seq_file *file, void *priv)
+{
+ struct mlx5_core_dev *dev = file->private;
+ int node;
+
+ mlx5_dma_pools_debugfs_print_header(file);
+
+ if (!dev->priv.frag_buf_node_pools)
+ return 0;
+
+ for_each_node_state(node, N_POSSIBLE) {
+ struct mlx5_frag_buf_node_pools *node_pools;
+
+ node_pools = dev->priv.frag_buf_node_pools[node];
+ if (!node_pools)
+ continue;
+
+ for (int i = 0; i < MLX5_FRAG_BUF_POOLS_NUM; i++) {
+ struct mlx5_dma_pool *pool = node_pools->pools[i];
+
+ if (!pool)
+ continue;
+
+ mlx5_dma_pool_debugfs_stats_print(file, pool);
+ }
+ }
+
+ return 0;
+}
+DEFINE_SHOW_ATTRIBUTE(mlx5_frag_buf_dma_pools_debugfs);
+
void mlx5_frag_buf_pools_cleanup(struct mlx5_core_dev *dev)
{
struct mlx5_priv *priv = &dev->priv;
int node;
+ debugfs_remove(priv->dbg.frag_buf_dma_pools_debugfs);
+ priv->dbg.frag_buf_dma_pools_debugfs = NULL;
+
for_each_node_state(node, N_POSSIBLE) {
struct mlx5_frag_buf_node_pools *node_pools;
@@ -296,6 +377,11 @@ int mlx5_frag_buf_pools_init(struct mlx5_core_dev *dev)
priv->frag_buf_node_pools[node] = node_pools;
}
+ priv->dbg.frag_buf_dma_pools_debugfs =
+ debugfs_create_file("frag_buf_dma_pools", 0444,
+ priv->dbg.dbg_root, dev,
+ &mlx5_frag_buf_dma_pools_debugfs_fops);
+
return 0;
}
diff --git a/include/linux/mlx5/driver.h b/include/linux/mlx5/driver.h
index 8b4d384125d1..9a4bb25d8e0a 100644
--- a/include/linux/mlx5/driver.h
+++ b/include/linux/mlx5/driver.h
@@ -547,6 +547,7 @@ struct mlx5_debugfs_entries {
struct dentry *eq_debugfs;
struct dentry *cq_debugfs;
struct dentry *cmdif_debugfs;
+ struct dentry *frag_buf_dma_pools_debugfs;
struct dentry *pages_debugfs;
struct dentry *lag_debugfs;
};
--
2.44.0
prev parent reply other threads:[~2026-05-14 10:50 UTC|newest]
Thread overview: 3+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-05-14 10:49 [PATCH net-next 0/2] net/mlx5: frag buffer improvements Tariq Toukan
2026-05-14 10:49 ` [PATCH net-next 1/2] net/mlx5: use numa_mem_id() for default frag buf allocations Tariq Toukan
2026-05-14 10:49 ` Tariq Toukan [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20260514104925.337570-3-tariqt@nvidia.com \
--to=tariqt@nvidia.com \
--cc=andrew+netdev@lunn.ch \
--cc=davem@davemloft.net \
--cc=dtatulea@nvidia.com \
--cc=edumazet@google.com \
--cc=gal@nvidia.com \
--cc=kuba@kernel.org \
--cc=leon@kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-rdma@vger.kernel.org \
--cc=mbloch@nvidia.com \
--cc=netdev@vger.kernel.org \
--cc=noren@nvidia.com \
--cc=pabeni@redhat.com \
--cc=saeedm@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox