From: Chiara Meiohas <cmeiohas@nvidia.com>
To: <leon@kernel.org>, <dsahern@gmail.com>, <stephen@networkplumber.org>
Cc: <michaelgur@nvidia.com>, <jgg@nvidia.com>,
<linux-rdma@vger.kernel.org>, <netdev@vger.kernel.org>,
Chiara Meiohas <cmeiohas@nvidia.com>
Subject: [PATCH iproute2-next] rdma: Align FRMR pool UAPI names with merged kernel UAPI
Date: Thu, 7 May 2026 21:46:09 +0300 [thread overview]
Message-ID: <20260507184609.3439875-1-cmeiohas@nvidia.com> (raw)
From: Michael Guralnik <michaelgur@nvidia.com>
The FRMR pools UAPI merged in kernel v7.0-rc1 commit dbd0472fd7a5
("RDMA/nldev: Expose kernel-internal FRMR pools in netlink")
uses different identifier names than what the iproute2 FRMR pools
series was developed against.
Update the vendored copy of RDMA UAPI and all references in the rdma
tool to match the names that actually shipped in the kernel.
Fixes: 93368ee34528 ("rdma: Update headers")
Signed-off-by: Michael Guralnik <michaelgur@nvidia.com>
Signed-off-by: Chiara Meiohas <cmeiohas@nvidia.com>
---
rdma/include/uapi/rdma/rdma_netlink.h | 30 ++++----
rdma/res-frmr-pools.c | 98 +++++++++++++--------------
rdma/res.h | 2 +-
3 files changed, 65 insertions(+), 65 deletions(-)
diff --git a/rdma/include/uapi/rdma/rdma_netlink.h b/rdma/include/uapi/rdma/rdma_netlink.h
index 8709e558b..4356ec4a1 100644
--- a/rdma/include/uapi/rdma/rdma_netlink.h
+++ b/rdma/include/uapi/rdma/rdma_netlink.h
@@ -308,9 +308,9 @@ enum rdma_nldev_command {
RDMA_NLDEV_CMD_MONITOR,
- RDMA_NLDEV_CMD_RES_FRMR_POOLS_GET, /* can dump */
+ RDMA_NLDEV_CMD_FRMR_POOLS_GET, /* can dump */
- RDMA_NLDEV_CMD_RES_FRMR_POOLS_SET,
+ RDMA_NLDEV_CMD_FRMR_POOLS_SET,
RDMA_NLDEV_NUM_OPS
};
@@ -590,19 +590,19 @@ enum rdma_nldev_attr {
/*
* FRMR Pools attributes
*/
- RDMA_NLDEV_ATTR_RES_FRMR_POOLS, /* nested table */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_ENTRY, /* nested table */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY, /* nested table */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ATS, /* u8 */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ACCESS_FLAGS, /* u32 */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_VENDOR_KEY, /* u64 */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_NUM_DMA_BLOCKS, /* u64 */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_QUEUE_HANDLES, /* u32 */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_MAX_IN_USE, /* u64 */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_IN_USE, /* u64 */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_AGING_PERIOD, /* u32 */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_PINNED, /* u32 */
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_KERNEL_VENDOR_KEY, /* u64 */
+ RDMA_NLDEV_ATTR_FRMR_POOLS, /* nested table */
+ RDMA_NLDEV_ATTR_FRMR_POOL_ENTRY, /* nested table */
+ RDMA_NLDEV_ATTR_FRMR_POOL_KEY, /* nested table */
+ RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ATS, /* u8 */
+ RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ACCESS_FLAGS, /* u32 */
+ RDMA_NLDEV_ATTR_FRMR_POOL_KEY_VENDOR_KEY, /* u64 */
+ RDMA_NLDEV_ATTR_FRMR_POOL_KEY_NUM_DMA_BLOCKS, /* u64 */
+ RDMA_NLDEV_ATTR_FRMR_POOL_QUEUE_HANDLES, /* u32 */
+ RDMA_NLDEV_ATTR_FRMR_POOL_MAX_IN_USE, /* u64 */
+ RDMA_NLDEV_ATTR_FRMR_POOL_IN_USE, /* u64 */
+ RDMA_NLDEV_ATTR_FRMR_POOLS_AGING_PERIOD, /* u32 */
+ RDMA_NLDEV_ATTR_FRMR_POOL_PINNED_HANDLES, /* u32 */
+ RDMA_NLDEV_ATTR_FRMR_POOL_KEY_KERNEL_VENDOR_KEY, /* u64 */
/*
* Always the end
diff --git a/rdma/res-frmr-pools.c b/rdma/res-frmr-pools.c
index abcd21884..d5faa5c14 100644
--- a/rdma/res-frmr-pools.c
+++ b/rdma/res-frmr-pools.c
@@ -80,83 +80,83 @@ static int res_frmr_pools_line(struct rd *rd, const char *name, int idx,
char key_str[FRMR_POOL_KEY_MAX_LEN];
struct frmr_pool_key key = { 0 };
- if (nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY]) {
+ if (nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_KEY]) {
if (mnl_attr_parse_nested(
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY],
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_KEY],
rd_attr_cb, key_tb) != MNL_CB_OK)
return MNL_CB_ERROR;
- if (key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ATS])
+ if (key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ATS])
key.ats = mnl_attr_get_u8(
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ATS]);
- if (key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ACCESS_FLAGS])
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ATS]);
+ if (key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ACCESS_FLAGS])
key.access_flags = mnl_attr_get_u32(
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ACCESS_FLAGS]);
- if (key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_VENDOR_KEY])
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ACCESS_FLAGS]);
+ if (key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_VENDOR_KEY])
key.vendor_key = mnl_attr_get_u64(
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_VENDOR_KEY]);
- if (key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_NUM_DMA_BLOCKS])
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_VENDOR_KEY]);
+ if (key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_NUM_DMA_BLOCKS])
key.num_dma_blocks = mnl_attr_get_u64(
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_NUM_DMA_BLOCKS]);
- if (key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_KERNEL_VENDOR_KEY])
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_NUM_DMA_BLOCKS]);
+ if (key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_KERNEL_VENDOR_KEY])
kernel_vendor_key = mnl_attr_get_u64(
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_KERNEL_VENDOR_KEY]);
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_KERNEL_VENDOR_KEY]);
if (rd_is_filtered_attr(
rd, "ats", key.ats,
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ATS]))
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ATS]))
goto out;
if (rd_is_filtered_attr(
rd, "access_flags", key.access_flags,
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ACCESS_FLAGS]))
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ACCESS_FLAGS]))
goto out;
if (rd_is_filtered_attr(
rd, "vendor_key", key.vendor_key,
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_VENDOR_KEY]))
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_VENDOR_KEY]))
goto out;
if (rd_is_filtered_attr(
rd, "num_dma_blocks", key.num_dma_blocks,
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_NUM_DMA_BLOCKS]))
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_NUM_DMA_BLOCKS]))
goto out;
}
- if (nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_QUEUE_HANDLES])
+ if (nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_QUEUE_HANDLES])
queue_handles = mnl_attr_get_u32(
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_QUEUE_HANDLES]);
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_QUEUE_HANDLES]);
if (rd_is_filtered_attr(
rd, "queue", queue_handles,
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_QUEUE_HANDLES]))
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_QUEUE_HANDLES]))
goto out;
- if (nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_IN_USE])
+ if (nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_IN_USE])
in_use = mnl_attr_get_u64(
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_IN_USE]);
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_IN_USE]);
if (rd_is_filtered_attr(rd, "in_use", in_use,
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_IN_USE]))
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_IN_USE]))
goto out;
- if (nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_MAX_IN_USE])
+ if (nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_MAX_IN_USE])
max_in_use = mnl_attr_get_u64(
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_MAX_IN_USE]);
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_MAX_IN_USE]);
if (rd_is_filtered_attr(
rd, "max_in_use", max_in_use,
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_MAX_IN_USE]))
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_MAX_IN_USE]))
goto out;
- if (nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_PINNED])
+ if (nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_PINNED_HANDLES])
pinned_handles = mnl_attr_get_u32(
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_PINNED]);
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_PINNED_HANDLES]);
if (rd_is_filtered_attr(rd, "pinned", pinned_handles,
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_PINNED]))
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_PINNED_HANDLES]))
goto out;
open_json_object(NULL);
print_dev(idx, name);
- if (nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY]) {
+ if (nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_KEY]) {
snprintf(key_str, sizeof(key_str),
"%" PRIx64 ":%" PRIx64 ":%x:%s",
key.vendor_key, key.num_dma_blocks,
@@ -166,30 +166,30 @@ static int res_frmr_pools_line(struct rd *rd, const char *name, int idx,
if (rd->show_details) {
res_print_u32(
"ats", key.ats,
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ATS]);
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ATS]);
res_print_u32(
"access_flags", key.access_flags,
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ACCESS_FLAGS]);
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ACCESS_FLAGS]);
res_print_u64(
"vendor_key", key.vendor_key,
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_VENDOR_KEY]);
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_VENDOR_KEY]);
res_print_u64(
"num_dma_blocks", key.num_dma_blocks,
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_NUM_DMA_BLOCKS]);
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_NUM_DMA_BLOCKS]);
res_print_u64(
"kernel_vendor_key", kernel_vendor_key,
- key_tb[RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_KERNEL_VENDOR_KEY]);
+ key_tb[RDMA_NLDEV_ATTR_FRMR_POOL_KEY_KERNEL_VENDOR_KEY]);
}
}
res_print_u32("queue", queue_handles,
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_QUEUE_HANDLES]);
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_QUEUE_HANDLES]);
res_print_u64("in_use", in_use,
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_IN_USE]);
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_IN_USE]);
res_print_u64("max_in_use", max_in_use,
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_MAX_IN_USE]);
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_MAX_IN_USE]);
res_print_u32("pinned", pinned_handles,
- nla_line[RDMA_NLDEV_ATTR_RES_FRMR_POOL_PINNED]);
+ nla_line[RDMA_NLDEV_ATTR_FRMR_POOL_PINNED_HANDLES]);
print_driver_table(rd, nla_line[RDMA_NLDEV_ATTR_DRIVER]);
close_json_object();
@@ -215,12 +215,12 @@ int res_frmr_pools_parse_cb(const struct nlmsghdr *nlh, void *data)
mnl_attr_parse(nlh, 0, rd_attr_cb, tb);
if (!tb[RDMA_NLDEV_ATTR_DEV_INDEX] || !tb[RDMA_NLDEV_ATTR_DEV_NAME] ||
- !tb[RDMA_NLDEV_ATTR_RES_FRMR_POOLS])
+ !tb[RDMA_NLDEV_ATTR_FRMR_POOLS])
return MNL_CB_ERROR;
name = mnl_attr_get_str(tb[RDMA_NLDEV_ATTR_DEV_NAME]);
idx = mnl_attr_get_u32(tb[RDMA_NLDEV_ATTR_DEV_INDEX]);
- nla_table = tb[RDMA_NLDEV_ATTR_RES_FRMR_POOLS];
+ nla_table = tb[RDMA_NLDEV_ATTR_FRMR_POOLS];
mnl_attr_for_each_nested(nla_entry, nla_table) {
struct nlattr *nla_line[RDMA_NLDEV_ATTR_MAX] = {};
@@ -256,10 +256,10 @@ static int res_frmr_pools_one_set_aging(struct rd *rd)
return -EINVAL;
}
- rd_prepare_msg(rd, RDMA_NLDEV_CMD_RES_FRMR_POOLS_SET, &seq,
+ rd_prepare_msg(rd, RDMA_NLDEV_CMD_FRMR_POOLS_SET, &seq,
(NLM_F_REQUEST | NLM_F_ACK));
mnl_attr_put_u32(rd->nlh, RDMA_NLDEV_ATTR_DEV_INDEX, rd->dev_idx);
- mnl_attr_put_u32(rd->nlh, RDMA_NLDEV_ATTR_RES_FRMR_POOL_AGING_PERIOD,
+ mnl_attr_put_u32(rd->nlh, RDMA_NLDEV_ATTR_FRMR_POOLS_AGING_PERIOD,
aging_period);
return rd_sendrecv_msg(rd, seq);
@@ -294,24 +294,24 @@ static int res_frmr_pools_one_set_pinned(struct rd *rd)
return -EINVAL;
}
- rd_prepare_msg(rd, RDMA_NLDEV_CMD_RES_FRMR_POOLS_SET, &seq,
+ rd_prepare_msg(rd, RDMA_NLDEV_CMD_FRMR_POOLS_SET, &seq,
(NLM_F_REQUEST | NLM_F_ACK));
mnl_attr_put_u32(rd->nlh, RDMA_NLDEV_ATTR_DEV_INDEX, rd->dev_idx);
- mnl_attr_put_u32(rd->nlh, RDMA_NLDEV_ATTR_RES_FRMR_POOL_PINNED,
+ mnl_attr_put_u32(rd->nlh, RDMA_NLDEV_ATTR_FRMR_POOL_PINNED_HANDLES,
pinned_value);
key_attr =
- mnl_attr_nest_start(rd->nlh, RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY);
- mnl_attr_put_u8(rd->nlh, RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ATS,
+ mnl_attr_nest_start(rd->nlh, RDMA_NLDEV_ATTR_FRMR_POOL_KEY);
+ mnl_attr_put_u8(rd->nlh, RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ATS,
pool_key.ats);
mnl_attr_put_u32(rd->nlh,
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_ACCESS_FLAGS,
+ RDMA_NLDEV_ATTR_FRMR_POOL_KEY_ACCESS_FLAGS,
pool_key.access_flags);
- mnl_attr_put_u64(rd->nlh, RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_VENDOR_KEY,
+ mnl_attr_put_u64(rd->nlh, RDMA_NLDEV_ATTR_FRMR_POOL_KEY_VENDOR_KEY,
pool_key.vendor_key);
mnl_attr_put_u64(rd->nlh,
- RDMA_NLDEV_ATTR_RES_FRMR_POOL_KEY_NUM_DMA_BLOCKS,
+ RDMA_NLDEV_ATTR_FRMR_POOL_KEY_NUM_DMA_BLOCKS,
pool_key.num_dma_blocks);
mnl_attr_nest_end(rd->nlh, key_attr);
diff --git a/rdma/res.h b/rdma/res.h
index 8d7b4a0bf..1f71115b9 100644
--- a/rdma/res.h
+++ b/rdma/res.h
@@ -200,7 +200,7 @@ struct filters frmr_pools_valid_filters[MAX_NUMBER_OF_FILTERS] = {
{ .name = "pinned", .is_number = true },
};
-RES_FUNC(res_frmr_pools, RDMA_NLDEV_CMD_RES_FRMR_POOLS_GET,
+RES_FUNC(res_frmr_pools, RDMA_NLDEV_CMD_FRMR_POOLS_GET,
frmr_pools_valid_filters, true, 0);
int res_frmr_pools_set(struct rd *rd);
--
2.38.1
next reply other threads:[~2026-05-07 18:47 UTC|newest]
Thread overview: 2+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-05-07 18:46 Chiara Meiohas [this message]
2026-05-07 20:50 ` [PATCH iproute2-next] rdma: Align FRMR pool UAPI names with merged kernel UAPI patchwork-bot+netdevbpf
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20260507184609.3439875-1-cmeiohas@nvidia.com \
--to=cmeiohas@nvidia.com \
--cc=dsahern@gmail.com \
--cc=jgg@nvidia.com \
--cc=leon@kernel.org \
--cc=linux-rdma@vger.kernel.org \
--cc=michaelgur@nvidia.com \
--cc=netdev@vger.kernel.org \
--cc=stephen@networkplumber.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox