netdev.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Stanislav Fomichev <stfomichev@gmail.com>
To: Daniel Borkmann <daniel@iogearbox.net>
Cc: netdev@vger.kernel.org, bpf@vger.kernel.org, kuba@kernel.org,
	davem@davemloft.net, razor@blackwall.org, pabeni@redhat.com,
	willemb@google.com, sdf@fomichev.me, john.fastabend@gmail.com,
	martin.lau@kernel.org, jordan@jrife.io,
	maciej.fijalkowski@intel.com, magnus.karlsson@intel.com,
	David Wei <dw@davidwei.uk>
Subject: Re: [PATCH net-next 13/20] xsk: Proxy pool management for mapped queues
Date: Mon, 22 Sep 2025 09:48:01 -0700	[thread overview]
Message-ID: <aNF9waxmQUipXe1_@mini-arch> (raw)
In-Reply-To: <20250919213153.103606-14-daniel@iogearbox.net>

On 09/19, Daniel Borkmann wrote:
> Similarly what we do for net_mp_{open,close}_rxq for mapped queues,
> proxy also the xsk_{reg,clear}_pool_at_qid via __netif_get_rx_queue_peer
> such that when a virtual netdev picked a mapped rxq, the request gets
> through to the real rxq in the physical netdev.
> 
> Change the function signatures for queue_id to unsigned int in order
> to pass the queue_id parameter into __netif_get_rx_queue_peer. The
> proxying is only relevant for queue_id < dev->real_num_rx_queues since
> right now its only supported for rxqs.
> 
> Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
> Co-developed-by: David Wei <dw@davidwei.uk>
> Signed-off-by: David Wei <dw@davidwei.uk>
> ---
>  include/net/xdp_sock_drv.h |  4 ++--
>  net/xdp/xsk.c              | 16 +++++++++++-----
>  net/xdp/xsk.h              |  5 ++---
>  3 files changed, 15 insertions(+), 10 deletions(-)
> 
> diff --git a/include/net/xdp_sock_drv.h b/include/net/xdp_sock_drv.h
> index 47120666d8d6..709af292cba7 100644
> --- a/include/net/xdp_sock_drv.h
> +++ b/include/net/xdp_sock_drv.h
> @@ -29,7 +29,7 @@ bool xsk_tx_peek_desc(struct xsk_buff_pool *pool, struct xdp_desc *desc);
>  u32 xsk_tx_peek_release_desc_batch(struct xsk_buff_pool *pool, u32 max);
>  void xsk_tx_release(struct xsk_buff_pool *pool);
>  struct xsk_buff_pool *xsk_get_pool_from_qid(struct net_device *dev,
> -					    u16 queue_id);
> +					    unsigned int queue_id);
>  void xsk_set_rx_need_wakeup(struct xsk_buff_pool *pool);
>  void xsk_set_tx_need_wakeup(struct xsk_buff_pool *pool);
>  void xsk_clear_rx_need_wakeup(struct xsk_buff_pool *pool);
> @@ -286,7 +286,7 @@ static inline void xsk_tx_release(struct xsk_buff_pool *pool)
>  }
>  
>  static inline struct xsk_buff_pool *
> -xsk_get_pool_from_qid(struct net_device *dev, u16 queue_id)
> +xsk_get_pool_from_qid(struct net_device *dev, unsigned int queue_id)
>  {
>  	return NULL;
>  }
> diff --git a/net/xdp/xsk.c b/net/xdp/xsk.c
> index cf40c70ee59f..b9efa6d8a112 100644
> --- a/net/xdp/xsk.c
> +++ b/net/xdp/xsk.c
> @@ -23,6 +23,8 @@
>  #include <linux/netdevice.h>
>  #include <linux/rculist.h>
>  #include <linux/vmalloc.h>
> +
> +#include <net/netdev_queues.h>
>  #include <net/xdp_sock_drv.h>
>  #include <net/busy_poll.h>
>  #include <net/netdev_lock.h>
> @@ -111,19 +113,20 @@ bool xsk_uses_need_wakeup(struct xsk_buff_pool *pool)
>  EXPORT_SYMBOL(xsk_uses_need_wakeup);
>  
>  struct xsk_buff_pool *xsk_get_pool_from_qid(struct net_device *dev,
> -					    u16 queue_id)
> +					    unsigned int queue_id)
>  {
>  	if (queue_id < dev->real_num_rx_queues)
>  		return dev->_rx[queue_id].pool;
>  	if (queue_id < dev->real_num_tx_queues)
>  		return dev->_tx[queue_id].pool;
> -
>  	return NULL;
>  }
>  EXPORT_SYMBOL(xsk_get_pool_from_qid);
>  
> -void xsk_clear_pool_at_qid(struct net_device *dev, u16 queue_id)
> +void xsk_clear_pool_at_qid(struct net_device *dev, unsigned int queue_id)
>  {
> +	if (queue_id < dev->real_num_rx_queues)
> +		__netif_get_rx_queue_peer(&dev, &queue_id);
>  	if (queue_id < dev->num_rx_queues)
>  		dev->_rx[queue_id].pool = NULL;
>  	if (queue_id < dev->num_tx_queues)
> @@ -135,7 +138,7 @@ void xsk_clear_pool_at_qid(struct net_device *dev, u16 queue_id)
>   * This might also change during run time.
>   */
>  int xsk_reg_pool_at_qid(struct net_device *dev, struct xsk_buff_pool *pool,
> -			u16 queue_id)
> +			unsigned int queue_id)
>  {
>  	if (queue_id >= max_t(unsigned int,
>  			      dev->real_num_rx_queues,
> @@ -143,6 +146,10 @@ int xsk_reg_pool_at_qid(struct net_device *dev, struct xsk_buff_pool *pool,
>  		return -EINVAL;
>  	if (xsk_get_pool_from_qid(dev, queue_id))
>  		return -EBUSY;
> +	if (queue_id < dev->real_num_rx_queues)
> +		__netif_get_rx_queue_peer(&dev, &queue_id);
> +	if (xsk_get_pool_from_qid(dev, queue_id))
> +		return -EBUSY;
>  
>  	pool->netdev = dev;
>  	pool->queue_id = queue_id;

I feel like both of the above are also gonna be problematic wrt netdev
lock. The callers lock the netdev, the callers will also have
to resolve the virtual->real queue mapping. Hacking up the
queue/netdev deep in the call stack in a few places is not gonna work.

Maybe also add assert for the (new) netdev lock to __netif_get_rx_queue_peer
to trigger these.

  reply	other threads:[~2025-09-22 16:48 UTC|newest]

Thread overview: 64+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2025-09-19 21:31 [PATCH net-next 00/20] netkit: Support for io_uring zero-copy and AF_XDP Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 01/20] net, ynl: Add bind-queue operation Daniel Borkmann
2025-09-22 16:04   ` Stanislav Fomichev
2025-09-22 16:13     ` Daniel Borkmann
2025-09-23  1:17   ` Jakub Kicinski
2025-09-23 16:13     ` David Wei
2025-09-19 21:31 ` [PATCH net-next 02/20] net: Add peer to netdev_rx_queue Daniel Borkmann
2025-09-23  1:22   ` Jakub Kicinski
2025-09-23 15:56     ` David Wei
2025-09-19 21:31 ` [PATCH net-next 03/20] net: Add ndo_queue_create callback Daniel Borkmann
2025-09-22 16:04   ` Stanislav Fomichev
2025-09-22 16:14     ` Daniel Borkmann
2025-09-23 15:58     ` David Wei
2025-09-23  1:22   ` Jakub Kicinski
2025-09-23 15:58     ` David Wei
2025-09-19 21:31 ` [PATCH net-next 04/20] net: Add ndo_{peer,unpeer}_queues callback Daniel Borkmann
2025-09-23  1:23   ` Jakub Kicinski
2025-09-23 16:06     ` David Wei
2025-09-23 16:26       ` Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 05/20] net, ynl: Implement netdev_nl_bind_queue_doit Daniel Borkmann
2025-09-22 16:06   ` Stanislav Fomichev
2025-09-23  1:26     ` Jakub Kicinski
2025-09-23 16:06       ` David Wei
2025-09-19 21:31 ` [PATCH net-next 06/20] net, ynl: Add peer info to queue-get response Daniel Borkmann
2025-09-23  1:32   ` Jakub Kicinski
2025-09-23 16:08     ` David Wei
2025-09-19 21:31 ` [PATCH net-next 07/20] net, ethtool: Disallow mapped real rxqs to be resized Daniel Borkmann
2025-09-23  1:34   ` Jakub Kicinski
2025-09-23  1:38     ` Jakub Kicinski
2025-09-23 16:08       ` David Wei
2025-09-19 21:31 ` [PATCH net-next 08/20] net: Proxy net_mp_{open,close}_rxq for mapped queues Daniel Borkmann
2025-09-22 16:35   ` Stanislav Fomichev
2025-09-19 21:31 ` [PATCH net-next 09/20] xsk: Move NETDEV_XDP_ACT_ZC into generic header Daniel Borkmann
2025-09-22 15:59   ` Maciej Fijalkowski
2025-09-19 21:31 ` [PATCH net-next 10/20] xsk: Move pool registration into single function Daniel Borkmann
2025-09-22 16:01   ` Maciej Fijalkowski
2025-09-22 16:15     ` Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 11/20] xsk: Add small helper xp_pool_bindable Daniel Borkmann
2025-09-22 16:03   ` Maciej Fijalkowski
2025-09-22 16:17     ` Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 12/20] xsk: Change xsk_rcv_check to check netdev/queue_id from pool Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 13/20] xsk: Proxy pool management for mapped queues Daniel Borkmann
2025-09-22 16:48   ` Stanislav Fomichev [this message]
2025-09-22 17:01     ` Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 14/20] netkit: Add single device mode for netkit Daniel Borkmann
2025-09-27  1:10   ` Jordan Rife
2025-09-29  7:55     ` Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 15/20] netkit: Document fast vs slowpath members via macros Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 16/20] netkit: Implement rtnl_link_ops->alloc Daniel Borkmann
2025-09-27  1:17   ` Jordan Rife
2025-09-29  7:50     ` Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 17/20] netkit: Implement ndo_queue_create Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 18/20] netkit: Add io_uring zero-copy support for TCP Daniel Borkmann
2025-09-22  3:17   ` zf
2025-09-22 16:23     ` Daniel Borkmann
2025-09-19 21:31 ` [PATCH net-next 19/20] netkit: Add xsk support for af_xdp applications Daniel Borkmann
2025-09-23 11:42   ` Toke Høiland-Jørgensen
2025-09-24 10:41     ` Daniel Borkmann
2025-09-26  8:55       ` Toke Høiland-Jørgensen
2025-09-19 21:31 ` [PATCH net-next 20/20] tools, ynl: Add queue binding ynl sample application Daniel Borkmann
2025-09-22 17:09   ` Stanislav Fomichev
2025-09-23 16:12     ` David Wei
2025-09-22 12:05 ` [PATCH net-next 00/20] netkit: Support for io_uring zero-copy and AF_XDP Nikolay Aleksandrov
2025-09-23  1:59 ` Jakub Kicinski

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=aNF9waxmQUipXe1_@mini-arch \
    --to=stfomichev@gmail.com \
    --cc=bpf@vger.kernel.org \
    --cc=daniel@iogearbox.net \
    --cc=davem@davemloft.net \
    --cc=dw@davidwei.uk \
    --cc=john.fastabend@gmail.com \
    --cc=jordan@jrife.io \
    --cc=kuba@kernel.org \
    --cc=maciej.fijalkowski@intel.com \
    --cc=magnus.karlsson@intel.com \
    --cc=martin.lau@kernel.org \
    --cc=netdev@vger.kernel.org \
    --cc=pabeni@redhat.com \
    --cc=razor@blackwall.org \
    --cc=sdf@fomichev.me \
    --cc=willemb@google.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).