netdev.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Maciej Fijalkowski <maciej.fijalkowski@intel.com>
To: Jason Xing <kerneljasonxing@gmail.com>
Cc: <davem@davemloft.net>, <edumazet@google.com>, <kuba@kernel.org>,
	<pabeni@redhat.com>, <bjorn@kernel.org>,
	<magnus.karlsson@intel.com>, <jonathan.lemon@gmail.com>,
	<sdf@fomichev.me>, <ast@kernel.org>, <daniel@iogearbox.net>,
	<hawk@kernel.org>, <john.fastabend@gmail.com>,
	<bpf@vger.kernel.org>, <netdev@vger.kernel.org>,
	Jason Xing <kernelxing@tencent.com>
Subject: Re: [PATCH net-next 2/3] xsk: add the atomic parameter around cq in generic path
Date: Mon, 24 Nov 2025 20:10:09 +0100	[thread overview]
Message-ID: <aSStkfeUrYRAKZeQ@boxer> (raw)
In-Reply-To: <20251124080858.89593-3-kerneljasonxing@gmail.com>

On Mon, Nov 24, 2025 at 04:08:57PM +0800, Jason Xing wrote:
> From: Jason Xing <kernelxing@tencent.com>
> 
> No functional changes here. Add a new parameter as a prep to help
> completion queue in copy mode convert into atomic type in the rest of
> this series. The patch also keeps the unified interface.

Jason,

anything used in ZC should not get a penalty from changes developed to
improve copy mode. I'd rather suggest separate functions rather than
branches within shared routines.

> 
> Signed-off-by: Jason Xing <kernelxing@tencent.com>
> ---
>  net/xdp/xsk.c       |  8 ++++----
>  net/xdp/xsk_queue.h | 31 +++++++++++++++++++------------
>  2 files changed, 23 insertions(+), 16 deletions(-)
> 
> diff --git a/net/xdp/xsk.c b/net/xdp/xsk.c
> index bcfd400e9cf8..4e95b894f218 100644
> --- a/net/xdp/xsk.c
> +++ b/net/xdp/xsk.c
> @@ -276,7 +276,7 @@ static int __xsk_rcv(struct xdp_sock *xs, struct xdp_buff *xdp, u32 len)
>  		xs->rx_dropped++;
>  		return -ENOMEM;
>  	}
> -	if (xskq_prod_nb_free(xs->rx, num_desc) < num_desc) {
> +	if (xskq_prod_nb_free(xs->rx, num_desc, false) < num_desc) {
>  		xs->rx_queue_full++;
>  		return -ENOBUFS;
>  	}
> @@ -519,7 +519,7 @@ u32 xsk_tx_peek_release_desc_batch(struct xsk_buff_pool *pool, u32 nb_pkts)
>  	 * packets. This avoids having to implement any buffering in
>  	 * the Tx path.
>  	 */
> -	nb_pkts = xskq_prod_nb_free(pool->cq, nb_pkts);
> +	nb_pkts = xskq_prod_nb_free(pool->cq, nb_pkts, false);
>  	if (!nb_pkts)
>  		goto out;
>  
> @@ -551,7 +551,7 @@ static int xsk_cq_reserve_locked(struct xsk_buff_pool *pool)
>  	int ret;
>  
>  	spin_lock(&pool->cq_cached_prod_lock);
> -	ret = xskq_prod_reserve(pool->cq);
> +	ret = xskq_prod_reserve(pool->cq, false);
>  	spin_unlock(&pool->cq_cached_prod_lock);
>  
>  	return ret;
> @@ -588,7 +588,7 @@ static void xsk_cq_submit_addr_locked(struct xsk_buff_pool *pool,
>  static void xsk_cq_cancel_locked(struct xsk_buff_pool *pool, u32 n)
>  {
>  	spin_lock(&pool->cq_cached_prod_lock);
> -	xskq_prod_cancel_n(pool->cq, n);
> +	xskq_prod_cancel_n(pool->cq, n, false);
>  	spin_unlock(&pool->cq_cached_prod_lock);
>  }
>  
> diff --git a/net/xdp/xsk_queue.h b/net/xdp/xsk_queue.h
> index 44cc01555c0b..7b4d9b954584 100644
> --- a/net/xdp/xsk_queue.h
> +++ b/net/xdp/xsk_queue.h
> @@ -378,37 +378,44 @@ static inline u32 xskq_get_prod(struct xsk_queue *q)
>  	return READ_ONCE(q->ring->producer);
>  }
>  
> -static inline u32 xskq_prod_nb_free(struct xsk_queue *q, u32 max)
> +static inline u32 xskq_prod_nb_free(struct xsk_queue *q, u32 max, bool atomic)
>  {
> -	u32 free_entries = q->nentries - (q->cached_prod - q->cached_cons);
> +	u32 cached_prod = atomic ? atomic_read(&q->cached_prod_atomic) : q->cached_prod;
> +	u32 free_entries = q->nentries - (cached_prod - q->cached_cons);
>  
>  	if (free_entries >= max)
>  		return max;
>  
>  	/* Refresh the local tail pointer */
>  	q->cached_cons = READ_ONCE(q->ring->consumer);
> -	free_entries = q->nentries - (q->cached_prod - q->cached_cons);
> +	free_entries = q->nentries - (cached_prod - q->cached_cons);
>  
>  	return free_entries >= max ? max : free_entries;
>  }
>  
> -static inline bool xskq_prod_is_full(struct xsk_queue *q)
> +static inline bool xskq_prod_is_full(struct xsk_queue *q, bool atomic)
>  {
> -	return xskq_prod_nb_free(q, 1) ? false : true;
> +	return xskq_prod_nb_free(q, 1, atomic) ? false : true;
>  }
>  
> -static inline void xskq_prod_cancel_n(struct xsk_queue *q, u32 cnt)
> +static inline void xskq_prod_cancel_n(struct xsk_queue *q, u32 cnt, bool atomic)
>  {
> -	q->cached_prod -= cnt;
> +	if (atomic)
> +		atomic_sub(cnt, &q->cached_prod_atomic);
> +	else
> +		q->cached_prod -= cnt;
>  }
>  
> -static inline int xskq_prod_reserve(struct xsk_queue *q)
> +static inline int xskq_prod_reserve(struct xsk_queue *q, bool atomic)
>  {
> -	if (xskq_prod_is_full(q))
> +	if (xskq_prod_is_full(q, atomic))
>  		return -ENOSPC;
>  
>  	/* A, matches D */
> -	q->cached_prod++;
> +	if (atomic)
> +		atomic_inc(&q->cached_prod_atomic);
> +	else
> +		q->cached_prod++;
>  	return 0;
>  }
>  
> @@ -416,7 +423,7 @@ static inline int xskq_prod_reserve_addr(struct xsk_queue *q, u64 addr)
>  {
>  	struct xdp_umem_ring *ring = (struct xdp_umem_ring *)q->ring;
>  
> -	if (xskq_prod_is_full(q))
> +	if (xskq_prod_is_full(q, false))
>  		return -ENOSPC;
>  
>  	/* A, matches D */
> @@ -450,7 +457,7 @@ static inline int xskq_prod_reserve_desc(struct xsk_queue *q,
>  	struct xdp_rxtx_ring *ring = (struct xdp_rxtx_ring *)q->ring;
>  	u32 idx;
>  
> -	if (xskq_prod_is_full(q))
> +	if (xskq_prod_is_full(q, false))
>  		return -ENOBUFS;
>  
>  	/* A, matches D */
> -- 
> 2.41.3
> 

  reply	other threads:[~2025-11-24 19:10 UTC|newest]

Thread overview: 6+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2025-11-24  8:08 [PATCH net-next 0/3] xsk: introduce atomic for cq in generic Jason Xing
2025-11-24  8:08 ` [PATCH net-next 1/3] xsk: add atomic cached_prod for copy mode Jason Xing
2025-11-24  8:08 ` [PATCH net-next 2/3] xsk: add the atomic parameter around cq in generic path Jason Xing
2025-11-24 19:10   ` Maciej Fijalkowski [this message]
2025-11-24 23:43     ` Jason Xing
2025-11-24  8:08 ` [PATCH net-next 3/3] xsk: convert cq from spin lock protection into atomic operations Jason Xing

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=aSStkfeUrYRAKZeQ@boxer \
    --to=maciej.fijalkowski@intel.com \
    --cc=ast@kernel.org \
    --cc=bjorn@kernel.org \
    --cc=bpf@vger.kernel.org \
    --cc=daniel@iogearbox.net \
    --cc=davem@davemloft.net \
    --cc=edumazet@google.com \
    --cc=hawk@kernel.org \
    --cc=john.fastabend@gmail.com \
    --cc=jonathan.lemon@gmail.com \
    --cc=kerneljasonxing@gmail.com \
    --cc=kernelxing@tencent.com \
    --cc=kuba@kernel.org \
    --cc=magnus.karlsson@intel.com \
    --cc=netdev@vger.kernel.org \
    --cc=pabeni@redhat.com \
    --cc=sdf@fomichev.me \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).