public inbox for netdev@vger.kernel.org
 help / color / mirror / Atom feed
From: "Tantilov, Emil S" <emil.s.tantilov@intel.com>
To: Alexander Lobakin <aleksander.lobakin@intel.com>,
	<intel-wired-lan@lists.osuosl.org>
Cc: Tony Nguyen <anthony.l.nguyen@intel.com>,
	Przemek Kitszel <przemyslaw.kitszel@intel.com>,
	Andrew Lunn <andrew+netdev@lunn.ch>,
	"David S. Miller" <davem@davemloft.net>,
	Eric Dumazet <edumazet@google.com>,
	"Jakub Kicinski" <kuba@kernel.org>,
	Paolo Abeni <pabeni@redhat.com>, Simon Horman <horms@kernel.org>,
	Kohei Enju <kohei@enjuk.jp>,
	Jacob Keller <jacob.e.keller@intel.com>,
	Aleksandr Loktionov <aleksandr.loktionov@intel.com>,
	<nxne.cnse.osdt.itp.upstreaming@intel.com>,
	<netdev@vger.kernel.org>, <linux-kernel@vger.kernel.org>
Subject: Re: [Intel-wired-lan] [PATCH iwl-next v3 2/5] libeth: handle creating pools with unreadable buffers
Date: Thu, 5 Mar 2026 14:04:52 -0800	[thread overview]
Message-ID: <b21bbcdc-ec87-4f83-8449-e627b6353281@intel.com> (raw)
In-Reply-To: <20260224174618.2780516-3-aleksander.lobakin@intel.com>



On 2/24/2026 9:46 AM, Alexander Lobakin wrote:
> libeth uses netmems for quite some time already, so in order to
> support unreadable frags / memory providers, it only needs to set
> PP_FLAG_ALLOW_UNREADABLE_NETMEM when needed.
> Also add a couple sanity checks to make sure the driver didn't mess
> up the configuration options and, in case when an MP is installed,
> return the truesize always equal to PAGE_SIZE, so that
> libeth_rx_alloc() will never try to allocate frags. Memory providers
> manage buffers on their own and expect 1:1 buffer / HW Rx descriptor
> association.
> 
> Bonus: mention in the libeth_sqe_type description that
> LIBETH_SQE_EMPTY should also be used for netmem Tx SQEs -- they
> don't need DMA unmapping.
> 
> Reviewed-by: Jacob Keller <jacob.e.keller@intel.com>
> Reviewed-by: Aleksandr Loktionov <aleksandr.loktionov@intel.com>
> Signed-off-by: Alexander Lobakin <aleksander.lobakin@intel.com>
> ---
>   include/net/libeth/tx.h                |  2 +-
>   drivers/net/ethernet/intel/libeth/rx.c | 42 ++++++++++++++++++++++++++
>   2 files changed, 43 insertions(+), 1 deletion(-)
> 
> diff --git a/include/net/libeth/tx.h b/include/net/libeth/tx.h
> index c3db5c6f1641..a66fc2b3a114 100644
> --- a/include/net/libeth/tx.h
> +++ b/include/net/libeth/tx.h
> @@ -12,7 +12,7 @@
>   
>   /**
>    * enum libeth_sqe_type - type of &libeth_sqe to act on Tx completion
> - * @LIBETH_SQE_EMPTY: unused/empty OR XDP_TX/XSk frame, no action required
> + * @LIBETH_SQE_EMPTY: empty OR netmem/XDP_TX/XSk frame, no action required
>    * @LIBETH_SQE_CTX: context descriptor with empty SQE, no action required
>    * @LIBETH_SQE_SLAB: kmalloc-allocated buffer, unmap and kfree()
>    * @LIBETH_SQE_FRAG: mapped skb frag, only unmap DMA
> diff --git a/drivers/net/ethernet/intel/libeth/rx.c b/drivers/net/ethernet/intel/libeth/rx.c
> index 8874b714cdcc..11e6e8f353ef 100644
> --- a/drivers/net/ethernet/intel/libeth/rx.c
> +++ b/drivers/net/ethernet/intel/libeth/rx.c
> @@ -6,6 +6,7 @@
>   #include <linux/export.h>
>   
>   #include <net/libeth/rx.h>
> +#include <net/netdev_queues.h>
>   
>   /* Rx buffer management */
>   
> @@ -139,9 +140,47 @@ static bool libeth_rx_page_pool_params_zc(struct libeth_fq *fq,
>   	fq->buf_len = clamp(mtu, LIBETH_RX_BUF_STRIDE, max);
>   	fq->truesize = fq->buf_len;
>   
> +	/*
> +	 * Allow frags only for kernel pages. `fq->truesize == pp->max_len`
> +	 * will always fall back to regular page_pool_alloc_netmems()
> +	 * regardless of the MTU / FQ buffer size.
> +	 */
> +	if (pp->flags & PP_FLAG_ALLOW_UNREADABLE_NETMEM)
> +		fq->truesize = pp->max_len;
> +
>   	return true;
>   }
>   
> +/**
> + * libeth_rx_page_pool_check_unread - check input params for unreadable MPs
> + * @fq: buffer queue to check
> + * @pp: &page_pool_params for the queue
> + *
> + * Make sure we don't create an invalid pool with full-frame unreadable
> + * buffers, bidirectional unreadable buffers or so, and configure the
> + * ZC payload pool accordingly.
> + *
> + * Return: true on success, false on invalid input params.
> + */
> +static bool libeth_rx_page_pool_check_unread(const struct libeth_fq *fq,
> +					     struct page_pool_params *pp)
> +{
> +	if (!netif_rxq_has_unreadable_mp(pp->netdev, pp->queue_idx))
> +		return true;

This is causing a crash on IDPF:

[  420.570632] BUG: kernel NULL pointer dereference, address: 
00000000000000e8
[  420.570684] #PF: supervisor read access in kernel mode
[  420.570712] #PF: error_code(0x0000) - not-present page
[  420.570739] PGD 0
[  420.570757] Oops: Oops: 0000 [#1] SMP NOPTI
[  420.570784] CPU: 35 UID: 0 PID: 1058 Comm: kworker/u258:8 Kdump: 
loaded Tainted: G S         OE       7.0.0-rc1-next-devq-030526+ #34 
PREEMPT(full)
[  420.570844] Tainted: [S]=CPU_OUT_OF_SPEC, [O]=OOT_MODULE, 
[E]=UNSIGNED_MODULE
[  420.570872] Hardware name: Intel Corporation 
M50CYP2SBSTD/M50CYP2SBSTD, BIOS SE5C6200.86B.0027.P10.2201070222 01/07/2022
[  420.570912] Workqueue: idpf-0000:83:00.0-vc_event idpf_vc_event_task 
[idpf]
[  420.570967] RIP: 0010:netif_rxq_has_unreadable_mp+0xf/0x30
[  420.571004] Code: 2e 0f 1f 84 00 00 00 00 00 66 90 90 90 90 90 90 90 
90 90 90 90 90 90 90 90 90 90 f3 0f 1e fa 0f 1f 44 00 00 89 f6 48 c1 e6 
08 <48> 03 b7 e8 00 00 00 48 83 be c8 00 00 00 00 0f 95 c0 e9 8a 25 2b
[  420.571078] RSP: 0018:ff6bf03aa83e7bd8 EFLAGS: 00010246
[  420.571109] RAX: ff3579f9888ed0d0 RBX: ff6bf03aa83e7c78 RCX: 
0000000000000000
[  420.571145] RDX: 0000000000000000 RSI: 0000000000000000 RDI: 
0000000000000000
[  420.571178] RBP: ff6bf03aa83e7be0 R08: 0000000000000040 R09: 
000000004b1a4093
[  420.571213] R10: 0000000000000003 R11: ff3579f9bed90ee0 R12: 
ff3579f9d58c6060
[  420.571247] R13: ff6bf03aa83e7d88 R14: 0000000000000001 R15: 
ff3579f9d58c6050
[  420.571281] FS:  0000000000000000(0000) GS:ff357a195e9db000(0000) 
knlGS:0000000000000000
[  420.571320] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[  420.571350] CR2: 00000000000000e8 CR3: 00000023b2e2a006 CR4: 
0000000000773ef0
[  420.571384] PKRU: 55555554
[  420.571402] Call Trace:
[  420.571419]  <TASK>
[  420.571436]  libeth_rx_fq_create+0x7c/0x380 [libeth]
[  420.571479]  libie_ctlq_init+0x304/0x460 [libie_cp]
[  420.571516]  libie_ctlq_xn_init+0x29/0x230 [libie_cp]
[  420.571550]  idpf_init_dflt_mbx+0xa7/0x170 [idpf]
[  420.571611]  idpf_vc_event_task+0x15d/0x2f0 [idpf]
[  420.571655]  process_one_work+0x226/0x730
[  420.571699]  worker_thread+0x19e/0x340
[  420.571729]  ? __pfx_worker_thread+0x10/0x10
[  420.571760]  kthread+0xf4/0x130
[  420.571785]  ? __pfx_kthread+0x10/0x10
[  420.571813]  ret_from_fork+0x32c/0x410
[  420.571844]  ? __pfx_kthread+0x10/0x10
[  420.571871]  ret_from_fork_asm+0x1a/0x30
[  420.571909]  </TASK>

The driver will call idpf_init_dflt_mbx() before the netdevs are created.

Thanks,
Emil

> +
> +	/* For now, the core stack doesn't allow XDP with unreadable frags */
> +	if (fq->xdp)
> +		return false;
> +
> +	/* It should be either a header pool or a ZC payload pool */
> +	if (fq->type == LIBETH_FQE_HDR)
> +		return !fq->hsplit;
> +
> +	pp->flags |= PP_FLAG_ALLOW_UNREADABLE_NETMEM;
> +
> +	return fq->hsplit;
> +}
> +
>   /**
>    * libeth_rx_fq_create - create a PP with the default libeth settings
>    * @fq: buffer queue struct to fill
> @@ -165,6 +204,9 @@ int libeth_rx_fq_create(struct libeth_fq *fq, struct napi_struct *napi)
>   	struct page_pool *pool;
>   	int ret;
>   
> +	if (!libeth_rx_page_pool_check_unread(fq, &pp))
> +		return -EINVAL;
> +
>   	pp.dma_dir = fq->xdp ? DMA_BIDIRECTIONAL : DMA_FROM_DEVICE;
>   
>   	if (!fq->hsplit)


  reply	other threads:[~2026-03-05 22:05 UTC|newest]

Thread overview: 10+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-02-24 17:46 [PATCH iwl-next v3 0/5] ice: add support for devmem/io_uring Rx and Tx Alexander Lobakin
2026-02-24 17:46 ` [PATCH iwl-next v3 1/5] libeth: pass Rx queue index to PP when creating a fill queue Alexander Lobakin
2026-02-24 18:53   ` [Intel-wired-lan] " Paul Menzel
2026-03-03 15:42     ` Alexander Lobakin
2026-02-24 17:46 ` [PATCH iwl-next v3 2/5] libeth: handle creating pools with unreadable buffers Alexander Lobakin
2026-03-05 22:04   ` Tantilov, Emil S [this message]
2026-03-06 11:57     ` [Intel-wired-lan] " Alexander Lobakin
2026-02-24 17:46 ` [PATCH iwl-next v3 3/5] ice: migrate to netdev ops lock Alexander Lobakin
2026-02-24 17:46 ` [PATCH iwl-next v3 4/5] ice: implement Rx queue management ops Alexander Lobakin
2026-02-24 17:46 ` [PATCH iwl-next v3 5/5] ice: add support for transmitting unreadable frags Alexander Lobakin

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=b21bbcdc-ec87-4f83-8449-e627b6353281@intel.com \
    --to=emil.s.tantilov@intel.com \
    --cc=aleksander.lobakin@intel.com \
    --cc=aleksandr.loktionov@intel.com \
    --cc=andrew+netdev@lunn.ch \
    --cc=anthony.l.nguyen@intel.com \
    --cc=davem@davemloft.net \
    --cc=edumazet@google.com \
    --cc=horms@kernel.org \
    --cc=intel-wired-lan@lists.osuosl.org \
    --cc=jacob.e.keller@intel.com \
    --cc=kohei@enjuk.jp \
    --cc=kuba@kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=netdev@vger.kernel.org \
    --cc=nxne.cnse.osdt.itp.upstreaming@intel.com \
    --cc=pabeni@redhat.com \
    --cc=przemyslaw.kitszel@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox