* [PATCH V3 net] net: mana: Fix error handling in mana_create_txq/rxq's NAPI cleanup
@ 2024-08-29 8:36 Souradeep Chakrabarti
2024-08-29 15:39 ` Haiyang Zhang
` (2 more replies)
0 siblings, 3 replies; 5+ messages in thread
From: Souradeep Chakrabarti @ 2024-08-29 8:36 UTC (permalink / raw)
To: kys, haiyangz, wei.liu, decui, davem, longli, ssengar,
linux-hyperv, netdev, linux-kernel, linux-rdma
Cc: schakrabarti, Souradeep Chakrabarti, stable
Currently napi_disable() gets called during rxq and txq cleanup,
even before napi is enabled and hrtimer is initialized. It causes
kernel panic.
? page_fault_oops+0x136/0x2b0
? page_counter_cancel+0x2e/0x80
? do_user_addr_fault+0x2f2/0x640
? refill_obj_stock+0xc4/0x110
? exc_page_fault+0x71/0x160
? asm_exc_page_fault+0x27/0x30
? __mmdrop+0x10/0x180
? __mmdrop+0xec/0x180
? hrtimer_active+0xd/0x50
hrtimer_try_to_cancel+0x2c/0xf0
hrtimer_cancel+0x15/0x30
napi_disable+0x65/0x90
mana_destroy_rxq+0x4c/0x2f0
mana_create_rxq.isra.0+0x56c/0x6d0
? mana_uncfg_vport+0x50/0x50
mana_alloc_queues+0x21b/0x320
? skb_dequeue+0x5f/0x80
Cc: stable@vger.kernel.org
Fixes: e1b5683ff62e ("net: mana: Move NAPI from EQ to CQ")
Signed-off-by: Souradeep Chakrabarti <schakrabarti@linux.microsoft.com>
Reviewed-by: Haiyang Zhang <haiyangz@microsoft.com>
Reviewed-by: Shradha Gupta <shradhagupta@linux.microsoft.com>
---
V3 -> V2:
Instead of using napi internal attribute, using an atomic
attribute to verify napi is initialized for a particular txq / rxq.
V2 -> V1:
Addressed the comment on cleaning up napi for the queues,
where queue creation was successful.
---
drivers/net/ethernet/microsoft/mana/mana_en.c | 30 ++++++++++++-------
include/net/mana/mana.h | 4 +++
2 files changed, 24 insertions(+), 10 deletions(-)
diff --git a/drivers/net/ethernet/microsoft/mana/mana_en.c b/drivers/net/ethernet/microsoft/mana/mana_en.c
index 39f56973746d..bd303c89cfa6 100644
--- a/drivers/net/ethernet/microsoft/mana/mana_en.c
+++ b/drivers/net/ethernet/microsoft/mana/mana_en.c
@@ -1872,10 +1872,12 @@ static void mana_destroy_txq(struct mana_port_context *apc)
for (i = 0; i < apc->num_queues; i++) {
napi = &apc->tx_qp[i].tx_cq.napi;
- napi_synchronize(napi);
- napi_disable(napi);
- netif_napi_del(napi);
-
+ if (atomic_read(&apc->tx_qp[i].txq.napi_initialized)) {
+ napi_synchronize(napi);
+ napi_disable(napi);
+ netif_napi_del(napi);
+ atomic_set(&apc->tx_qp[i].txq.napi_initialized, 0);
+ }
mana_destroy_wq_obj(apc, GDMA_SQ, apc->tx_qp[i].tx_object);
mana_deinit_cq(apc, &apc->tx_qp[i].tx_cq);
@@ -1931,6 +1933,7 @@ static int mana_create_txq(struct mana_port_context *apc,
txq->ndev = net;
txq->net_txq = netdev_get_tx_queue(net, i);
txq->vp_offset = apc->tx_vp_offset;
+ atomic_set(&txq->napi_initialized, 0);
skb_queue_head_init(&txq->pending_skbs);
memset(&spec, 0, sizeof(spec));
@@ -1997,6 +2000,7 @@ static int mana_create_txq(struct mana_port_context *apc,
netif_napi_add_tx(net, &cq->napi, mana_poll);
napi_enable(&cq->napi);
+ atomic_set(&txq->napi_initialized, 1);
mana_gd_ring_cq(cq->gdma_cq, SET_ARM_BIT);
}
@@ -2023,14 +2027,18 @@ static void mana_destroy_rxq(struct mana_port_context *apc,
napi = &rxq->rx_cq.napi;
- if (validate_state)
- napi_synchronize(napi);
+ if (atomic_read(&rxq->napi_initialized)) {
- napi_disable(napi);
+ if (validate_state)
+ napi_synchronize(napi);
- xdp_rxq_info_unreg(&rxq->xdp_rxq);
+ napi_disable(napi);
- netif_napi_del(napi);
+ netif_napi_del(napi);
+ atomic_set(&rxq->napi_initialized, 0);
+ }
+
+ xdp_rxq_info_unreg(&rxq->xdp_rxq);
mana_destroy_wq_obj(apc, GDMA_RQ, rxq->rxobj);
@@ -2199,6 +2207,7 @@ static struct mana_rxq *mana_create_rxq(struct mana_port_context *apc,
rxq->num_rx_buf = RX_BUFFERS_PER_QUEUE;
rxq->rxq_idx = rxq_idx;
rxq->rxobj = INVALID_MANA_HANDLE;
+ atomic_set(&rxq->napi_initialized, 0);
mana_get_rxbuf_cfg(ndev->mtu, &rxq->datasize, &rxq->alloc_size,
&rxq->headroom);
@@ -2286,6 +2295,8 @@ static struct mana_rxq *mana_create_rxq(struct mana_port_context *apc,
napi_enable(&cq->napi);
+ atomic_set(&rxq->napi_initialized, 1);
+
mana_gd_ring_cq(cq->gdma_cq, SET_ARM_BIT);
out:
if (!err)
@@ -2336,7 +2347,6 @@ static void mana_destroy_vport(struct mana_port_context *apc)
rxq = apc->rxqs[rxq_idx];
if (!rxq)
continue;
-
mana_destroy_rxq(apc, rxq, true);
apc->rxqs[rxq_idx] = NULL;
}
diff --git a/include/net/mana/mana.h b/include/net/mana/mana.h
index 7caa334f4888..be75abd63dc8 100644
--- a/include/net/mana/mana.h
+++ b/include/net/mana/mana.h
@@ -98,6 +98,8 @@ struct mana_txq {
atomic_t pending_sends;
+ atomic_t napi_initialized;
+
struct mana_stats_tx stats;
};
@@ -335,6 +337,8 @@ struct mana_rxq {
bool xdp_flush;
int xdp_rc; /* XDP redirect return code */
+ atomic_t napi_initialized;
+
struct page_pool *page_pool;
/* MUST BE THE LAST MEMBER:
--
2.34.1
^ permalink raw reply related [flat|nested] 5+ messages in thread
* RE: [PATCH V3 net] net: mana: Fix error handling in mana_create_txq/rxq's NAPI cleanup
2024-08-29 8:36 [PATCH V3 net] net: mana: Fix error handling in mana_create_txq/rxq's NAPI cleanup Souradeep Chakrabarti
@ 2024-08-29 15:39 ` Haiyang Zhang
2024-08-29 16:55 ` Haiyang Zhang
2024-08-30 6:13 ` Shradha Gupta
2 siblings, 0 replies; 5+ messages in thread
From: Haiyang Zhang @ 2024-08-29 15:39 UTC (permalink / raw)
To: Souradeep Chakrabarti, KY Srinivasan, wei.liu@kernel.org,
Dexuan Cui, davem@davemloft.net, Long Li,
ssengar@linux.microsoft.com, linux-hyperv@vger.kernel.org,
netdev@vger.kernel.org, linux-kernel@vger.kernel.org,
linux-rdma@vger.kernel.org
Cc: Souradeep Chakrabarti, stable@vger.kernel.org
> -----Original Message-----
> From: Souradeep Chakrabarti <schakrabarti@linux.microsoft.com>
> Sent: Thursday, August 29, 2024 4:37 AM
> To: KY Srinivasan <kys@microsoft.com>; Haiyang Zhang
> <haiyangz@microsoft.com>; wei.liu@kernel.org; Dexuan Cui
> <decui@microsoft.com>; davem@davemloft.net; Long Li
> <longli@microsoft.com>; ssengar@linux.microsoft.com; linux-
> hyperv@vger.kernel.org; netdev@vger.kernel.org; linux-
> kernel@vger.kernel.org; linux-rdma@vger.kernel.org
> Cc: Souradeep Chakrabarti <schakrabarti@microsoft.com>; Souradeep
> Chakrabarti <schakrabarti@linux.microsoft.com>; stable@vger.kernel.org
> Subject: [PATCH V3 net] net: mana: Fix error handling in
> mana_create_txq/rxq's NAPI cleanup
>
> Currently napi_disable() gets called during rxq and txq cleanup,
> even before napi is enabled and hrtimer is initialized. It causes
> kernel panic.
>
> ? page_fault_oops+0x136/0x2b0
> ? page_counter_cancel+0x2e/0x80
> ? do_user_addr_fault+0x2f2/0x640
> ? refill_obj_stock+0xc4/0x110
> ? exc_page_fault+0x71/0x160
> ? asm_exc_page_fault+0x27/0x30
> ? __mmdrop+0x10/0x180
> ? __mmdrop+0xec/0x180
> ? hrtimer_active+0xd/0x50
> hrtimer_try_to_cancel+0x2c/0xf0
> hrtimer_cancel+0x15/0x30
> napi_disable+0x65/0x90
> mana_destroy_rxq+0x4c/0x2f0
> mana_create_rxq.isra.0+0x56c/0x6d0
> ? mana_uncfg_vport+0x50/0x50
> mana_alloc_queues+0x21b/0x320
> ? skb_dequeue+0x5f/0x80
>
> Cc: stable@vger.kernel.org
> Fixes: e1b5683ff62e ("net: mana: Move NAPI from EQ to CQ")
> Signed-off-by: Souradeep Chakrabarti <schakrabarti@linux.microsoft.com>
> Reviewed-by: Haiyang Zhang <haiyangz@microsoft.com>
> Reviewed-by: Shradha Gupta <shradhagupta@linux.microsoft.com>
> ---
> V3 -> V2:
> Instead of using napi internal attribute, using an atomic
> attribute to verify napi is initialized for a particular txq / rxq.
>
> V2 -> V1:
> Addressed the comment on cleaning up napi for the queues,
> where queue creation was successful.
> ---
> drivers/net/ethernet/microsoft/mana/mana_en.c | 30 ++++++++++++-------
> include/net/mana/mana.h | 4 +++
> 2 files changed, 24 insertions(+), 10 deletions(-)
>
> diff --git a/drivers/net/ethernet/microsoft/mana/mana_en.c
> b/drivers/net/ethernet/microsoft/mana/mana_en.c
> index 39f56973746d..bd303c89cfa6 100644
> --- a/drivers/net/ethernet/microsoft/mana/mana_en.c
> +++ b/drivers/net/ethernet/microsoft/mana/mana_en.c
> @@ -1872,10 +1872,12 @@ static void mana_destroy_txq(struct
> mana_port_context *apc)
>
> for (i = 0; i < apc->num_queues; i++) {
> napi = &apc->tx_qp[i].tx_cq.napi;
> - napi_synchronize(napi);
> - napi_disable(napi);
> - netif_napi_del(napi);
> -
> + if (atomic_read(&apc->tx_qp[i].txq.napi_initialized)) {
> + napi_synchronize(napi);
> + napi_disable(napi);
> + netif_napi_del(napi);
> + atomic_set(&apc->tx_qp[i].txq.napi_initialized, 0);
> + }
> mana_destroy_wq_obj(apc, GDMA_SQ, apc->tx_qp[i].tx_object);
>
> mana_deinit_cq(apc, &apc->tx_qp[i].tx_cq);
> @@ -1931,6 +1933,7 @@ static int mana_create_txq(struct mana_port_context
> *apc,
> txq->ndev = net;
> txq->net_txq = netdev_get_tx_queue(net, i);
> txq->vp_offset = apc->tx_vp_offset;
> + atomic_set(&txq->napi_initialized, 0);
> skb_queue_head_init(&txq->pending_skbs);
>
> memset(&spec, 0, sizeof(spec));
> @@ -1997,6 +2000,7 @@ static int mana_create_txq(struct mana_port_context
> *apc,
>
> netif_napi_add_tx(net, &cq->napi, mana_poll);
> napi_enable(&cq->napi);
> + atomic_set(&txq->napi_initialized, 1);
>
> mana_gd_ring_cq(cq->gdma_cq, SET_ARM_BIT);
> }
> @@ -2023,14 +2027,18 @@ static void mana_destroy_rxq(struct
> mana_port_context *apc,
>
> napi = &rxq->rx_cq.napi;
>
> - if (validate_state)
> - napi_synchronize(napi);
> + if (atomic_read(&rxq->napi_initialized)) {
>
> - napi_disable(napi);
> + if (validate_state)
> + napi_synchronize(napi);
>
> - xdp_rxq_info_unreg(&rxq->xdp_rxq);
> + napi_disable(napi);
>
> - netif_napi_del(napi);
> + netif_napi_del(napi);
> + atomic_set(&rxq->napi_initialized, 0);
> + }
> +
> + xdp_rxq_info_unreg(&rxq->xdp_rxq);
>
> mana_destroy_wq_obj(apc, GDMA_RQ, rxq->rxobj);
>
> @@ -2199,6 +2207,7 @@ static struct mana_rxq *mana_create_rxq(struct
> mana_port_context *apc,
> rxq->num_rx_buf = RX_BUFFERS_PER_QUEUE;
> rxq->rxq_idx = rxq_idx;
> rxq->rxobj = INVALID_MANA_HANDLE;
> + atomic_set(&rxq->napi_initialized, 0);
>
> mana_get_rxbuf_cfg(ndev->mtu, &rxq->datasize, &rxq->alloc_size,
> &rxq->headroom);
> @@ -2286,6 +2295,8 @@ static struct mana_rxq *mana_create_rxq(struct
> mana_port_context *apc,
>
> napi_enable(&cq->napi);
>
> + atomic_set(&rxq->napi_initialized, 1);
> +
> mana_gd_ring_cq(cq->gdma_cq, SET_ARM_BIT);
> out:
> if (!err)
> @@ -2336,7 +2347,6 @@ static void mana_destroy_vport(struct
> mana_port_context *apc)
> rxq = apc->rxqs[rxq_idx];
> if (!rxq)
> continue;
> -
> mana_destroy_rxq(apc, rxq, true);
> apc->rxqs[rxq_idx] = NULL;
> }
> diff --git a/include/net/mana/mana.h b/include/net/mana/mana.h
> index 7caa334f4888..be75abd63dc8 100644
> --- a/include/net/mana/mana.h
> +++ b/include/net/mana/mana.h
> @@ -98,6 +98,8 @@ struct mana_txq {
>
> atomic_t pending_sends;
>
> + atomic_t napi_initialized;
> +
> struct mana_stats_tx stats;
> };
>
> @@ -335,6 +337,8 @@ struct mana_rxq {
> bool xdp_flush;
> int xdp_rc; /* XDP redirect return code */
>
> + atomic_t napi_initialized;
> +
> struct page_pool *page_pool;
>
> /* MUST BE THE LAST MEMBER:
> --
> 2.34.1
I agree with the idea that adds napi_initialized for the
state.
But the Rx/Tx queue creation/removal are either on the same
thread, or protected by rtnl_lock, correct? So napi_initialized
should be a regular variable, not atomic_t.
Thanks,
- Haiyang
^ permalink raw reply [flat|nested] 5+ messages in thread
* RE: [PATCH V3 net] net: mana: Fix error handling in mana_create_txq/rxq's NAPI cleanup
2024-08-29 8:36 [PATCH V3 net] net: mana: Fix error handling in mana_create_txq/rxq's NAPI cleanup Souradeep Chakrabarti
2024-08-29 15:39 ` Haiyang Zhang
@ 2024-08-29 16:55 ` Haiyang Zhang
2024-08-30 6:13 ` Shradha Gupta
2 siblings, 0 replies; 5+ messages in thread
From: Haiyang Zhang @ 2024-08-29 16:55 UTC (permalink / raw)
To: Souradeep Chakrabarti, KY Srinivasan, wei.liu@kernel.org,
Dexuan Cui, davem@davemloft.net, Long Li,
ssengar@linux.microsoft.com, linux-hyperv@vger.kernel.org,
netdev@vger.kernel.org, linux-kernel@vger.kernel.org,
linux-rdma@vger.kernel.org
Cc: Souradeep Chakrabarti, stable@vger.kernel.org
> -----Original Message-----
> From: Souradeep Chakrabarti <schakrabarti@linux.microsoft.com>
> Sent: Thursday, August 29, 2024 4:37 AM
> To: KY Srinivasan <kys@microsoft.com>; Haiyang Zhang
> <haiyangz@microsoft.com>; wei.liu@kernel.org; Dexuan Cui
> <decui@microsoft.com>; davem@davemloft.net; Long Li
> <longli@microsoft.com>; ssengar@linux.microsoft.com; linux-
> hyperv@vger.kernel.org; netdev@vger.kernel.org; linux-
> kernel@vger.kernel.org; linux-rdma@vger.kernel.org
> Cc: Souradeep Chakrabarti <schakrabarti@microsoft.com>; Souradeep
> Chakrabarti <schakrabarti@linux.microsoft.com>; stable@vger.kernel.org
> Subject: [PATCH V3 net] net: mana: Fix error handling in
> mana_create_txq/rxq's NAPI cleanup
>
> Currently napi_disable() gets called during rxq and txq cleanup,
> even before napi is enabled and hrtimer is initialized. It causes
> kernel panic.
>
> ? page_fault_oops+0x136/0x2b0
> ? page_counter_cancel+0x2e/0x80
> ? do_user_addr_fault+0x2f2/0x640
> ? refill_obj_stock+0xc4/0x110
> ? exc_page_fault+0x71/0x160
> ? asm_exc_page_fault+0x27/0x30
> ? __mmdrop+0x10/0x180
> ? __mmdrop+0xec/0x180
> ? hrtimer_active+0xd/0x50
> hrtimer_try_to_cancel+0x2c/0xf0
> hrtimer_cancel+0x15/0x30
> napi_disable+0x65/0x90
> mana_destroy_rxq+0x4c/0x2f0
> mana_create_rxq.isra.0+0x56c/0x6d0
> ? mana_uncfg_vport+0x50/0x50
> mana_alloc_queues+0x21b/0x320
> ? skb_dequeue+0x5f/0x80
>
> Cc: stable@vger.kernel.org
> Fixes: e1b5683ff62e ("net: mana: Move NAPI from EQ to CQ")
> Signed-off-by: Souradeep Chakrabarti <schakrabarti@linux.microsoft.com>
> Reviewed-by: Haiyang Zhang <haiyangz@microsoft.com>
> Reviewed-by: Shradha Gupta <shradhagupta@linux.microsoft.com>
> ---
> V3 -> V2:
> Instead of using napi internal attribute, using an atomic
> attribute to verify napi is initialized for a particular txq / rxq.
>
> V2 -> V1:
> Addressed the comment on cleaning up napi for the queues,
> where queue creation was successful.
> ---
> drivers/net/ethernet/microsoft/mana/mana_en.c | 30 ++++++++++++-------
> include/net/mana/mana.h | 4 +++
> 2 files changed, 24 insertions(+), 10 deletions(-)
>
> diff --git a/drivers/net/ethernet/microsoft/mana/mana_en.c
> b/drivers/net/ethernet/microsoft/mana/mana_en.c
> index 39f56973746d..bd303c89cfa6 100644
> --- a/drivers/net/ethernet/microsoft/mana/mana_en.c
> +++ b/drivers/net/ethernet/microsoft/mana/mana_en.c
> @@ -1872,10 +1872,12 @@ static void mana_destroy_txq(struct
> mana_port_context *apc)
>
> for (i = 0; i < apc->num_queues; i++) {
> napi = &apc->tx_qp[i].tx_cq.napi;
> - napi_synchronize(napi);
> - napi_disable(napi);
> - netif_napi_del(napi);
> -
> + if (atomic_read(&apc->tx_qp[i].txq.napi_initialized)) {
> + napi_synchronize(napi);
> + napi_disable(napi);
> + netif_napi_del(napi);
> + atomic_set(&apc->tx_qp[i].txq.napi_initialized, 0);
> + }
> mana_destroy_wq_obj(apc, GDMA_SQ, apc->tx_qp[i].tx_object);
>
> mana_deinit_cq(apc, &apc->tx_qp[i].tx_cq);
> @@ -1931,6 +1933,7 @@ static int mana_create_txq(struct mana_port_context
> *apc,
> txq->ndev = net;
> txq->net_txq = netdev_get_tx_queue(net, i);
> txq->vp_offset = apc->tx_vp_offset;
> + atomic_set(&txq->napi_initialized, 0);
> skb_queue_head_init(&txq->pending_skbs);
>
> memset(&spec, 0, sizeof(spec));
> @@ -1997,6 +2000,7 @@ static int mana_create_txq(struct mana_port_context
> *apc,
>
> netif_napi_add_tx(net, &cq->napi, mana_poll);
> napi_enable(&cq->napi);
> + atomic_set(&txq->napi_initialized, 1);
>
> mana_gd_ring_cq(cq->gdma_cq, SET_ARM_BIT);
> }
> @@ -2023,14 +2027,18 @@ static void mana_destroy_rxq(struct
> mana_port_context *apc,
>
> napi = &rxq->rx_cq.napi;
>
> - if (validate_state)
> - napi_synchronize(napi);
> + if (atomic_read(&rxq->napi_initialized)) {
>
> - napi_disable(napi);
> + if (validate_state)
> + napi_synchronize(napi);
>
> - xdp_rxq_info_unreg(&rxq->xdp_rxq);
> + napi_disable(napi);
>
> - netif_napi_del(napi);
> + netif_napi_del(napi);
> + atomic_set(&rxq->napi_initialized, 0);
> + }
> +
> + xdp_rxq_info_unreg(&rxq->xdp_rxq);
>
> mana_destroy_wq_obj(apc, GDMA_RQ, rxq->rxobj);
>
> @@ -2199,6 +2207,7 @@ static struct mana_rxq *mana_create_rxq(struct
> mana_port_context *apc,
> rxq->num_rx_buf = RX_BUFFERS_PER_QUEUE;
> rxq->rxq_idx = rxq_idx;
> rxq->rxobj = INVALID_MANA_HANDLE;
> + atomic_set(&rxq->napi_initialized, 0);
>
> mana_get_rxbuf_cfg(ndev->mtu, &rxq->datasize, &rxq->alloc_size,
> &rxq->headroom);
> @@ -2286,6 +2295,8 @@ static struct mana_rxq *mana_create_rxq(struct
> mana_port_context *apc,
>
> napi_enable(&cq->napi);
>
> + atomic_set(&rxq->napi_initialized, 1);
> +
> mana_gd_ring_cq(cq->gdma_cq, SET_ARM_BIT);
> out:
> if (!err)
> @@ -2336,7 +2347,6 @@ static void mana_destroy_vport(struct
> mana_port_context *apc)
> rxq = apc->rxqs[rxq_idx];
> if (!rxq)
> continue;
> -
> mana_destroy_rxq(apc, rxq, true);
Also, this line removal is not necessary.
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH V3 net] net: mana: Fix error handling in mana_create_txq/rxq's NAPI cleanup
2024-08-29 8:36 [PATCH V3 net] net: mana: Fix error handling in mana_create_txq/rxq's NAPI cleanup Souradeep Chakrabarti
2024-08-29 15:39 ` Haiyang Zhang
2024-08-29 16:55 ` Haiyang Zhang
@ 2024-08-30 6:13 ` Shradha Gupta
2024-08-30 17:36 ` Haiyang Zhang
2 siblings, 1 reply; 5+ messages in thread
From: Shradha Gupta @ 2024-08-30 6:13 UTC (permalink / raw)
To: Souradeep Chakrabarti
Cc: kys, haiyangz, wei.liu, decui, davem, longli, ssengar,
linux-hyperv, netdev, linux-kernel, linux-rdma, schakrabarti,
stable
On Thu, Aug 29, 2024 at 01:36:50AM -0700, Souradeep Chakrabarti wrote:
> Currently napi_disable() gets called during rxq and txq cleanup,
> even before napi is enabled and hrtimer is initialized. It causes
> kernel panic.
>
> ? page_fault_oops+0x136/0x2b0
> ? page_counter_cancel+0x2e/0x80
> ? do_user_addr_fault+0x2f2/0x640
> ? refill_obj_stock+0xc4/0x110
> ? exc_page_fault+0x71/0x160
> ? asm_exc_page_fault+0x27/0x30
> ? __mmdrop+0x10/0x180
> ? __mmdrop+0xec/0x180
> ? hrtimer_active+0xd/0x50
> hrtimer_try_to_cancel+0x2c/0xf0
> hrtimer_cancel+0x15/0x30
> napi_disable+0x65/0x90
> mana_destroy_rxq+0x4c/0x2f0
> mana_create_rxq.isra.0+0x56c/0x6d0
> ? mana_uncfg_vport+0x50/0x50
> mana_alloc_queues+0x21b/0x320
> ? skb_dequeue+0x5f/0x80
>
> Cc: stable@vger.kernel.org
> Fixes: e1b5683ff62e ("net: mana: Move NAPI from EQ to CQ")
> Signed-off-by: Souradeep Chakrabarti <schakrabarti@linux.microsoft.com>
> Reviewed-by: Haiyang Zhang <haiyangz@microsoft.com>
> Reviewed-by: Shradha Gupta <shradhagupta@linux.microsoft.com>
> ---
> V3 -> V2:
> Instead of using napi internal attribute, using an atomic
> attribute to verify napi is initialized for a particular txq / rxq.
>
> V2 -> V1:
> Addressed the comment on cleaning up napi for the queues,
> where queue creation was successful.
> ---
> drivers/net/ethernet/microsoft/mana/mana_en.c | 30 ++++++++++++-------
> include/net/mana/mana.h | 4 +++
> 2 files changed, 24 insertions(+), 10 deletions(-)
>
> diff --git a/drivers/net/ethernet/microsoft/mana/mana_en.c b/drivers/net/ethernet/microsoft/mana/mana_en.c
> index 39f56973746d..bd303c89cfa6 100644
> --- a/drivers/net/ethernet/microsoft/mana/mana_en.c
> +++ b/drivers/net/ethernet/microsoft/mana/mana_en.c
> @@ -1872,10 +1872,12 @@ static void mana_destroy_txq(struct mana_port_context *apc)
>
> for (i = 0; i < apc->num_queues; i++) {
> napi = &apc->tx_qp[i].tx_cq.napi;
> - napi_synchronize(napi);
> - napi_disable(napi);
> - netif_napi_del(napi);
> -
> + if (atomic_read(&apc->tx_qp[i].txq.napi_initialized)) {
> + napi_synchronize(napi);
> + napi_disable(napi);
> + netif_napi_del(napi);
> + atomic_set(&apc->tx_qp[i].txq.napi_initialized, 0);
> + }
> mana_destroy_wq_obj(apc, GDMA_SQ, apc->tx_qp[i].tx_object);
>
> mana_deinit_cq(apc, &apc->tx_qp[i].tx_cq);
> @@ -1931,6 +1933,7 @@ static int mana_create_txq(struct mana_port_context *apc,
> txq->ndev = net;
> txq->net_txq = netdev_get_tx_queue(net, i);
> txq->vp_offset = apc->tx_vp_offset;
> + atomic_set(&txq->napi_initialized, 0);
> skb_queue_head_init(&txq->pending_skbs);
>
> memset(&spec, 0, sizeof(spec));
> @@ -1997,6 +2000,7 @@ static int mana_create_txq(struct mana_port_context *apc,
>
> netif_napi_add_tx(net, &cq->napi, mana_poll);
> napi_enable(&cq->napi);
> + atomic_set(&txq->napi_initialized, 1);
>
> mana_gd_ring_cq(cq->gdma_cq, SET_ARM_BIT);
> }
> @@ -2023,14 +2027,18 @@ static void mana_destroy_rxq(struct mana_port_context *apc,
>
> napi = &rxq->rx_cq.napi;
>
> - if (validate_state)
> - napi_synchronize(napi);
> + if (atomic_read(&rxq->napi_initialized)) {
>
> - napi_disable(napi);
> + if (validate_state)
> + napi_synchronize(napi);
Is this validate_state flag still needed? The new napi_initialized
variable will make sure the napi_synchronize() is called only for rxqs
that have napi_enabled.
Regards,
Shradha.
>
> - xdp_rxq_info_unreg(&rxq->xdp_rxq);
> + napi_disable(napi);
>
> - netif_napi_del(napi);
> + netif_napi_del(napi);
> + atomic_set(&rxq->napi_initialized, 0);
> + }
> +
> + xdp_rxq_info_unreg(&rxq->xdp_rxq);
>
> mana_destroy_wq_obj(apc, GDMA_RQ, rxq->rxobj);
>
> @@ -2199,6 +2207,7 @@ static struct mana_rxq *mana_create_rxq(struct mana_port_context *apc,
> rxq->num_rx_buf = RX_BUFFERS_PER_QUEUE;
> rxq->rxq_idx = rxq_idx;
> rxq->rxobj = INVALID_MANA_HANDLE;
> + atomic_set(&rxq->napi_initialized, 0);
>
> mana_get_rxbuf_cfg(ndev->mtu, &rxq->datasize, &rxq->alloc_size,
> &rxq->headroom);
> @@ -2286,6 +2295,8 @@ static struct mana_rxq *mana_create_rxq(struct mana_port_context *apc,
>
> napi_enable(&cq->napi);
>
> + atomic_set(&rxq->napi_initialized, 1);
> +
> mana_gd_ring_cq(cq->gdma_cq, SET_ARM_BIT);
> out:
> if (!err)
> @@ -2336,7 +2347,6 @@ static void mana_destroy_vport(struct mana_port_context *apc)
> rxq = apc->rxqs[rxq_idx];
> if (!rxq)
> continue;
> -
> mana_destroy_rxq(apc, rxq, true);
> apc->rxqs[rxq_idx] = NULL;
> }
> diff --git a/include/net/mana/mana.h b/include/net/mana/mana.h
> index 7caa334f4888..be75abd63dc8 100644
> --- a/include/net/mana/mana.h
> +++ b/include/net/mana/mana.h
> @@ -98,6 +98,8 @@ struct mana_txq {
>
> atomic_t pending_sends;
>
> + atomic_t napi_initialized;
> +
> struct mana_stats_tx stats;
> };
>
> @@ -335,6 +337,8 @@ struct mana_rxq {
> bool xdp_flush;
> int xdp_rc; /* XDP redirect return code */
>
> + atomic_t napi_initialized;
> +
> struct page_pool *page_pool;
>
> /* MUST BE THE LAST MEMBER:
> --
> 2.34.1
>
>
^ permalink raw reply [flat|nested] 5+ messages in thread
* RE: [PATCH V3 net] net: mana: Fix error handling in mana_create_txq/rxq's NAPI cleanup
2024-08-30 6:13 ` Shradha Gupta
@ 2024-08-30 17:36 ` Haiyang Zhang
0 siblings, 0 replies; 5+ messages in thread
From: Haiyang Zhang @ 2024-08-30 17:36 UTC (permalink / raw)
To: Shradha Gupta, Souradeep Chakrabarti
Cc: KY Srinivasan, wei.liu@kernel.org, Dexuan Cui,
davem@davemloft.net, Long Li, ssengar@linux.microsoft.com,
linux-hyperv@vger.kernel.org, netdev@vger.kernel.org,
linux-kernel@vger.kernel.org, linux-rdma@vger.kernel.org,
Souradeep Chakrabarti, stable@vger.kernel.org
> -----Original Message-----
> From: Shradha Gupta <shradhagupta@linux.microsoft.com>
> Sent: Friday, August 30, 2024 2:14 AM
> To: Souradeep Chakrabarti <schakrabarti@linux.microsoft.com>
> Cc: KY Srinivasan <kys@microsoft.com>; Haiyang Zhang
> <haiyangz@microsoft.com>; wei.liu@kernel.org; Dexuan Cui
> <decui@microsoft.com>; davem@davemloft.net; Long Li
> <longli@microsoft.com>; ssengar@linux.microsoft.com; linux-
> hyperv@vger.kernel.org; netdev@vger.kernel.org; linux-
> kernel@vger.kernel.org; linux-rdma@vger.kernel.org; Souradeep Chakrabarti
> <schakrabarti@microsoft.com>; stable@vger.kernel.org
> Subject: Re: [PATCH V3 net] net: mana: Fix error handling in
> mana_create_txq/rxq's NAPI cleanup
>
> On Thu, Aug 29, 2024 at 01:36:50AM -0700, Souradeep Chakrabarti wrote:
> > Currently napi_disable() gets called during rxq and txq cleanup,
> > even before napi is enabled and hrtimer is initialized. It causes
> > kernel panic.
> >
> > ? page_fault_oops+0x136/0x2b0
> > ? page_counter_cancel+0x2e/0x80
> > ? do_user_addr_fault+0x2f2/0x640
> > ? refill_obj_stock+0xc4/0x110
> > ? exc_page_fault+0x71/0x160
> > ? asm_exc_page_fault+0x27/0x30
> > ? __mmdrop+0x10/0x180
> > ? __mmdrop+0xec/0x180
> > ? hrtimer_active+0xd/0x50
> > hrtimer_try_to_cancel+0x2c/0xf0
> > hrtimer_cancel+0x15/0x30
> > napi_disable+0x65/0x90
> > mana_destroy_rxq+0x4c/0x2f0
> > mana_create_rxq.isra.0+0x56c/0x6d0
> > ? mana_uncfg_vport+0x50/0x50
> > mana_alloc_queues+0x21b/0x320
> > ? skb_dequeue+0x5f/0x80
> >
> > Cc: stable@vger.kernel.org
> > Fixes: e1b5683ff62e ("net: mana: Move NAPI from EQ to CQ")
> > Signed-off-by: Souradeep Chakrabarti <schakrabarti@linux.microsoft.com>
> > Reviewed-by: Haiyang Zhang <haiyangz@microsoft.com>
> > Reviewed-by: Shradha Gupta <shradhagupta@linux.microsoft.com>
> > ---
> > V3 -> V2:
> > Instead of using napi internal attribute, using an atomic
> > attribute to verify napi is initialized for a particular txq / rxq.
> >
> > V2 -> V1:
> > Addressed the comment on cleaning up napi for the queues,
> > where queue creation was successful.
> > ---
> > drivers/net/ethernet/microsoft/mana/mana_en.c | 30 ++++++++++++-------
> > include/net/mana/mana.h | 4 +++
> > 2 files changed, 24 insertions(+), 10 deletions(-)
> >
> > diff --git a/drivers/net/ethernet/microsoft/mana/mana_en.c
> b/drivers/net/ethernet/microsoft/mana/mana_en.c
> > index 39f56973746d..bd303c89cfa6 100644
> > --- a/drivers/net/ethernet/microsoft/mana/mana_en.c
> > +++ b/drivers/net/ethernet/microsoft/mana/mana_en.c
> > @@ -1872,10 +1872,12 @@ static void mana_destroy_txq(struct
> mana_port_context *apc)
> >
> > for (i = 0; i < apc->num_queues; i++) {
> > napi = &apc->tx_qp[i].tx_cq.napi;
> > - napi_synchronize(napi);
> > - napi_disable(napi);
> > - netif_napi_del(napi);
> > -
> > + if (atomic_read(&apc->tx_qp[i].txq.napi_initialized)) {
> > + napi_synchronize(napi);
> > + napi_disable(napi);
> > + netif_napi_del(napi);
> > + atomic_set(&apc->tx_qp[i].txq.napi_initialized, 0);
> > + }
> > mana_destroy_wq_obj(apc, GDMA_SQ, apc->tx_qp[i].tx_object);
> >
> > mana_deinit_cq(apc, &apc->tx_qp[i].tx_cq);
> > @@ -1931,6 +1933,7 @@ static int mana_create_txq(struct
> mana_port_context *apc,
> > txq->ndev = net;
> > txq->net_txq = netdev_get_tx_queue(net, i);
> > txq->vp_offset = apc->tx_vp_offset;
> > + atomic_set(&txq->napi_initialized, 0);
> > skb_queue_head_init(&txq->pending_skbs);
> >
> > memset(&spec, 0, sizeof(spec));
> > @@ -1997,6 +2000,7 @@ static int mana_create_txq(struct
> mana_port_context *apc,
> >
> > netif_napi_add_tx(net, &cq->napi, mana_poll);
> > napi_enable(&cq->napi);
> > + atomic_set(&txq->napi_initialized, 1);
> >
> > mana_gd_ring_cq(cq->gdma_cq, SET_ARM_BIT);
> > }
> > @@ -2023,14 +2027,18 @@ static void mana_destroy_rxq(struct
> mana_port_context *apc,
> >
> > napi = &rxq->rx_cq.napi;
> >
> > - if (validate_state)
> > - napi_synchronize(napi);
> > + if (atomic_read(&rxq->napi_initialized)) {
> >
> > - napi_disable(napi);
> > + if (validate_state)
> > + napi_synchronize(napi);
>
> Is this validate_state flag still needed? The new napi_initialized
> variable will make sure the napi_synchronize() is called only for rxqs
> that have napi_enabled.
>
Yes, the validate_state flag for mana_destroy_rxq() can be combined/renamed.
@Souradeep Chakrabarti, As Jakub suggested in V2, you may:
Rename the parameter validate_state to napi_initialized. The two callers
below can still call like before.
1 mana_en.c mana_create_rxq 2296 mana_destroy_rxq(apc, rxq, false);
2 mana_en.c mana_destroy_vport 2340 mana_destroy_rxq(apc, rxq, true);
So you don't need to add napi_initialized struct mana_rxq.
In mana_destroy_rxq() you can have code like this
if (napi_initialized) {
napi = &rxq->rx_cq.napi;
napi_synchronize(napi);
napi_disable(napi);
xdp_rxq_info_unreg(&rxq->xdp_rxq);
netif_napi_del(napi);
}
For mana_destroy_txq, it looks more complex, your adding napi_initialized
to struct mana_txq is good.
Thanks,
- Haiyang
^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2024-08-30 17:36 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2024-08-29 8:36 [PATCH V3 net] net: mana: Fix error handling in mana_create_txq/rxq's NAPI cleanup Souradeep Chakrabarti
2024-08-29 15:39 ` Haiyang Zhang
2024-08-29 16:55 ` Haiyang Zhang
2024-08-30 6:13 ` Shradha Gupta
2024-08-30 17:36 ` Haiyang Zhang
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).