* [PATCH 0/1] ovpn: tcp - defer TX from softirq to workqueue
@ 2026-05-01 14:54 Dao Zhong Ma
2026-05-01 14:54 ` [PATCH 1/1] " Dao Zhong Ma
0 siblings, 1 reply; 4+ messages in thread
From: Dao Zhong Ma @ 2026-05-01 14:54 UTC (permalink / raw)
To: linux-kernel, netdev, antonio
Cc: sd, andrew+netdev, davem, edumazet, kuba, pabeni, Dao Zhong Ma
Hi,
I observed system hangs when using OpenVPN across suspend/resume
under heavy network load.
The issue can be reproduced with:
1. Establish an OpenVPN connection
2. Generate traffic with iperf3
3. Add CPU/memory pressure with stress-ng
4. Suspend and resume the system
After resume, the system may become unresponsive. SysRq output shows
tasks spinning on sk->sk_lock.slock, and RCU stall warnings are
observed.
Kernel logs and stack traces:
kernel: watchdog: BUG: soft lockup - CPU#4 stuck for 22s! [stress-ng-cpu:11506]
kernel: CPU#4 Utilization every 4000ms during lockup:
kernel: #1: 0% system, 100% softirq, 1% hardirq, 0% idle
kernel: #2: 0% system, 100% softirq, 0% hardirq, 0% idle
kernel: #3: 0% system, 100% softirq, 1% hardirq, 0% idle
kernel: #4: 0% system, 100% softirq, 0% hardirq, 0% idle
kernel: #5: 0% system, 100% softirq, 0% hardirq, 0% idle
kernel: CPU: 4 UID: 1000 PID: 11506 Comm: stress-ng-cpu Tainted: G U 7.0.2-2-cachyos #1 PREEMPT 136cb9373f01c1def72b1b5def9dd1fcc7884035
kernel: Tainted: [U]=USER
kernel: Hardware name: Framework Laptop 13 (AMD Ryzen 7040Series)/FRANMDCP05, BIOS 03.18 01/08/2026
kernel: RIP: 0010:native_queued_spin_lock_slowpath+0x5a/0x260
kernel: Code: 08 0f 92 c0 b9 ff 00 ff ff 23 0f 89 c2 c1 e2 08 09 ca 81 fa 00 01 00 00 73 5a 85 c9 74 0e 66 90 80 3f 00 74 07 f3 90 80 3f 00 <75> f9 66 c7 07 01 00 65 48 ff 05 3f c6 d8 01 5b 41 5e 41 5f 5d e9
kernel: RSP: 0000:ffffcf548d9bb758 EFLAGS: 00000202
kernel: RAX: 0000000000000000 RBX: ffff8ec75f9f4d48 RCX: 0000000000000001
kernel: RDX: 0000000000000001 RSI: 0000000000000001 RDI: ffff8ec75f9f4d48
kernel: RBP: 0000000000000000 R08: 0000000000000000 R09: ffff8ec8547abd80
kernel: R10: 0000000000000020 R11: 0000000000000020 R12: ffff8ec75f9f4c00
kernel: R13: ffff8ec862d9d400 R14: ffff8ec753ade400 R15: ffff8ec771548800
kernel: FS: 00007f9b4fcd0b00(0000) GS:ffff8eccfe99c000(0000) knlGS:0000000000000000
kernel: CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
kernel: CR2: 00007ffb7e1f8e38 CR3: 0000000166ced000 CR4: 0000000000f50ef0
kernel: PKRU: 55555554
kernel: Call Trace:
kernel: <TASK>
kernel: _raw_spin_lock+0x2a/0x40
kernel: ovpn_tcp_send_skb+0x4c/0x130 [ovpn 03847a96bef64ed5cc535ef1935e53bb388bcced]
kernel: ovpn_encrypt_post+0x95/0x1f0 [ovpn 03847a96bef64ed5cc535ef1935e53bb388bcced]
kernel: ovpn_send+0x15f/0x230 [ovpn 03847a96bef64ed5cc535ef1935e53bb388bcced]
kernel: ovpn_net_xmit+0x30f/0x560 [ovpn 03847a96bef64ed5cc535ef1935e53bb388bcced]
kernel: dev_hard_start_xmit+0x93/0x140
kernel: __dev_queue_xmit+0x99b/0xa80
kernel: ? __dev_queue_xmit+0x6f/0xa80
kernel: ? srso_alias_return_thunk+0x5/0xfbef5
kernel: ? nf_hook_slow+0x4b/0xd0
kernel: ? __pfx_ip_finish_output+0x10/0x10
kernel: ? srso_alias_return_thunk+0x5/0xfbef5
kernel: ? nf_hook+0x9c/0xb0
kernel: ip_finish_output2+0x189/0x320
kernel: dst_output+0x99/0xd0
kernel: __ip_queue_xmit+0x209/0x350
kernel: __tcp_transmit_skb+0x4cb/0xba0
kernel: ? srso_alias_return_thunk+0x5/0xfbef5
kernel: ? __alloc_skb+0x2b6/0x340
kernel: ? srso_alias_return_thunk+0x5/0xfbef5
kernel: ? skb_split+0x5f/0x3c0
kernel: tcp_write_xmit+0xad3/0x1a50
kernel: ? __pfx_tcp_write_timer+0x10/0x10
kernel: tcp_send_loss_probe+0x94/0x2b0
kernel: ? __pfx_tcp_write_timer+0x10/0x10
kernel: tcp_write_timer+0x67/0xb0
kernel: __run_timer_base+0x33d/0x500
kernel: tmigr_handle_remote+0x3da/0x3e0
kernel: run_timer_softirq.cold+0x47/0x394
kernel: ? srso_alias_return_thunk+0x5/0xfbef5
kernel: ? __pfx_tick_nohz_handler+0x10/0x10
kernel: ? srso_alias_return_thunk+0x5/0xfbef5
kernel: ? ktime_get+0x48/0xa0
kernel: ? srso_alias_return_thunk+0x5/0xfbef5
kernel: ? clockevents_program_event+0xaa/0xf0
kernel: ? srso_alias_return_thunk+0x5/0xfbef5
kernel: ? srso_alias_return_thunk+0x5/0xfbef5
kernel: irq_exit_rcu+0x1ad/0x290
kernel: sysvec_apic_timer_interrupt+0x33/0x80
kernel: asm_sysvec_apic_timer_interrupt+0x1a/0x20
kernel: RIP: 0033:0x559fc1ece66f
kernel: Code: 44 89 c8 c0 e8 03 44 31 c0 83 e0 01 66 41 d1 e8 f7 d8 66 25 08 84 44 31 c0 45 89 c8 41 c0 e8 04 41 31 c0 41 83 e0 01 41 f7 d8 <66> 41 81 e0 08 84 66 d1 e8 41 31 c0 44 89 c8 c0 e8 05 44 31 c0 83
kernel: RSP: 002b:00007ffe1344e860 EFLAGS: 00000297
kernel: RAX: 00000000ffff5912 RBX: 000000000000000a RCX: 00007ffe1344ebcb
kernel: RDX: 00007ffe1344ea27 RSI: 00007ffe1344ec60 RDI: 0000559fc318d450
kernel: RBP: 00007ffe1344edf0 R08: 00000000ffffffff R09: 00000000000000b4
kernel: R10: 0000559fc333b3e0 R11: 0000000000000000 R12: 00007f9b4fcc6258
kernel: R13: 0000000000000000 R14: 0000559fc316ed00 R15: 0000559fc3337e00
kernel: </TASK>
Dao Zhong Ma (1):
ovpn: tcp - defer TX from softirq to workqueue to avoid RCU stalls
drivers/net/ovpn/tcp.c | 80 +++++++++++++++++++++++++++++++-----------
1 file changed, 59 insertions(+), 21 deletions(-)
--
2.54.0
^ permalink raw reply [flat|nested] 4+ messages in thread
* [PATCH 1/1] ovpn: tcp - defer TX from softirq to workqueue
2026-05-01 14:54 [PATCH 0/1] ovpn: tcp - defer TX from softirq to workqueue Dao Zhong Ma
@ 2026-05-01 14:54 ` Dao Zhong Ma
2026-05-01 23:12 ` Jakub Kicinski
2026-05-03 21:14 ` Antonio Quartulli
0 siblings, 2 replies; 4+ messages in thread
From: Dao Zhong Ma @ 2026-05-01 14:54 UTC (permalink / raw)
To: linux-kernel, netdev, antonio
Cc: sd, andrew+netdev, davem, edumazet, kuba, pabeni, Dao Zhong Ma
ovpn_tcp_send_skb() holds sk->sk_lock.slock while performing the full TCP
send in softirq context. This can hold the spinlock for a long time
(large skb), blocking lock_sock() users. This can starve the RCU GP
kthread and trigger RCU stalls warnings and hung tasks.
Defer the TCP send operation to process context:
- In interrupt context, only enqueue the skb under the spinlock
schedule tcp_tx_work.
- In process context, dequeue and flush the send queue under lock_sock()
This reduces the softirq critical section to a short duration, allowing
lock_sock() users to make progress and preventing RCU stalls.
Signed-off-by: Dao Zhong Ma <cz1346219@gmail.com>
---
drivers/net/ovpn/tcp.c | 80 +++++++++++++++++++++++++++++++-----------
1 file changed, 59 insertions(+), 21 deletions(-)
diff --git a/drivers/net/ovpn/tcp.c b/drivers/net/ovpn/tcp.c
index 65054cc84be5..d75ad0c22a30 100644
--- a/drivers/net/ovpn/tcp.c
+++ b/drivers/net/ovpn/tcp.c
@@ -6,6 +6,7 @@
* Author: Antonio Quartulli <antonio@openvpn.net>
*/
+#include <linux/interrupt.h>
#include <linux/skbuff.h>
#include <net/hotdata.h>
#include <net/inet_common.h>
@@ -312,6 +313,40 @@ static void ovpn_tcp_send_sock(struct ovpn_peer *peer, struct sock *sk)
peer->tcp.tx_in_progress = false;
}
+/* Caller must hold sk->sk_lock.slock. */
+static bool ovpn_tcp_queue_skb(struct ovpn_peer *peer, struct sk_buff *skb)
+{
+ if (skb_queue_len(&peer->tcp.out_queue) >=
+ READ_ONCE(net_hotdata.max_backlog)) {
+ dev_dstats_tx_dropped(peer->ovpn->dev);
+ kfree_skb(skb);
+ return false;
+ }
+
+ __skb_queue_tail(&peer->tcp.out_queue, skb);
+ return true;
+}
+
+/* Caller must hold sk->sk_lock.slock and own the socket. */
+static void ovpn_tcp_tx_flush(struct ovpn_peer *peer, struct sock *sk)
+{
+ struct sk_buff *skb;
+
+ if (peer->tcp.out_msg.skb)
+ ovpn_tcp_send_sock(peer, sk);
+
+ while (!peer->tcp.out_msg.skb) {
+ skb = __skb_dequeue(&peer->tcp.out_queue);
+ if (!skb)
+ break;
+
+ peer->tcp.out_msg.skb = skb;
+ peer->tcp.out_msg.len = skb->len;
+ peer->tcp.out_msg.offset = 0;
+ ovpn_tcp_send_sock(peer, sk);
+ }
+}
+
void ovpn_tcp_tx_work(struct work_struct *work)
{
struct ovpn_socket *sock;
@@ -320,7 +355,7 @@ void ovpn_tcp_tx_work(struct work_struct *work)
lock_sock(sock->sk);
if (sock->peer)
- ovpn_tcp_send_sock(sock->peer, sock->sk);
+ ovpn_tcp_tx_flush(sock->peer, sock->sk);
release_sock(sock->sk);
}
@@ -345,32 +380,38 @@ static void ovpn_tcp_send_sock_skb(struct ovpn_peer *peer, struct sock *sk,
void ovpn_tcp_send_skb(struct ovpn_peer *peer, struct sock *sk,
struct sk_buff *skb)
{
+ struct ovpn_socket *sock;
u16 len = skb->len;
+ bool queued;
*(__be16 *)__skb_push(skb, sizeof(u16)) = htons(len);
- spin_lock_nested(&sk->sk_lock.slock, OVPN_TCP_DEPTH_NESTING);
- if (sock_owned_by_user(sk)) {
- if (skb_queue_len(&peer->tcp.out_queue) >=
- READ_ONCE(net_hotdata.max_backlog)) {
- dev_dstats_tx_dropped(peer->ovpn->dev);
- kfree_skb(skb);
- goto unlock;
- }
- __skb_queue_tail(&peer->tcp.out_queue, skb);
- } else {
- ovpn_tcp_send_sock_skb(peer, sk, skb);
+ if (unlikely(in_interrupt())) {
+ spin_lock_nested(&sk->sk_lock.slock, OVPN_TCP_DEPTH_NESTING);
+ queued = ovpn_tcp_queue_skb(peer, skb);
+ spin_unlock(&sk->sk_lock.slock);
+ if (!queued)
+ return;
+
+ rcu_read_lock();
+ sock = rcu_dereference_sk_user_data(sk);
+ if (sock)
+ schedule_work(&sock->tcp_tx_work);
+ rcu_read_unlock();
+ return;
}
-unlock:
- spin_unlock(&sk->sk_lock.slock);
+
+ lock_sock_nested(sk, OVPN_TCP_DEPTH_NESTING);
+ queued = ovpn_tcp_queue_skb(peer, skb);
+ if (queued)
+ ovpn_tcp_tx_flush(peer, sk);
+ release_sock(sk);
}
static void ovpn_tcp_release(struct sock *sk)
{
- struct sk_buff_head queue;
struct ovpn_socket *sock;
struct ovpn_peer *peer;
- struct sk_buff *skb;
rcu_read_lock();
sock = rcu_dereference_sk_user_data(sk);
@@ -390,11 +431,7 @@ static void ovpn_tcp_release(struct sock *sk)
}
rcu_read_unlock();
- __skb_queue_head_init(&queue);
- skb_queue_splice_init(&peer->tcp.out_queue, &queue);
-
- while ((skb = __skb_dequeue(&queue)))
- ovpn_tcp_send_sock_skb(peer, sk, skb);
+ ovpn_tcp_tx_flush(peer, sk);
peer->tcp.sk_cb.prot->release_cb(sk);
ovpn_peer_put(peer);
@@ -653,3 +690,4 @@ void __init ovpn_tcp_init(void)
&inet6_stream_ops);
#endif
}
+
--
2.54.0
^ permalink raw reply related [flat|nested] 4+ messages in thread
* Re: [PATCH 1/1] ovpn: tcp - defer TX from softirq to workqueue
2026-05-01 14:54 ` [PATCH 1/1] " Dao Zhong Ma
@ 2026-05-01 23:12 ` Jakub Kicinski
2026-05-03 21:14 ` Antonio Quartulli
1 sibling, 0 replies; 4+ messages in thread
From: Jakub Kicinski @ 2026-05-01 23:12 UTC (permalink / raw)
To: Dao Zhong Ma
Cc: linux-kernel, netdev, antonio, sd, andrew+netdev, davem, edumazet,
pabeni
On Fri, 1 May 2026 22:54:25 +0800 Dao Zhong Ma wrote:
> ovpn_tcp_send_skb() holds sk->sk_lock.slock while performing the full TCP
> send in softirq context. This can hold the spinlock for a long time
> (large skb), blocking lock_sock() users. This can starve the RCU GP
> kthread and trigger RCU stalls warnings and hung tasks.
>
> Defer the TCP send operation to process context:
> - In interrupt context, only enqueue the skb under the spinlock
> schedule tcp_tx_work.
> - In process context, dequeue and flush the send queue under lock_sock()
>
> This reduces the softirq critical section to a short duration, allowing
> lock_sock() users to make progress and preventing RCU stalls.
This appears to break the ovpn selftests
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [PATCH 1/1] ovpn: tcp - defer TX from softirq to workqueue
2026-05-01 14:54 ` [PATCH 1/1] " Dao Zhong Ma
2026-05-01 23:12 ` Jakub Kicinski
@ 2026-05-03 21:14 ` Antonio Quartulli
1 sibling, 0 replies; 4+ messages in thread
From: Antonio Quartulli @ 2026-05-03 21:14 UTC (permalink / raw)
To: Dao Zhong Ma, linux-kernel, netdev
Cc: sd, andrew+netdev, davem, edumazet, kuba, pabeni, Ralf Lici
Hi,
On 01/05/2026 16:54, Dao Zhong Ma wrote:
> ovpn_tcp_send_skb() holds sk->sk_lock.slock while performing the full TCP
> send in softirq context. This can hold the spinlock for a long time
> (large skb), blocking lock_sock() users. This can starve the RCU GP
Are you sure that the stall is triggered by just processing 1 large skb?
> kthread and trigger RCU stalls warnings and hung tasks.
Can you trigger the stall/warning easily?
Any steps to reproduce?
>
> Defer the TCP send operation to process context:
> - In interrupt context, only enqueue the skb under the spinlock
> schedule tcp_tx_work.
> - In process context, dequeue and flush the send queue under lock_sock()
>
> This reduces the softirq critical section to a short duration, allowing
> lock_sock() users to make progress and preventing RCU stalls.
>
> Signed-off-by: Dao Zhong Ma <cz1346219@gmail.com>
As Jakub pointed out, this is breaking the selftests, therefore you
should first double check what's going wrong.
On top of that, could you please run some performance tests? I think
this is going to hit quite a penalty.
Especially if you have multiple peers and you defer all their traffic to
the same workqueue.
Regards,
--
Antonio Quartulli
OpenVPN Inc.
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2026-05-03 21:14 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2026-05-01 14:54 [PATCH 0/1] ovpn: tcp - defer TX from softirq to workqueue Dao Zhong Ma
2026-05-01 14:54 ` [PATCH 1/1] " Dao Zhong Ma
2026-05-01 23:12 ` Jakub Kicinski
2026-05-03 21:14 ` Antonio Quartulli
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox