* [PATCH net v4 0/2] tcp: fix listener wakeup after reuseport migration
@ 2026-04-22 2:45 Zhenzhong Wu
2026-04-22 2:45 ` [PATCH net v4 1/2] tcp: call sk_data_ready() after listener migration Zhenzhong Wu
2026-04-22 2:45 ` [PATCH net v4 2/2] selftests/bpf: check epoll readiness during reuseport migration Zhenzhong Wu
0 siblings, 2 replies; 7+ messages in thread
From: Zhenzhong Wu @ 2026-04-22 2:45 UTC (permalink / raw)
To: netdev
Cc: edumazet, ncardwell, kuniyu, davem, dsahern, kuba, pabeni, horms,
shuah, tamird, linux-kernel, linux-kselftest, Zhenzhong Wu
This series fixes a missing wakeup when inet_csk_listen_stop() migrates
an established child socket from a closing listener to another socket
in the same SO_REUSEPORT group after the child has already been queued
for accept.
The target listener receives the migrated accept-queue entry via
inet_csk_reqsk_queue_add(), but its waiters are not notified.
Nonblocking accept() still succeeds because it checks the accept queue
directly, but readiness-based waiters can remain asleep until another
connection generates a wakeup.
Patch 1 notifies the target listener after a successful migration in
inet_csk_listen_stop() and protects the post-queue_add() nsk accesses
with rcu_read_lock()/rcu_read_unlock().
Patch 2 extends the existing migrate_reuseport BPF selftest with epoll
readiness checks inside migrate_dance(), around shutdown() where the
migration happens. The test now verifies that the target listener is
not ready before migration and becomes ready immediately after it, for
both TCP_ESTABLISHED and TCP_SYN_RECV. TCP_NEW_SYN_RECV remains
excluded because it still depends on later handshake completion.
Testing:
- On a local unpatched kernel, the focused migrate_reuseport test
fails for the listener-migration cases and passes for the
TCP_NEW_SYN_RECV cases:
not ok 1 IPv4 TCP_ESTABLISHED inet_csk_listen_stop
not ok 2 IPv4 TCP_SYN_RECV inet_csk_listen_stop
ok 3 IPv4 TCP_NEW_SYN_RECV reqsk_timer_handler
ok 4 IPv4 TCP_NEW_SYN_RECV inet_csk_complete_hashdance
not ok 5 IPv6 TCP_ESTABLISHED inet_csk_listen_stop
not ok 6 IPv6 TCP_SYN_RECV inet_csk_listen_stop
ok 7 IPv6 TCP_NEW_SYN_RECV reqsk_timer_handler
ok 8 IPv6 TCP_NEW_SYN_RECV inet_csk_complete_hashdance
- On a patched kernel booted under QEMU, the full migrate_reuseport
selftest passes:
ok 1 IPv4 TCP_ESTABLISHED inet_csk_listen_stop
ok 2 IPv4 TCP_SYN_RECV inet_csk_listen_stop
ok 3 IPv4 TCP_NEW_SYN_RECV reqsk_timer_handler
ok 4 IPv4 TCP_NEW_SYN_RECV inet_csk_complete_hashdance
ok 5 IPv6 TCP_ESTABLISHED inet_csk_listen_stop
ok 6 IPv6 TCP_SYN_RECV inet_csk_listen_stop
ok 7 IPv6 TCP_NEW_SYN_RECV reqsk_timer_handler
ok 8 IPv6 TCP_NEW_SYN_RECV inet_csk_complete_hashdance
SELFTEST_RC=0
---
v4:
- fix patch 2 to route shutdown() failures through the epoll cleanup path
- guard patch 2's epoll cleanup when no epoll fd was created
v3:
https://lore.kernel.org/netdev/20260421123106.142299-1-jt26wzz@gmail.com/
v2:
https://lore.kernel.org/netdev/20260418181333.1713389-1-jt26wzz@gmail.com/
v1:
https://lore.kernel.org/netdev/20260418041633.691435-1-jt26wzz@gmail.com/
Zhenzhong Wu (2):
tcp: call sk_data_ready() after listener migration
selftests/bpf: check epoll readiness during reuseport migration
net/ipv4/inet_connection_sock.c | 3 ++
.../bpf/prog_tests/migrate_reuseport.c | 49 ++++++++++++++++---
2 files changed, 45 insertions(+), 7 deletions(-)
base-commit: 52bcb57a4e8a0865a76c587c2451906342ae1b2d
--
2.43.0
^ permalink raw reply [flat|nested] 7+ messages in thread
* [PATCH net v4 1/2] tcp: call sk_data_ready() after listener migration
2026-04-22 2:45 [PATCH net v4 0/2] tcp: fix listener wakeup after reuseport migration Zhenzhong Wu
@ 2026-04-22 2:45 ` Zhenzhong Wu
2026-04-22 2:45 ` [PATCH net v4 2/2] selftests/bpf: check epoll readiness during reuseport migration Zhenzhong Wu
1 sibling, 0 replies; 7+ messages in thread
From: Zhenzhong Wu @ 2026-04-22 2:45 UTC (permalink / raw)
To: netdev
Cc: edumazet, ncardwell, kuniyu, davem, dsahern, kuba, pabeni, horms,
shuah, tamird, linux-kernel, linux-kselftest, Zhenzhong Wu,
stable
When inet_csk_listen_stop() migrates an established child socket from
a closing listener to another socket in the same SO_REUSEPORT group,
the target listener gets a new accept-queue entry via
inet_csk_reqsk_queue_add(), but that path never notifies the target
listener's waiters. A nonblocking accept() still works because it
checks the queue directly, but poll()/epoll_wait() waiters and
blocking accept() callers can also remain asleep indefinitely.
Call READ_ONCE(nsk->sk_data_ready)(nsk) after a successful migration
in inet_csk_listen_stop().
However, after inet_csk_reqsk_queue_add() succeeds, the ref acquired
in reuseport_migrate_sock() is effectively transferred to
nreq->rsk_listener. Another CPU can then dequeue nreq via accept()
or listener shutdown, hit reqsk_put(), and drop that listener ref.
Since listeners are SOCK_RCU_FREE, wrap the post-queue_add()
dereferences of nsk in rcu_read_lock()/rcu_read_unlock(), which also
covers the existing sock_net(nsk) access in that path.
The reqsk_timer_handler() path does not need the same changes for two
reasons: half-open requests become readable only after the final ACK,
where tcp_child_process() already wakes the listener; and once nreq is
visible via inet_ehash_insert(), the success path no longer touches
nsk directly.
Fixes: 54b92e841937 ("tcp: Migrate TCP_ESTABLISHED/TCP_SYN_RECV sockets in accept queues.")
Cc: stable@vger.kernel.org
Suggested-by: Eric Dumazet <edumazet@google.com>
Reviewed-by: Kuniyuki Iwashima <kuniyu@google.com>
Signed-off-by: Zhenzhong Wu <jt26wzz@gmail.com>
---
net/ipv4/inet_connection_sock.c | 3 +++
1 file changed, 3 insertions(+)
diff --git a/net/ipv4/inet_connection_sock.c b/net/ipv4/inet_connection_sock.c
index 4ac3ae1bc..928654c34 100644
--- a/net/ipv4/inet_connection_sock.c
+++ b/net/ipv4/inet_connection_sock.c
@@ -1479,16 +1479,19 @@ void inet_csk_listen_stop(struct sock *sk)
if (nreq) {
refcount_set(&nreq->rsk_refcnt, 1);
+ rcu_read_lock();
if (inet_csk_reqsk_queue_add(nsk, nreq, child)) {
__NET_INC_STATS(sock_net(nsk),
LINUX_MIB_TCPMIGRATEREQSUCCESS);
reqsk_migrate_reset(req);
+ READ_ONCE(nsk->sk_data_ready)(nsk);
} else {
__NET_INC_STATS(sock_net(nsk),
LINUX_MIB_TCPMIGRATEREQFAILURE);
reqsk_migrate_reset(nreq);
__reqsk_free(nreq);
}
+ rcu_read_unlock();
/* inet_csk_reqsk_queue_add() has already
* called inet_child_forget() on failure case.
--
2.43.0
^ permalink raw reply related [flat|nested] 7+ messages in thread
* [PATCH net v4 2/2] selftests/bpf: check epoll readiness during reuseport migration
2026-04-22 2:45 [PATCH net v4 0/2] tcp: fix listener wakeup after reuseport migration Zhenzhong Wu
2026-04-22 2:45 ` [PATCH net v4 1/2] tcp: call sk_data_ready() after listener migration Zhenzhong Wu
@ 2026-04-22 2:45 ` Zhenzhong Wu
2026-04-22 3:03 ` Tamir Duberstein
1 sibling, 1 reply; 7+ messages in thread
From: Zhenzhong Wu @ 2026-04-22 2:45 UTC (permalink / raw)
To: netdev
Cc: edumazet, ncardwell, kuniyu, davem, dsahern, kuba, pabeni, horms,
shuah, tamird, linux-kernel, linux-kselftest, Zhenzhong Wu
Inside migrate_dance(), add epoll checks around shutdown() to
verify that the target listener is not ready before shutdown()
and becomes ready immediately after shutdown() triggers migration.
Cover TCP_ESTABLISHED and TCP_SYN_RECV. Exclude TCP_NEW_SYN_RECV
as it depends on later handshake completion.
Suggested-by: Kuniyuki Iwashima <kuniyu@google.com>
Reviewed-by: Kuniyuki Iwashima <kuniyu@google.com>
Signed-off-by: Zhenzhong Wu <jt26wzz@gmail.com>
---
.../bpf/prog_tests/migrate_reuseport.c | 49 ++++++++++++++++---
1 file changed, 42 insertions(+), 7 deletions(-)
diff --git a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
index 653b0a20f..c62907732 100644
--- a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
+++ b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
@@ -7,24 +7,29 @@
* 3. call listen() for 1 server socket. (migration target)
* 4. update a map to migrate all child sockets
* to the last server socket (migrate_map[cookie] = 4)
- * 5. call shutdown() for first 4 server sockets
+ * 5. for TCP_ESTABLISHED and TCP_SYN_RECV cases, verify via epoll
+ * that the last server socket is not ready before migration.
+ * 6. call shutdown() for first 4 server sockets
* and migrate the requests in the accept queue
* to the last server socket.
- * 6. call listen() for the second server socket.
- * 7. call shutdown() for the last server
+ * 7. for TCP_ESTABLISHED and TCP_SYN_RECV cases, verify via epoll
+ * that the last server socket is ready after migration.
+ * 8. call listen() for the second server socket.
+ * 9. call shutdown() for the last server
* and migrate the requests in the accept queue
* to the second server socket.
- * 8. call listen() for the last server.
- * 9. call shutdown() for the second server
+ * 10. call listen() for the last server.
+ * 11. call shutdown() for the second server
* and migrate the requests in the accept queue
* to the last server socket.
- * 10. call accept() for the last server socket.
+ * 12. call accept() for the last server socket.
*
* Author: Kuniyuki Iwashima <kuniyu@amazon.co.jp>
*/
#include <bpf/bpf.h>
#include <bpf/libbpf.h>
+#include <sys/epoll.h>
#include "test_progs.h"
#include "test_migrate_reuseport.skel.h"
@@ -350,21 +355,51 @@ static int update_maps(struct migrate_reuseport_test_case *test_case,
static int migrate_dance(struct migrate_reuseport_test_case *test_case)
{
+ struct epoll_event ev = {
+ .events = EPOLLIN,
+ };
+ int epoll = -1, nfds;
int i, err;
+ if (test_case->state != BPF_TCP_NEW_SYN_RECV) {
+ epoll = epoll_create1(0);
+ if (!ASSERT_NEQ(epoll, -1, "epoll_create1"))
+ return -1;
+
+ ev.data.fd = test_case->servers[MIGRATED_TO];
+ if (!ASSERT_OK(epoll_ctl(epoll, EPOLL_CTL_ADD,
+ test_case->servers[MIGRATED_TO], &ev),
+ "epoll_ctl"))
+ goto close_epoll;
+
+ nfds = epoll_wait(epoll, &ev, 1, 0);
+ if (!ASSERT_EQ(nfds, 0, "epoll_wait 1"))
+ goto close_epoll;
+ }
+
/* Migrate TCP_ESTABLISHED and TCP_SYN_RECV requests
* to the last listener based on eBPF.
*/
for (i = 0; i < MIGRATED_TO; i++) {
err = shutdown(test_case->servers[i], SHUT_RDWR);
if (!ASSERT_OK(err, "shutdown"))
- return -1;
+ goto close_epoll;
}
/* No dance for TCP_NEW_SYN_RECV to migrate based on eBPF */
if (test_case->state == BPF_TCP_NEW_SYN_RECV)
return 0;
+ nfds = epoll_wait(epoll, &ev, 1, 0);
+ if (!ASSERT_EQ(nfds, 1, "epoll_wait 2")) {
+close_epoll:
+ if (epoll >= 0)
+ close(epoll);
+ return -1;
+ }
+
+ close(epoll);
+
/* Note that we use the second listener instead of the
* first one here.
*
--
2.43.0
^ permalink raw reply related [flat|nested] 7+ messages in thread
* Re: [PATCH net v4 2/2] selftests/bpf: check epoll readiness during reuseport migration
2026-04-22 2:45 ` [PATCH net v4 2/2] selftests/bpf: check epoll readiness during reuseport migration Zhenzhong Wu
@ 2026-04-22 3:03 ` Tamir Duberstein
2026-04-22 6:34 ` Zhenzhong Wu
0 siblings, 1 reply; 7+ messages in thread
From: Tamir Duberstein @ 2026-04-22 3:03 UTC (permalink / raw)
To: Zhenzhong Wu
Cc: netdev, edumazet, ncardwell, kuniyu, davem, dsahern, kuba, pabeni,
horms, shuah, linux-kernel, linux-kselftest
On Tue, Apr 21, 2026 at 7:46 PM Zhenzhong Wu <jt26wzz@gmail.com> wrote:
>
> Inside migrate_dance(), add epoll checks around shutdown() to
> verify that the target listener is not ready before shutdown()
> and becomes ready immediately after shutdown() triggers migration.
>
> Cover TCP_ESTABLISHED and TCP_SYN_RECV. Exclude TCP_NEW_SYN_RECV
> as it depends on later handshake completion.
>
> Suggested-by: Kuniyuki Iwashima <kuniyu@google.com>
> Reviewed-by: Kuniyuki Iwashima <kuniyu@google.com>
> Signed-off-by: Zhenzhong Wu <jt26wzz@gmail.com>
> ---
> .../bpf/prog_tests/migrate_reuseport.c | 49 ++++++++++++++++---
> 1 file changed, 42 insertions(+), 7 deletions(-)
>
> diff --git a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> index 653b0a20f..c62907732 100644
> --- a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> +++ b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> @@ -7,24 +7,29 @@
> * 3. call listen() for 1 server socket. (migration target)
> * 4. update a map to migrate all child sockets
> * to the last server socket (migrate_map[cookie] = 4)
> - * 5. call shutdown() for first 4 server sockets
> + * 5. for TCP_ESTABLISHED and TCP_SYN_RECV cases, verify via epoll
> + * that the last server socket is not ready before migration.
> + * 6. call shutdown() for first 4 server sockets
> * and migrate the requests in the accept queue
> * to the last server socket.
> - * 6. call listen() for the second server socket.
> - * 7. call shutdown() for the last server
> + * 7. for TCP_ESTABLISHED and TCP_SYN_RECV cases, verify via epoll
> + * that the last server socket is ready after migration.
> + * 8. call listen() for the second server socket.
> + * 9. call shutdown() for the last server
> * and migrate the requests in the accept queue
> * to the second server socket.
> - * 8. call listen() for the last server.
> - * 9. call shutdown() for the second server
> + * 10. call listen() for the last server.
> + * 11. call shutdown() for the second server
> * and migrate the requests in the accept queue
> * to the last server socket.
> - * 10. call accept() for the last server socket.
> + * 12. call accept() for the last server socket.
> *
> * Author: Kuniyuki Iwashima <kuniyu@amazon.co.jp>
> */
>
> #include <bpf/bpf.h>
> #include <bpf/libbpf.h>
> +#include <sys/epoll.h>
>
> #include "test_progs.h"
> #include "test_migrate_reuseport.skel.h"
> @@ -350,21 +355,51 @@ static int update_maps(struct migrate_reuseport_test_case *test_case,
>
> static int migrate_dance(struct migrate_reuseport_test_case *test_case)
> {
> + struct epoll_event ev = {
> + .events = EPOLLIN,
> + };
> + int epoll = -1, nfds;
> int i, err;
>
> + if (test_case->state != BPF_TCP_NEW_SYN_RECV) {
> + epoll = epoll_create1(0);
> + if (!ASSERT_NEQ(epoll, -1, "epoll_create1"))
> + return -1;
> +
> + ev.data.fd = test_case->servers[MIGRATED_TO];
> + if (!ASSERT_OK(epoll_ctl(epoll, EPOLL_CTL_ADD,
> + test_case->servers[MIGRATED_TO], &ev),
> + "epoll_ctl"))
> + goto close_epoll;
> +
> + nfds = epoll_wait(epoll, &ev, 1, 0);
> + if (!ASSERT_EQ(nfds, 0, "epoll_wait 1"))
> + goto close_epoll;
> + }
> +
> /* Migrate TCP_ESTABLISHED and TCP_SYN_RECV requests
> * to the last listener based on eBPF.
> */
> for (i = 0; i < MIGRATED_TO; i++) {
> err = shutdown(test_case->servers[i], SHUT_RDWR);
> if (!ASSERT_OK(err, "shutdown"))
> - return -1;
> + goto close_epoll;
> }
>
> /* No dance for TCP_NEW_SYN_RECV to migrate based on eBPF */
> if (test_case->state == BPF_TCP_NEW_SYN_RECV)
> return 0;
>
> + nfds = epoll_wait(epoll, &ev, 1, 0);
> + if (!ASSERT_EQ(nfds, 1, "epoll_wait 2")) {
> +close_epoll:
> + if (epoll >= 0)
> + close(epoll);
> + return -1;
> + }
I think you can avoid this goto using `__attribute__((cleanup(...))`:
static void close_fd(int *fd)
{
if (*fd >= 0)
close(*fd);
}
#define __close_fd __attribute__((cleanup(close_fd)))
then in the test:
__close_fd int epoll = -1;
there's an example of this in
tools/testing/selftests/bpf/prog_tests/socket_helpers.h.
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PATCH net v4 2/2] selftests/bpf: check epoll readiness during reuseport migration
2026-04-22 3:03 ` Tamir Duberstein
@ 2026-04-22 6:34 ` Zhenzhong Wu
2026-04-22 6:40 ` Kuniyuki Iwashima
0 siblings, 1 reply; 7+ messages in thread
From: Zhenzhong Wu @ 2026-04-22 6:34 UTC (permalink / raw)
To: Tamir Duberstein
Cc: netdev, edumazet, ncardwell, kuniyu, davem, dsahern, kuba, pabeni,
horms, shuah, linux-kernel, linux-kselftest
Thanks Tamir, I will fold this into v5 - there's already a precedent
in socket_helpers.h, so I will just include it from there.
On Wed, Apr 22, 2026 at 11:04 AM Tamir Duberstein <tamird@kernel.org> wrote:
>
> On Tue, Apr 21, 2026 at 7:46 PM Zhenzhong Wu <jt26wzz@gmail.com> wrote:
> >
> > Inside migrate_dance(), add epoll checks around shutdown() to
> > verify that the target listener is not ready before shutdown()
> > and becomes ready immediately after shutdown() triggers migration.
> >
> > Cover TCP_ESTABLISHED and TCP_SYN_RECV. Exclude TCP_NEW_SYN_RECV
> > as it depends on later handshake completion.
> >
> > Suggested-by: Kuniyuki Iwashima <kuniyu@google.com>
> > Reviewed-by: Kuniyuki Iwashima <kuniyu@google.com>
> > Signed-off-by: Zhenzhong Wu <jt26wzz@gmail.com>
> > ---
> > .../bpf/prog_tests/migrate_reuseport.c | 49 ++++++++++++++++---
> > 1 file changed, 42 insertions(+), 7 deletions(-)
> >
> > diff --git a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> > index 653b0a20f..c62907732 100644
> > --- a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> > +++ b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> > @@ -7,24 +7,29 @@
> > * 3. call listen() for 1 server socket. (migration target)
> > * 4. update a map to migrate all child sockets
> > * to the last server socket (migrate_map[cookie] = 4)
> > - * 5. call shutdown() for first 4 server sockets
> > + * 5. for TCP_ESTABLISHED and TCP_SYN_RECV cases, verify via epoll
> > + * that the last server socket is not ready before migration.
> > + * 6. call shutdown() for first 4 server sockets
> > * and migrate the requests in the accept queue
> > * to the last server socket.
> > - * 6. call listen() for the second server socket.
> > - * 7. call shutdown() for the last server
> > + * 7. for TCP_ESTABLISHED and TCP_SYN_RECV cases, verify via epoll
> > + * that the last server socket is ready after migration.
> > + * 8. call listen() for the second server socket.
> > + * 9. call shutdown() for the last server
> > * and migrate the requests in the accept queue
> > * to the second server socket.
> > - * 8. call listen() for the last server.
> > - * 9. call shutdown() for the second server
> > + * 10. call listen() for the last server.
> > + * 11. call shutdown() for the second server
> > * and migrate the requests in the accept queue
> > * to the last server socket.
> > - * 10. call accept() for the last server socket.
> > + * 12. call accept() for the last server socket.
> > *
> > * Author: Kuniyuki Iwashima <kuniyu@amazon.co.jp>
> > */
> >
> > #include <bpf/bpf.h>
> > #include <bpf/libbpf.h>
> > +#include <sys/epoll.h>
> >
> > #include "test_progs.h"
> > #include "test_migrate_reuseport.skel.h"
> > @@ -350,21 +355,51 @@ static int update_maps(struct migrate_reuseport_test_case *test_case,
> >
> > static int migrate_dance(struct migrate_reuseport_test_case *test_case)
> > {
> > + struct epoll_event ev = {
> > + .events = EPOLLIN,
> > + };
> > + int epoll = -1, nfds;
> > int i, err;
> >
> > + if (test_case->state != BPF_TCP_NEW_SYN_RECV) {
> > + epoll = epoll_create1(0);
> > + if (!ASSERT_NEQ(epoll, -1, "epoll_create1"))
> > + return -1;
> > +
> > + ev.data.fd = test_case->servers[MIGRATED_TO];
> > + if (!ASSERT_OK(epoll_ctl(epoll, EPOLL_CTL_ADD,
> > + test_case->servers[MIGRATED_TO], &ev),
> > + "epoll_ctl"))
> > + goto close_epoll;
> > +
> > + nfds = epoll_wait(epoll, &ev, 1, 0);
> > + if (!ASSERT_EQ(nfds, 0, "epoll_wait 1"))
> > + goto close_epoll;
> > + }
> > +
> > /* Migrate TCP_ESTABLISHED and TCP_SYN_RECV requests
> > * to the last listener based on eBPF.
> > */
> > for (i = 0; i < MIGRATED_TO; i++) {
> > err = shutdown(test_case->servers[i], SHUT_RDWR);
> > if (!ASSERT_OK(err, "shutdown"))
> > - return -1;
> > + goto close_epoll;
> > }
> >
> > /* No dance for TCP_NEW_SYN_RECV to migrate based on eBPF */
> > if (test_case->state == BPF_TCP_NEW_SYN_RECV)
> > return 0;
> >
> > + nfds = epoll_wait(epoll, &ev, 1, 0);
> > + if (!ASSERT_EQ(nfds, 1, "epoll_wait 2")) {
> > +close_epoll:
> > + if (epoll >= 0)
> > + close(epoll);
> > + return -1;
> > + }
>
> I think you can avoid this goto using `__attribute__((cleanup(...))`:
>
> static void close_fd(int *fd)
> {
> if (*fd >= 0)
> close(*fd);
> }
>
> #define __close_fd __attribute__((cleanup(close_fd)))
>
> then in the test:
>
> __close_fd int epoll = -1;
>
> there's an example of this in
> tools/testing/selftests/bpf/prog_tests/socket_helpers.h.
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PATCH net v4 2/2] selftests/bpf: check epoll readiness during reuseport migration
2026-04-22 6:34 ` Zhenzhong Wu
@ 2026-04-22 6:40 ` Kuniyuki Iwashima
2026-04-22 7:14 ` Zhenzhong Wu
0 siblings, 1 reply; 7+ messages in thread
From: Kuniyuki Iwashima @ 2026-04-22 6:40 UTC (permalink / raw)
To: Zhenzhong Wu
Cc: Tamir Duberstein, netdev, edumazet, ncardwell, davem, dsahern,
kuba, pabeni, horms, shuah, linux-kernel, linux-kselftest
On Tue, Apr 21, 2026 at 11:34 PM Zhenzhong Wu <jt26wzz@gmail.com> wrote:
>
> Thanks Tamir, I will fold this into v5 - there's already a precedent
> in socket_helpers.h, so I will just include it from there.
The test does not use cleanup in other places,
no need to mix them up.
>
> On Wed, Apr 22, 2026 at 11:04 AM Tamir Duberstein <tamird@kernel.org> wrote:
> >
> > On Tue, Apr 21, 2026 at 7:46 PM Zhenzhong Wu <jt26wzz@gmail.com> wrote:
> > >
> > > Inside migrate_dance(), add epoll checks around shutdown() to
> > > verify that the target listener is not ready before shutdown()
> > > and becomes ready immediately after shutdown() triggers migration.
> > >
> > > Cover TCP_ESTABLISHED and TCP_SYN_RECV. Exclude TCP_NEW_SYN_RECV
> > > as it depends on later handshake completion.
> > >
> > > Suggested-by: Kuniyuki Iwashima <kuniyu@google.com>
> > > Reviewed-by: Kuniyuki Iwashima <kuniyu@google.com>
> > > Signed-off-by: Zhenzhong Wu <jt26wzz@gmail.com>
> > > ---
> > > .../bpf/prog_tests/migrate_reuseport.c | 49 ++++++++++++++++---
> > > 1 file changed, 42 insertions(+), 7 deletions(-)
> > >
> > > diff --git a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> > > index 653b0a20f..c62907732 100644
> > > --- a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> > > +++ b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> > > @@ -7,24 +7,29 @@
> > > * 3. call listen() for 1 server socket. (migration target)
> > > * 4. update a map to migrate all child sockets
> > > * to the last server socket (migrate_map[cookie] = 4)
> > > - * 5. call shutdown() for first 4 server sockets
> > > + * 5. for TCP_ESTABLISHED and TCP_SYN_RECV cases, verify via epoll
> > > + * that the last server socket is not ready before migration.
> > > + * 6. call shutdown() for first 4 server sockets
> > > * and migrate the requests in the accept queue
> > > * to the last server socket.
> > > - * 6. call listen() for the second server socket.
> > > - * 7. call shutdown() for the last server
> > > + * 7. for TCP_ESTABLISHED and TCP_SYN_RECV cases, verify via epoll
> > > + * that the last server socket is ready after migration.
> > > + * 8. call listen() for the second server socket.
> > > + * 9. call shutdown() for the last server
> > > * and migrate the requests in the accept queue
> > > * to the second server socket.
> > > - * 8. call listen() for the last server.
> > > - * 9. call shutdown() for the second server
> > > + * 10. call listen() for the last server.
> > > + * 11. call shutdown() for the second server
> > > * and migrate the requests in the accept queue
> > > * to the last server socket.
> > > - * 10. call accept() for the last server socket.
> > > + * 12. call accept() for the last server socket.
> > > *
> > > * Author: Kuniyuki Iwashima <kuniyu@amazon.co.jp>
> > > */
> > >
> > > #include <bpf/bpf.h>
> > > #include <bpf/libbpf.h>
> > > +#include <sys/epoll.h>
> > >
> > > #include "test_progs.h"
> > > #include "test_migrate_reuseport.skel.h"
> > > @@ -350,21 +355,51 @@ static int update_maps(struct migrate_reuseport_test_case *test_case,
> > >
> > > static int migrate_dance(struct migrate_reuseport_test_case *test_case)
> > > {
> > > + struct epoll_event ev = {
> > > + .events = EPOLLIN,
> > > + };
> > > + int epoll = -1, nfds;
> > > int i, err;
> > >
> > > + if (test_case->state != BPF_TCP_NEW_SYN_RECV) {
> > > + epoll = epoll_create1(0);
> > > + if (!ASSERT_NEQ(epoll, -1, "epoll_create1"))
> > > + return -1;
> > > +
> > > + ev.data.fd = test_case->servers[MIGRATED_TO];
> > > + if (!ASSERT_OK(epoll_ctl(epoll, EPOLL_CTL_ADD,
> > > + test_case->servers[MIGRATED_TO], &ev),
> > > + "epoll_ctl"))
> > > + goto close_epoll;
> > > +
> > > + nfds = epoll_wait(epoll, &ev, 1, 0);
> > > + if (!ASSERT_EQ(nfds, 0, "epoll_wait 1"))
> > > + goto close_epoll;
> > > + }
> > > +
> > > /* Migrate TCP_ESTABLISHED and TCP_SYN_RECV requests
> > > * to the last listener based on eBPF.
> > > */
> > > for (i = 0; i < MIGRATED_TO; i++) {
> > > err = shutdown(test_case->servers[i], SHUT_RDWR);
> > > if (!ASSERT_OK(err, "shutdown"))
> > > - return -1;
> > > + goto close_epoll;
> > > }
> > >
> > > /* No dance for TCP_NEW_SYN_RECV to migrate based on eBPF */
> > > if (test_case->state == BPF_TCP_NEW_SYN_RECV)
> > > return 0;
> > >
> > > + nfds = epoll_wait(epoll, &ev, 1, 0);
> > > + if (!ASSERT_EQ(nfds, 1, "epoll_wait 2")) {
> > > +close_epoll:
> > > + if (epoll >= 0)
> > > + close(epoll);
> > > + return -1;
> > > + }
> >
> > I think you can avoid this goto using `__attribute__((cleanup(...))`:
> >
> > static void close_fd(int *fd)
> > {
> > if (*fd >= 0)
> > close(*fd);
> > }
> >
> > #define __close_fd __attribute__((cleanup(close_fd)))
> >
> > then in the test:
> >
> > __close_fd int epoll = -1;
> >
> > there's an example of this in
> > tools/testing/selftests/bpf/prog_tests/socket_helpers.h.
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PATCH net v4 2/2] selftests/bpf: check epoll readiness during reuseport migration
2026-04-22 6:40 ` Kuniyuki Iwashima
@ 2026-04-22 7:14 ` Zhenzhong Wu
0 siblings, 0 replies; 7+ messages in thread
From: Zhenzhong Wu @ 2026-04-22 7:14 UTC (permalink / raw)
To: Kuniyuki Iwashima
Cc: Tamir Duberstein, netdev, edumazet, ncardwell, davem, dsahern,
kuba, pabeni, horms, shuah, linux-kernel, linux-kselftest
Thanks both. Keeping the goto for consistency with the rest of the test.
On Wed, Apr 22, 2026 at 2:40 PM Kuniyuki Iwashima <kuniyu@google.com> wrote:
>
> On Tue, Apr 21, 2026 at 11:34 PM Zhenzhong Wu <jt26wzz@gmail.com> wrote:
> >
> > Thanks Tamir, I will fold this into v5 - there's already a precedent
> > in socket_helpers.h, so I will just include it from there.
>
> The test does not use cleanup in other places,
> no need to mix them up.
>
>
> >
> > On Wed, Apr 22, 2026 at 11:04 AM Tamir Duberstein <tamird@kernel.org> wrote:
> > >
> > > On Tue, Apr 21, 2026 at 7:46 PM Zhenzhong Wu <jt26wzz@gmail.com> wrote:
> > > >
> > > > Inside migrate_dance(), add epoll checks around shutdown() to
> > > > verify that the target listener is not ready before shutdown()
> > > > and becomes ready immediately after shutdown() triggers migration.
> > > >
> > > > Cover TCP_ESTABLISHED and TCP_SYN_RECV. Exclude TCP_NEW_SYN_RECV
> > > > as it depends on later handshake completion.
> > > >
> > > > Suggested-by: Kuniyuki Iwashima <kuniyu@google.com>
> > > > Reviewed-by: Kuniyuki Iwashima <kuniyu@google.com>
> > > > Signed-off-by: Zhenzhong Wu <jt26wzz@gmail.com>
> > > > ---
> > > > .../bpf/prog_tests/migrate_reuseport.c | 49 ++++++++++++++++---
> > > > 1 file changed, 42 insertions(+), 7 deletions(-)
> > > >
> > > > diff --git a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> > > > index 653b0a20f..c62907732 100644
> > > > --- a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> > > > +++ b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
> > > > @@ -7,24 +7,29 @@
> > > > * 3. call listen() for 1 server socket. (migration target)
> > > > * 4. update a map to migrate all child sockets
> > > > * to the last server socket (migrate_map[cookie] = 4)
> > > > - * 5. call shutdown() for first 4 server sockets
> > > > + * 5. for TCP_ESTABLISHED and TCP_SYN_RECV cases, verify via epoll
> > > > + * that the last server socket is not ready before migration.
> > > > + * 6. call shutdown() for first 4 server sockets
> > > > * and migrate the requests in the accept queue
> > > > * to the last server socket.
> > > > - * 6. call listen() for the second server socket.
> > > > - * 7. call shutdown() for the last server
> > > > + * 7. for TCP_ESTABLISHED and TCP_SYN_RECV cases, verify via epoll
> > > > + * that the last server socket is ready after migration.
> > > > + * 8. call listen() for the second server socket.
> > > > + * 9. call shutdown() for the last server
> > > > * and migrate the requests in the accept queue
> > > > * to the second server socket.
> > > > - * 8. call listen() for the last server.
> > > > - * 9. call shutdown() for the second server
> > > > + * 10. call listen() for the last server.
> > > > + * 11. call shutdown() for the second server
> > > > * and migrate the requests in the accept queue
> > > > * to the last server socket.
> > > > - * 10. call accept() for the last server socket.
> > > > + * 12. call accept() for the last server socket.
> > > > *
> > > > * Author: Kuniyuki Iwashima <kuniyu@amazon.co.jp>
> > > > */
> > > >
> > > > #include <bpf/bpf.h>
> > > > #include <bpf/libbpf.h>
> > > > +#include <sys/epoll.h>
> > > >
> > > > #include "test_progs.h"
> > > > #include "test_migrate_reuseport.skel.h"
> > > > @@ -350,21 +355,51 @@ static int update_maps(struct migrate_reuseport_test_case *test_case,
> > > >
> > > > static int migrate_dance(struct migrate_reuseport_test_case *test_case)
> > > > {
> > > > + struct epoll_event ev = {
> > > > + .events = EPOLLIN,
> > > > + };
> > > > + int epoll = -1, nfds;
> > > > int i, err;
> > > >
> > > > + if (test_case->state != BPF_TCP_NEW_SYN_RECV) {
> > > > + epoll = epoll_create1(0);
> > > > + if (!ASSERT_NEQ(epoll, -1, "epoll_create1"))
> > > > + return -1;
> > > > +
> > > > + ev.data.fd = test_case->servers[MIGRATED_TO];
> > > > + if (!ASSERT_OK(epoll_ctl(epoll, EPOLL_CTL_ADD,
> > > > + test_case->servers[MIGRATED_TO], &ev),
> > > > + "epoll_ctl"))
> > > > + goto close_epoll;
> > > > +
> > > > + nfds = epoll_wait(epoll, &ev, 1, 0);
> > > > + if (!ASSERT_EQ(nfds, 0, "epoll_wait 1"))
> > > > + goto close_epoll;
> > > > + }
> > > > +
> > > > /* Migrate TCP_ESTABLISHED and TCP_SYN_RECV requests
> > > > * to the last listener based on eBPF.
> > > > */
> > > > for (i = 0; i < MIGRATED_TO; i++) {
> > > > err = shutdown(test_case->servers[i], SHUT_RDWR);
> > > > if (!ASSERT_OK(err, "shutdown"))
> > > > - return -1;
> > > > + goto close_epoll;
> > > > }
> > > >
> > > > /* No dance for TCP_NEW_SYN_RECV to migrate based on eBPF */
> > > > if (test_case->state == BPF_TCP_NEW_SYN_RECV)
> > > > return 0;
> > > >
> > > > + nfds = epoll_wait(epoll, &ev, 1, 0);
> > > > + if (!ASSERT_EQ(nfds, 1, "epoll_wait 2")) {
> > > > +close_epoll:
> > > > + if (epoll >= 0)
> > > > + close(epoll);
> > > > + return -1;
> > > > + }
> > >
> > > I think you can avoid this goto using `__attribute__((cleanup(...))`:
> > >
> > > static void close_fd(int *fd)
> > > {
> > > if (*fd >= 0)
> > > close(*fd);
> > > }
> > >
> > > #define __close_fd __attribute__((cleanup(close_fd)))
> > >
> > > then in the test:
> > >
> > > __close_fd int epoll = -1;
> > >
> > > there's an example of this in
> > > tools/testing/selftests/bpf/prog_tests/socket_helpers.h.
^ permalink raw reply [flat|nested] 7+ messages in thread
end of thread, other threads:[~2026-04-22 7:14 UTC | newest]
Thread overview: 7+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2026-04-22 2:45 [PATCH net v4 0/2] tcp: fix listener wakeup after reuseport migration Zhenzhong Wu
2026-04-22 2:45 ` [PATCH net v4 1/2] tcp: call sk_data_ready() after listener migration Zhenzhong Wu
2026-04-22 2:45 ` [PATCH net v4 2/2] selftests/bpf: check epoll readiness during reuseport migration Zhenzhong Wu
2026-04-22 3:03 ` Tamir Duberstein
2026-04-22 6:34 ` Zhenzhong Wu
2026-04-22 6:40 ` Kuniyuki Iwashima
2026-04-22 7:14 ` Zhenzhong Wu
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox