* [PATCH] selftests/net: deflake GRO tests and fix return value and output
@ 2025-02-18 16:45 Kevin Krakauer
2025-02-21 1:04 ` Jakub Kicinski
2025-02-25 13:16 ` Petr Machata
0 siblings, 2 replies; 6+ messages in thread
From: Kevin Krakauer @ 2025-02-18 16:45 UTC (permalink / raw)
To: netdev, linux-kselftest
Cc: David S. Miller, Eric Dumazet, Jakub Kicinski, Paolo Abeni,
Simon Horman, Shuah Khan, linux-kernel, Kevin Krakauer
GRO tests are timing dependent and can easily flake. This is partially
mitigated in gro.sh by giving each subtest 3 chances to pass. However,
this still flakes on some machines.
Set the device's napi_defer_hard_irqs to 50 so that GRO is less likely
to immediately flush. This already happened in setup_loopback.sh, but
wasn't added to setup_veth.sh. This accounts for most of the reduction
in flakiness.
We also increase the number of chances for success from 3 to 6.
`gro.sh -t <test>` now returns a passing/failing exit code as expected.
gro.c:main no longer erroneously claims a test passes when running as a
server.
Tested: Ran `gro.sh -t large` 100 times with and without this change.
Passed 100/100 with and 64/100 without. Ran inside strace to increase
flakiness.
Signed-off-by: Kevin Krakauer <krakauer@google.com>
---
tools/testing/selftests/net/gro.c | 8 +++++---
tools/testing/selftests/net/gro.sh | 5 +++--
tools/testing/selftests/net/setup_veth.sh | 1 +
3 files changed, 9 insertions(+), 5 deletions(-)
diff --git a/tools/testing/selftests/net/gro.c b/tools/testing/selftests/net/gro.c
index b2184847e388..d5824eadea10 100644
--- a/tools/testing/selftests/net/gro.c
+++ b/tools/testing/selftests/net/gro.c
@@ -1318,11 +1318,13 @@ int main(int argc, char **argv)
read_MAC(src_mac, smac);
read_MAC(dst_mac, dmac);
- if (tx_socket)
+ if (tx_socket) {
gro_sender();
- else
+ } else {
+ /* Only the receiver exit status determines test success. */
gro_receiver();
+ fprintf(stderr, "Gro::%s test passed.\n", testname);
+ }
- fprintf(stderr, "Gro::%s test passed.\n", testname);
return 0;
}
diff --git a/tools/testing/selftests/net/gro.sh b/tools/testing/selftests/net/gro.sh
index 02c21ff4ca81..703173f8c8a9 100755
--- a/tools/testing/selftests/net/gro.sh
+++ b/tools/testing/selftests/net/gro.sh
@@ -21,7 +21,7 @@ run_test() {
# Each test is run 3 times to deflake, because given the receive timing,
# not all packets that should coalesce will be considered in the same flow
# on every try.
- for tries in {1..3}; do
+ for tries in {1..6}; do
# Actual test starts here
ip netns exec $server_ns ./gro "${ARGS[@]}" "--rx" "--iface" "server" \
1>>log.txt &
@@ -100,5 +100,6 @@ trap cleanup EXIT
if [[ "${test}" == "all" ]]; then
run_all_tests
else
- run_test "${proto}" "${test}"
+ exit_code=$(run_test "${proto}" "${test}")
+ exit $exit_code
fi;
diff --git a/tools/testing/selftests/net/setup_veth.sh b/tools/testing/selftests/net/setup_veth.sh
index 1f78a87f6f37..9882ad730c24 100644
--- a/tools/testing/selftests/net/setup_veth.sh
+++ b/tools/testing/selftests/net/setup_veth.sh
@@ -12,6 +12,7 @@ setup_veth_ns() {
[[ -e /var/run/netns/"${ns_name}" ]] || ip netns add "${ns_name}"
echo 1000000 > "/sys/class/net/${ns_dev}/gro_flush_timeout"
+ echo 50 > "/sys/class/net/${ns_dev}/napi_defer_hard_irqs"
ip link set dev "${ns_dev}" netns "${ns_name}" mtu 65535
ip -netns "${ns_name}" link set dev "${ns_dev}" up
--
2.48.1
^ permalink raw reply related [flat|nested] 6+ messages in thread
* Re: [PATCH] selftests/net: deflake GRO tests and fix return value and output
2025-02-18 16:45 [PATCH] selftests/net: deflake GRO tests and fix return value and output Kevin Krakauer
@ 2025-02-21 1:04 ` Jakub Kicinski
2025-02-23 15:19 ` Kevin Krakauer
2025-02-25 13:16 ` Petr Machata
1 sibling, 1 reply; 6+ messages in thread
From: Jakub Kicinski @ 2025-02-21 1:04 UTC (permalink / raw)
To: Kevin Krakauer
Cc: netdev, linux-kselftest, David S. Miller, Eric Dumazet,
Paolo Abeni, Simon Horman, Shuah Khan, linux-kernel
On Tue, 18 Feb 2025 08:45:55 -0800 Kevin Krakauer wrote:
> GRO tests are timing dependent and can easily flake. This is partially
> mitigated in gro.sh by giving each subtest 3 chances to pass. However,
> this still flakes on some machines.
To be clear - are you running this over veth or a real device?
> Set the device's napi_defer_hard_irqs to 50 so that GRO is less likely
> to immediately flush. This already happened in setup_loopback.sh, but
> wasn't added to setup_veth.sh. This accounts for most of the reduction
> in flakiness.
That doesn't make intuitive sense to me. If we already defer flushes
why do we need to also defer IRQs?
> We also increase the number of chances for success from 3 to 6.
>
> `gro.sh -t <test>` now returns a passing/failing exit code as expected.
>
> gro.c:main no longer erroneously claims a test passes when running as a
> server.
>
> Tested: Ran `gro.sh -t large` 100 times with and without this change.
> Passed 100/100 with and 64/100 without. Ran inside strace to increase
> flakiness.
>
> Signed-off-by: Kevin Krakauer <krakauer@google.com>
> ---
> tools/testing/selftests/net/gro.c | 8 +++++---
> tools/testing/selftests/net/gro.sh | 5 +++--
> tools/testing/selftests/net/setup_veth.sh | 1 +
> 3 files changed, 9 insertions(+), 5 deletions(-)
>
> diff --git a/tools/testing/selftests/net/gro.c b/tools/testing/selftests/net/gro.c
> index b2184847e388..d5824eadea10 100644
> --- a/tools/testing/selftests/net/gro.c
> +++ b/tools/testing/selftests/net/gro.c
> @@ -1318,11 +1318,13 @@ int main(int argc, char **argv)
> read_MAC(src_mac, smac);
> read_MAC(dst_mac, dmac);
>
> - if (tx_socket)
> + if (tx_socket) {
> gro_sender();
> - else
> + } else {
> + /* Only the receiver exit status determines test success. */
> gro_receiver();
> + fprintf(stderr, "Gro::%s test passed.\n", testname);
> + }
>
> - fprintf(stderr, "Gro::%s test passed.\n", testname);
That seems quite separate to the stability fix?
> return 0;
> }
> diff --git a/tools/testing/selftests/net/gro.sh b/tools/testing/selftests/net/gro.sh
> index 02c21ff4ca81..703173f8c8a9 100755
> --- a/tools/testing/selftests/net/gro.sh
> +++ b/tools/testing/selftests/net/gro.sh
> @@ -21,7 +21,7 @@ run_test() {
> # Each test is run 3 times to deflake, because given the receive timing,
> # not all packets that should coalesce will be considered in the same flow
> # on every try.
> - for tries in {1..3}; do
> + for tries in {1..6}; do
> # Actual test starts here
> ip netns exec $server_ns ./gro "${ARGS[@]}" "--rx" "--iface" "server" \
> 1>>log.txt &
> @@ -100,5 +100,6 @@ trap cleanup EXIT
> if [[ "${test}" == "all" ]]; then
> run_all_tests
> else
> - run_test "${proto}" "${test}"
> + exit_code=$(run_test "${proto}" "${test}")
> + exit $exit_code
Also separate from stability?
Let's split the patch up into logically separate changes.
> fi;
> diff --git a/tools/testing/selftests/net/setup_veth.sh b/tools/testing/selftests/net/setup_veth.sh
> index 1f78a87f6f37..9882ad730c24 100644
> --- a/tools/testing/selftests/net/setup_veth.sh
> +++ b/tools/testing/selftests/net/setup_veth.sh
> @@ -12,6 +12,7 @@ setup_veth_ns() {
>
> [[ -e /var/run/netns/"${ns_name}" ]] || ip netns add "${ns_name}"
> echo 1000000 > "/sys/class/net/${ns_dev}/gro_flush_timeout"
> + echo 50 > "/sys/class/net/${ns_dev}/napi_defer_hard_irqs"
> ip link set dev "${ns_dev}" netns "${ns_name}" mtu 65535
> ip -netns "${ns_name}" link set dev "${ns_dev}" up
>
--
pw-bot: cr
^ permalink raw reply [flat|nested] 6+ messages in thread
* [PATCH] selftests/net: deflake GRO tests and fix return value and output
2025-02-21 1:04 ` Jakub Kicinski
@ 2025-02-23 15:19 ` Kevin Krakauer
2025-02-24 20:48 ` Jakub Kicinski
0 siblings, 1 reply; 6+ messages in thread
From: Kevin Krakauer @ 2025-02-23 15:19 UTC (permalink / raw)
To: kuba
Cc: davem, edumazet, horms, krakauer, linux-kernel, linux-kselftest,
netdev, pabeni, shuah
Thanks for the review! I'll split this up. Do you think it's better as two
patchsets -- one for stability/deflaking, one for return value and output
cleanup -- or as a single patchset with several commits?
> To be clear - are you running this over veth or a real device?
Over a veth.
>> Set the device's napi_defer_hard_irqs to 50 so that GRO is less likely
>> to immediately flush. This already happened in setup_loopback.sh, but
>> wasn't added to setup_veth.sh. This accounts for most of the reduction
>> in flakiness.
>
>That doesn't make intuitive sense to me. If we already defer flushes
>why do we need to also defer IRQs?
Yep, the behavior here is weird. I ran `gro.sh -t large` 1000 times with each of
the following setups (all inside strace to increase flakiness):
- gro_flush_timeout=1ms, napi_defer_hard_irqs=0 --> failed to GRO 29 times
- gro_flush_timeout=5ms, napi_defer_hard_irqs=0 --> failed to GRO 45 times
- gro_flush_timeout=50ms, napi_defer_hard_irqs=0 --> failed to GRO 35 times
- gro_flush_timeout=1ms, napi_defer_hard_irqs=1 --> failed to GRO 0 times
- gro_flush_timeout=1ms, napi_defer_hard_irqs=50 --> failed to GRO 0 times
napi_defer_hard_irqs is clearly having an effect. And deferring once is enough.
I believe that deferring IRQs prevents anything else from causing a GRO flush
before gro_flush_timeout expires. While waiting for the timeout to expire, an
incoming packet can cause napi_complete_done and thus napi_gro_flush to run.
Outgoing packets from the veth can also cause this: veth_xmit calls
__veth_xdp_flush, which only actually does anything when IRQs are enabled.
So napi_defer_hard_irqs=1 seems sufficient to allow the full gro_flush_timeout
to expire before flushing GRO.
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [PATCH] selftests/net: deflake GRO tests and fix return value and output
2025-02-23 15:19 ` Kevin Krakauer
@ 2025-02-24 20:48 ` Jakub Kicinski
2025-02-25 20:05 ` Kevin Krakauer
0 siblings, 1 reply; 6+ messages in thread
From: Jakub Kicinski @ 2025-02-24 20:48 UTC (permalink / raw)
To: Kevin Krakauer
Cc: davem, edumazet, horms, linux-kernel, linux-kselftest, netdev,
pabeni, shuah
On Sun, 23 Feb 2025 07:19:49 -0800 Kevin Krakauer wrote:
> Thanks for the review! I'll split this up. Do you think it's better as two
> patchsets -- one for stability/deflaking, one for return value and output
> cleanup -- or as a single patchset with several commits?
Should be fine either way, they will both end up in net-next.
One patchset may be easier to merge, as we can't CI-test two
conflicting series on the list.
> > To be clear - are you running this over veth or a real device?
>
> Over a veth.
>
> >> Set the device's napi_defer_hard_irqs to 50 so that GRO is less likely
> >> to immediately flush. This already happened in setup_loopback.sh, but
> >> wasn't added to setup_veth.sh. This accounts for most of the reduction
> >> in flakiness.
> >
> >That doesn't make intuitive sense to me. If we already defer flushes
> >why do we need to also defer IRQs?
>
> Yep, the behavior here is weird. I ran `gro.sh -t large` 1000 times with each of
> the following setups (all inside strace to increase flakiness):
>
> - gro_flush_timeout=1ms, napi_defer_hard_irqs=0 --> failed to GRO 29 times
> - gro_flush_timeout=5ms, napi_defer_hard_irqs=0 --> failed to GRO 45 times
> - gro_flush_timeout=50ms, napi_defer_hard_irqs=0 --> failed to GRO 35 times
> - gro_flush_timeout=1ms, napi_defer_hard_irqs=1 --> failed to GRO 0 times
> - gro_flush_timeout=1ms, napi_defer_hard_irqs=50 --> failed to GRO 0 times
>
> napi_defer_hard_irqs is clearly having an effect. And deferring once is enough.
> I believe that deferring IRQs prevents anything else from causing a GRO flush
> before gro_flush_timeout expires. While waiting for the timeout to expire, an
> incoming packet can cause napi_complete_done and thus napi_gro_flush to run.
> Outgoing packets from the veth can also cause this: veth_xmit calls
> __veth_xdp_flush, which only actually does anything when IRQs are enabled.
>
> So napi_defer_hard_irqs=1 seems sufficient to allow the full gro_flush_timeout
> to expire before flushing GRO.
With msec-long deferrals we'll flush due to jiffies change. At least
that explains a bit. Could you maybe try lower timeouts than 1msec?
Previously we'd just keep partially-completed packets in GRO for up
to 1msec, now we'll delay all packet processing for 1msec, that's a lot.
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [PATCH] selftests/net: deflake GRO tests and fix return value and output
2025-02-18 16:45 [PATCH] selftests/net: deflake GRO tests and fix return value and output Kevin Krakauer
2025-02-21 1:04 ` Jakub Kicinski
@ 2025-02-25 13:16 ` Petr Machata
1 sibling, 0 replies; 6+ messages in thread
From: Petr Machata @ 2025-02-25 13:16 UTC (permalink / raw)
To: Kevin Krakauer
Cc: netdev, linux-kselftest, David S. Miller, Eric Dumazet,
Jakub Kicinski, Paolo Abeni, Simon Horman, Shuah Khan,
linux-kernel
Kevin Krakauer <krakauer@google.com> writes:
> diff --git a/tools/testing/selftests/net/gro.sh b/tools/testing/selftests/net/gro.sh
> index 02c21ff4ca81..703173f8c8a9 100755
> --- a/tools/testing/selftests/net/gro.sh
> +++ b/tools/testing/selftests/net/gro.sh
> @@ -21,7 +21,7 @@ run_test() {
> # Each test is run 3 times to deflake, because given the receive timing,
- # Each test is run 3 times to deflake, because given the receive timing,
+ # Each test is run 6 times to deflake, because given the receive timing,
> # not all packets that should coalesce will be considered in the same flow
> # on every try.
> - for tries in {1..3}; do
> + for tries in {1..6}; do
> # Actual test starts here
> ip netns exec $server_ns ./gro "${ARGS[@]}" "--rx" "--iface" "server" \
> 1>>log.txt &
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [PATCH] selftests/net: deflake GRO tests and fix return value and output
2025-02-24 20:48 ` Jakub Kicinski
@ 2025-02-25 20:05 ` Kevin Krakauer
0 siblings, 0 replies; 6+ messages in thread
From: Kevin Krakauer @ 2025-02-25 20:05 UTC (permalink / raw)
To: Jakub Kicinski
Cc: davem, edumazet, horms, linux-kernel, linux-kselftest, netdev,
pabeni, shuah
On Mon, Feb 24, 2025 at 12:48:30PM -0800, Jakub Kicinski wrote:
> With msec-long deferrals we'll flush due to jiffies change. At least
> that explains a bit. Could you maybe try lower timeouts than 1msec?
> Previously we'd just keep partially-completed packets in GRO for up
> to 1msec, now we'll delay all packet processing for 1msec, that's a lot.
Results again with each test run 1000 times:
gro_flush_timeout=50us napi_defer_hard_irqs=1 --> failed to GRO 0 times
gro_flush_timeout=100us napi_defer_hard_irqs=1 --> failed to GRO 0 times
gro_flush_timeout=50us napi_defer_hard_irqs=0 --> failed to GRO 36 times
gro_flush_timeout=100us napi_defer_hard_irqs=0 --> failed to GRO 46 times
100us with 1 defer seems to work fine and is well below the duration of
a jiffy. So we'll usually be testing the "default" GRO path and only
occasionally the jiffy-update path. I'll make these the numbers in the
revised patch unless someone thinks otherwise.
^ permalink raw reply [flat|nested] 6+ messages in thread
end of thread, other threads:[~2025-02-25 20:05 UTC | newest]
Thread overview: 6+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2025-02-18 16:45 [PATCH] selftests/net: deflake GRO tests and fix return value and output Kevin Krakauer
2025-02-21 1:04 ` Jakub Kicinski
2025-02-23 15:19 ` Kevin Krakauer
2025-02-24 20:48 ` Jakub Kicinski
2025-02-25 20:05 ` Kevin Krakauer
2025-02-25 13:16 ` Petr Machata
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).