From: Yonghong Song <yhs@fb.com>
To: Mykola Lysenko <mykolal@fb.com>
Cc: Andrii Nakryiko <andrii.nakryiko@gmail.com>,
bpf <bpf@vger.kernel.org>, Alexei Starovoitov <ast@kernel.org>,
Andrii Nakryiko <andrii@kernel.org>,
Daniel Borkmann <daniel@iogearbox.net>
Subject: Re: [PATCH bpf-next] Improve BPF test stability (related to perf events and scheduling)
Date: Tue, 1 Mar 2022 20:53:21 -0800 [thread overview]
Message-ID: <ab6970fc-eaa0-9462-9614-893085c42331@fb.com> (raw)
In-Reply-To: <8A284AAF-84BA-48A4-BF4D-7BDB2426DAB8@fb.com>
On 2/28/22 7:45 PM, Mykola Lysenko wrote:
>
>
>> On Feb 22, 2022, at 8:32 PM, Yonghong Song <yhs@fb.com> wrote:
>>
>>
>>
>> On 2/22/22 7:13 PM, Andrii Nakryiko wrote:
>>> On Tue, Feb 22, 2022 at 12:35 PM Mykola Lysenko <mykolal@fb.com> wrote:
>>>>
>>>> Thanks for the review Andrii!
>>>>
>>>>> On Feb 19, 2022, at 8:39 PM, Andrii Nakryiko <andrii.nakryiko@gmail.com> wrote:
>>>>>
>>>>> On Fri, Feb 18, 2022 at 4:30 PM Mykola Lysenko <mykolal@fb.com> wrote:
>>>>>>
>>>>>> In send_signal, replace sleep with dummy cpu intensive computation
>>>>>> to increase probability of child process being scheduled. Add few
>>>>>> more asserts.
>>>>>>
>>>>>> In find_vma, reduce sample_freq as higher values may be rejected in
>>>>>> some qemu setups, remove usleep and increase length of cpu intensive
>>>>>> computation.
>>>>>>
>>>>>> In bpf_cookie, perf_link and perf_branches, reduce sample_freq as
>>>>>> higher values may be rejected in some qemu setups
>>>>>>
>>>>>> Signed-off-by: Mykola Lysenko <mykolal@fb.com>
>>>>>> ---
>>>>>> .../testing/selftests/bpf/prog_tests/bpf_cookie.c | 2 +-
>>>>>> tools/testing/selftests/bpf/prog_tests/find_vma.c | 5 ++---
>>>>>> .../selftests/bpf/prog_tests/perf_branches.c | 4 ++--
>>>>>> tools/testing/selftests/bpf/prog_tests/perf_link.c | 2 +-
>>>>>> .../testing/selftests/bpf/prog_tests/send_signal.c | 14 ++++++++++----
>>>>>> 5 files changed, 16 insertions(+), 11 deletions(-)
>>>>>>
>>>>>> diff --git a/tools/testing/selftests/bpf/prog_tests/bpf_cookie.c b/tools/testing/selftests/bpf/prog_tests/bpf_cookie.c
>>>>>> index cd10df6cd0fc..0612e79a9281 100644
>>>>>> --- a/tools/testing/selftests/bpf/prog_tests/bpf_cookie.c
>>>>>> +++ b/tools/testing/selftests/bpf/prog_tests/bpf_cookie.c
>>>>>> @@ -199,7 +199,7 @@ static void pe_subtest(struct test_bpf_cookie *skel)
>>>>>> attr.type = PERF_TYPE_SOFTWARE;
>>>>>> attr.config = PERF_COUNT_SW_CPU_CLOCK;
>>>>>> attr.freq = 1;
>>>>>> - attr.sample_freq = 4000;
>>>>>> + attr.sample_freq = 1000;
>>>>>> pfd = syscall(__NR_perf_event_open, &attr, -1, 0, -1, PERF_FLAG_FD_CLOEXEC);
>>>>>> if (!ASSERT_GE(pfd, 0, "perf_fd"))
>>>>>> goto cleanup;
>>>>>> diff --git a/tools/testing/selftests/bpf/prog_tests/find_vma.c b/tools/testing/selftests/bpf/prog_tests/find_vma.c
>>>>>> index b74b3c0c555a..acc41223a112 100644
>>>>>> --- a/tools/testing/selftests/bpf/prog_tests/find_vma.c
>>>>>> +++ b/tools/testing/selftests/bpf/prog_tests/find_vma.c
>>>>>> @@ -30,7 +30,7 @@ static int open_pe(void)
>>>>>> attr.type = PERF_TYPE_HARDWARE;
>>>>>> attr.config = PERF_COUNT_HW_CPU_CYCLES;
>>>>>> attr.freq = 1;
>>>>>> - attr.sample_freq = 4000;
>>>>>> + attr.sample_freq = 1000;
>>>>>> pfd = syscall(__NR_perf_event_open, &attr, 0, -1, -1, PERF_FLAG_FD_CLOEXEC);
>>>>>>
>>>>>> return pfd >= 0 ? pfd : -errno;
>>>>>> @@ -57,7 +57,7 @@ static void test_find_vma_pe(struct find_vma *skel)
>>>>>> if (!ASSERT_OK_PTR(link, "attach_perf_event"))
>>>>>> goto cleanup;
>>>>>>
>>>>>> - for (i = 0; i < 1000000; ++i)
>>>>>> + for (i = 0; i < 1000000000; ++i)
>>>>>
>>>>> 1bln seems excessive... maybe 10mln would be enough?
>>>>
>>>> See explanation for send_signal test case below
>>>>
>>>>>
>>>>>> ++j;
>>>>>>
>>>>>> test_and_reset_skel(skel, -EBUSY /* in nmi, irq_work is busy */);
>>>>>
>>>>> [...]
>>>>>
>>>>>> diff --git a/tools/testing/selftests/bpf/prog_tests/send_signal.c b/tools/testing/selftests/bpf/prog_tests/send_signal.c
>>>>>> index 776916b61c40..841217bd1df6 100644
>>>>>> --- a/tools/testing/selftests/bpf/prog_tests/send_signal.c
>>>>>> +++ b/tools/testing/selftests/bpf/prog_tests/send_signal.c
>>>>>> @@ -4,11 +4,12 @@
>>>>>> #include <sys/resource.h>
>>>>>> #include "test_send_signal_kern.skel.h"
>>>>>>
>>>>>> -int sigusr1_received = 0;
>>>>>> +int sigusr1_received;
>>>>>> +volatile int volatile_variable;
>>>>>
>>>>> please make them static
>>>>
>>>> sure
>>>>
>>>>>
>>>>>>
>>>>>> static void sigusr1_handler(int signum)
>>>>>> {
>>>>>> - sigusr1_received++;
>>>>>> + sigusr1_received = 1;
>>>>>> }
>>>>>>
>>>>>> static void test_send_signal_common(struct perf_event_attr *attr,
>>>>>> @@ -42,7 +43,9 @@ static void test_send_signal_common(struct perf_event_attr *attr,
>>>>>> int old_prio;
>>>>>>
>>>>>> /* install signal handler and notify parent */
>>>>>> + errno = 0;
>>>>>> signal(SIGUSR1, sigusr1_handler);
>>>>>> + ASSERT_OK(errno, "signal");
>>>>>
>>>>> just ASSERT_OK(signal(...), "signal");
>>>>
>>>> I am fine to merge signal and ASSERT lines, but will substitute with condition "signal(SIGUSR1, sigusr1_handler) != SIG_ERR”, sounds good?
>>>>
>>> Ah, signal is a bit special with return values. Yeah,
>>> ASSERT_NEQ(signal(...), SIG_ERR, "signal") sounds good.
>>>>>
>>>>>>
>>>>>> close(pipe_c2p[0]); /* close read */
>>>>>> close(pipe_p2c[1]); /* close write */
>>>>>> @@ -63,9 +66,12 @@ static void test_send_signal_common(struct perf_event_attr *attr,
>>>>>> ASSERT_EQ(read(pipe_p2c[0], buf, 1), 1, "pipe_read");
>>>>>>
>>>>>> /* wait a little for signal handler */
>>>>>> - sleep(1);
>>>>>> + for (int i = 0; i < 1000000000; i++)
>>>>>
>>>>> same about 1bln
>>>>
>>>> With 10mln and 100 test runs I got 86 failures
>>>> 100mln - 63 failures
>>>> 1bln - 0 failures on 100 runs
>>>>
>>>> Now, there is performance concern for this test. Running
>>>>
>>>> time sudo ./test_progs -t send_signal/send_signal_nmi_thread
>>>>
>>>> With 1bln takes ~4s
>>>> 100mln - 1s.
>>>> Unchanged test with sleep(1); takes ~2s.
>>>>
>>>> On the other hand 300mln runs ~2s, and only fails 1 time per 100 runs. As 300mln does not regress performance comparing to the current “sleep(1)” implementation, I propose to go with it. What do you think?
>>> I think if we need to burn multiple seconds of CPU to make the test
>>> reliable, then we should either rework or disable/remove the test. In
>>> CI those billions of iterations will be much slower. And even waiting
>>> for 4 seconds for just one test is painful.
>>> Yonghong, WDYT? Should we just drop thi test? It has caused us a bunch
>>> of flakiness and maintenance burden without actually catching any
>>> issues. Maybe it's better to just get rid of it?
>>
>> Could we try to set affinity for the child process here?
>> See perf_branches.c:
>>
>> ...
>> /* generate some branches on cpu 0 */
>> CPU_ZERO(&cpu_set);
>> CPU_SET(0, &cpu_set);
>> err = pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set);
>> if (CHECK(err, "set_affinity", "cpu #0, err %d\n", err))
>> goto out_destroy;
>> /* spin the loop for a while (random high number) */
>> for (i = 0; i < 1000000; ++i)
>> ++j;
>> ...
>>
>> Binding the process (single thread) to a particular cpu can
>> prevent other non-binding processes from migrating to this
>> cpu and boost the chance for NMI triggered on this cpu.
>> This could be the reason perf_branches.c (and a few other tests)
>> does.
>>
>> In send_signal case, the cpu affinity probably should
>> set to cpu 1 as cpu 0 has been pinned by previous tests for
>> the main process and I didn't see it 'unpinned'
>> (by setaffinity to ALL cpus).
>> This is inconvenient.
>>
>> So the following is my suggestion:
>> 1. abstract the above 'pthread_setaffinity_np to
>> a helper to set affinity to a particular cpu as
>> this function has been used in several cases.
>> 2. create a new helper to undo setaffinity (set cpu
>> mask to all available cpus) so we can pair it
>> with pthread_setaffinity_np helper in prog_tests/...
>> files.
>> 3. clean up prog_tests/... files which have pthread_setaffinity_np.
>> 4. use helpers 1/2 with loop bound 1000000 for send_signal test.
>> The implementation here will be consistent with
>> other NMI tests. Hopefully the test can consistent
>> pass similar to other NMI tests.
>>
>> WDYT?
>
> Hi Yonghong,
>
> I have tried this approach in the send_signal test without much success unfortunately (different CPUs and configurations options). It is required though for perf_branches test, yet to understand why.
Thanks for experiments. I looked at the code again. Indeed
pthread_setaffinity_np is not needed for send_signal test. This is
because we use perf_event_open pid/cpu config like below:
pid > 0 and cpu == -1
This measures the specified process/thread on any CPU.
For perf_branches, pthread_setaffinity_np is needed since it uses
the perf_evnet_open pid/cpu config like below:
pid == -1 and cpu >= 0
This measures all processes/threads on the specified CPU.
This requires CAP_SYS_ADMIN capabil‐
ity or a /proc/sys/kernel/perf_event_paranoid value of
less than 1.
>
> In the V2 of this patch, I used modified approach when we will stop crunching volatile variable when needed condition became true. I hope this will be an acceptable middle ground in this case.
My current setup is using qemu on a physical server and cannot reproduce
the issue. So I created another setup which uses qemu on a VM itseld and
can actually reproduce the issue. Replacing the sleep(1) with
for (int i = 0; i < 1000000000; i++) /* 1billion */
j++; /* volatile int j */
seems fixing the issue. But your change
for (int i = 0; i < 100000000 && !sigusr1_received; i++)
volatile_variable /= i + 1;
works too and I tested it and in most cases the time for the subtest
is 0.8x or 0.9x seconds. Sometimes it can be < 0.5 seconds, and
occasionally it may be 1.0x seconds. Overall, this is definitely
an improvement for fixing flakiness and better runtime.
>
> Thanks!
>
>>
>>>>
>>>>>
>>>>>> + volatile_variable++;
>>>>>>
>>>>>> buf[0] = sigusr1_received ? '2' : '0';
>>>>>> + ASSERT_EQ(sigusr1_received, 1, "sigusr1_received");
>>>>>> +
>>>>>> ASSERT_EQ(write(pipe_c2p[1], buf, 1), 1, "pipe_write");
>>>>>>
>>>>>> /* wait for parent notification and exit */
>>>>>> @@ -110,9 +116,9 @@ static void test_send_signal_common(struct perf_event_attr *attr,
>>>>>> ASSERT_EQ(read(pipe_c2p[0], buf, 1), 1, "pipe_read");
>>>>>>
>>>>>> /* trigger the bpf send_signal */
>>>>>> + skel->bss->signal_thread = signal_thread;
>>>>>> skel->bss->pid = pid;
>>>>>> skel->bss->sig = SIGUSR1;
>>>>>> - skel->bss->signal_thread = signal_thread;
>>>>>>
>>>>>> /* notify child that bpf program can send_signal now */
>>>>>> ASSERT_EQ(write(pipe_p2c[1], buf, 1), 1, "pipe_write");
>>>>>> --
>>>>>> 2.30.2
>
prev parent reply other threads:[~2022-03-02 4:53 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-02-19 0:30 [PATCH bpf-next] Improve BPF test stability (related to perf events and scheduling) Mykola Lysenko
2022-02-19 1:47 ` Song Liu
2022-02-22 20:00 ` Mykola Lysenko
2022-02-20 4:39 ` Andrii Nakryiko
2022-02-22 20:35 ` Mykola Lysenko
2022-02-23 3:13 ` Andrii Nakryiko
2022-02-23 4:32 ` Yonghong Song
2022-02-24 0:52 ` Mykola Lysenko
2022-03-01 3:45 ` Mykola Lysenko
2022-03-02 4:53 ` Yonghong Song [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=ab6970fc-eaa0-9462-9614-893085c42331@fb.com \
--to=yhs@fb.com \
--cc=andrii.nakryiko@gmail.com \
--cc=andrii@kernel.org \
--cc=ast@kernel.org \
--cc=bpf@vger.kernel.org \
--cc=daniel@iogearbox.net \
--cc=mykolal@fb.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox