From: Song Liu <songliubraving@fb.com>
To: Andrii Nakryiko <andrii.nakryiko@gmail.com>
Cc: Song Liu <song@kernel.org>, bpf <bpf@vger.kernel.org>,
Networking <netdev@vger.kernel.org>,
open list <linux-kernel@vger.kernel.org>,
"Alexei Starovoitov" <ast@kernel.org>,
Daniel Borkmann <daniel@iogearbox.net>,
"Andrii Nakryiko" <andrii@kernel.org>,
Kernel Team <Kernel-team@fb.com>,
"Peter Ziljstra" <peterz@infradead.org>, X86 ML <x86@kernel.org>
Subject: Re: [PATCH v2 bpf-next 0/7] bpf_prog_pack allocator
Date: Fri, 17 Dec 2021 17:13:15 +0000 [thread overview]
Message-ID: <5AD9F449-6462-4501-9D1D-407956103DD4@fb.com> (raw)
In-Reply-To: <CAEf4Bzb3sbf5Ddq4FaBsZpyiqhoFD+PxxbZHP6ips6h01EuNYg@mail.gmail.com>
> On Dec 17, 2021, at 8:43 AM, Andrii Nakryiko <andrii.nakryiko@gmail.com> wrote:
>
> On Fri, Dec 17, 2021 at 8:42 AM Andrii Nakryiko
> <andrii.nakryiko@gmail.com> wrote:
>>
>> On Thu, Dec 16, 2021 at 5:53 PM Song Liu <songliubraving@fb.com> wrote:
>>>
>>>
>>>
>>>> On Dec 16, 2021, at 12:06 PM, Andrii Nakryiko <andrii.nakryiko@gmail.com> wrote:
>>>>
>>>> On Tue, Dec 14, 2021 at 10:01 PM Song Liu <song@kernel.org> wrote:
>>>>>
>>>>> Changes v1 => v2:
>>>>> 1. Use text_poke instead of writing through linear mapping. (Peter)
>>>>> 2. Avoid making changes to non-x86_64 code.
>>>>>
>>>>> Most BPF programs are small, but they consume a page each. For systems
>>>>> with busy traffic and many BPF programs, this could also add significant
>>>>> pressure to instruction TLB.
>>>>>
>>>>> This set tries to solve this problem with customized allocator that pack
>>>>> multiple programs into a huge page.
>>>>>
>>>>> Patches 1-5 prepare the work. Patch 6 contains key logic of the allocator.
>>>>> Patch 7 uses this allocator in x86_64 jit compiler.
>>>>>
>>>>
>>>> There are test failures, please see [0]. But I was also wondering if
>>>> there could be an explicit selftest added to validate that all this
>>>> huge page machinery is actually activated and working as expected?
>>>
>>> We can enable some debug option that dumps the page table. Then from the
>>> page table, we can confirm the programs are running on a huge page. This
>>> only works on x86_64 though. WDYT?
>>>
>>
>> I don't know what exactly is involved, so it's hard to say. Ideally
>> whatever we do doesn't complicate our CI setup. Can we use BPF tracing
>> magic to check this from inside the kernel somehow?
>>
>
> But I don't feel strongly about this, if it's hard to detect, it's
> fine to not have a specific test (especially that it's very
> architecture-specific)
It will be more or less architecture-specific, as we need somehow walk
the page table (with debug option or with BPF iterator). I will try
something.
Thanks,
Song
>
>>> Thanks,
>>> Song
>>>
>>>
>>>>
>>>> [0] https://github.com/kernel-patches/bpf/runs/4530372387?check_suite_focus=true
>>>>
>>>>> Song Liu (7):
>>>>> x86/Kconfig: select HAVE_ARCH_HUGE_VMALLOC with HAVE_ARCH_HUGE_VMAP
>>>>> bpf: use bytes instead of pages for bpf_jit_[charge|uncharge]_modmem
>>>>> bpf: use size instead of pages in bpf_binary_header
>>>>> bpf: add a pointer of bpf_binary_header to bpf_prog
>>>>> x86/alternative: introduce text_poke_jit
>>>>> bpf: introduce bpf_prog_pack allocator
>>>>> bpf, x86_64: use bpf_prog_pack allocator
>>>>>
>>>>> arch/x86/Kconfig | 1 +
>>>>> arch/x86/include/asm/text-patching.h | 1 +
>>>>> arch/x86/kernel/alternative.c | 28 ++++
>>>>> arch/x86/net/bpf_jit_comp.c | 93 ++++++++++--
>>>>> include/linux/bpf.h | 4 +-
>>>>> include/linux/filter.h | 23 ++-
>>>>> kernel/bpf/core.c | 213 ++++++++++++++++++++++++---
>>>>> kernel/bpf/trampoline.c | 6 +-
>>>>> 8 files changed, 328 insertions(+), 41 deletions(-)
>>>>>
>>>>> --
>>>>> 2.30.2
>>>
next prev parent reply other threads:[~2021-12-17 17:13 UTC|newest]
Thread overview: 19+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-12-15 6:00 [PATCH v2 bpf-next 0/7] bpf_prog_pack allocator Song Liu
2021-12-15 6:00 ` [PATCH v2 bpf-next 1/7] x86/Kconfig: select HAVE_ARCH_HUGE_VMALLOC with HAVE_ARCH_HUGE_VMAP Song Liu
2021-12-15 6:00 ` [PATCH v2 bpf-next 2/7] bpf: use bytes instead of pages for bpf_jit_[charge|uncharge]_modmem Song Liu
2021-12-15 8:56 ` Peter Zijlstra
2021-12-16 19:45 ` Song Liu
2021-12-15 6:00 ` [PATCH v2 bpf-next 3/7] bpf: use size instead of pages in bpf_binary_header Song Liu
2021-12-15 6:00 ` [PATCH v2 bpf-next 4/7] bpf: add a pointer of bpf_binary_header to bpf_prog Song Liu
2021-12-15 6:01 ` [PATCH v2 bpf-next 5/7] x86/alternative: introduce text_poke_jit Song Liu
2021-12-15 9:06 ` Peter Zijlstra
2021-12-15 9:17 ` Peter Zijlstra
2021-12-16 19:40 ` Song Liu
2021-12-15 6:01 ` [PATCH v2 bpf-next 6/7] bpf: introduce bpf_prog_pack allocator Song Liu
2021-12-15 6:01 ` [PATCH v2 bpf-next 7/7] bpf, x86_64: use " Song Liu
2021-12-16 20:06 ` [PATCH v2 bpf-next 0/7] " Andrii Nakryiko
2021-12-17 1:53 ` Song Liu
2021-12-17 16:42 ` Andrii Nakryiko
2021-12-17 16:43 ` Andrii Nakryiko
2021-12-17 17:13 ` Song Liu [this message]
2021-12-17 17:16 ` Andrii Nakryiko
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=5AD9F449-6462-4501-9D1D-407956103DD4@fb.com \
--to=songliubraving@fb.com \
--cc=Kernel-team@fb.com \
--cc=andrii.nakryiko@gmail.com \
--cc=andrii@kernel.org \
--cc=ast@kernel.org \
--cc=bpf@vger.kernel.org \
--cc=daniel@iogearbox.net \
--cc=linux-kernel@vger.kernel.org \
--cc=netdev@vger.kernel.org \
--cc=peterz@infradead.org \
--cc=song@kernel.org \
--cc=x86@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox