From: Yonghong Song <yonghong.song@linux.dev>
To: Hou Tao <houtao@huaweicloud.com>, bpf@vger.kernel.org
Cc: Alexei Starovoitov <ast@kernel.org>,
Andrii Nakryiko <andrii@kernel.org>,
Daniel Borkmann <daniel@iogearbox.net>,
kernel-team@fb.com, Martin KaFai Lau <martin.lau@kernel.org>
Subject: Re: [PATCH bpf-next v4 4/7] bpf: Refill only one percpu element in memalloc
Date: Wed, 20 Dec 2023 11:55:03 -0800 [thread overview]
Message-ID: <c5eb0959-288c-4932-a1f1-d8192aa7edbe@linux.dev> (raw)
In-Reply-To: <2b92dfac-1ce1-4981-c2ec-a432e4dd24a5@huaweicloud.com>
On 12/19/23 3:31 AM, Hou Tao wrote:
> Hi,
>
> On 12/18/2023 2:30 PM, Yonghong Song wrote:
>> Typically for percpu map element or data structure, once allocated,
>> most operations are lookup or in-place update. Deletion are really
>> rare. Currently, for percpu data strcture, 4 elements will be
>> refilled if the size is <= 256. Let us just do with one element
>> for percpu data. For example, for size 256 and 128 cpus, the
>> potential saving will be 3 * 256 * 128 * 128 = 12MB.
>>
>> Acked-by: Hou Tao <houtao1@huawei.com>
>> Signed-off-by: Yonghong Song <yonghong.song@linux.dev>
>> ---
>> kernel/bpf/memalloc.c | 13 +++++++++----
>> 1 file changed, 9 insertions(+), 4 deletions(-)
>>
>> diff --git a/kernel/bpf/memalloc.c b/kernel/bpf/memalloc.c
>> index 50ab2fecc005..f37998662146 100644
>> --- a/kernel/bpf/memalloc.c
>> +++ b/kernel/bpf/memalloc.c
>> @@ -485,11 +485,16 @@ static void init_refill_work(struct bpf_mem_cache *c)
>>
>> static void prefill_mem_cache(struct bpf_mem_cache *c, int cpu)
>> {
>> - /* To avoid consuming memory assume that 1st run of bpf
>> - * prog won't be doing more than 4 map_update_elem from
>> - * irq disabled region
>> + int cnt = 1;
>> +
>> + /* To avoid consuming memory, for non-percpu allocation, assume that
>> + * 1st run of bpf prog won't be doing more than 4 map_update_elem from
>> + * irq disabled region if unit size is less than or equal to 256.
>> + * For all other cases, let us just do one allocation.
>> */
>> - alloc_bulk(c, c->unit_size <= 256 ? 4 : 1, cpu_to_node(cpu), false);
>> + if (!c->percpu_size && c->unit_size <= 256)
>> + cnt = 4;
>> + alloc_bulk(c, cnt, cpu_to_node(cpu), false);
>> }
> Another thought about this patch. When the prefilled element is
> allocated by the invocation of bpf_percpu_obj_new(), the prefill will
> trigger again and this time it will allocate c->batch elements. For
> 256-bytes unit_size, c->batch will be 64, so the maximal memory
> consumption under 128-cpus host will be: 64 * 256 * 128 * 128 = 256MB
Actually, it should be 48 * 256 * 128 * 128 in the worst case, due to
c->batch = max((c->high_watermark - c->low_watermark) / 4 * 3, 1);
But in reality, for percpu allocation, we probably won't have allocation
for all 128 cpus.
But your point is taken, for percpu allocation, we will have much less
allocations, so we should not use current lower_watermark/upper_watermark used
for non-percpu allocations. So for percpu allocation, I suggest to do
lower_watermark = 1;
upper_watermark = 3;
c->batch = 1;
Thanks!
> when there is one allocation of bpf_percpu_obj_new() on each CPU. And my
> question is that should we adjust the low_watermark and high_watermark
> accordingly for per-cpu allocation to reduce the memory consumption ?
>>
>> static int check_obj_size(struct bpf_mem_cache *c, unsigned int idx)
next prev parent reply other threads:[~2023-12-20 19:55 UTC|newest]
Thread overview: 14+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-12-18 6:30 [PATCH bpf-next v4 0/7] bpf: Reduce memory usage for bpf_global_percpu_ma Yonghong Song
2023-12-18 6:30 ` [PATCH bpf-next v4 1/7] bpf: Avoid unnecessary extra percpu memory allocation Yonghong Song
2023-12-18 6:30 ` [PATCH bpf-next v4 2/7] bpf: Add objcg to bpf_mem_alloc Yonghong Song
2023-12-19 3:03 ` Hou Tao
2023-12-18 6:30 ` [PATCH bpf-next v4 3/7] bpf: Allow per unit prefill for non-fix-size percpu memory allocator Yonghong Song
2023-12-19 3:04 ` Hou Tao
2023-12-20 4:37 ` Alexei Starovoitov
2023-12-20 17:57 ` Yonghong Song
2023-12-18 6:30 ` [PATCH bpf-next v4 4/7] bpf: Refill only one percpu element in memalloc Yonghong Song
2023-12-19 11:31 ` Hou Tao
2023-12-20 19:55 ` Yonghong Song [this message]
2023-12-18 6:30 ` [PATCH bpf-next v4 5/7] bpf: Limit up to 512 bytes for bpf_global_percpu_ma allocation Yonghong Song
2023-12-18 6:31 ` [PATCH bpf-next v4 6/7] selftests/bpf: Cope with 512 bytes limit with bpf_global_percpu_ma Yonghong Song
2023-12-18 6:31 ` [PATCH bpf-next v4 7/7] selftests/bpf: Add a selftest with > 512-byte percpu allocation size Yonghong Song
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=c5eb0959-288c-4932-a1f1-d8192aa7edbe@linux.dev \
--to=yonghong.song@linux.dev \
--cc=andrii@kernel.org \
--cc=ast@kernel.org \
--cc=bpf@vger.kernel.org \
--cc=daniel@iogearbox.net \
--cc=houtao@huaweicloud.com \
--cc=kernel-team@fb.com \
--cc=martin.lau@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox