From: Waiman Long <longman@redhat.com>
To: Jesper Dangaard Brouer <hawk@kernel.org>,
tj@kernel.org, hannes@cmpxchg.org, lizefan.x@bytedance.com,
cgroups@vger.kernel.org, yosryahmed@google.com
Cc: netdev@vger.kernel.org, linux-mm@kvack.org,
shakeel.butt@linux.dev, kernel-team@cloudflare.com,
Arnaldo Carvalho de Melo <acme@kernel.org>,
Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Subject: Re: [PATCH v1] cgroup/rstat: add cgroup_rstat_cpu_lock helpers and tracepoints
Date: Fri, 3 May 2024 10:30:02 -0400 [thread overview]
Message-ID: <ca46435a-f67a-4d85-bf8d-b5d3289b6185@redhat.com> (raw)
In-Reply-To: <42a6d218-206b-4f87-a8fa-ef42d107fb23@kernel.org>
On 5/3/24 10:00, Jesper Dangaard Brouer wrote:
>> I may have mistakenly thinking the lock hold time refers to just the
>> cpu_lock. Your reported times here are about the cgroup_rstat_lock.
>> Right? If so, the numbers make sense to me.
>>
>
> True, my reported number here are about the cgroup_rstat_lock.
> Glad to hear, we are more aligned then 🙂
>
> Given I just got some prod machines online with this patch
> cgroup_rstat_cpu_lock tracepoints, I can give you some early results,
> about hold-time for the cgroup_rstat_cpu_lock.
>
> From this oneliner bpftrace commands:
>
> sudo bpftrace -e '
> tracepoint:cgroup:cgroup_rstat_cpu_lock_contended {
> @start[tid]=nsecs; @cnt[probe]=count()}
> tracepoint:cgroup:cgroup_rstat_cpu_locked {
> $now=nsecs;
> if (args->contended) {
> @wait_per_cpu_ns=hist($now-@start[tid]);
> delete(@start[tid]);}
> @cnt[probe]=count(); @locked[tid]=$now}
> tracepoint:cgroup:cgroup_rstat_cpu_unlock {
> $now=nsecs;
> @locked_per_cpu_ns=hist($now-@locked[tid]);
> delete(@locked[tid]);
> @cnt[probe]=count()}
> interval:s:1 {time("%H:%M:%S "); print(@wait_per_cpu_ns);
> print(@locked_per_cpu_ns); print(@cnt); clear(@cnt);}'
>
> Results from one 1 sec period:
>
> 13:39:55 @wait_per_cpu_ns:
> [512, 1K) 3 | |
> [1K, 2K) 12 |@ |
> [2K, 4K) 390
> |@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@|
> [4K, 8K) 70 |@@@@@@@@@ |
> [8K, 16K) 24 |@@@ |
> [16K, 32K) 183 |@@@@@@@@@@@@@@@@@@@@@@@@ |
> [32K, 64K) 11 |@ |
>
> @locked_per_cpu_ns:
> [256, 512) 75592 |@ |
> [512, 1K) 2537357
> |@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@|
> [1K, 2K) 528615 |@@@@@@@@@@ |
> [2K, 4K) 168519 |@@@ |
> [4K, 8K) 162039 |@@@ |
> [8K, 16K) 100730 |@@ |
> [16K, 32K) 42276 | |
> [32K, 64K) 1423 | |
> [64K, 128K) 89 | |
>
> @cnt[tracepoint:cgroup:cgroup_rstat_cpu_lock_contended]: 3 /sec
> @cnt[tracepoint:cgroup:cgroup_rstat_cpu_unlock]: 3200 /sec
> @cnt[tracepoint:cgroup:cgroup_rstat_cpu_locked]: 3200 /sec
>
>
> So, we see "flush-code-path" per-CPU-holding @locked_per_cpu_ns isn't
> exceeding 128 usec.
>
> My latency requirements, to avoid RX-queue overflow, with 1024 slots,
> running at 25 Gbit/s, is 27.6 usec with small packets, and 500 usec
> (0.5ms) with MTU size packets. This is very close to my latency
> requirements.
Thanks for sharing the data.
This is more aligned with what I would have expected. Still, a high up
to 128 usec is still on the high side. I remembered during my latency
testing when I worked on cpu_lock latency patch, it was in the 2 digit
range. Perhaps there are other sources of noise or the update list is
really long. Anyway, it may be a bit hard to reach the 27.6 usec target
for small packets.
Cheers,
Longman
next prev parent reply other threads:[~2024-05-03 14:30 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-05-01 14:04 [PATCH v1] cgroup/rstat: add cgroup_rstat_cpu_lock helpers and tracepoints Jesper Dangaard Brouer
2024-05-01 14:24 ` Waiman Long
2024-05-01 17:22 ` Jesper Dangaard Brouer
2024-05-01 18:41 ` Waiman Long
2024-05-02 11:23 ` Jesper Dangaard Brouer
2024-05-02 18:19 ` Waiman Long
2024-05-03 14:00 ` Jesper Dangaard Brouer
2024-05-03 14:30 ` Waiman Long [this message]
2024-05-03 19:18 ` Shakeel Butt
2024-05-06 12:03 ` Jesper Dangaard Brouer
2024-05-06 16:22 ` Shakeel Butt
2024-05-06 16:28 ` Ivan Babrou
2024-05-06 19:45 ` Shakeel Butt
2024-05-06 19:54 ` Jesper Dangaard Brouer
2024-05-02 19:44 ` Shakeel Butt
2024-05-03 12:58 ` Jesper Dangaard Brouer
2024-05-03 18:11 ` Shakeel Butt
2024-05-14 5:18 ` Jesper Dangaard Brouer
2024-05-14 5:55 ` Tejun Heo
2024-05-14 16:59 ` Waiman Long
2024-05-15 16:58 ` Tejun Heo
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=ca46435a-f67a-4d85-bf8d-b5d3289b6185@redhat.com \
--to=longman@redhat.com \
--cc=acme@kernel.org \
--cc=bigeasy@linutronix.de \
--cc=cgroups@vger.kernel.org \
--cc=hannes@cmpxchg.org \
--cc=hawk@kernel.org \
--cc=kernel-team@cloudflare.com \
--cc=linux-mm@kvack.org \
--cc=lizefan.x@bytedance.com \
--cc=netdev@vger.kernel.org \
--cc=shakeel.butt@linux.dev \
--cc=tj@kernel.org \
--cc=yosryahmed@google.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).