From: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com>
To: "Rogozhkin, Dmitry V" <dmitry.v.rogozhkin@intel.com>
Cc: "peterz@infradead.org" <peterz@infradead.org>,
"Intel-gfx@lists.freedesktop.org"
<Intel-gfx@lists.freedesktop.org>
Subject: Re: [RFC v3 00/11] i915 PMU and engine busy stats
Date: Wed, 13 Sep 2017 10:01:14 +0100 [thread overview]
Message-ID: <234289d7-99e5-f78c-a34f-e28682e30576@linux.intel.com> (raw)
In-Reply-To: <1505224750.6296.13.camel@intel.com>
On 12/09/2017 23:01, Rogozhkin, Dmitry V wrote:
> On Tue, 2017-09-12 at 15:54 +0100, Tvrtko Ursulin wrote:
>> On 12/09/2017 03:03, Rogozhkin, Dmitry V wrote:
>>> Hi,
>>>
>>> Just tried v3 series. perf-stat works fine. From the IGT tests which I wrote for i915 PMU
>>> (https://patchwork.freedesktop.org/series/29313/) all pass (assuming pmu.enabled will be exposed
>>> in debugfs) except cpu_online subtest. And this is pretty interesting - see details below.
>>>
>>> Ok, be prepared for the long mail:)...
>>>
>>> So, cpu_online subtest loads RCS0 engine 100% and starts to put CPUs offline one by one starting
>>> from CPU0 (don't forget to have CONFIG_BOOTPARAM_HOTPLUG_CPU0=y in .config). Test expectation is
>>> that i915 PMU will report almost 100% (with 2% tolerance) busyness of RCS0. Test requests metric
>>> just twice: before running on RCS0 and right after. This fails as follows:
>>>
>>> Executed on rcs0 for 32004262184us
>>> i915/rcs0-busy/: 2225723999us
>>> (perf_pmu:6325) CRITICAL: Test assertion failure function test_cpu_online, file perf_pmu.c:719:
>>> (perf_pmu:6325) CRITICAL: Failed assertion: perf_elapsed(&pm.metrics[0]) > (1-USAGE_TOLERANCE) * elapsed_ns(&start, &now)
>>> Stack trace:
>>> #0 [__igt_fail_assert+0xf1]
>>> #1 [__real_main773+0xff1]
>>> #2 [main+0x35]
>>> #3 [__libc_start_main+0xf5]
>>> #4 [_start+0x29]
>>> #5 [<unknown>+0x29]
>>> Subtest cpu_online failed.
>>> **** DEBUG ****
>>> (perf_pmu:6325) DEBUG: Test requirement passed: is_hotplug_cpu0()
>>> (perf_pmu:6325) INFO: perf_init: enabled 1 metrics from 1 requested
>>> (perf_pmu:6325) ioctl-wrappers-DEBUG: Test requirement passed: gem_has_ring(fd, ring)
>>> (perf_pmu:6325) INFO: Executed on rcs0 for 32004262184us
>>> (perf_pmu:6325) INFO: i915/rcs0-busy/: 2225723999us
>>> (perf_pmu:6325) CRITICAL: Test assertion failure function test_cpu_online, file perf_pmu.c:719:
>>> (perf_pmu:6325) CRITICAL: Failed assertion: perf_elapsed(&pm.metrics[0]) > (1-USAGE_TOLERANCE) * elapsed_ns(&start, &now)
>>>
>>> Now. Looks like that by itself PMU context migration works. For example, if you will comment out
>>> "perf_pmu_migrate_context(&pmu->base, cpu, target)" you will get:
>>>
>>> Executed on rcs0 for 32004434918us
>>> i915/rcs0-busy/: 76623707us
>>>
>>> Compare with previous:
>>> Executed on rcs0 for 32004262184us
>>> i915/rcs0-busy/: 2225723999us
>>>
>>> This test passed on the previous set of patches, I mean Tvrtko's v2 series + my patches.
>>>
>>> So, it seems we are loosing counter values somehow. I saw in the patches that this place really was modified - you
>>> have added subtraction from initial counter value:
>>> static void i915_pmu_event_read(struct perf_event *event)
>>> {
>>>
>>> local64_set(&event->count,
>>> __i915_pmu_event_read(event) -
>>> local64_read(&event->hw.prev_count));
>>> }
>>>
>>> But looks like the problem is that with the PMU context migration we get sequence of events start/stop (or maybe
>>> add/del) which eventually call our i915_pmu_enable/disable. Here is the dmesg log with the obvious printk:
>>>
>>> [ 153.971096] [IGT] perf_pmu: starting subtest cpu_online
>>> [ 153.971151] i915_pmu_enable: event->hw.prev_count=0
>>> [ 154.036015] i915_pmu_disable: event->hw.prev_count=0
>>> [ 154.048027] i915_pmu_enable: event->hw.prev_count=0
>>> [ 154.049343] smpboot: CPU 0 is now offline
>>> [ 155.059028] smpboot: Booting Node 0 Processor 0 APIC 0x0
>>> [ 155.155078] smpboot: CPU 1 is now offline
>>> [ 156.161026] x86: Booting SMP configuration:
>>> [ 156.161027] smpboot: Booting Node 0 Processor 1 APIC 0x2
>>> [ 156.197065] IRQ 122: no longer affine to CPU2
>>> [ 156.198087] smpboot: CPU 2 is now offline
>>> [ 157.208028] smpboot: Booting Node 0 Processor 2 APIC 0x4
>>> [ 157.263093] smpboot: CPU 3 is now offline
>>> [ 158.273026] smpboot: Booting Node 0 Processor 3 APIC 0x6
>>> [ 158.310026] i915_pmu_disable: event->hw.prev_count=76648307
>>> [ 158.319020] i915_pmu_enable: event->hw.prev_count=76648307
>>> [ 158.319098] IRQ 124: no longer affine to CPU4
>>> [ 158.320368] smpboot: CPU 4 is now offline
>>> [ 159.326030] smpboot: Booting Node 0 Processor 4 APIC 0x1
>>> [ 159.365306] smpboot: CPU 5 is now offline
>>> [ 160.371030] smpboot: Booting Node 0 Processor 5 APIC 0x3
>>> [ 160.421077] IRQ 125: no longer affine to CPU6
>>> [ 160.422093] smpboot: CPU 6 is now offline
>>> [ 161.429030] smpboot: Booting Node 0 Processor 6 APIC 0x5
>>> [ 161.467091] smpboot: CPU 7 is now offline
>>> [ 162.473027] smpboot: Booting Node 0 Processor 7 APIC 0x7
>>> [ 162.527019] i915_pmu_disable: event->hw.prev_count=4347548222
>>> [ 162.546017] i915_pmu_enable: event->hw.prev_count=4347548222
>>> [ 162.547317] smpboot: CPU 0 is now offline
>>> [ 163.553028] smpboot: Booting Node 0 Processor 0 APIC 0x0
>>> [ 163.621089] smpboot: CPU 1 is now offline
>>> [ 164.627028] x86: Booting SMP configuration:
>>> [ 164.627029] smpboot: Booting Node 0 Processor 1 APIC 0x2
>>> [ 164.669308] smpboot: CPU 2 is now offline
>>> [ 165.679025] smpboot: Booting Node 0 Processor 2 APIC 0x4
>>> [ 165.717089] smpboot: CPU 3 is now offline
>>> [ 166.723025] smpboot: Booting Node 0 Processor 3 APIC 0x6
>>> [ 166.775016] i915_pmu_disable: event->hw.prev_count=8574197312
>>> [ 166.787016] i915_pmu_enable: event->hw.prev_count=8574197312
>>> [ 166.788309] smpboot: CPU 4 is now offline
>>> [ 167.794025] smpboot: Booting Node 0 Processor 4 APIC 0x1
>>> [ 167.837114] smpboot: CPU 5 is now offline
>>> [ 168.847025] smpboot: Booting Node 0 Processor 5 APIC 0x3
>>> [ 168.889312] smpboot: CPU 6 is now offline
>>> [ 169.899030] smpboot: Booting Node 0 Processor 6 APIC 0x5
>>> [ 169.944104] smpboot: CPU 7 is now offline
>>> [ 170.954032] smpboot: Booting Node 0 Processor 7 APIC 0x7
>>> [ 171.000016] i915_pmu_disable: event->hw.prev_count=12815138319
>>> [ 171.008017] i915_pmu_enable: event->hw.prev_count=12815138319
>>> [ 171.009304] smpboot: CPU 0 is now offline
>>> [ 172.017028] smpboot: Booting Node 0 Processor 0 APIC 0x0
>>> [ 172.096104] smpboot: CPU 1 is now offline
>>> [ 173.106025] x86: Booting SMP configuration:
>>> [ 173.106026] smpboot: Booting Node 0 Processor 1 APIC 0x2
>>> [ 173.147078] smpboot: CPU 2 is now offline
>>> [ 174.153025] smpboot: Booting Node 0 Processor 2 APIC 0x4
>>> [ 174.192093] smpboot: CPU 3 is now offline
>>> [ 175.198028] smpboot: Booting Node 0 Processor 3 APIC 0x6
>>> [ 175.229042] i915_pmu_disable: event->hw.prev_count=17035889079
>>> [ 175.242030] i915_pmu_enable: event->hw.prev_count=17035889079
>>> [ 175.242163] IRQ fixup: irq 120 move in progress, old vector 131
>>> [ 175.242165] IRQ fixup: irq 121 move in progress, old vector 147
>>> [ 175.242171] IRQ 124: no longer affine to CPU4
>>> [ 175.243435] smpboot: CPU 4 is now offline
>>> [ 176.248040] smpboot: Booting Node 0 Processor 4 APIC 0x1
>>> [ 176.285328] smpboot: CPU 5 is now offline
>>> [ 177.296039] smpboot: Booting Node 0 Processor 5 APIC 0x3
>>> [ 177.325067] IRQ 125: no longer affine to CPU6
>>> [ 177.326087] smpboot: CPU 6 is now offline
>>> [ 178.335036] smpboot: Booting Node 0 Processor 6 APIC 0x5
>>> [ 178.377063] IRQ 122: no longer affine to CPU7
>>> [ 178.378086] smpboot: CPU 7 is now offline
>>> [ 179.388028] smpboot: Booting Node 0 Processor 7 APIC 0x7
>>> [ 179.454030] i915_pmu_disable: event->hw.prev_count=21269856967
>>> [ 179.470026] i915_pmu_enable: event->hw.prev_count=21269856967
>>> [ 179.471110] smpboot: CPU 0 is now offline
>>> [ 180.481028] smpboot: Booting Node 0 Processor 0 APIC 0x0
>>> [ 180.551075] smpboot: CPU 1 is now offline
>>> [ 181.558029] x86: Booting SMP configuration:
>>> [ 181.558030] smpboot: Booting Node 0 Processor 1 APIC 0x2
>>> [ 181.595096] smpboot: CPU 2 is now offline
>>> [ 182.605029] smpboot: Booting Node 0 Processor 2 APIC 0x4
>>> [ 182.657084] smpboot: CPU 3 is now offline
>>> [ 183.668030] smpboot: Booting Node 0 Processor 3 APIC 0x6
>>> [ 183.709017] i915_pmu_disable: event->hw.prev_count=25497358644
>>> [ 183.727016] i915_pmu_enable: event->hw.prev_count=25497358644
>>> [ 183.728305] smpboot: CPU 4 is now offline
>>> [ 184.734027] smpboot: Booting Node 0 Processor 4 APIC 0x1
>>> [ 184.767090] smpboot: CPU 5 is now offline
>>> [ 185.777036] smpboot: Booting Node 0 Processor 5 APIC 0x3
>>> [ 185.823096] smpboot: CPU 6 is now offline
>>> [ 186.829051] smpboot: Booting Node 0 Processor 6 APIC 0x5
>>> [ 186.856350] smpboot: CPU 7 is now offline
>>> [ 187.862051] smpboot: Booting Node 0 Processor 7 APIC 0x7
>>> [ 187.871216] [IGT] perf_pmu: exiting, ret=99
>>> [ 187.889199] Console: switching to colour frame buffer device 240x67
>>> [ 187.889583] i915_pmu_disable: event->hw.prev_count=29754080941
>>>
>>> And the result which I got in userspace for this run were
>>> Executed on rcs0 for 32003587981us
>>> i915/rcs0-busy/: 2247436461us
>>>
>>> After that I decided to roll back the change with counting values which I mentioned before, i.e.:
>>> static void i915_pmu_event_read(struct perf_event *event)
>>> {
>>>
>>> local64_set(&event->count,
>>> __i915_pmu_event_read(event) /*-
>>> local64_read(&event->hw.prev_count)*/);
>>> }
>>>
>>> And I got test PASSED :):
>>> Executed on rcs0 for 32002282603us
>>> i915/rcs0-busy/: 31998855052us
>>> Subtest cpu_online: SUCCESS (33.950s)
>>>
>>> At this point I need to go home :). Maybe you will have time to look into this issue? If not, I will continue
>>> tomorrow.
>>
>> I forgot to run this test since I did not have the kernel feature
>> enabled. But yes, now that I tried it, it is failing.
>>
>> What is happening is that event del (so counter stop as well) is getting
>> called when the CPU goes offline, followed by add->start, and the
>> initial counter value then gets reloaded.
>>
>> I don't see a way for i915 to distinguish between userspace
>> starting/stopping the event, and perf core doing the same in the CPU
>> migration process. Perhaps Peter could help here?
>>
>> I am storing the initial counter value when the counter is started so
>> that I can report it's relative value. In other words the change from
>> event start to stop. Perhaps that is not correct and should be left to
>> userspace to handle?
>>
>> Otherwise we have counters like energy use, and even engine busyness,
>> which will/can already be at some large value before PMU monitoring
>> starts. Which makes things like "perf stat -a -I <command>", or even
>> just normal "perf stat <command>", attribute all previous usage (from
>> before the command profiling started) to the reported stats.
>>
>
> Actually that's pretty easy to fix. The following patch is doing that:
>
> diff --git a/drivers/gpu/drm/i915/i915_pmu.c
> b/drivers/gpu/drm/i915/i915_pmu.c
> index bce4951..277098d 100644
> --- a/drivers/gpu/drm/i915/i915_pmu.c
> +++ b/drivers/gpu/drm/i915/i915_pmu.c
> @@ -390,10 +390,18 @@ static u64 __i915_pmu_event_read(struct perf_event
> *event)
>
> static void i915_pmu_event_read(struct perf_event *event)
> {
> + struct hw_perf_event *hwc = &event->hw;
> + u64 prev_raw_count, new_raw_count;
>
> - local64_set(&event->count,
> - __i915_pmu_event_read(event) -
> - local64_read(&event->hw.prev_count));
> +again:
> + prev_raw_count = local64_read(&hwc->prev_count);
> + new_raw_count = __i915_pmu_event_read(event);
> +
> + if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
> + new_raw_count) != prev_raw_count)
> + goto again;
> +
> + local64_add(new_raw_count - prev_raw_count, &event->count);
> }
>
>
> I believe you need to squash it to the major i915 PMU enabling one.
>
> So, the idea is:
> 1. event->count contains current counter value, it is ever growing for
> the particular event _instance_, i.e. even if event is stopped/started
> or added/deleted, till this event exists (not destroyed) it holds ever
> growing value
> 2. Since it is ever growing, in the read() we always add a _delta_ to
> the event count where start point is when event got enabled (started)
> 3. On PMU context migration to another CPU we will be issued a call to
> del(PERF_EF_UPDATE). Thus, here is the trick:
> 3.1. The first thing we do we _update_ event count adding to the count
> everything we gathered on the previous CPU
> 3.2. The second thing - we update event->hw.prev_count to the new value.
> Next time we will add delta counting from it
>
> With this all tests I have for PMU passed. The code above is taken from
> arch/x86/events/intel/cstate.c as is. So, that's really how it should
> be.
Makes sense, simple solution which I overlooked. :I
I've sent a v6 with this fix and the updated v4 changelog. Just managed
to do it initially against the wrong message in the thread, hopefully
patchwork is not confused beyond repair now.
> And... I should say that was the last technical problem I saw for i915
> PMU implementation. With it puzzle looks complete for me. :).
So time to poke Chris I guess.
Regards,
Tvrtko
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx
next prev parent reply other threads:[~2017-09-13 9:01 UTC|newest]
Thread overview: 56+ messages / expand[flat|nested] mbox.gz Atom feed top
2017-09-11 15:25 [RFC v3 00/11] i915 PMU and engine busy stats Tvrtko Ursulin
2017-09-11 15:25 ` [RFC 01/11] drm/i915: Convert intel_rc6_residency_us to ns Tvrtko Ursulin
2017-09-14 19:48 ` Chris Wilson
2017-09-11 15:25 ` [RFC 02/11] drm/i915: Add intel_energy_uJ Tvrtko Ursulin
2017-09-14 19:49 ` Chris Wilson
2017-09-15 9:18 ` Tvrtko Ursulin
2017-09-14 20:36 ` Ville Syrjälä
2017-09-15 6:56 ` Tvrtko Ursulin
2017-09-15 8:51 ` Chris Wilson
2017-09-15 10:07 ` Tvrtko Ursulin
2017-09-15 10:34 ` Ville Syrjälä
2017-09-15 10:38 ` Chris Wilson
2017-09-15 11:16 ` Tvrtko Ursulin
2017-09-11 15:25 ` [RFC 03/11] drm/i915: Extract intel_get_cagf Tvrtko Ursulin
2017-09-14 19:51 ` Chris Wilson
2017-09-11 15:25 ` [RFC 04/11] drm/i915/pmu: Expose a PMU interface for perf queries Tvrtko Ursulin
2017-09-12 2:06 ` Rogozhkin, Dmitry V
2017-09-12 14:59 ` Tvrtko Ursulin
2017-09-13 8:57 ` [RFC v6 " Tvrtko Ursulin
2017-09-13 10:34 ` [RFC v7 " Tvrtko Ursulin
2017-09-15 0:00 ` Rogozhkin, Dmitry V
2017-09-15 7:57 ` Tvrtko Ursulin
2017-09-14 19:46 ` [RFC " Chris Wilson
2017-09-11 15:25 ` [RFC 05/11] drm/i915/pmu: Suspend sampling when GPU is idle Tvrtko Ursulin
2017-09-13 10:34 ` [RFC v5 " Tvrtko Ursulin
2017-09-14 19:57 ` Chris Wilson
2017-09-15 9:22 ` Tvrtko Ursulin
2017-09-11 15:25 ` [RFC 06/11] drm/i915: Wrap context schedule notification Tvrtko Ursulin
2017-09-11 15:25 ` [RFC 07/11] drm/i915: Engine busy time tracking Tvrtko Ursulin
2017-09-14 20:16 ` Chris Wilson
2017-09-15 9:45 ` Tvrtko Ursulin
2017-09-11 15:25 ` [RFC 08/11] drm/i915: Export engine busy stats in debugfs Tvrtko Ursulin
2017-09-14 20:17 ` Chris Wilson
2017-09-15 9:46 ` Tvrtko Ursulin
2017-09-11 15:25 ` [RFC 09/11] drm/i915/pmu: Wire up engine busy stats to PMU Tvrtko Ursulin
2017-09-11 15:25 ` [RFC 10/11] drm/i915: Export engine stats API to other users Tvrtko Ursulin
2017-09-12 18:35 ` Ben Widawsky
2017-09-14 20:26 ` Chris Wilson
2017-09-15 9:49 ` Tvrtko Ursulin
2017-09-19 19:50 ` Ben Widawsky
2017-09-19 20:11 ` Rogozhkin, Dmitry V
2017-09-29 10:59 ` Joonas Lahtinen
2017-09-11 15:25 ` [RFC 11/11] drm/i915: Gate engine stats collection with a static key Tvrtko Ursulin
2017-09-13 12:18 ` [RFC v3 " Tvrtko Ursulin
2017-09-14 20:22 ` Chris Wilson
2017-09-15 9:51 ` Tvrtko Ursulin
2017-09-11 15:50 ` ✗ Fi.CI.BAT: warning for i915 PMU and engine busy stats (rev3) Patchwork
2017-09-12 2:03 ` [RFC v3 00/11] i915 PMU and engine busy stats Rogozhkin, Dmitry V
2017-09-12 14:54 ` Tvrtko Ursulin
2017-09-12 22:01 ` Rogozhkin, Dmitry V
2017-09-13 8:54 ` [RFC v6 04/11] drm/i915/pmu: Expose a PMU interface for perf queries Tvrtko Ursulin
2017-09-13 9:01 ` Tvrtko Ursulin [this message]
2017-09-13 9:34 ` ✗ Fi.CI.BAT: warning for i915 PMU and engine busy stats (rev4) Patchwork
2017-09-13 10:46 ` ✗ Fi.CI.BAT: failure for i915 PMU and engine busy stats (rev6) Patchwork
2017-09-13 13:27 ` ✓ Fi.CI.BAT: success for i915 PMU and engine busy stats (rev7) Patchwork
2017-09-13 21:24 ` ✓ Fi.CI.IGT: " Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=234289d7-99e5-f78c-a34f-e28682e30576@linux.intel.com \
--to=tvrtko.ursulin@linux.intel.com \
--cc=Intel-gfx@lists.freedesktop.org \
--cc=dmitry.v.rogozhkin@intel.com \
--cc=peterz@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox