* [RFC/PATCH] perf/core: Use POLLHUP for a pinned event in error @ 2025-03-17 6:17 Namhyung Kim 2025-03-17 7:47 ` [tip: perf/core] perf/core: Use POLLHUP for pinned events " tip-bot2 for Namhyung Kim 2025-06-02 14:10 ` [RFC/PATCH] perf/core: Use POLLHUP for a pinned event " Lai, Yi 0 siblings, 2 replies; 6+ messages in thread From: Namhyung Kim @ 2025-03-17 6:17 UTC (permalink / raw) To: Peter Zijlstra, Ingo Molnar Cc: Kan Liang, Mark Rutland, Alexander Shishkin, Arnaldo Carvalho de Melo, LKML Pinned events can go to an error state when they are failed to be scheduled in the context. And they won't generate samples anymore and silently ignored until it's recovered by PERF_EVENT_IOC_ENABLE or something (and of course the condition also should be changed so that they can be scheduled in). But then users should know about the state change. Currently there's no mechanism to notify users when they go to an error state. One way to do this is to issue POLLHUP event to poll(2) to handle this. Reading events in an error state would return 0 (EOF) and it matches to the behavior of POLLHUP according to the man page. Users should remove the fd of the event from pollfd after getting POLLHUP, otherwise it'll be returned repeatedly. Signed-off-by: Namhyung Kim <namhyung@kernel.org> --- kernel/events/core.c | 9 +++++++++ 1 file changed, 9 insertions(+) diff --git a/kernel/events/core.c b/kernel/events/core.c index 2533fc32d890eacd..cef1f5c60f642d21 100644 --- a/kernel/events/core.c +++ b/kernel/events/core.c @@ -3984,6 +3984,11 @@ static int merge_sched_in(struct perf_event *event, void *data) if (event->attr.pinned) { perf_cgroup_event_disable(event, ctx); perf_event_set_state(event, PERF_EVENT_STATE_ERROR); + + if (*perf_event_fasync(event)) + event->pending_kill = POLL_HUP; + + perf_event_wakeup(event); } else { struct perf_cpu_pmu_context *cpc = this_cpc(event->pmu_ctx->pmu); @@ -5925,6 +5930,10 @@ static __poll_t perf_poll(struct file *file, poll_table *wait) if (is_event_hup(event)) return events; + if (unlikely(READ_ONCE(event->state) == PERF_EVENT_STATE_ERROR && + event->attr.pinned)) + return events; + /* * Pin the event->rb by taking event->mmap_mutex; otherwise * perf_event_set_output() can swizzle our rb and make us miss wakeups. -- 2.49.0.rc1.451.g8f38331e32-goog ^ permalink raw reply related [flat|nested] 6+ messages in thread
* [tip: perf/core] perf/core: Use POLLHUP for pinned events in error 2025-03-17 6:17 [RFC/PATCH] perf/core: Use POLLHUP for a pinned event in error Namhyung Kim @ 2025-03-17 7:47 ` tip-bot2 for Namhyung Kim 2025-06-02 14:10 ` [RFC/PATCH] perf/core: Use POLLHUP for a pinned event " Lai, Yi 1 sibling, 0 replies; 6+ messages in thread From: tip-bot2 for Namhyung Kim @ 2025-03-17 7:47 UTC (permalink / raw) To: linux-tip-commits Cc: Namhyung Kim, Ingo Molnar, Peter Zijlstra, Arnaldo Carvalho de Melo, H. Peter Anvin, Linus Torvalds, x86, linux-kernel The following commit has been merged into the perf/core branch of tip: Commit-ID: f4b07fd62d4d11d57a15cb4ae01b3833282eb8f6 Gitweb: https://git.kernel.org/tip/f4b07fd62d4d11d57a15cb4ae01b3833282eb8f6 Author: Namhyung Kim <namhyung@kernel.org> AuthorDate: Sun, 16 Mar 2025 23:17:45 -07:00 Committer: Ingo Molnar <mingo@kernel.org> CommitterDate: Mon, 17 Mar 2025 08:31:03 +01:00 perf/core: Use POLLHUP for pinned events in error Pinned performance events can enter an error state when they fail to be scheduled in the context due to a failed constraint or some other conflict or condition. In error state these events won't generate any samples anymore and are silently ignored until they are recovered by PERF_EVENT_IOC_ENABLE, or the condition can also change so that they can be scheduled in. Tooling should be allowed to know about the state change, but currently there's no mechanism to notify tooling when events enter an error state. One way to do this is to issue a POLLHUP event to poll(2) to handle this. Reading events in an error state would return 0 (EOF) and it matches to the behavior of POLLHUP according to the man page. Tooling should remove the fd of the event from pollfd after getting POLLHUP, otherwise it'll be returned repeatedly. [ mingo: Clarified the changelog ] Signed-off-by: Namhyung Kim <namhyung@kernel.org> Signed-off-by: Ingo Molnar <mingo@kernel.org> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Arnaldo Carvalho de Melo <acme@redhat.com> Cc: "H. Peter Anvin" <hpa@zytor.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Link: https://lore.kernel.org/r/20250317061745.1777584-1-namhyung@kernel.org --- kernel/events/core.c | 9 +++++++++ 1 file changed, 9 insertions(+) diff --git a/kernel/events/core.c b/kernel/events/core.c index 2533fc3..ace1bcc 100644 --- a/kernel/events/core.c +++ b/kernel/events/core.c @@ -3984,6 +3984,11 @@ static int merge_sched_in(struct perf_event *event, void *data) if (event->attr.pinned) { perf_cgroup_event_disable(event, ctx); perf_event_set_state(event, PERF_EVENT_STATE_ERROR); + + if (*perf_event_fasync(event)) + event->pending_kill = POLL_HUP; + + perf_event_wakeup(event); } else { struct perf_cpu_pmu_context *cpc = this_cpc(event->pmu_ctx->pmu); @@ -5925,6 +5930,10 @@ static __poll_t perf_poll(struct file *file, poll_table *wait) if (is_event_hup(event)) return events; + if (unlikely(READ_ONCE(event->state) == PERF_EVENT_STATE_ERROR && + event->attr.pinned)) + return events; + /* * Pin the event->rb by taking event->mmap_mutex; otherwise * perf_event_set_output() can swizzle our rb and make us miss wakeups. ^ permalink raw reply related [flat|nested] 6+ messages in thread
* Re: [RFC/PATCH] perf/core: Use POLLHUP for a pinned event in error 2025-03-17 6:17 [RFC/PATCH] perf/core: Use POLLHUP for a pinned event in error Namhyung Kim 2025-03-17 7:47 ` [tip: perf/core] perf/core: Use POLLHUP for pinned events " tip-bot2 for Namhyung Kim @ 2025-06-02 14:10 ` Lai, Yi 2025-06-02 17:32 ` Namhyung Kim 1 sibling, 1 reply; 6+ messages in thread From: Lai, Yi @ 2025-06-02 14:10 UTC (permalink / raw) To: Namhyung Kim Cc: Peter Zijlstra, Ingo Molnar, Kan Liang, Mark Rutland, Alexander Shishkin, Arnaldo Carvalho de Melo, LKML, yi1.lai Hi Namhyung Kim, Greetings! I used Syzkaller and found that there is WARNING: locking bug in perf_event_wakeup in linux-next next-20250530. After bisection and the first bad commit is: " f4b07fd62d4d perf/core: Use POLLHUP for pinned events in error " All detailed into can be found at: https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup Syzkaller repro code: https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.c Syzkaller repro syscall steps: https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.prog Syzkaller report: https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.report Kconfig(make olddefconfig): https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/kconfig_origin Bisect info: https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/bisect_info.log bzImage: https://github.com/laifryiee/syzkaller_logs/raw/refs/heads/main/250601_162355_perf_event_wakeup/bzImage_next-20250530 Issue dmesg: https://github.com/laifryiee/syzkaller_logs/blob/main/250601_162355_perf_event_wakeup/next-20250530_dmesg.log " [ 39.913691] ============================= [ 39.914157] [ BUG: Invalid wait context ] [ 39.914623] 6.15.0-next-20250530-next-2025053 #1 Not tainted [ 39.915271] ----------------------------- [ 39.915731] repro/837 is trying to lock: [ 39.916191] ffff88801acfabd8 (&event->waitq){....}-{3:3}, at: __wake_up+0x26/0x60 [ 39.917182] other info that might help us debug this: [ 39.917761] context-{5:5} [ 39.918079] 4 locks held by repro/837: [ 39.918530] #0: ffffffff8725cd00 (rcu_read_lock){....}-{1:3}, at: __perf_event_task_sched_in+0xd1/0xbc0 [ 39.919612] #1: ffff88806ca3c6f8 (&cpuctx_lock){....}-{2:2}, at: __perf_event_task_sched_in+0x1a7/0xbc0 [ 39.920748] #2: ffff88800d91fc18 (&ctx->lock){....}-{2:2}, at: __perf_event_task_sched_in+0x1f9/0xbc0 [ 39.921819] #3: ffffffff8725cd00 (rcu_read_lock){....}-{1:3}, at: perf_event_wakeup+0x6c/0x470 [ 39.922823] stack backtrace: [ 39.923171] CPU: 0 UID: 0 PID: 837 Comm: repro Not tainted 6.15.0-next-20250530-next-2025053 #1 PREEMPT(voluntary) [ 39.923196] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS rel-1.16.3-0-ga6ed6b701f0a-prebuilt.qemu.o4 [ 39.923214] Call Trace: [ 39.923221] <TASK> [ 39.923228] dump_stack_lvl+0xea/0x150 [ 39.923256] dump_stack+0x19/0x20 [ 39.923276] __lock_acquire+0xb22/0x22a0 [ 39.923308] ? x86_pmu_commit_txn+0x195/0x2b0 [ 39.923339] ? __lock_acquire+0x412/0x22a0 [ 39.923375] lock_acquire+0x170/0x310 [ 39.923407] ? __wake_up+0x26/0x60 [ 39.923448] _raw_spin_lock_irqsave+0x52/0x80 [ 39.923473] ? __wake_up+0x26/0x60 [ 39.923504] __wake_up+0x26/0x60 [ 39.923537] perf_event_wakeup+0x14a/0x470 [ 39.923571] merge_sched_in+0x846/0x15c0 [ 39.923610] visit_groups_merge.constprop.0.isra.0+0x952/0x1420 [ 39.923653] ? __pfx_visit_groups_merge.constprop.0.isra.0+0x10/0x10 [ 39.923688] ? sched_clock_noinstr+0x12/0x20 [ 39.923724] ? __sanitizer_cov_trace_const_cmp1+0x1e/0x30 [ 39.923766] ctx_sched_in+0x471/0xa20 [ 39.923804] ? __pfx_ctx_sched_in+0x10/0x10 [ 39.923838] ? __sanitizer_cov_trace_const_cmp4+0x1a/0x20 [ 39.923878] perf_event_sched_in+0x47/0xa0 [ 39.923912] __perf_event_task_sched_in+0x3fc/0xbc0 [ 39.923951] ? __pfx___perf_event_task_sched_in+0x10/0x10 [ 39.923984] ? __this_cpu_preempt_check+0x21/0x30 [ 39.924012] ? __sanitizer_cov_trace_cmp8+0x1c/0x30 [ 39.924046] ? xfd_validate_state+0x14f/0x1b0 [ 39.924081] finish_task_switch.isra.0+0x525/0x990 [ 39.924117] ? lock_unpin_lock+0xdc/0x170 [ 39.924152] __schedule+0xef3/0x3840 [ 39.924185] ? __pfx___schedule+0x10/0x10 [ 39.924218] ? ktime_get_coarse_real_ts64+0xad/0xf0 [ 39.924259] schedule+0xf6/0x3d0 [ 39.924285] exit_to_user_mode_loop+0x7a/0x110 [ 39.924315] do_syscall_64+0x284/0x2e0 [ 39.924340] entry_SYSCALL_64_after_hwframe+0x76/0x7e [ 39.924360] RIP: 0033:0x7ff14103ee5d [ 39.924381] Code: ff c3 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 8 [ 39.924400] RSP: 002b:00007fffb2745578 EFLAGS: 00000202 ORIG_RAX: 0000000000000038 [ 39.924418] RAX: 0000000000000346 RBX: 0000000000000000 RCX: 00007ff14103ee5d [ 39.924431] RDX: 0000000000000000 RSI: 0000000000000000 RDI: 0000000000082000 [ 39.924443] RBP: 00007fffb27455c0 R08: 0000000000000000 R09: 0000000000000000 [ 39.924456] R10: 0000000000000000 R11: 0000000000000202 R12: 00007fffb27459a8 [ 39.924468] R13: 0000000000404e78 R14: 0000000000406e08 R15: 00007ff141389000 [ 39.924497] </TASK> [ 40.307815] coredump: 804(repro): over core_pipe_limit, skipping core dump [ 40.472093] coredump: 795(repro): over core_pipe_limit, skipping core dump [ 40.545575] coredump: 799(repro): over core_pipe_limit, skipping core dump [ 40.948915] coredump: 833(repro): over core_pipe_limit, skipping core dump [ 40.989336] coredump: 811(repro): over core_pipe_limit, skipping core dump [ 42.121469] coredump: 857(repro): over core_pipe_limit, skipping core dump " Hope this cound be insightful to you. Regards, Yi Lai --- If you don't need the following environment to reproduce the problem or if you already have one reproduced environment, please ignore the following information. How to reproduce: git clone https://gitlab.com/xupengfe/repro_vm_env.git cd repro_vm_env tar -xvf repro_vm_env.tar.gz cd repro_vm_env; ./start3.sh // it needs qemu-system-x86_64 and I used v7.1.0 // start3.sh will load bzImage_2241ab53cbb5cdb08a6b2d4688feb13971058f65 v6.2-rc5 kernel // You could change the bzImage_xxx as you want // Maybe you need to remove line "-drive if=pflash,format=raw,readonly=on,file=./OVMF_CODE.fd \" for different qemu version You could use below command to log in, there is no password for root. ssh -p 10023 root@localhost After login vm(virtual machine) successfully, you could transfer reproduced binary to the vm by below way, and reproduce the problem in vm: gcc -pthread -o repro repro.c scp -P 10023 repro root@localhost:/root/ Get the bzImage for target kernel: Please use target kconfig and copy it to kernel_src/.config make olddefconfig make -jx bzImage //x should equal or less than cpu num your pc has Fill the bzImage file into above start3.sh to load the target kernel in vm. Tips: If you already have qemu-system-x86_64, please ignore below info. If you want to install qemu v7.1.0 version: git clone https://github.com/qemu/qemu.git cd qemu git checkout -f v7.1.0 mkdir build cd build yum install -y ninja-build.x86_64 yum -y install libslirp-devel.x86_64 ../configure --target-list=x86_64-softmmu --enable-kvm --enable-vnc --enable-gtk --enable-sdl --enable-usb-redir --enable-slirp make make install On Sun, Mar 16, 2025 at 11:17:45PM -0700, Namhyung Kim wrote: > Pinned events can go to an error state when they are failed to be > scheduled in the context. And they won't generate samples anymore > and silently ignored until it's recovered by PERF_EVENT_IOC_ENABLE or > something (and of course the condition also should be changed so that > they can be scheduled in). But then users should know about the state > change. > > Currently there's no mechanism to notify users when they go to an error > state. > > One way to do this is to issue POLLHUP event to poll(2) to handle this. > Reading events in an error state would return 0 (EOF) and it matches to > the behavior of POLLHUP according to the man page. > > Users should remove the fd of the event from pollfd after getting > POLLHUP, otherwise it'll be returned repeatedly. > > Signed-off-by: Namhyung Kim <namhyung@kernel.org> > --- > kernel/events/core.c | 9 +++++++++ > 1 file changed, 9 insertions(+) > > diff --git a/kernel/events/core.c b/kernel/events/core.c > index 2533fc32d890eacd..cef1f5c60f642d21 100644 > --- a/kernel/events/core.c > +++ b/kernel/events/core.c > @@ -3984,6 +3984,11 @@ static int merge_sched_in(struct perf_event *event, void *data) > if (event->attr.pinned) { > perf_cgroup_event_disable(event, ctx); > perf_event_set_state(event, PERF_EVENT_STATE_ERROR); > + > + if (*perf_event_fasync(event)) > + event->pending_kill = POLL_HUP; > + > + perf_event_wakeup(event); > } else { > struct perf_cpu_pmu_context *cpc = this_cpc(event->pmu_ctx->pmu); > > @@ -5925,6 +5930,10 @@ static __poll_t perf_poll(struct file *file, poll_table *wait) > if (is_event_hup(event)) > return events; > > + if (unlikely(READ_ONCE(event->state) == PERF_EVENT_STATE_ERROR && > + event->attr.pinned)) > + return events; > + > /* > * Pin the event->rb by taking event->mmap_mutex; otherwise > * perf_event_set_output() can swizzle our rb and make us miss wakeups. > -- > 2.49.0.rc1.451.g8f38331e32-goog > ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [RFC/PATCH] perf/core: Use POLLHUP for a pinned event in error 2025-06-02 14:10 ` [RFC/PATCH] perf/core: Use POLLHUP for a pinned event " Lai, Yi @ 2025-06-02 17:32 ` Namhyung Kim 2025-06-03 1:48 ` Lai, Yi 0 siblings, 1 reply; 6+ messages in thread From: Namhyung Kim @ 2025-06-02 17:32 UTC (permalink / raw) To: Lai, Yi Cc: Peter Zijlstra, Ingo Molnar, Kan Liang, Mark Rutland, Alexander Shishkin, Arnaldo Carvalho de Melo, LKML, yi1.lai Hello, On Mon, Jun 02, 2025 at 10:10:47PM +0800, Lai, Yi wrote: > Hi Namhyung Kim, > > Greetings! > > I used Syzkaller and found that there is WARNING: locking bug in perf_event_wakeup in linux-next next-20250530. > > After bisection and the first bad commit is: > " > f4b07fd62d4d perf/core: Use POLLHUP for pinned events in error > " > > All detailed into can be found at: > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup > Syzkaller repro code: > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.c > Syzkaller repro syscall steps: > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.prog > Syzkaller report: > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.report > Kconfig(make olddefconfig): > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/kconfig_origin > Bisect info: > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/bisect_info.log > bzImage: > https://github.com/laifryiee/syzkaller_logs/raw/refs/heads/main/250601_162355_perf_event_wakeup/bzImage_next-20250530 > Issue dmesg: > https://github.com/laifryiee/syzkaller_logs/blob/main/250601_162355_perf_event_wakeup/next-20250530_dmesg.log > > " > [ 39.913691] ============================= > [ 39.914157] [ BUG: Invalid wait context ] > [ 39.914623] 6.15.0-next-20250530-next-2025053 #1 Not tainted > [ 39.915271] ----------------------------- > [ 39.915731] repro/837 is trying to lock: > [ 39.916191] ffff88801acfabd8 (&event->waitq){....}-{3:3}, at: __wake_up+0x26/0x60 > [ 39.917182] other info that might help us debug this: > [ 39.917761] context-{5:5} > [ 39.918079] 4 locks held by repro/837: > [ 39.918530] #0: ffffffff8725cd00 (rcu_read_lock){....}-{1:3}, at: __perf_event_task_sched_in+0xd1/0xbc0 > [ 39.919612] #1: ffff88806ca3c6f8 (&cpuctx_lock){....}-{2:2}, at: __perf_event_task_sched_in+0x1a7/0xbc0 > [ 39.920748] #2: ffff88800d91fc18 (&ctx->lock){....}-{2:2}, at: __perf_event_task_sched_in+0x1f9/0xbc0 > [ 39.921819] #3: ffffffff8725cd00 (rcu_read_lock){....}-{1:3}, at: perf_event_wakeup+0x6c/0x470 > [ 39.922823] stack backtrace: > [ 39.923171] CPU: 0 UID: 0 PID: 837 Comm: repro Not tainted 6.15.0-next-20250530-next-2025053 #1 PREEMPT(voluntary) > [ 39.923196] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS rel-1.16.3-0-ga6ed6b701f0a-prebuilt.qemu.o4 > [ 39.923214] Call Trace: > [ 39.923221] <TASK> > [ 39.923228] dump_stack_lvl+0xea/0x150 > [ 39.923256] dump_stack+0x19/0x20 > [ 39.923276] __lock_acquire+0xb22/0x22a0 > [ 39.923308] ? x86_pmu_commit_txn+0x195/0x2b0 > [ 39.923339] ? __lock_acquire+0x412/0x22a0 > [ 39.923375] lock_acquire+0x170/0x310 > [ 39.923407] ? __wake_up+0x26/0x60 > [ 39.923448] _raw_spin_lock_irqsave+0x52/0x80 > [ 39.923473] ? __wake_up+0x26/0x60 > [ 39.923504] __wake_up+0x26/0x60 > [ 39.923537] perf_event_wakeup+0x14a/0x470 > [ 39.923571] merge_sched_in+0x846/0x15c0 > [ 39.923610] visit_groups_merge.constprop.0.isra.0+0x952/0x1420 > [ 39.923653] ? __pfx_visit_groups_merge.constprop.0.isra.0+0x10/0x10 > [ 39.923688] ? sched_clock_noinstr+0x12/0x20 > [ 39.923724] ? __sanitizer_cov_trace_const_cmp1+0x1e/0x30 > [ 39.923766] ctx_sched_in+0x471/0xa20 > [ 39.923804] ? __pfx_ctx_sched_in+0x10/0x10 > [ 39.923838] ? __sanitizer_cov_trace_const_cmp4+0x1a/0x20 > [ 39.923878] perf_event_sched_in+0x47/0xa0 > [ 39.923912] __perf_event_task_sched_in+0x3fc/0xbc0 > [ 39.923951] ? __pfx___perf_event_task_sched_in+0x10/0x10 > [ 39.923984] ? __this_cpu_preempt_check+0x21/0x30 > [ 39.924012] ? __sanitizer_cov_trace_cmp8+0x1c/0x30 > [ 39.924046] ? xfd_validate_state+0x14f/0x1b0 > [ 39.924081] finish_task_switch.isra.0+0x525/0x990 > [ 39.924117] ? lock_unpin_lock+0xdc/0x170 > [ 39.924152] __schedule+0xef3/0x3840 > [ 39.924185] ? __pfx___schedule+0x10/0x10 > [ 39.924218] ? ktime_get_coarse_real_ts64+0xad/0xf0 > [ 39.924259] schedule+0xf6/0x3d0 > [ 39.924285] exit_to_user_mode_loop+0x7a/0x110 > [ 39.924315] do_syscall_64+0x284/0x2e0 > [ 39.924340] entry_SYSCALL_64_after_hwframe+0x76/0x7e > [ 39.924360] RIP: 0033:0x7ff14103ee5d > [ 39.924381] Code: ff c3 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 8 > [ 39.924400] RSP: 002b:00007fffb2745578 EFLAGS: 00000202 ORIG_RAX: 0000000000000038 > [ 39.924418] RAX: 0000000000000346 RBX: 0000000000000000 RCX: 00007ff14103ee5d > [ 39.924431] RDX: 0000000000000000 RSI: 0000000000000000 RDI: 0000000000082000 > [ 39.924443] RBP: 00007fffb27455c0 R08: 0000000000000000 R09: 0000000000000000 > [ 39.924456] R10: 0000000000000000 R11: 0000000000000202 R12: 00007fffb27459a8 > [ 39.924468] R13: 0000000000404e78 R14: 0000000000406e08 R15: 00007ff141389000 > [ 39.924497] </TASK> > [ 40.307815] coredump: 804(repro): over core_pipe_limit, skipping core dump > [ 40.472093] coredump: 795(repro): over core_pipe_limit, skipping core dump > [ 40.545575] coredump: 799(repro): over core_pipe_limit, skipping core dump > [ 40.948915] coredump: 833(repro): over core_pipe_limit, skipping core dump > [ 40.989336] coredump: 811(repro): over core_pipe_limit, skipping core dump > [ 42.121469] coredump: 857(repro): over core_pipe_limit, skipping core dump > " > > Hope this cound be insightful to you. > > Regards, > Yi Lai > > --- > > If you don't need the following environment to reproduce the problem or if you > already have one reproduced environment, please ignore the following information. > > How to reproduce: > git clone https://gitlab.com/xupengfe/repro_vm_env.git > cd repro_vm_env > tar -xvf repro_vm_env.tar.gz > cd repro_vm_env; ./start3.sh // it needs qemu-system-x86_64 and I used v7.1.0 > // start3.sh will load bzImage_2241ab53cbb5cdb08a6b2d4688feb13971058f65 v6.2-rc5 kernel > // You could change the bzImage_xxx as you want > // Maybe you need to remove line "-drive if=pflash,format=raw,readonly=on,file=./OVMF_CODE.fd \" for different qemu version > You could use below command to log in, there is no password for root. > ssh -p 10023 root@localhost > > After login vm(virtual machine) successfully, you could transfer reproduced > binary to the vm by below way, and reproduce the problem in vm: > gcc -pthread -o repro repro.c > scp -P 10023 repro root@localhost:/root/ > > Get the bzImage for target kernel: > Please use target kconfig and copy it to kernel_src/.config > make olddefconfig > make -jx bzImage //x should equal or less than cpu num your pc has > > Fill the bzImage file into above start3.sh to load the target kernel in vm. > > > Tips: > If you already have qemu-system-x86_64, please ignore below info. > If you want to install qemu v7.1.0 version: > git clone https://github.com/qemu/qemu.git > cd qemu > git checkout -f v7.1.0 > mkdir build > cd build > yum install -y ninja-build.x86_64 > yum -y install libslirp-devel.x86_64 > ../configure --target-list=x86_64-softmmu --enable-kvm --enable-vnc --enable-gtk --enable-sdl --enable-usb-redir --enable-slirp > make > make install Thanks for the detailed instruction. I was able to reproduce it with your setup. The below patch fixes it for me. Can you please double check if it works well for you? Thanks, Namhyung ---8--- diff --git a/kernel/events/core.c b/kernel/events/core.c index f34c99f8ce8f446a..e22eb88eb105b95b 100644 --- a/kernel/events/core.c +++ b/kernel/events/core.c @@ -3995,7 +3995,8 @@ static int merge_sched_in(struct perf_event *event, void *data) if (*perf_event_fasync(event)) event->pending_kill = POLL_ERR; - perf_event_wakeup(event); + event->pending_wakeup = 1; + irq_work_queue(&event->pending_irq); } else { struct perf_cpu_pmu_context *cpc = this_cpc(event->pmu_ctx->pmu); -- 2.49.0.1204.g71687c7c1d-goog ^ permalink raw reply related [flat|nested] 6+ messages in thread
* Re: [RFC/PATCH] perf/core: Use POLLHUP for a pinned event in error 2025-06-02 17:32 ` Namhyung Kim @ 2025-06-03 1:48 ` Lai, Yi 2025-06-03 4:49 ` Namhyung Kim 0 siblings, 1 reply; 6+ messages in thread From: Lai, Yi @ 2025-06-03 1:48 UTC (permalink / raw) To: Namhyung Kim Cc: Peter Zijlstra, Ingo Molnar, Kan Liang, Mark Rutland, Alexander Shishkin, Arnaldo Carvalho de Melo, LKML, yi1.lai On Mon, Jun 02, 2025 at 10:32:08AM -0700, Namhyung Kim wrote: > Hello, > > On Mon, Jun 02, 2025 at 10:10:47PM +0800, Lai, Yi wrote: > > Hi Namhyung Kim, > > > > Greetings! > > > > I used Syzkaller and found that there is WARNING: locking bug in perf_event_wakeup in linux-next next-20250530. > > > > After bisection and the first bad commit is: > > " > > f4b07fd62d4d perf/core: Use POLLHUP for pinned events in error > > " > > > > All detailed into can be found at: > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup > > Syzkaller repro code: > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.c > > Syzkaller repro syscall steps: > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.prog > > Syzkaller report: > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.report > > Kconfig(make olddefconfig): > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/kconfig_origin > > Bisect info: > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/bisect_info.log > > bzImage: > > https://github.com/laifryiee/syzkaller_logs/raw/refs/heads/main/250601_162355_perf_event_wakeup/bzImage_next-20250530 > > Issue dmesg: > > https://github.com/laifryiee/syzkaller_logs/blob/main/250601_162355_perf_event_wakeup/next-20250530_dmesg.log > > > > " > > [ 39.913691] ============================= > > [ 39.914157] [ BUG: Invalid wait context ] > > [ 39.914623] 6.15.0-next-20250530-next-2025053 #1 Not tainted > > [ 39.915271] ----------------------------- > > [ 39.915731] repro/837 is trying to lock: > > [ 39.916191] ffff88801acfabd8 (&event->waitq){....}-{3:3}, at: __wake_up+0x26/0x60 > > [ 39.917182] other info that might help us debug this: > > [ 39.917761] context-{5:5} > > [ 39.918079] 4 locks held by repro/837: > > [ 39.918530] #0: ffffffff8725cd00 (rcu_read_lock){....}-{1:3}, at: __perf_event_task_sched_in+0xd1/0xbc0 > > [ 39.919612] #1: ffff88806ca3c6f8 (&cpuctx_lock){....}-{2:2}, at: __perf_event_task_sched_in+0x1a7/0xbc0 > > [ 39.920748] #2: ffff88800d91fc18 (&ctx->lock){....}-{2:2}, at: __perf_event_task_sched_in+0x1f9/0xbc0 > > [ 39.921819] #3: ffffffff8725cd00 (rcu_read_lock){....}-{1:3}, at: perf_event_wakeup+0x6c/0x470 > > [ 39.922823] stack backtrace: > > [ 39.923171] CPU: 0 UID: 0 PID: 837 Comm: repro Not tainted 6.15.0-next-20250530-next-2025053 #1 PREEMPT(voluntary) > > [ 39.923196] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS rel-1.16.3-0-ga6ed6b701f0a-prebuilt.qemu.o4 > > [ 39.923214] Call Trace: > > [ 39.923221] <TASK> > > [ 39.923228] dump_stack_lvl+0xea/0x150 > > [ 39.923256] dump_stack+0x19/0x20 > > [ 39.923276] __lock_acquire+0xb22/0x22a0 > > [ 39.923308] ? x86_pmu_commit_txn+0x195/0x2b0 > > [ 39.923339] ? __lock_acquire+0x412/0x22a0 > > [ 39.923375] lock_acquire+0x170/0x310 > > [ 39.923407] ? __wake_up+0x26/0x60 > > [ 39.923448] _raw_spin_lock_irqsave+0x52/0x80 > > [ 39.923473] ? __wake_up+0x26/0x60 > > [ 39.923504] __wake_up+0x26/0x60 > > [ 39.923537] perf_event_wakeup+0x14a/0x470 > > [ 39.923571] merge_sched_in+0x846/0x15c0 > > [ 39.923610] visit_groups_merge.constprop.0.isra.0+0x952/0x1420 > > [ 39.923653] ? __pfx_visit_groups_merge.constprop.0.isra.0+0x10/0x10 > > [ 39.923688] ? sched_clock_noinstr+0x12/0x20 > > [ 39.923724] ? __sanitizer_cov_trace_const_cmp1+0x1e/0x30 > > [ 39.923766] ctx_sched_in+0x471/0xa20 > > [ 39.923804] ? __pfx_ctx_sched_in+0x10/0x10 > > [ 39.923838] ? __sanitizer_cov_trace_const_cmp4+0x1a/0x20 > > [ 39.923878] perf_event_sched_in+0x47/0xa0 > > [ 39.923912] __perf_event_task_sched_in+0x3fc/0xbc0 > > [ 39.923951] ? __pfx___perf_event_task_sched_in+0x10/0x10 > > [ 39.923984] ? __this_cpu_preempt_check+0x21/0x30 > > [ 39.924012] ? __sanitizer_cov_trace_cmp8+0x1c/0x30 > > [ 39.924046] ? xfd_validate_state+0x14f/0x1b0 > > [ 39.924081] finish_task_switch.isra.0+0x525/0x990 > > [ 39.924117] ? lock_unpin_lock+0xdc/0x170 > > [ 39.924152] __schedule+0xef3/0x3840 > > [ 39.924185] ? __pfx___schedule+0x10/0x10 > > [ 39.924218] ? ktime_get_coarse_real_ts64+0xad/0xf0 > > [ 39.924259] schedule+0xf6/0x3d0 > > [ 39.924285] exit_to_user_mode_loop+0x7a/0x110 > > [ 39.924315] do_syscall_64+0x284/0x2e0 > > [ 39.924340] entry_SYSCALL_64_after_hwframe+0x76/0x7e > > [ 39.924360] RIP: 0033:0x7ff14103ee5d > > [ 39.924381] Code: ff c3 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 8 > > [ 39.924400] RSP: 002b:00007fffb2745578 EFLAGS: 00000202 ORIG_RAX: 0000000000000038 > > [ 39.924418] RAX: 0000000000000346 RBX: 0000000000000000 RCX: 00007ff14103ee5d > > [ 39.924431] RDX: 0000000000000000 RSI: 0000000000000000 RDI: 0000000000082000 > > [ 39.924443] RBP: 00007fffb27455c0 R08: 0000000000000000 R09: 0000000000000000 > > [ 39.924456] R10: 0000000000000000 R11: 0000000000000202 R12: 00007fffb27459a8 > > [ 39.924468] R13: 0000000000404e78 R14: 0000000000406e08 R15: 00007ff141389000 > > [ 39.924497] </TASK> > > [ 40.307815] coredump: 804(repro): over core_pipe_limit, skipping core dump > > [ 40.472093] coredump: 795(repro): over core_pipe_limit, skipping core dump > > [ 40.545575] coredump: 799(repro): over core_pipe_limit, skipping core dump > > [ 40.948915] coredump: 833(repro): over core_pipe_limit, skipping core dump > > [ 40.989336] coredump: 811(repro): over core_pipe_limit, skipping core dump > > [ 42.121469] coredump: 857(repro): over core_pipe_limit, skipping core dump > > " > > > > Hope this cound be insightful to you. > > > > Regards, > > Yi Lai > > > > --- > > > > If you don't need the following environment to reproduce the problem or if you > > already have one reproduced environment, please ignore the following information. > > > > How to reproduce: > > git clone https://gitlab.com/xupengfe/repro_vm_env.git > > cd repro_vm_env > > tar -xvf repro_vm_env.tar.gz > > cd repro_vm_env; ./start3.sh // it needs qemu-system-x86_64 and I used v7.1.0 > > // start3.sh will load bzImage_2241ab53cbb5cdb08a6b2d4688feb13971058f65 v6.2-rc5 kernel > > // You could change the bzImage_xxx as you want > > // Maybe you need to remove line "-drive if=pflash,format=raw,readonly=on,file=./OVMF_CODE.fd \" for different qemu version > > You could use below command to log in, there is no password for root. > > ssh -p 10023 root@localhost > > > > After login vm(virtual machine) successfully, you could transfer reproduced > > binary to the vm by below way, and reproduce the problem in vm: > > gcc -pthread -o repro repro.c > > scp -P 10023 repro root@localhost:/root/ > > > > Get the bzImage for target kernel: > > Please use target kconfig and copy it to kernel_src/.config > > make olddefconfig > > make -jx bzImage //x should equal or less than cpu num your pc has > > > > Fill the bzImage file into above start3.sh to load the target kernel in vm. > > > > > > Tips: > > If you already have qemu-system-x86_64, please ignore below info. > > If you want to install qemu v7.1.0 version: > > git clone https://github.com/qemu/qemu.git > > cd qemu > > git checkout -f v7.1.0 > > mkdir build > > cd build > > yum install -y ninja-build.x86_64 > > yum -y install libslirp-devel.x86_64 > > ../configure --target-list=x86_64-softmmu --enable-kvm --enable-vnc --enable-gtk --enable-sdl --enable-usb-redir --enable-slirp > > make > > make install > > Thanks for the detailed instruction. I was able to reproduce it with > your setup. The below patch fixes it for me. Can you please double > check if it works well for you? > > Thanks, > Namhyung > After applying following patch on top of latest linux-next, issue cannot be reproduced. Regards, Yi Lai > ---8--- > > diff --git a/kernel/events/core.c b/kernel/events/core.c > index f34c99f8ce8f446a..e22eb88eb105b95b 100644 > --- a/kernel/events/core.c > +++ b/kernel/events/core.c > @@ -3995,7 +3995,8 @@ static int merge_sched_in(struct perf_event *event, void *data) > if (*perf_event_fasync(event)) > event->pending_kill = POLL_ERR; > > - perf_event_wakeup(event); > + event->pending_wakeup = 1; > + irq_work_queue(&event->pending_irq); > } else { > struct perf_cpu_pmu_context *cpc = this_cpc(event->pmu_ctx->pmu); > > -- > 2.49.0.1204.g71687c7c1d-goog > ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [RFC/PATCH] perf/core: Use POLLHUP for a pinned event in error 2025-06-03 1:48 ` Lai, Yi @ 2025-06-03 4:49 ` Namhyung Kim 0 siblings, 0 replies; 6+ messages in thread From: Namhyung Kim @ 2025-06-03 4:49 UTC (permalink / raw) To: Lai, Yi Cc: Peter Zijlstra, Ingo Molnar, Kan Liang, Mark Rutland, Alexander Shishkin, Arnaldo Carvalho de Melo, LKML, yi1.lai On Tue, Jun 03, 2025 at 09:48:26AM +0800, Lai, Yi wrote: > On Mon, Jun 02, 2025 at 10:32:08AM -0700, Namhyung Kim wrote: > > Hello, > > > > On Mon, Jun 02, 2025 at 10:10:47PM +0800, Lai, Yi wrote: > > > Hi Namhyung Kim, > > > > > > Greetings! > > > > > > I used Syzkaller and found that there is WARNING: locking bug in perf_event_wakeup in linux-next next-20250530. > > > > > > After bisection and the first bad commit is: > > > " > > > f4b07fd62d4d perf/core: Use POLLHUP for pinned events in error > > > " > > > > > > All detailed into can be found at: > > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup > > > Syzkaller repro code: > > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.c > > > Syzkaller repro syscall steps: > > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.prog > > > Syzkaller report: > > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/repro.report > > > Kconfig(make olddefconfig): > > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/kconfig_origin > > > Bisect info: > > > https://github.com/laifryiee/syzkaller_logs/tree/main/250601_162355_perf_event_wakeup/bisect_info.log > > > bzImage: > > > https://github.com/laifryiee/syzkaller_logs/raw/refs/heads/main/250601_162355_perf_event_wakeup/bzImage_next-20250530 > > > Issue dmesg: > > > https://github.com/laifryiee/syzkaller_logs/blob/main/250601_162355_perf_event_wakeup/next-20250530_dmesg.log > > > > > > " > > > [ 39.913691] ============================= > > > [ 39.914157] [ BUG: Invalid wait context ] > > > [ 39.914623] 6.15.0-next-20250530-next-2025053 #1 Not tainted > > > [ 39.915271] ----------------------------- > > > [ 39.915731] repro/837 is trying to lock: > > > [ 39.916191] ffff88801acfabd8 (&event->waitq){....}-{3:3}, at: __wake_up+0x26/0x60 > > > [ 39.917182] other info that might help us debug this: > > > [ 39.917761] context-{5:5} > > > [ 39.918079] 4 locks held by repro/837: > > > [ 39.918530] #0: ffffffff8725cd00 (rcu_read_lock){....}-{1:3}, at: __perf_event_task_sched_in+0xd1/0xbc0 > > > [ 39.919612] #1: ffff88806ca3c6f8 (&cpuctx_lock){....}-{2:2}, at: __perf_event_task_sched_in+0x1a7/0xbc0 > > > [ 39.920748] #2: ffff88800d91fc18 (&ctx->lock){....}-{2:2}, at: __perf_event_task_sched_in+0x1f9/0xbc0 > > > [ 39.921819] #3: ffffffff8725cd00 (rcu_read_lock){....}-{1:3}, at: perf_event_wakeup+0x6c/0x470 > > > [ 39.922823] stack backtrace: > > > [ 39.923171] CPU: 0 UID: 0 PID: 837 Comm: repro Not tainted 6.15.0-next-20250530-next-2025053 #1 PREEMPT(voluntary) > > > [ 39.923196] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS rel-1.16.3-0-ga6ed6b701f0a-prebuilt.qemu.o4 > > > [ 39.923214] Call Trace: > > > [ 39.923221] <TASK> > > > [ 39.923228] dump_stack_lvl+0xea/0x150 > > > [ 39.923256] dump_stack+0x19/0x20 > > > [ 39.923276] __lock_acquire+0xb22/0x22a0 > > > [ 39.923308] ? x86_pmu_commit_txn+0x195/0x2b0 > > > [ 39.923339] ? __lock_acquire+0x412/0x22a0 > > > [ 39.923375] lock_acquire+0x170/0x310 > > > [ 39.923407] ? __wake_up+0x26/0x60 > > > [ 39.923448] _raw_spin_lock_irqsave+0x52/0x80 > > > [ 39.923473] ? __wake_up+0x26/0x60 > > > [ 39.923504] __wake_up+0x26/0x60 > > > [ 39.923537] perf_event_wakeup+0x14a/0x470 > > > [ 39.923571] merge_sched_in+0x846/0x15c0 > > > [ 39.923610] visit_groups_merge.constprop.0.isra.0+0x952/0x1420 > > > [ 39.923653] ? __pfx_visit_groups_merge.constprop.0.isra.0+0x10/0x10 > > > [ 39.923688] ? sched_clock_noinstr+0x12/0x20 > > > [ 39.923724] ? __sanitizer_cov_trace_const_cmp1+0x1e/0x30 > > > [ 39.923766] ctx_sched_in+0x471/0xa20 > > > [ 39.923804] ? __pfx_ctx_sched_in+0x10/0x10 > > > [ 39.923838] ? __sanitizer_cov_trace_const_cmp4+0x1a/0x20 > > > [ 39.923878] perf_event_sched_in+0x47/0xa0 > > > [ 39.923912] __perf_event_task_sched_in+0x3fc/0xbc0 > > > [ 39.923951] ? __pfx___perf_event_task_sched_in+0x10/0x10 > > > [ 39.923984] ? __this_cpu_preempt_check+0x21/0x30 > > > [ 39.924012] ? __sanitizer_cov_trace_cmp8+0x1c/0x30 > > > [ 39.924046] ? xfd_validate_state+0x14f/0x1b0 > > > [ 39.924081] finish_task_switch.isra.0+0x525/0x990 > > > [ 39.924117] ? lock_unpin_lock+0xdc/0x170 > > > [ 39.924152] __schedule+0xef3/0x3840 > > > [ 39.924185] ? __pfx___schedule+0x10/0x10 > > > [ 39.924218] ? ktime_get_coarse_real_ts64+0xad/0xf0 > > > [ 39.924259] schedule+0xf6/0x3d0 > > > [ 39.924285] exit_to_user_mode_loop+0x7a/0x110 > > > [ 39.924315] do_syscall_64+0x284/0x2e0 > > > [ 39.924340] entry_SYSCALL_64_after_hwframe+0x76/0x7e > > > [ 39.924360] RIP: 0033:0x7ff14103ee5d > > > [ 39.924381] Code: ff c3 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 8 > > > [ 39.924400] RSP: 002b:00007fffb2745578 EFLAGS: 00000202 ORIG_RAX: 0000000000000038 > > > [ 39.924418] RAX: 0000000000000346 RBX: 0000000000000000 RCX: 00007ff14103ee5d > > > [ 39.924431] RDX: 0000000000000000 RSI: 0000000000000000 RDI: 0000000000082000 > > > [ 39.924443] RBP: 00007fffb27455c0 R08: 0000000000000000 R09: 0000000000000000 > > > [ 39.924456] R10: 0000000000000000 R11: 0000000000000202 R12: 00007fffb27459a8 > > > [ 39.924468] R13: 0000000000404e78 R14: 0000000000406e08 R15: 00007ff141389000 > > > [ 39.924497] </TASK> > > > [ 40.307815] coredump: 804(repro): over core_pipe_limit, skipping core dump > > > [ 40.472093] coredump: 795(repro): over core_pipe_limit, skipping core dump > > > [ 40.545575] coredump: 799(repro): over core_pipe_limit, skipping core dump > > > [ 40.948915] coredump: 833(repro): over core_pipe_limit, skipping core dump > > > [ 40.989336] coredump: 811(repro): over core_pipe_limit, skipping core dump > > > [ 42.121469] coredump: 857(repro): over core_pipe_limit, skipping core dump > > > " > > > > > > Hope this cound be insightful to you. > > > > > > Regards, > > > Yi Lai > > > > > > --- > > > > > > If you don't need the following environment to reproduce the problem or if you > > > already have one reproduced environment, please ignore the following information. > > > > > > How to reproduce: > > > git clone https://gitlab.com/xupengfe/repro_vm_env.git > > > cd repro_vm_env > > > tar -xvf repro_vm_env.tar.gz > > > cd repro_vm_env; ./start3.sh // it needs qemu-system-x86_64 and I used v7.1.0 > > > // start3.sh will load bzImage_2241ab53cbb5cdb08a6b2d4688feb13971058f65 v6.2-rc5 kernel > > > // You could change the bzImage_xxx as you want > > > // Maybe you need to remove line "-drive if=pflash,format=raw,readonly=on,file=./OVMF_CODE.fd \" for different qemu version > > > You could use below command to log in, there is no password for root. > > > ssh -p 10023 root@localhost > > > > > > After login vm(virtual machine) successfully, you could transfer reproduced > > > binary to the vm by below way, and reproduce the problem in vm: > > > gcc -pthread -o repro repro.c > > > scp -P 10023 repro root@localhost:/root/ > > > > > > Get the bzImage for target kernel: > > > Please use target kconfig and copy it to kernel_src/.config > > > make olddefconfig > > > make -jx bzImage //x should equal or less than cpu num your pc has > > > > > > Fill the bzImage file into above start3.sh to load the target kernel in vm. > > > > > > > > > Tips: > > > If you already have qemu-system-x86_64, please ignore below info. > > > If you want to install qemu v7.1.0 version: > > > git clone https://github.com/qemu/qemu.git > > > cd qemu > > > git checkout -f v7.1.0 > > > mkdir build > > > cd build > > > yum install -y ninja-build.x86_64 > > > yum -y install libslirp-devel.x86_64 > > > ../configure --target-list=x86_64-softmmu --enable-kvm --enable-vnc --enable-gtk --enable-sdl --enable-usb-redir --enable-slirp > > > make > > > make install > > > > Thanks for the detailed instruction. I was able to reproduce it with > > your setup. The below patch fixes it for me. Can you please double > > check if it works well for you? > > > > Thanks, > > Namhyung > > > > After applying following patch on top of latest linux-next, issue > cannot be reproduced. Thanks for confirming the fix. I'll add your Tested-by and send a formal patch. Thanks, Namhyung > > ---8--- > > > > diff --git a/kernel/events/core.c b/kernel/events/core.c > > index f34c99f8ce8f446a..e22eb88eb105b95b 100644 > > --- a/kernel/events/core.c > > +++ b/kernel/events/core.c > > @@ -3995,7 +3995,8 @@ static int merge_sched_in(struct perf_event *event, void *data) > > if (*perf_event_fasync(event)) > > event->pending_kill = POLL_ERR; > > > > - perf_event_wakeup(event); > > + event->pending_wakeup = 1; > > + irq_work_queue(&event->pending_irq); > > } else { > > struct perf_cpu_pmu_context *cpc = this_cpc(event->pmu_ctx->pmu); > > > > -- > > 2.49.0.1204.g71687c7c1d-goog > > ^ permalink raw reply [flat|nested] 6+ messages in thread
end of thread, other threads:[~2025-06-03 4:49 UTC | newest] Thread overview: 6+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2025-03-17 6:17 [RFC/PATCH] perf/core: Use POLLHUP for a pinned event in error Namhyung Kim 2025-03-17 7:47 ` [tip: perf/core] perf/core: Use POLLHUP for pinned events " tip-bot2 for Namhyung Kim 2025-06-02 14:10 ` [RFC/PATCH] perf/core: Use POLLHUP for a pinned event " Lai, Yi 2025-06-02 17:32 ` Namhyung Kim 2025-06-03 1:48 ` Lai, Yi 2025-06-03 4:49 ` Namhyung Kim
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).