* Deadlock in usb subsystem on shutdown, 6.18.3+
@ 2026-01-14 0:21 Ben Greear
2026-01-14 2:45 ` Hillf Danton
0 siblings, 1 reply; 9+ messages in thread
From: Ben Greear @ 2026-01-14 0:21 UTC (permalink / raw)
To: LKML; +Cc: linux-usb
Hello,
We caught a deadlock that appears to be in the USB code during shutdown.
We do a lot of reboots and normally all goes well, so I don't think we
can reliably reproduce the problem.
INFO: task systemd-shutdow:1 blocked for more than 180 seconds.
Tainted: G S O 6.18.3+ #33
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:systemd-shutdow state:D stack:0 pid:1 tgid:1 ppid:0 task_flags:0x400100 flags:0x00080001
Call Trace:
<TASK>
__schedule+0x46b/0x1140
schedule+0x23/0xc0
schedule_preempt_disabled+0x11/0x20
__mutex_lock.constprop.0+0x4f7/0x9a0
device_shutdown+0xa0/0x220
kernel_restart+0x36/0x90
__do_sys_reboot+0x127/0x220
? do_writev+0x76/0x110
? do_writev+0x76/0x110
do_syscall_64+0x50/0x6d0
entry_SYSCALL_64_after_hwframe+0x4b/0x53
RIP: 0033:0x7fad03531087
RSP: 002b:00007ffe137cf918 EFLAGS: 00000246 ORIG_RAX: 00000000000000a9
RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fad03531087
RDX: 0000000001234567 RSI: 0000000028121969 RDI: 00000000fee1dead
RBP: 00007ffe137cfac0 R08: 0000000000000069 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000
R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000000
</TASK>
INFO: task systemd-shutdow:1 is blocked on a mutex likely owned by task kworker/4:1:16648.
INFO: task kworker/4:2:1520 blocked for more than 360 seconds.
Tainted: G S O 6.18.3+ #33
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/4:2 state:D stack:0 pid:1520 tgid:1520 ppid:2 task_flags:0x4288060 flags:0x00080000
Workqueue: events __usb_queue_reset_device
Call Trace:
<TASK>
__schedule+0x46b/0x1140
? schedule_timeout+0x79/0xf0
schedule+0x23/0xc0
usb_kill_urb+0x7b/0xc0
? housekeeping_affine+0x30/0x30
usb_start_wait_urb+0xd6/0x160
usb_control_msg+0xe2/0x140
hub_port_init+0x647/0xf70
usb_reset_and_verify_device+0x191/0x4a0
? device_release_driver_internal+0x4a/0x200
usb_reset_device+0x138/0x280
__usb_queue_reset_device+0x35/0x50
process_one_work+0x17e/0x390
worker_thread+0x2c8/0x3e0
? process_one_work+0x390/0x390
kthread+0xf7/0x1f0
? kthreads_online_cpu+0x100/0x100
? kthreads_online_cpu+0x100/0x100
ret_from_fork+0x114/0x140
? kthreads_online_cpu+0x100/0x100
ret_from_fork_asm+0x11/0x20
</TASK>
INFO: task kworker/4:1:16648 blocked for more than 360 seconds.
Tainted: G S O 6.18.3+ #33
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/4:1 state:D stack:0 pid:16648 tgid:16648 ppid:2 task_flags:0x4288060 flags:0x00080000
Workqueue: events __usb_queue_reset_device
Call Trace:
<TASK>
__schedule+0x46b/0x1140
schedule+0x23/0xc0
usb_kill_urb+0x7b/0xc0
? housekeeping_affine+0x30/0x30
usb_hcd_flush_endpoint+0xb2/0x160
usb_disable_interface+0xbb/0xe0
usb_unbind_interface+0x11e/0x290
? kernfs_remove_by_name_ns+0xb0/0xd0
device_release_driver_internal+0x197/0x200
usb_forced_unbind_intf+0x4d/0xa0
usb_reset_device+0xe1/0x280
__usb_queue_reset_device+0x35/0x50
process_one_work+0x17e/0x390
worker_thread+0x2c8/0x3e0
? process_one_work+0x390/0x390
kthread+0xf7/0x1f0
? kthreads_online_cpu+0x100/0x100
? kthreads_online_cpu+0x100/0x100
ret_from_fork+0x114/0x140
? kthreads_online_cpu+0x100/0x100
ret_from_fork_asm+0x11/0x20
</TASK>
INFO: task kworker/6:1:16649 blocked for more than 360 seconds.
Tainted: G S O 6.18.3+ #33
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/6:1 state:D stack:0 pid:16649 tgid:16649 ppid:2 task_flags:0x4208060 flags:0x00080000
Workqueue: usb_hub_wq hub_event
Call Trace:
<TASK>
__schedule+0x46b/0x1140
schedule+0x23/0xc0
schedule_preempt_disabled+0x11/0x20
__mutex_lock.constprop.0+0x4f7/0x9a0
? update_load_avg+0x7f/0x790
? update_curr+0x14d/0x180
hub_event+0x172/0x1b30
? finish_task_switch.isra.0+0x8d/0x270
? __schedule+0x473/0x1140
process_one_work+0x17e/0x390
worker_thread+0x2c8/0x3e0
? process_one_work+0x390/0x390
kthread+0xf7/0x1f0
? kthreads_online_cpu+0x100/0x100
? kthreads_online_cpu+0x100/0x100
ret_from_fork+0x114/0x140
? kthreads_online_cpu+0x100/0x100
ret_from_fork_asm+0x11/0x20
</TASK>
INFO: task kworker/6:1:16649 is blocked on a mutex likely owned by task kworker/4:2:1520.
NMI backtrace for cpu 14
CPU: 14 UID: 0 PID: 141 Comm: khungtaskd Kdump: loaded Tainted: G S O 6.18.3+ #33 PREEMPT(full)
Tainted: [S]=CPU_OUT_OF_SPEC, [O]=OOT_MODULE
Hardware name: Default string /Default string, BIOS 5.27 08/08/2025
Call Trace:
<TASK>
dump_stack_lvl+0x60/0x80
nmi_cpu_backtrace+0xca/0x100
? lapic_can_unplug_cpu+0xa0/0xa0
nmi_trigger_cpumask_backtrace+0xc5/0xe0
watchdog+0x6ce/0x6f0
? proc_dohung_task_timeout_secs+0x30/0x30
kthread+0xf7/0x1f0
? kthreads_online_cpu+0x100/0x100
? kthreads_online_cpu+0x100/0x100
ret_from_fork+0x114/0x140
? kthreads_online_cpu+0x100/0x100
ret_from_fork_asm+0x11/0x20
</TASK>
Sending NMI from CPU 14 to CPUs 0-13,15-19:
NMI backtrace for cpu 0 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 1 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 2 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 3 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 4 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 5 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 6 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 7 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 8 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 9 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 10 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 11 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 12 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 13 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 15 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 16 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 17 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 18 skipped: idling at intel_idle+0x50/0x80
NMI backtrace for cpu 19 skipped: idling at intel_idle+0x50/0x80
Thanks,
Ben
--
Ben Greear <greearb@candelatech.com>
Candela Technologies Inc http://www.candelatech.com
^ permalink raw reply [flat|nested] 9+ messages in thread* Re: Deadlock in usb subsystem on shutdown, 6.18.3+ 2026-01-14 0:21 Deadlock in usb subsystem on shutdown, 6.18.3+ Ben Greear @ 2026-01-14 2:45 ` Hillf Danton 2026-01-14 14:36 ` Ben Greear 0 siblings, 1 reply; 9+ messages in thread From: Hillf Danton @ 2026-01-14 2:45 UTC (permalink / raw) To: Ben Greear; +Cc: LKML, linux-usb On Tue, 13 Jan 2026 16:21:07 -0800 Ben Greear wrote: > Hello, > > We caught a deadlock that appears to be in the USB code during shutdown. > We do a lot of reboots and normally all goes well, so I don't think we > can reliably reproduce the problem. > > INFO: task systemd-shutdow:1 blocked for more than 180 seconds. > Tainted: G S O 6.18.3+ #33 > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. > task:systemd-shutdow state:D stack:0 pid:1 tgid:1 ppid:0 task_flags:0x400100 flags:0x00080001 > Call Trace: > <TASK> > __schedule+0x46b/0x1140 > schedule+0x23/0xc0 > schedule_preempt_disabled+0x11/0x20 > __mutex_lock.constprop.0+0x4f7/0x9a0 > device_shutdown+0xa0/0x220 > kernel_restart+0x36/0x90 > __do_sys_reboot+0x127/0x220 > ? do_writev+0x76/0x110 > ? do_writev+0x76/0x110 > do_syscall_64+0x50/0x6d0 > entry_SYSCALL_64_after_hwframe+0x4b/0x53 > RIP: 0033:0x7fad03531087 > RSP: 002b:00007ffe137cf918 EFLAGS: 00000246 ORIG_RAX: 00000000000000a9 > RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fad03531087 > RDX: 0000000001234567 RSI: 0000000028121969 RDI: 00000000fee1dead > RBP: 00007ffe137cfac0 R08: 0000000000000069 R09: 0000000000000000 > R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000 > R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000000 > </TASK> > INFO: task systemd-shutdow:1 is blocked on a mutex likely owned by task kworker/4:1:16648. This explains why the shutdown stalled. > INFO: task kworker/4:2:1520 blocked for more than 360 seconds. > Tainted: G S O 6.18.3+ #33 > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. > task:kworker/4:2 state:D stack:0 pid:1520 tgid:1520 ppid:2 task_flags:0x4288060 flags:0x00080000 > Workqueue: events __usb_queue_reset_device > Call Trace: > <TASK> > __schedule+0x46b/0x1140 > ? schedule_timeout+0x79/0xf0 > schedule+0x23/0xc0 > usb_kill_urb+0x7b/0xc0 > ? housekeeping_affine+0x30/0x30 > usb_start_wait_urb+0xd6/0x160 > usb_control_msg+0xe2/0x140 > hub_port_init+0x647/0xf70 > usb_reset_and_verify_device+0x191/0x4a0 > ? device_release_driver_internal+0x4a/0x200 > usb_reset_device+0x138/0x280 > __usb_queue_reset_device+0x35/0x50 > process_one_work+0x17e/0x390 > worker_thread+0x2c8/0x3e0 > ? process_one_work+0x390/0x390 > kthread+0xf7/0x1f0 > ? kthreads_online_cpu+0x100/0x100 > ? kthreads_online_cpu+0x100/0x100 > ret_from_fork+0x114/0x140 > ? kthreads_online_cpu+0x100/0x100 > ret_from_fork_asm+0x11/0x20 > </TASK> > INFO: task kworker/4:1:16648 blocked for more than 360 seconds. > Tainted: G S O 6.18.3+ #33 > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. > task:kworker/4:1 state:D stack:0 pid:16648 tgid:16648 ppid:2 task_flags:0x4288060 flags:0x00080000 > Workqueue: events __usb_queue_reset_device > Call Trace: > <TASK> > __schedule+0x46b/0x1140 > schedule+0x23/0xc0 > usb_kill_urb+0x7b/0xc0 Kworker failed to kill urb within 300 seconds, so we know the underlying usb hardware failed to response within 300s. That said, the deadlock in the subject line is incorrect, but task hung due to hardware glitch. > ? housekeeping_affine+0x30/0x30 > usb_hcd_flush_endpoint+0xb2/0x160 > usb_disable_interface+0xbb/0xe0 > usb_unbind_interface+0x11e/0x290 > ? kernfs_remove_by_name_ns+0xb0/0xd0 > device_release_driver_internal+0x197/0x200 > usb_forced_unbind_intf+0x4d/0xa0 > usb_reset_device+0xe1/0x280 > __usb_queue_reset_device+0x35/0x50 > process_one_work+0x17e/0x390 > worker_thread+0x2c8/0x3e0 > ? process_one_work+0x390/0x390 > kthread+0xf7/0x1f0 > ? kthreads_online_cpu+0x100/0x100 > ? kthreads_online_cpu+0x100/0x100 > ret_from_fork+0x114/0x140 > ? kthreads_online_cpu+0x100/0x100 > ret_from_fork_asm+0x11/0x20 > </TASK> > INFO: task kworker/6:1:16649 blocked for more than 360 seconds. > Tainted: G S O 6.18.3+ #33 > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. > task:kworker/6:1 state:D stack:0 pid:16649 tgid:16649 ppid:2 task_flags:0x4208060 flags:0x00080000 > Workqueue: usb_hub_wq hub_event > Call Trace: > <TASK> > __schedule+0x46b/0x1140 > schedule+0x23/0xc0 > schedule_preempt_disabled+0x11/0x20 > __mutex_lock.constprop.0+0x4f7/0x9a0 > ? update_load_avg+0x7f/0x790 > ? update_curr+0x14d/0x180 > hub_event+0x172/0x1b30 > ? finish_task_switch.isra.0+0x8d/0x270 > ? __schedule+0x473/0x1140 > process_one_work+0x17e/0x390 > worker_thread+0x2c8/0x3e0 > ? process_one_work+0x390/0x390 > kthread+0xf7/0x1f0 > ? kthreads_online_cpu+0x100/0x100 > ? kthreads_online_cpu+0x100/0x100 > ret_from_fork+0x114/0x140 > ? kthreads_online_cpu+0x100/0x100 > ret_from_fork_asm+0x11/0x20 > </TASK> > INFO: task kworker/6:1:16649 is blocked on a mutex likely owned by task kworker/4:2:1520. > NMI backtrace for cpu 14 > CPU: 14 UID: 0 PID: 141 Comm: khungtaskd Kdump: loaded Tainted: G S O 6.18.3+ #33 PREEMPT(full) > Tainted: [S]=CPU_OUT_OF_SPEC, [O]=OOT_MODULE > Hardware name: Default string /Default string, BIOS 5.27 08/08/2025 > Call Trace: > <TASK> > dump_stack_lvl+0x60/0x80 > nmi_cpu_backtrace+0xca/0x100 > ? lapic_can_unplug_cpu+0xa0/0xa0 > nmi_trigger_cpumask_backtrace+0xc5/0xe0 > watchdog+0x6ce/0x6f0 > ? proc_dohung_task_timeout_secs+0x30/0x30 > kthread+0xf7/0x1f0 > ? kthreads_online_cpu+0x100/0x100 > ? kthreads_online_cpu+0x100/0x100 > ret_from_fork+0x114/0x140 > ? kthreads_online_cpu+0x100/0x100 > ret_from_fork_asm+0x11/0x20 > </TASK> > Sending NMI from CPU 14 to CPUs 0-13,15-19: > NMI backtrace for cpu 0 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 1 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 2 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 3 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 4 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 5 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 6 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 7 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 8 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 9 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 10 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 11 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 12 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 13 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 15 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 16 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 17 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 18 skipped: idling at intel_idle+0x50/0x80 > NMI backtrace for cpu 19 skipped: idling at intel_idle+0x50/0x80 > > Thanks, > Ben > > -- > Ben Greear <greearb@candelatech.com> > Candela Technologies Inc http://www.candelatech.com ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: Deadlock in usb subsystem on shutdown, 6.18.3+ 2026-01-14 2:45 ` Hillf Danton @ 2026-01-14 14:36 ` Ben Greear 2026-01-14 15:13 ` Alan Stern 0 siblings, 1 reply; 9+ messages in thread From: Ben Greear @ 2026-01-14 14:36 UTC (permalink / raw) To: Hillf Danton; +Cc: LKML, linux-usb On 1/13/26 18:45, Hillf Danton wrote: > On Tue, 13 Jan 2026 16:21:07 -0800 Ben Greear wrote: >> Hello, >> >> We caught a deadlock that appears to be in the USB code during shutdown. >> We do a lot of reboots and normally all goes well, so I don't think we >> can reliably reproduce the problem. >> >> INFO: task systemd-shutdow:1 blocked for more than 180 seconds. >> Tainted: G S O 6.18.3+ #33 >> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. >> task:systemd-shutdow state:D stack:0 pid:1 tgid:1 ppid:0 task_flags:0x400100 flags:0x00080001 >> Call Trace: >> <TASK> >> __schedule+0x46b/0x1140 >> schedule+0x23/0xc0 >> schedule_preempt_disabled+0x11/0x20 >> __mutex_lock.constprop.0+0x4f7/0x9a0 >> device_shutdown+0xa0/0x220 >> kernel_restart+0x36/0x90 >> __do_sys_reboot+0x127/0x220 >> ? do_writev+0x76/0x110 >> ? do_writev+0x76/0x110 >> do_syscall_64+0x50/0x6d0 >> entry_SYSCALL_64_after_hwframe+0x4b/0x53 >> RIP: 0033:0x7fad03531087 >> RSP: 002b:00007ffe137cf918 EFLAGS: 00000246 ORIG_RAX: 00000000000000a9 >> RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fad03531087 >> RDX: 0000000001234567 RSI: 0000000028121969 RDI: 00000000fee1dead >> RBP: 00007ffe137cfac0 R08: 0000000000000069 R09: 0000000000000000 >> R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000 >> R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000000 >> </TASK> >> INFO: task systemd-shutdow:1 is blocked on a mutex likely owned by task kworker/4:1:16648. > > This explains why the shutdown stalled. > >> INFO: task kworker/4:2:1520 blocked for more than 360 seconds. >> Tainted: G S O 6.18.3+ #33 >> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. >> task:kworker/4:2 state:D stack:0 pid:1520 tgid:1520 ppid:2 task_flags:0x4288060 flags:0x00080000 >> Workqueue: events __usb_queue_reset_device >> Call Trace: >> <TASK> >> __schedule+0x46b/0x1140 >> ? schedule_timeout+0x79/0xf0 >> schedule+0x23/0xc0 >> usb_kill_urb+0x7b/0xc0 >> ? housekeeping_affine+0x30/0x30 >> usb_start_wait_urb+0xd6/0x160 >> usb_control_msg+0xe2/0x140 >> hub_port_init+0x647/0xf70 >> usb_reset_and_verify_device+0x191/0x4a0 >> ? device_release_driver_internal+0x4a/0x200 >> usb_reset_device+0x138/0x280 >> __usb_queue_reset_device+0x35/0x50 >> process_one_work+0x17e/0x390 >> worker_thread+0x2c8/0x3e0 >> ? process_one_work+0x390/0x390 >> kthread+0xf7/0x1f0 >> ? kthreads_online_cpu+0x100/0x100 >> ? kthreads_online_cpu+0x100/0x100 >> ret_from_fork+0x114/0x140 >> ? kthreads_online_cpu+0x100/0x100 >> ret_from_fork_asm+0x11/0x20 >> </TASK> >> INFO: task kworker/4:1:16648 blocked for more than 360 seconds. >> Tainted: G S O 6.18.3+ #33 >> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. >> task:kworker/4:1 state:D stack:0 pid:16648 tgid:16648 ppid:2 task_flags:0x4288060 flags:0x00080000 >> Workqueue: events __usb_queue_reset_device >> Call Trace: >> <TASK> >> __schedule+0x46b/0x1140 >> schedule+0x23/0xc0 >> usb_kill_urb+0x7b/0xc0 > > Kworker failed to kill urb within 300 seconds, so we know the underlying usb > hardware failed to response within 300s. > > That said, the deadlock in the subject line is incorrect, but task hung due > to hardware glitch. In the case where hardware is not responding, shouldn't we just consider it dead and move on instead of deadlocking the whole OS? In this case, the system was un-plugged from a KVM (usb mouse & keyboard) right around time of shutdown, so I guess that would explain why the USB device didn't respond. Thanks, Ben -- Ben Greear <greearb@candelatech.com> Candela Technologies Inc http://www.candelatech.com ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: Deadlock in usb subsystem on shutdown, 6.18.3+ 2026-01-14 14:36 ` Ben Greear @ 2026-01-14 15:13 ` Alan Stern 2026-01-14 17:51 ` Ben Greear 0 siblings, 1 reply; 9+ messages in thread From: Alan Stern @ 2026-01-14 15:13 UTC (permalink / raw) To: Ben Greear; +Cc: Hillf Danton, LKML, linux-usb On Wed, Jan 14, 2026 at 06:36:41AM -0800, Ben Greear wrote: > On 1/13/26 18:45, Hillf Danton wrote: > > On Tue, 13 Jan 2026 16:21:07 -0800 Ben Greear wrote: > > > Hello, > > > > > > We caught a deadlock that appears to be in the USB code during shutdown. > > > We do a lot of reboots and normally all goes well, so I don't think we > > > can reliably reproduce the problem. > > > > > > INFO: task systemd-shutdow:1 blocked for more than 180 seconds. > > > Tainted: G S O 6.18.3+ #33 > > > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. > > > task:systemd-shutdow state:D stack:0 pid:1 tgid:1 ppid:0 task_flags:0x400100 flags:0x00080001 > > > Call Trace: > > > <TASK> > > > __schedule+0x46b/0x1140 > > > schedule+0x23/0xc0 > > > schedule_preempt_disabled+0x11/0x20 > > > __mutex_lock.constprop.0+0x4f7/0x9a0 > > > device_shutdown+0xa0/0x220 > > > kernel_restart+0x36/0x90 > > > __do_sys_reboot+0x127/0x220 > > > ? do_writev+0x76/0x110 > > > ? do_writev+0x76/0x110 > > > do_syscall_64+0x50/0x6d0 > > > entry_SYSCALL_64_after_hwframe+0x4b/0x53 > > > RIP: 0033:0x7fad03531087 > > > RSP: 002b:00007ffe137cf918 EFLAGS: 00000246 ORIG_RAX: 00000000000000a9 > > > RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fad03531087 > > > RDX: 0000000001234567 RSI: 0000000028121969 RDI: 00000000fee1dead > > > RBP: 00007ffe137cfac0 R08: 0000000000000069 R09: 0000000000000000 > > > R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000 > > > R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000000 > > > </TASK> > > > INFO: task systemd-shutdow:1 is blocked on a mutex likely owned by task kworker/4:1:16648. > > > > This explains why the shutdown stalled. > > > > > INFO: task kworker/4:2:1520 blocked for more than 360 seconds. > > > Tainted: G S O 6.18.3+ #33 > > > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. > > > task:kworker/4:2 state:D stack:0 pid:1520 tgid:1520 ppid:2 task_flags:0x4288060 flags:0x00080000 > > > Workqueue: events __usb_queue_reset_device > > > Call Trace: > > > <TASK> > > > __schedule+0x46b/0x1140 > > > ? schedule_timeout+0x79/0xf0 > > > schedule+0x23/0xc0 > > > usb_kill_urb+0x7b/0xc0 > > > ? housekeeping_affine+0x30/0x30 > > > usb_start_wait_urb+0xd6/0x160 > > > usb_control_msg+0xe2/0x140 > > > hub_port_init+0x647/0xf70 > > > usb_reset_and_verify_device+0x191/0x4a0 > > > ? device_release_driver_internal+0x4a/0x200 > > > usb_reset_device+0x138/0x280 > > > __usb_queue_reset_device+0x35/0x50 > > > process_one_work+0x17e/0x390 > > > worker_thread+0x2c8/0x3e0 > > > ? process_one_work+0x390/0x390 > > > kthread+0xf7/0x1f0 > > > ? kthreads_online_cpu+0x100/0x100 > > > ? kthreads_online_cpu+0x100/0x100 > > > ret_from_fork+0x114/0x140 > > > ? kthreads_online_cpu+0x100/0x100 > > > ret_from_fork_asm+0x11/0x20 > > > </TASK> > > > INFO: task kworker/4:1:16648 blocked for more than 360 seconds. > > > Tainted: G S O 6.18.3+ #33 > > > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. > > > task:kworker/4:1 state:D stack:0 pid:16648 tgid:16648 ppid:2 task_flags:0x4288060 flags:0x00080000 > > > Workqueue: events __usb_queue_reset_device > > > Call Trace: > > > <TASK> > > > __schedule+0x46b/0x1140 > > > schedule+0x23/0xc0 > > > usb_kill_urb+0x7b/0xc0 > > > > Kworker failed to kill urb within 300 seconds, so we know the underlying usb > > hardware failed to response within 300s. > > > > That said, the deadlock in the subject line is incorrect, but task hung due > > to hardware glitch. In fact, we do not know whether this was a hardware glitch or a software bug. > In the case where hardware is not responding, shouldn't we just consider it > dead and move on instead of deadlocking the whole OS? > > In this case, the system was un-plugged from a KVM (usb mouse & keyboard) > right around time of shutdown, so I guess that would explain why the USB device > didn't respond. You misunderstand. What's failing is the USB host controller on the computer, not the attached (or unplugged) USB device. If the host controller really had a hardware glitch then the host controller driver should have realized it and moved on. It seems to me at least as likely that the problem is caused by a bug in the host controller driver rather than anything wrong with the hardware. (Of course, it could be a combination of things going wrong: a glitch in the hardware that the driver wasn't expecting and is unable to cope with. But even in that case, the proper solution would be to fix the driver since we can't fix the hardware.) Unfortunately, we have no to tell from the log you collected which host controller driver encountered this problem. Nor, unless you can replicate the problem, any way to track down exactly where in that driver the bug is -- or even any way to tell whether a proposed fix actually solves the problem. Alan Stern ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: Deadlock in usb subsystem on shutdown, 6.18.3+ 2026-01-14 15:13 ` Alan Stern @ 2026-01-14 17:51 ` Ben Greear 2026-01-14 19:26 ` Alan Stern 0 siblings, 1 reply; 9+ messages in thread From: Ben Greear @ 2026-01-14 17:51 UTC (permalink / raw) To: Alan Stern; +Cc: Hillf Danton, LKML, linux-usb On 1/14/26 07:13, Alan Stern wrote: > On Wed, Jan 14, 2026 at 06:36:41AM -0800, Ben Greear wrote: >> On 1/13/26 18:45, Hillf Danton wrote: >>> On Tue, 13 Jan 2026 16:21:07 -0800 Ben Greear wrote: >>>> Hello, >>>> >>>> We caught a deadlock that appears to be in the USB code during shutdown. >>>> We do a lot of reboots and normally all goes well, so I don't think we >>>> can reliably reproduce the problem. >>>> >>>> INFO: task systemd-shutdow:1 blocked for more than 180 seconds. >>>> Tainted: G S O 6.18.3+ #33 >>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. >>>> task:systemd-shutdow state:D stack:0 pid:1 tgid:1 ppid:0 task_flags:0x400100 flags:0x00080001 >>>> Call Trace: >>>> <TASK> >>>> __schedule+0x46b/0x1140 >>>> schedule+0x23/0xc0 >>>> schedule_preempt_disabled+0x11/0x20 >>>> __mutex_lock.constprop.0+0x4f7/0x9a0 >>>> device_shutdown+0xa0/0x220 >>>> kernel_restart+0x36/0x90 >>>> __do_sys_reboot+0x127/0x220 >>>> ? do_writev+0x76/0x110 >>>> ? do_writev+0x76/0x110 >>>> do_syscall_64+0x50/0x6d0 >>>> entry_SYSCALL_64_after_hwframe+0x4b/0x53 >>>> RIP: 0033:0x7fad03531087 >>>> RSP: 002b:00007ffe137cf918 EFLAGS: 00000246 ORIG_RAX: 00000000000000a9 >>>> RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fad03531087 >>>> RDX: 0000000001234567 RSI: 0000000028121969 RDI: 00000000fee1dead >>>> RBP: 00007ffe137cfac0 R08: 0000000000000069 R09: 0000000000000000 >>>> R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000 >>>> R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000000 >>>> </TASK> >>>> INFO: task systemd-shutdow:1 is blocked on a mutex likely owned by task kworker/4:1:16648. >>> >>> This explains why the shutdown stalled. >>> >>>> INFO: task kworker/4:2:1520 blocked for more than 360 seconds. >>>> Tainted: G S O 6.18.3+ #33 >>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. >>>> task:kworker/4:2 state:D stack:0 pid:1520 tgid:1520 ppid:2 task_flags:0x4288060 flags:0x00080000 >>>> Workqueue: events __usb_queue_reset_device >>>> Call Trace: >>>> <TASK> >>>> __schedule+0x46b/0x1140 >>>> ? schedule_timeout+0x79/0xf0 >>>> schedule+0x23/0xc0 >>>> usb_kill_urb+0x7b/0xc0 >>>> ? housekeeping_affine+0x30/0x30 >>>> usb_start_wait_urb+0xd6/0x160 >>>> usb_control_msg+0xe2/0x140 >>>> hub_port_init+0x647/0xf70 >>>> usb_reset_and_verify_device+0x191/0x4a0 >>>> ? device_release_driver_internal+0x4a/0x200 >>>> usb_reset_device+0x138/0x280 >>>> __usb_queue_reset_device+0x35/0x50 >>>> process_one_work+0x17e/0x390 >>>> worker_thread+0x2c8/0x3e0 >>>> ? process_one_work+0x390/0x390 >>>> kthread+0xf7/0x1f0 >>>> ? kthreads_online_cpu+0x100/0x100 >>>> ? kthreads_online_cpu+0x100/0x100 >>>> ret_from_fork+0x114/0x140 >>>> ? kthreads_online_cpu+0x100/0x100 >>>> ret_from_fork_asm+0x11/0x20 >>>> </TASK> >>>> INFO: task kworker/4:1:16648 blocked for more than 360 seconds. >>>> Tainted: G S O 6.18.3+ #33 >>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. >>>> task:kworker/4:1 state:D stack:0 pid:16648 tgid:16648 ppid:2 task_flags:0x4288060 flags:0x00080000 >>>> Workqueue: events __usb_queue_reset_device >>>> Call Trace: >>>> <TASK> >>>> __schedule+0x46b/0x1140 >>>> schedule+0x23/0xc0 >>>> usb_kill_urb+0x7b/0xc0 >>> >>> Kworker failed to kill urb within 300 seconds, so we know the underlying usb >>> hardware failed to response within 300s. >>> >>> That said, the deadlock in the subject line is incorrect, but task hung due >>> to hardware glitch. > > In fact, we do not know whether this was a hardware glitch or a software > bug. > >> In the case where hardware is not responding, shouldn't we just consider it >> dead and move on instead of deadlocking the whole OS? >> >> In this case, the system was un-plugged from a KVM (usb mouse & keyboard) >> right around time of shutdown, so I guess that would explain why the USB device >> didn't respond. > > You misunderstand. What's failing is the USB host controller on the > computer, not the attached (or unplugged) USB device. If the host > controller really had a hardware glitch then the host controller driver > should have realized it and moved on. It seems to me at least as likely > that the problem is caused by a bug in the host controller driver rather > than anything wrong with the hardware. > > (Of course, it could be a combination of things going wrong: a glitch in > the hardware that the driver wasn't expecting and is unable to cope > with. But even in that case, the proper solution would be to fix the > driver since we can't fix the hardware.) > > Unfortunately, we have no to tell from the log you collected which host > controller driver encountered this problem. Nor, unless you can > replicate the problem, any way to track down exactly where in that > driver the bug is -- or even any way to tell whether a proposed fix > actually solves the problem. > > Alan Stern The system was in the final stage of shutdown, so all we have in this case is serial console output. If we set console to more verbose mode, would that provide what you need? Or maybe 'dmesg' from when system boots? We can try to reproduce, but need some clues about what to provide to make progress. Thanks, Ben -- Ben Greear <greearb@candelatech.com> Candela Technologies Inc http://www.candelatech.com ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: Deadlock in usb subsystem on shutdown, 6.18.3+ 2026-01-14 17:51 ` Ben Greear @ 2026-01-14 19:26 ` Alan Stern 2026-01-14 20:14 ` Oliver Neukum 0 siblings, 1 reply; 9+ messages in thread From: Alan Stern @ 2026-01-14 19:26 UTC (permalink / raw) To: Ben Greear; +Cc: Hillf Danton, LKML, linux-usb On Wed, Jan 14, 2026 at 09:51:48AM -0800, Ben Greear wrote: > On 1/14/26 07:13, Alan Stern wrote: ... > > > > > task:kworker/4:2 state:D stack:0 pid:1520 tgid:1520 ppid:2 task_flags:0x4288060 flags:0x00080000 > > > > > Workqueue: events __usb_queue_reset_device > > > > > Call Trace: > > > > > <TASK> > > > > > __schedule+0x46b/0x1140 > > > > > ? schedule_timeout+0x79/0xf0 > > > > > schedule+0x23/0xc0 > > > > > usb_kill_urb+0x7b/0xc0 > > > > > ? housekeeping_affine+0x30/0x30 > > > > > usb_start_wait_urb+0xd6/0x160 > > > > > usb_control_msg+0xe2/0x140 > > > > > hub_port_init+0x647/0xf70 > > > > > usb_reset_and_verify_device+0x191/0x4a0 > > > > > ? device_release_driver_internal+0x4a/0x200 > > > > > usb_reset_device+0x138/0x280 > > > > > __usb_queue_reset_device+0x35/0x50 > > > > > process_one_work+0x17e/0x390 > > > > > worker_thread+0x2c8/0x3e0 > > > > > ? process_one_work+0x390/0x390 > > > > > kthread+0xf7/0x1f0 > > Unfortunately, we have no to tell from the log you collected which host > > controller driver encountered this problem. Nor, unless you can > > replicate the problem, any way to track down exactly where in that > > driver the bug is -- or even any way to tell whether a proposed fix > > actually solves the problem. > > > > Alan Stern > > The system was in the final stage of shutdown, so all we have in this case is serial > console output. If we set console to more verbose mode, would that provide what > you need? Maybe; it's hard to tell. You'd have to enable dynamic debugging for the usbcore module and set the console to show debug-level messages. > Or maybe 'dmesg' from when system boots? I suspect that nothing from before the start of the shutdown would be useful. > We can try to reproduce, but need some clues about what to provide to make progress. The stack dump above shows that a kernel thread get stuck inside of usb_reset_device(). The first thing we would need to know is which device the thread is trying to reset. Adding a dev_info() line near the start of usb_reset_device() would get us that information and you could capture it in the log. (Along with that, we'd need to see the output from "lsusb -t".) Knowing the device, we would know what host controller the device is attached to. The trick will then be to figure what's going wrong with that host controller's driver. It won't be easy, especially if the problem only occurs while the system is shutting down. Alan Stern ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: Deadlock in usb subsystem on shutdown, 6.18.3+ 2026-01-14 19:26 ` Alan Stern @ 2026-01-14 20:14 ` Oliver Neukum 2026-01-20 17:29 ` Ben Greear 0 siblings, 1 reply; 9+ messages in thread From: Oliver Neukum @ 2026-01-14 20:14 UTC (permalink / raw) To: Alan Stern, Ben Greear; +Cc: Hillf Danton, LKML, linux-usb On 14.01.26 20:26, Alan Stern wrote: > On Wed, Jan 14, 2026 at 09:51:48AM -0800, Ben Greear wrote: >> On 1/14/26 07:13, Alan Stern wrote: > ... >>>>>> task:kworker/4:2 state:D stack:0 pid:1520 tgid:1520 ppid:2 task_flags:0x4288060 flags:0x00080000 >>>>>> Workqueue: events __usb_queue_reset_device >>>>>> Call Trace: >>>>>> <TASK> >>>>>> __schedule+0x46b/0x1140 >>>>>> ? schedule_timeout+0x79/0xf0 >>>>>> schedule+0x23/0xc0 >>>>>> usb_kill_urb+0x7b/0xc0 >>>>>> ? housekeeping_affine+0x30/0x30 >>>>>> usb_start_wait_urb+0xd6/0x160 >>>>>> usb_control_msg+0xe2/0x140 >>>>>> hub_port_init+0x647/0xf70 >>>>>> usb_reset_and_verify_device+0x191/0x4a0 >>>>>> ? device_release_driver_internal+0x4a/0x200 >>>>>> usb_reset_device+0x138/0x280 >>>>>> __usb_queue_reset_device+0x35/0x50 >>>>>> process_one_work+0x17e/0x390 >>>>>> worker_thread+0x2c8/0x3e0 >>>>>> ? process_one_work+0x390/0x390 >>>>>> kthread+0xf7/0x1f0 > >>> Unfortunately, we have no to tell from the log you collected which host >>> controller driver encountered this problem. Nor, unless you can >>> replicate the problem, any way to track down exactly where in that >>> driver the bug is -- or even any way to tell whether a proposed fix >>> actually solves the problem. >>> >>> Alan Stern >> >> The system was in the final stage of shutdown, so all we have in this case is serial >> console output. If we set console to more verbose mode, would that provide what >> you need? > > Maybe; it's hard to tell. You'd have to enable dynamic debugging for > the usbcore module and set the console to show debug-level messages. > >> Or maybe 'dmesg' from when system boots? > > I suspect that nothing from before the start of the shutdown would be > useful. > >> We can try to reproduce, but need some clues about what to provide to make progress. > > The stack dump above shows that a kernel thread get stuck inside of > usb_reset_device(). The first thing we would need to know is which > device the thread is trying to reset. Adding a dev_info() line near the > start of usb_reset_device() would get us that information and you could > capture it in the log. (Along with that, we'd need to see the output > from "lsusb -t".) > > Knowing the device, we would know what host controller the device is > attached to. The trick will then be to figure what's going wrong with > that host controller's driver. It won't be easy, especially if the > problem only occurs while the system is shutting down. Something called usb_queue_reset_device() Workqueue: events __usb_queue_reset_device Call Trace: <TASK> __schedule+0x46b/0x1140 ? schedule_timeout+0x79/0xf0 schedule+0x23/0xc0 usb_kill_urb+0x7b/0xc0 ? housekeeping_affine+0x30/0x30 usb_start_wait_urb+0xd6/0x160 usb_control_msg+0xe2/0x140 hub_port_init+0x647/0xf70 usb_reset_and_verify_device+0x191/0x4a0 ? device_release_driver_internal+0x4a/0x200 usb_reset_device+0x138/0x280 __usb_queue_reset_device+0x35/0x50 process_one_work+0x17e/0x390 That is a fairly rare method. We also know that "the system was un-plugged from a KVM (usb mouse & keyboard)" One of the few places usb_queue_reset_device() is used is hid_io_error() Do you still know which port that KVM had been plugged into? I usually would not run a chain of logic with so many assumptions, but if you cannot reproduce I see no alternative. Regards Oliver ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: Deadlock in usb subsystem on shutdown, 6.18.3+ 2026-01-14 20:14 ` Oliver Neukum @ 2026-01-20 17:29 ` Ben Greear 2026-01-29 9:28 ` Michal Pecio 0 siblings, 1 reply; 9+ messages in thread From: Ben Greear @ 2026-01-20 17:29 UTC (permalink / raw) To: Oliver Neukum, Alan Stern; +Cc: Hillf Danton, LKML, linux-usb On 1/14/26 12:14, Oliver Neukum wrote: > On 14.01.26 20:26, Alan Stern wrote: >> On Wed, Jan 14, 2026 at 09:51:48AM -0800, Ben Greear wrote: >>> On 1/14/26 07:13, Alan Stern wrote: >> ... >>>>>>> task:kworker/4:2 state:D stack:0 pid:1520 tgid:1520 ppid:2 task_flags:0x4288060 flags:0x00080000 >>>>>>> Workqueue: events __usb_queue_reset_device >>>>>>> Call Trace: >>>>>>> <TASK> >>>>>>> __schedule+0x46b/0x1140 >>>>>>> ? schedule_timeout+0x79/0xf0 >>>>>>> schedule+0x23/0xc0 >>>>>>> usb_kill_urb+0x7b/0xc0 >>>>>>> ? housekeeping_affine+0x30/0x30 >>>>>>> usb_start_wait_urb+0xd6/0x160 >>>>>>> usb_control_msg+0xe2/0x140 >>>>>>> hub_port_init+0x647/0xf70 >>>>>>> usb_reset_and_verify_device+0x191/0x4a0 >>>>>>> ? device_release_driver_internal+0x4a/0x200 >>>>>>> usb_reset_device+0x138/0x280 >>>>>>> __usb_queue_reset_device+0x35/0x50 >>>>>>> process_one_work+0x17e/0x390 >>>>>>> worker_thread+0x2c8/0x3e0 >>>>>>> ? process_one_work+0x390/0x390 >>>>>>> kthread+0xf7/0x1f0 >> >>>> Unfortunately, we have no to tell from the log you collected which host >>>> controller driver encountered this problem. Nor, unless you can >>>> replicate the problem, any way to track down exactly where in that >>>> driver the bug is -- or even any way to tell whether a proposed fix >>>> actually solves the problem. >>>> >>>> Alan Stern >>> >>> The system was in the final stage of shutdown, so all we have in this case is serial >>> console output. If we set console to more verbose mode, would that provide what >>> you need? >> >> Maybe; it's hard to tell. You'd have to enable dynamic debugging for >> the usbcore module and set the console to show debug-level messages. >> >>> Or maybe 'dmesg' from when system boots? >> >> I suspect that nothing from before the start of the shutdown would be >> useful. >> >>> We can try to reproduce, but need some clues about what to provide to make progress. >> >> The stack dump above shows that a kernel thread get stuck inside of >> usb_reset_device(). The first thing we would need to know is which >> device the thread is trying to reset. Adding a dev_info() line near the >> start of usb_reset_device() would get us that information and you could >> capture it in the log. (Along with that, we'd need to see the output >> from "lsusb -t".) >> >> Knowing the device, we would know what host controller the device is >> attached to. The trick will then be to figure what's going wrong with >> that host controller's driver. It won't be easy, especially if the >> problem only occurs while the system is shutting down. > > Something called usb_queue_reset_device() > > Workqueue: events __usb_queue_reset_device > Call Trace: > <TASK> > __schedule+0x46b/0x1140 > ? schedule_timeout+0x79/0xf0 > schedule+0x23/0xc0 > usb_kill_urb+0x7b/0xc0 > ? housekeeping_affine+0x30/0x30 > usb_start_wait_urb+0xd6/0x160 > usb_control_msg+0xe2/0x140 > hub_port_init+0x647/0xf70 > usb_reset_and_verify_device+0x191/0x4a0 > ? device_release_driver_internal+0x4a/0x200 > usb_reset_device+0x138/0x280 > __usb_queue_reset_device+0x35/0x50 > process_one_work+0x17e/0x390 > > That is a fairly rare method. > We also know that > "the system was un-plugged from a KVM (usb mouse & keyboard)" > One of the few places usb_queue_reset_device() is used > is hid_io_error() > > Do you still know which port that KVM had been plugged > into? I usually would not run a chain of logic with so > many assumptions, but if you cannot reproduce I see > no alternative. Hello, We are not able to reproduce the problem. If we are able to reproduce, we'll take more precise notes and provide what details we can find. Thanks, Ben -- Ben Greear <greearb@candelatech.com> Candela Technologies Inc http://www.candelatech.com ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: Deadlock in usb subsystem on shutdown, 6.18.3+ 2026-01-20 17:29 ` Ben Greear @ 2026-01-29 9:28 ` Michal Pecio 0 siblings, 0 replies; 9+ messages in thread From: Michal Pecio @ 2026-01-29 9:28 UTC (permalink / raw) To: Ben Greear; +Cc: Oliver Neukum, Alan Stern, Hillf Danton, LKML, linux-usb On Tue, 20 Jan 2026 09:29:56 -0800, Ben Greear wrote: > >>> On 1/14/26 07:13, Alan Stern wrote: > >>>> Unfortunately, we have no to tell from the log you collected > >>>> which host controller driver encountered this problem. Nor, > >>>> unless you can replicate the problem, any way to track down > >>>> exactly where in that driver the bug is -- or even any way to > >>>> tell whether a proposed fix actually solves the problem. Considering this looks like x86-64 I think we can make an educated guess that the usual suspect was involved. The guess could be confirmed by running 'lsusb -v' to see if there is anything *other* than xhci-hcd on this machine at all. I have seen shutdown hangs many times when I broke this driver in various ways that caused URBs not to be given back. The unlink mechanism is complex, with many steps to possibly go wrong. > We are not able to reproduce the problem. If we are able to > reproduce, we'll take more precise notes and provide what details > we can find. Chances are you may not see it again, but if you want to hunt for this bug I would suggest setting the system to enable dynamic debug before every shutdown. Particularly xhci_hcd module debug if you can confirm that this is the HCD in use. It is possible to have those symptoms at shutdown due to a problem that already began before shutdown, but maybe went unnoticed. Regards, Michal ^ permalink raw reply [flat|nested] 9+ messages in thread
end of thread, other threads:[~2026-01-29 9:28 UTC | newest] Thread overview: 9+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2026-01-14 0:21 Deadlock in usb subsystem on shutdown, 6.18.3+ Ben Greear 2026-01-14 2:45 ` Hillf Danton 2026-01-14 14:36 ` Ben Greear 2026-01-14 15:13 ` Alan Stern 2026-01-14 17:51 ` Ben Greear 2026-01-14 19:26 ` Alan Stern 2026-01-14 20:14 ` Oliver Neukum 2026-01-20 17:29 ` Ben Greear 2026-01-29 9:28 ` Michal Pecio
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox