From: Johan Hovold <johan@kernel.org>
To: Jens Axboe <axboe@kernel.dk>, Ming Lei <ming.lei@redhat.com>,
Christoph Hellwig <hch@lst.de>
Cc: Alan Stern <stern@rowland.harvard.edu>,
linux-block@vger.kernel.org, linux-usb@vger.kernel.org,
usb-storage@lists.one-eyed-alien.net,
linux-kernel@vger.kernel.org, regressions@lists.linux.dev
Subject: Lockdep splat on UMS disconnect (6.13-rc1)
Date: Wed, 4 Dec 2024 12:26:53 +0100 [thread overview]
Message-ID: <Z1A8fai9_fQFhs1s@hovoldconsulting.com> (raw)
Hi,
I'm seeing a new lockdep splat with 6.13-rc1 on resume after having
disconnected a USB mass storage device while suspended. This does not
seem to show up with 6.12.
On one attempt to reproduce this with a ThinkPad T14s, the NVMe was
hosed after resume as well so I fear that this is not just about missing
annotation.
I don't have time to look into this myself right now so I'm reporting in
the hope that someone familiar with these paths would be able to
pinpoint any potential regression without too much effort.
Johan
#regzbot introduced: v6.12..v6.13-rc1
[ 59.529878] Restarting tasks ... done.
[ 59.534619] usb 5-1: USB disconnect, device number 2
[ 59.536537] random: crng reseeded on system resumption
[ 59.541059] xhci-hcd xhci-hcd.7.auto: remove, state 1
[ 59.547012] PM: suspend exit
[ 59.584871] ======================================================
[ 59.584873] WARNING: possible circular locking dependency detected
[ 59.584875] 6.13.0-rc1 #11 Tainted: G W
[ 59.584877] ------------------------------------------------------
[ 59.584878] kworker/0:2/126 is trying to acquire lock:
[ 59.584880] ffff1a9a948b83e0 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: __flush_work+0x1c0/0x504
[ 59.584894]
but task is already holding lock:
[ 59.584895] ffff1a9a948b00e0 (&q->q_usage_counter(queue)#3){++++}-{0:0}, at: sd_remove+0x38/0x88
[ 59.584905]
which lock already depends on the new lock.
[ 59.584906]
the existing dependency chain (in reverse order) is:
[ 59.584907]
-> #3 (&q->q_usage_counter(queue)#3){++++}-{0:0}:
[ 59.584912] blk_queue_enter+0x174/0x1c4
[ 59.584917] blk_mq_alloc_request+0xa0/0xcc
[ 59.584921] scsi_execute_cmd+0xbc/0x3c8
[ 59.584924] read_capacity_10+0xcc/0x1e0
[ 59.584927] sd_revalidate_disk.isra.0+0xad4/0x26fc
[ 59.584930] sd_probe+0x2a0/0x42c
[ 59.584933] really_probe+0xc0/0x388
[ 59.584938] __driver_probe_device+0x7c/0x160
[ 59.584941] driver_probe_device+0x40/0x110
[ 59.584944] __device_attach_driver+0xbc/0x158
[ 59.584948] bus_for_each_drv+0x84/0xe0
[ 59.584950] __device_attach_async_helper+0xb0/0x10c
[ 59.584954] async_run_entry_fn+0x34/0x14c
[ 59.584956] process_one_work+0x20c/0x610
[ 59.584960] worker_thread+0x23c/0x378
[ 59.584963] kthread+0x124/0x128
[ 59.584965] ret_from_fork+0x10/0x20
[ 59.584968]
-> #2 (&q->limits_lock){+.+.}-{4:4}:
[ 59.584972] __mutex_lock+0xa8/0x468
[ 59.584976] mutex_lock_nested+0x24/0x30
[ 59.584978] nvme_update_ns_info_block+0x1a4/0x6c4 [nvme_core]
[ 59.584990] nvme_update_ns_info+0x9c/0xf0 [nvme_core]
[ 59.584995] nvme_alloc_ns+0x2f0/0x9f0 [nvme_core]
[ 59.584999] nvme_scan_ns+0x278/0x378 [nvme_core]
[ 59.585004] nvme_scan_ns_async+0x2c/0x70 [nvme_core]
[ 59.585008] async_run_entry_fn+0x34/0x14c
[ 59.585010] process_one_work+0x20c/0x610
[ 59.585014] worker_thread+0x23c/0x378
[ 59.585017] kthread+0x124/0x128
[ 59.585019] ret_from_fork+0x10/0x20
[ 59.585021]
-> #1 (&q->q_usage_counter(io)){++++}-{0:0}:
[ 59.585025] blk_mq_submit_bio+0x82c/0x9ac
[ 59.585028] __submit_bio+0x78/0x168
[ 59.585030] submit_bio_noacct_nocheck+0x2d4/0x3b8
[ 59.585033] submit_bio_noacct+0x148/0x3f8
[ 59.585035] submit_bio+0x12c/0x200
[ 59.585037] submit_bh_wbc+0x144/0x1c0
[ 59.585042] __block_write_full_folio+0x210/0x4cc
[ 59.585045] block_write_full_folio+0xc8/0x14c
[ 59.585048] write_cache_pages+0x64/0xd4
[ 59.585054] blkdev_writepages+0x4c/0x88
[ 59.585057] do_writepages+0x84/0x2b8
[ 59.585059] __writeback_single_inode+0x64/0x5d8
[ 59.585062] writeback_sb_inodes+0x244/0x500
[ 59.585065] __writeback_inodes_wb+0x50/0x108
[ 59.585068] wb_writeback+0x3a8/0x4a0
[ 59.585070] wb_workfn+0x2e4/0x5fc
[ 59.585073] process_one_work+0x20c/0x610
[ 59.585076] worker_thread+0x23c/0x378
[ 59.585079] kthread+0x124/0x128
[ 59.585082] ret_from_fork+0x10/0x20
[ 59.585085]
-> #0 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}:
[ 60.565855] __lock_acquire+0x1344/0x20e8
[ 60.565860] lock_acquire+0x1c8/0x354
[ 60.565863] __flush_work+0x1e8/0x504
[ 60.565868] flush_delayed_work+0x50/0x84
[ 60.565874] wb_shutdown+0x9c/0x10c
[ 60.565881] bdi_unregister+0x88/0x1e8
[ 60.565887] del_gendisk+0x330/0x370
[ 60.565892] sd_remove+0x38/0x88
[ 60.565897] device_remove+0x70/0x80
[ 60.565902] device_release_driver_internal+0x1e4/0x240
[ 60.565908] device_release_driver+0x18/0x24
[ 60.565913] bus_remove_device+0xd0/0x148
[ 60.565918] device_del+0x148/0x374
[ 60.565921] __scsi_remove_device+0x124/0x18c
[ 60.565925] scsi_forget_host+0x78/0x80
[ 60.565930] scsi_remove_host+0x74/0x194
[ 60.565934] usb_stor_disconnect+0x74/0xfc [usb_storage]
[ 60.565942] usb_unbind_interface+0x80/0x2bc
[ 60.565946] device_remove+0x70/0x80
[ 60.565951] device_release_driver_internal+0x1e4/0x240
[ 60.565957] device_release_driver+0x18/0x24
[ 60.565962] bus_remove_device+0xd0/0x148
[ 60.565967] device_del+0x148/0x374
[ 60.565970] usb_disable_device+0xf8/0x244
[ 60.565976] usb_disconnect+0xe4/0x314
[ 60.565980] hub_event+0xbe8/0x1950
[ 60.565985] process_one_work+0x20c/0x610
[ 60.565990] worker_thread+0x23c/0x378
[ 60.565995] kthread+0x124/0x128
[ 60.565998] ret_from_fork+0x10/0x20
[ 60.566003]
other info that might help us debug this:
[ 60.566004] Chain exists of:
(work_completion)(&(&wb->dwork)->work) --> &q->limits_lock --> &q->q_usage_counter(queue)#3
[ 60.566014] Possible unsafe locking scenario:
[ 60.566015] CPU0 CPU1
[ 60.566016] ---- ----
[ 60.566018] lock(&q->q_usage_counter(queue)#3);
[ 60.566022] lock(&q->limits_lock);
[ 60.566026] lock(&q->q_usage_counter(queue)#3);
[ 60.566030] lock((work_completion)(&(&wb->dwork)->work));
[ 60.566033]
*** DEADLOCK ***
[ 60.566035] 9 locks held by kworker/0:2/126:
[ 60.566038] #0: ffff1a9a81cb8148 ((wq_completion)usb_hub_wq){+.+.}-{0:0}, at: process_one_work+0x190/0x610
[ 60.566049] #1: ffff800081c63de0 ((work_completion)(&hub->events)){+.+.}-{0:0}, at: process_one_work+0x1b8/0x610
[ 60.566060] #2: ffff1a9a9960f990 (&dev->mutex){....}-{4:4}, at: hub_event+0x5c/0x1950
[ 60.566070] #3: ffff1a9a94f8a990 (&dev->mutex){....}-{4:4}, at: usb_disconnect+0x84/0x314
[ 60.566080] #4: ffff1a9a94f8b160 (&dev->mutex){....}-{4:4}, at: device_release_driver_internal+0x40/0x240
[ 60.566092] #5: ffff1a9a9409c0e0 (&shost->scan_mutex){+.+.}-{4:4}, at: scsi_remove_host+0x24/0x194
[ 60.566101] #6: ffff1a9a94abf378 (&dev->mutex){....}-{4:4}, at: device_release_driver_internal+0x40/0x240
[ 60.566112] #7: ffff1a9a948b00e0 (&q->q_usage_counter(queue)#3){++++}-{0:0}, at: sd_remove+0x38/0x88
[ 60.566123] #8: ffffd15e3cd706c0 (rcu_read_lock){....}-{1:3}, at: __flush_work+0x48/0x504
[ 60.566134]
stack backtrace:
[ 60.566137] CPU: 0 UID: 0 PID: 126 Comm: kworker/0:2 Tainted: G W 6.13.0-rc1 #11
[ 60.566143] Tainted: [W]=WARN
[ 60.566145] Hardware name: Qualcomm CRD, BIOS 6.0.231221.BOOT.MXF.2.4-00348.1-HAMOA-1 12/21/2023
[ 60.566149] Workqueue: usb_hub_wq hub_event
[ 60.566155] Call trace:
[ 60.566158] show_stack+0x18/0x24 (C)
[ 60.566163] dump_stack_lvl+0x90/0xd0
[ 61.400051] dump_stack+0x18/0x24
[ 61.400055] print_circular_bug+0x298/0x37c
[ 61.400059] check_noncircular+0x15c/0x170
[ 61.400062] __lock_acquire+0x1344/0x20e8
[ 61.400066] lock_acquire+0x1c8/0x354
[ 61.400069] __flush_work+0x1e8/0x504
[ 61.400074] flush_delayed_work+0x50/0x84
[ 61.400080] wb_shutdown+0x9c/0x10c
[ 61.400086] bdi_unregister+0x88/0x1e8
[ 61.400091] del_gendisk+0x330/0x370
[ 61.400097] sd_remove+0x38/0x88
[ 61.400102] device_remove+0x70/0x80
[ 61.400107] device_release_driver_internal+0x1e4/0x240
[ 61.400113] device_release_driver+0x18/0x24
[ 61.400118] bus_remove_device+0xd0/0x148
[ 61.400123] device_del+0x148/0x374
[ 61.400126] __scsi_remove_device+0x124/0x18c
[ 61.400130] scsi_forget_host+0x78/0x80
[ 61.400135] scsi_remove_host+0x74/0x194
[ 61.400139] usb_stor_disconnect+0x74/0xfc [usb_storage]
[ 61.400144] usb_unbind_interface+0x80/0x2bc
[ 61.400148] device_remove+0x70/0x80
[ 61.400153] device_release_driver_internal+0x1e4/0x240
[ 61.400158] device_release_driver+0x18/0x24
[ 61.400164] bus_remove_device+0xd0/0x148
[ 61.400169] device_del+0x148/0x374
[ 61.400172] usb_disable_device+0xf8/0x244
[ 61.400178] usb_disconnect+0xe4/0x314
[ 61.400182] hub_event+0xbe8/0x1950
[ 61.400187] process_one_work+0x20c/0x610
[ 61.400192] worker_thread+0x23c/0x378
[ 61.400197] kthread+0x124/0x128
[ 61.400201] ret_from_fork+0x10/0x20
[ 61.508474] usb usb5: USB disconnect, device number 1
[ 61.618437] xhci-hcd xhci-hcd.7.auto: USB bus 5 deregistered
[ 61.618467] xhci-hcd xhci-hcd.7.auto: remove, state 4
[ 61.618472] usb usb4: USB disconnect, device number 1
[ 61.619930] xhci-hcd xhci-hcd.7.auto: USB bus 4 deregistered
next reply other threads:[~2024-12-04 11:26 UTC|newest]
Thread overview: 2+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-12-04 11:26 Johan Hovold [this message]
2024-12-04 12:14 ` Lockdep splat on UMS disconnect (6.13-rc1) Ming Lei
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=Z1A8fai9_fQFhs1s@hovoldconsulting.com \
--to=johan@kernel.org \
--cc=axboe@kernel.dk \
--cc=hch@lst.de \
--cc=linux-block@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-usb@vger.kernel.org \
--cc=ming.lei@redhat.com \
--cc=regressions@lists.linux.dev \
--cc=stern@rowland.harvard.edu \
--cc=usb-storage@lists.one-eyed-alien.net \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox