From: "Markus Blöchl" <markus.bloechl@ipetronik.com>
To: Keith Busch <kbusch@kernel.org>, Jens Axboe <axboe@kernel.dk>
Cc: Christoph Hellwig <hch@lst.de>, Sagi Grimberg <sagi@grimberg.me>,
linux-nvme@lists.infradead.org, linux-block@vger.kernel.org,
linux-kernel@vger.kernel.org, Stefan Roese <sr@denx.de>
Subject: [RFC PATCH] nvme: prevent hang on surprise removal of NVMe disk
Date: Mon, 14 Feb 2022 10:51:07 +0100 [thread overview]
Message-ID: <20220214095107.3t5en5a3tosaeoo6@ipetronik.com> (raw)
After the surprise removal of a mounted NVMe disk the pciehp task
reliably hangs forever with a trace similar to this one:
INFO: task irq/43-pciehp:64 blocked for more than 120 seconds.
Call Trace:
<TASK>
__bio_queue_enter
blk_mq_submit_bio
submit_bio_noacct
submit_bio_wait
blkdev_issue_flush
ext4_sync_fs
sync_filesystem
fsync_bdev
delete_partition
blk_drop_partitions
del_gendisk
nvme_ns_remove
nvme_remove_namespaces
nvme_remove
pci_device_remove
__device_release_driver
device_release_driver
pci_stop_bus_device
pci_stop_and_remove_bus_device
pciehp_unconfigure_device
pciehp_disable_slot
pciehp_handle_presence_or_link_change
pciehp_ist
</TASK>
I observed this with 5.15.5 from debian bullseye-backports and confirmed
with 5.17.0-rc3 but previous kernels may be affected as well.
I read that del_gendisk() prevents any new I/O only after
flushing and dropping all partitions.
But in case of a surprise removal any new blocking I/O must be prevented
first. I assume that nvme_set_queue_dying() is supposed to do that.
Is there any other mechanism in place which should achieve this?
Unfortunately I am not very familiar with the blk_mq infrastructure so
any comments and suggestions are very welcome.
Best regards,
Markus
Signed-off-by: Markus Blöchl <markus.bloechl@ipetronik.com>
---
drivers/nvme/host/core.c | 2 ++
1 file changed, 2 insertions(+)
diff --git a/drivers/nvme/host/core.c b/drivers/nvme/host/core.c
index 961a5f8a44d2..0654cbe9b80e 100644
--- a/drivers/nvme/host/core.c
+++ b/drivers/nvme/host/core.c
@@ -4573,6 +4573,8 @@ static void nvme_set_queue_dying(struct nvme_ns *ns)
if (test_and_set_bit(NVME_NS_DEAD, &ns->flags))
return;
+ set_bit(GD_DEAD, &ns->disk->state);
+
blk_set_queue_dying(ns->queue);
nvme_start_ns_queue(ns);
base-commit: f1baf68e1383f6ed93eb9cff2866d46562607a43
--
2.35.1
next reply other threads:[~2022-02-14 10:13 UTC|newest]
Thread overview: 12+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-02-14 9:51 Markus Blöchl [this message]
2022-02-15 15:22 ` [RFC PATCH] nvme: prevent hang on surprise removal of NVMe disk Keith Busch
2022-02-15 18:47 ` Christoph Hellwig
2022-02-15 19:14 ` Keith Busch
2022-02-15 19:17 ` Christoph Hellwig
2022-02-15 19:37 ` Keith Busch
2022-02-16 6:39 ` Hannes Reinecke
2022-02-16 11:18 ` Markus Blöchl
2022-02-16 11:32 ` Hannes Reinecke
2022-02-15 20:17 ` Christoph Hellwig
2022-02-16 12:59 ` Markus Blöchl
2022-02-16 13:33 ` Christoph Hellwig
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20220214095107.3t5en5a3tosaeoo6@ipetronik.com \
--to=markus.bloechl@ipetronik.com \
--cc=axboe@kernel.dk \
--cc=hch@lst.de \
--cc=kbusch@kernel.org \
--cc=linux-block@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-nvme@lists.infradead.org \
--cc=sagi@grimberg.me \
--cc=sr@denx.de \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox