From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 0C7726A8A3; Wed, 7 Feb 2024 21:26:50 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1707341210; cv=none; b=cdFYGF3vqWDLtc7/o0ndylK9z3FmrqXS+GQw/rEjKupJAs4WQn+CqSFz2vyYGxylL7uwNbv4lxljrXzo5/xp5NLfKvf6GMsNcFgP4qfgkTk603WA7/aTERA3RvFtt/+RCSia0MNFIqJdfSrT1T4TIpxDe4UVg/NTFtIdHmifgIg= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1707341210; c=relaxed/simple; bh=vDbkSJI7wzoSvESOacVLNT1hZSBIwV4Ql2xwrsn1nGk=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=Yjx4cxXHJHAGh47SGRzSrwDPCVk5V8UQ6LorkEEV8TD+ZetepCfc8U1awhUNxOQefiY7PTE8svmPvNJleo6BcQElRKvdaGVXIQoeD1ybJeJRngYJJQ1MahpR7bcHkp4qXE4xyHzFqNiPRvFhmLMrLnv+94lalCDwDCB+Gtb3hVE= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b=UWa3qHLK; arc=none smtp.client-ip=10.30.226.201 Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b="UWa3qHLK" Received: by smtp.kernel.org (Postfix) with ESMTPSA id 94A35C43394; Wed, 7 Feb 2024 21:26:48 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1707341209; bh=vDbkSJI7wzoSvESOacVLNT1hZSBIwV4Ql2xwrsn1nGk=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=UWa3qHLKUVx1/a9IInCk4FXb8O8SbeBODNw3RIFYvSOjVGNsyOIA17SndoOCCYSo9 vGOhYY3hp4iOn87Y3GcPQn88PcRS39hBMq+7corG0lrvLyLFQLZh6LxrZMUby9GBsr VHQ8SHDm3j7P2QNSIM77HYeqxhnrCYUcRvwRHpu+oSafc0qRyzARt4DDq4oG38ZVwV kcmcTe3hQ1380Qca5hPs6vpY0rn6e83OBMLAWEOg3ny3PdIWePWsdY8w9aVFkV0yWK n8iby4wml/ZFNPnDdv0snwDh7dOBdR1vV8jLopqwArG9u5WQWEsXIwDAx9nHitkSf4 oLfWtUIslaiJA== From: Sasha Levin To: linux-kernel@vger.kernel.org, stable@vger.kernel.org Cc: Daniel Wagner , Christoph Hellwig , Keith Busch , Sasha Levin , james.smart@broadcom.com, sagi@grimberg.me, kch@nvidia.com, linux-nvme@lists.infradead.org Subject: [PATCH AUTOSEL 5.15 20/23] nvmet-fc: avoid deadlock on delete association path Date: Wed, 7 Feb 2024 16:26:01 -0500 Message-ID: <20240207212611.3793-20-sashal@kernel.org> X-Mailer: git-send-email 2.43.0 In-Reply-To: <20240207212611.3793-1-sashal@kernel.org> References: <20240207212611.3793-1-sashal@kernel.org> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 X-stable: review X-Patchwork-Hint: Ignore X-stable-base: Linux 5.15.148 Content-Transfer-Encoding: 8bit From: Daniel Wagner [ Upstream commit 710c69dbaccdac312e32931abcb8499c1525d397 ] When deleting an association the shutdown path is deadlocking because we try to flush the nvmet_wq nested. Avoid this by deadlock by deferring the put work into its own work item. Reviewed-by: Christoph Hellwig Signed-off-by: Daniel Wagner Signed-off-by: Keith Busch Signed-off-by: Sasha Levin --- drivers/nvme/target/fc.c | 16 +++++++++++++--- 1 file changed, 13 insertions(+), 3 deletions(-) diff --git a/drivers/nvme/target/fc.c b/drivers/nvme/target/fc.c index 20d3013be08a..1ef075b159b9 100644 --- a/drivers/nvme/target/fc.c +++ b/drivers/nvme/target/fc.c @@ -111,6 +111,8 @@ struct nvmet_fc_tgtport { struct nvmet_fc_port_entry *pe; struct kref ref; u32 max_sg_cnt; + + struct work_struct put_work; }; struct nvmet_fc_port_entry { @@ -248,6 +250,13 @@ static int nvmet_fc_tgt_a_get(struct nvmet_fc_tgt_assoc *assoc); static void nvmet_fc_tgt_q_put(struct nvmet_fc_tgt_queue *queue); static int nvmet_fc_tgt_q_get(struct nvmet_fc_tgt_queue *queue); static void nvmet_fc_tgtport_put(struct nvmet_fc_tgtport *tgtport); +static void nvmet_fc_put_tgtport_work(struct work_struct *work) +{ + struct nvmet_fc_tgtport *tgtport = + container_of(work, struct nvmet_fc_tgtport, put_work); + + nvmet_fc_tgtport_put(tgtport); +} static int nvmet_fc_tgtport_get(struct nvmet_fc_tgtport *tgtport); static void nvmet_fc_handle_fcp_rqst(struct nvmet_fc_tgtport *tgtport, struct nvmet_fc_fcp_iod *fod); @@ -359,7 +368,7 @@ __nvmet_fc_finish_ls_req(struct nvmet_fc_ls_req_op *lsop) if (!lsop->req_queued) { spin_unlock_irqrestore(&tgtport->lock, flags); - goto out_puttgtport; + goto out_putwork; } list_del(&lsop->lsreq_list); @@ -372,8 +381,8 @@ __nvmet_fc_finish_ls_req(struct nvmet_fc_ls_req_op *lsop) (lsreq->rqstlen + lsreq->rsplen), DMA_BIDIRECTIONAL); -out_puttgtport: - nvmet_fc_tgtport_put(tgtport); +out_putwork: + queue_work(nvmet_wq, &tgtport->put_work); } static int @@ -1404,6 +1413,7 @@ nvmet_fc_register_targetport(struct nvmet_fc_port_info *pinfo, kref_init(&newrec->ref); ida_init(&newrec->assoc_cnt); newrec->max_sg_cnt = template->max_sgl_segments; + INIT_WORK(&newrec->put_work, nvmet_fc_put_tgtport_work); ret = nvmet_fc_alloc_ls_iodlist(newrec); if (ret) { -- 2.43.0