From: Mike Christie <michael.christie@oracle.com>
To: subbu.seetharaman@broadcom.com, ketan.mukadam@broadcom.com,
jitendra.bhivare@broadcom.com, lduncan@suse.com,
cleech@redhat.com, njavali@marvell.com, mrangankar@marvell.com,
GR-QLogic-Storage-Upstream@marvell.com, varun@chelsio.com,
martin.petersen@oracle.com, linux-scsi@vger.kernel.org,
james.bottomley@hansenpartnership.com
Subject: [PATCH 10/15] qedi: set scsi_host_template cmd_size
Date: Tue, 1 Dec 2020 15:29:51 -0600 [thread overview]
Message-ID: <1606858196-5421-11-git-send-email-michael.christie@oracle.com> (raw)
In-Reply-To: <1606858196-5421-1-git-send-email-michael.christie@oracle.com>
Use scsi_host_template cmd_size so the block/scsi-ml layers allocate
the iscsi structs for the driver.
Signed-off-by: Mike Christie <michael.christie@oracle.com>
---
drivers/scsi/qedi/qedi_iscsi.c | 113 ++++++++++++++++++-----------------------
1 file changed, 49 insertions(+), 64 deletions(-)
diff --git a/drivers/scsi/qedi/qedi_iscsi.c b/drivers/scsi/qedi/qedi_iscsi.c
index 08c0540..979b875 100644
--- a/drivers/scsi/qedi/qedi_iscsi.c
+++ b/drivers/scsi/qedi/qedi_iscsi.c
@@ -14,6 +14,9 @@
#include "qedi_iscsi.h"
#include "qedi_gbl.h"
+static int qedi_exit_cmd_priv(struct Scsi_Host *shost, struct scsi_cmnd *sc);
+static int qedi_init_cmd_priv(struct Scsi_Host *shost, struct scsi_cmnd *sc);
+
int qedi_recover_all_conns(struct qedi_ctx *qedi)
{
struct qedi_conn *qedi_conn;
@@ -59,6 +62,9 @@ struct scsi_host_template qedi_host_template = {
.dma_boundary = QEDI_HW_DMA_BOUNDARY,
.cmd_per_lun = 128,
.shost_attrs = qedi_shost_attrs,
+ .cmd_size = sizeof(struct qedi_cmd) + sizeof(struct iscsi_task),
+ .init_cmd_priv = qedi_init_cmd_priv,
+ .exit_cmd_priv = qedi_exit_cmd_priv,
};
static void qedi_conn_free_login_resources(struct qedi_ctx *qedi,
@@ -160,32 +166,17 @@ static int qedi_conn_alloc_login_resources(struct qedi_ctx *qedi,
return -ENOMEM;
}
-static void qedi_destroy_cmd_pool(struct qedi_ctx *qedi,
- struct iscsi_session *session)
+static void qedi_free_sget(struct qedi_ctx *qedi, struct qedi_cmd *cmd)
{
- int i;
+ if (!cmd->io_tbl.sge_tbl)
+ return;
- for (i = 0; i < session->cmds_max; i++) {
- struct iscsi_task *task = session->cmds[i];
- struct qedi_cmd *cmd = task->dd_data;
-
- if (cmd->io_tbl.sge_tbl)
- dma_free_coherent(&qedi->pdev->dev,
- QEDI_ISCSI_MAX_BDS_PER_CMD *
- sizeof(struct scsi_sge),
- cmd->io_tbl.sge_tbl,
- cmd->io_tbl.sge_tbl_dma);
-
- if (cmd->sense_buffer)
- dma_free_coherent(&qedi->pdev->dev,
- SCSI_SENSE_BUFFERSIZE,
- cmd->sense_buffer,
- cmd->sense_buffer_dma);
- }
+ dma_free_coherent(&qedi->pdev->dev,
+ QEDI_ISCSI_MAX_BDS_PER_CMD * sizeof(struct scsi_sge),
+ cmd->io_tbl.sge_tbl, cmd->io_tbl.sge_tbl_dma);
}
-static int qedi_alloc_sget(struct qedi_ctx *qedi, struct iscsi_session *session,
- struct qedi_cmd *cmd)
+static int qedi_alloc_sget(struct qedi_ctx *qedi, struct qedi_cmd *cmd)
{
struct qedi_io_bdt *io = &cmd->io_tbl;
struct scsi_sge *sge;
@@ -195,8 +186,8 @@ static int qedi_alloc_sget(struct qedi_ctx *qedi, struct iscsi_session *session,
sizeof(*sge),
&io->sge_tbl_dma, GFP_KERNEL);
if (!io->sge_tbl) {
- iscsi_session_printk(KERN_ERR, session,
- "Could not allocate BD table.\n");
+ shost_printk(KERN_ERR, qedi->shost,
+ "Could not allocate BD table.\n");
return -ENOMEM;
}
@@ -204,33 +195,47 @@ static int qedi_alloc_sget(struct qedi_ctx *qedi, struct iscsi_session *session,
return 0;
}
-static int qedi_setup_cmd_pool(struct qedi_ctx *qedi,
- struct iscsi_session *session)
+static int qedi_exit_cmd_priv(struct Scsi_Host *shost, struct scsi_cmnd *sc)
{
- int i;
+ struct qedi_ctx *qedi = iscsi_host_priv(shost);
+ struct iscsi_task *task = scsi_cmd_priv(sc);
+ struct qedi_cmd *cmd = task->dd_data;
- for (i = 0; i < session->cmds_max; i++) {
- struct iscsi_task *task = session->cmds[i];
- struct qedi_cmd *cmd = task->dd_data;
+ qedi_free_sget(qedi, cmd);
- task->hdr = &cmd->hdr;
- task->hdr_max = sizeof(struct iscsi_hdr);
+ if (cmd->sense_buffer)
+ dma_free_coherent(&qedi->pdev->dev, SCSI_SENSE_BUFFERSIZE,
+ cmd->sense_buffer, cmd->sense_buffer_dma);
+ return 0;
+}
+
+static int qedi_init_cmd_priv(struct Scsi_Host *shost, struct scsi_cmnd *sc)
+{
+ struct qedi_ctx *qedi = iscsi_host_priv(shost);
+ struct iscsi_task *task;
+ struct qedi_cmd *cmd;
- if (qedi_alloc_sget(qedi, session, cmd))
- goto free_sgets;
+ iscsi_init_cmd_priv(shost, sc);
- cmd->sense_buffer = dma_alloc_coherent(&qedi->pdev->dev,
- SCSI_SENSE_BUFFERSIZE,
- &cmd->sense_buffer_dma,
- GFP_KERNEL);
- if (!cmd->sense_buffer)
- goto free_sgets;
- }
+ task = scsi_cmd_priv(sc);
+ cmd = task->dd_data;
+ task->hdr = &cmd->hdr;
+ task->hdr_max = sizeof(struct iscsi_hdr);
+
+ if (qedi_alloc_sget(qedi, cmd))
+ goto free_sgets;
+
+ cmd->sense_buffer = dma_alloc_coherent(&qedi->pdev->dev,
+ SCSI_SENSE_BUFFERSIZE,
+ &cmd->sense_buffer_dma,
+ GFP_KERNEL);
+ if (!cmd->sense_buffer)
+ goto free_sgets;
return 0;
free_sgets:
- qedi_destroy_cmd_pool(qedi, session);
+ qedi_free_sget(qedi, cmd);
return -ENOMEM;
}
@@ -264,27 +269,7 @@ static int qedi_setup_cmd_pool(struct qedi_ctx *qedi,
return NULL;
}
- if (qedi_setup_cmd_pool(qedi, cls_session->dd_data)) {
- QEDI_ERR(&qedi->dbg_ctx,
- "Failed to setup cmd pool for ep=%p\n", qedi_ep);
- goto session_teardown;
- }
-
return cls_session;
-
-session_teardown:
- iscsi_session_teardown(cls_session);
- return NULL;
-}
-
-static void qedi_session_destroy(struct iscsi_cls_session *cls_session)
-{
- struct iscsi_session *session = cls_session->dd_data;
- struct Scsi_Host *shost = iscsi_session_to_shost(cls_session);
- struct qedi_ctx *qedi = iscsi_host_priv(shost);
-
- qedi_destroy_cmd_pool(qedi, session);
- iscsi_session_teardown(cls_session);
}
static struct iscsi_cls_conn *
@@ -1398,7 +1383,7 @@ struct iscsi_transport qedi_iscsi_transport = {
.caps = CAP_RECOVERY_L0 | CAP_HDRDGST | CAP_MULTI_R2T | CAP_DATADGST |
CAP_DATA_PATH_OFFLOAD | CAP_TEXT_NEGO,
.create_session = qedi_session_create,
- .destroy_session = qedi_session_destroy,
+ .destroy_session = iscsi_session_teardown,
.create_conn = qedi_conn_create,
.bind_conn = qedi_conn_bind,
.start_conn = qedi_conn_start,
@@ -1625,7 +1610,7 @@ void qedi_clear_session_ctx(struct iscsi_cls_session *cls_sess)
qedi_conn_destroy(qedi_conn->cls_conn);
- qedi_session_destroy(cls_sess);
+ iscsi_session_teardown(cls_sess);
}
void qedi_process_tcp_error(struct qedi_endpoint *ep,
--
1.8.3.1
next prev parent reply other threads:[~2020-12-01 21:31 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-12-01 21:29 [RFC PATCH 00/15] libiscsi: lock clean ups Mike Christie
2020-12-01 21:29 ` [PATCH 01/15] libiscsi: fix iscsi_prep_scsi_cmd_pdu error handling Mike Christie
2020-12-02 17:04 ` Lee Duncan
2020-12-01 21:29 ` [PATCH 02/15] libiscsi: drop taskqueuelock Mike Christie
2020-12-01 21:29 ` [PATCH 03/15] qla4xxx: use iscsi_is_session_online Mike Christie
2020-12-03 21:07 ` Lee Duncan
2020-12-01 21:29 ` [PATCH 04/15] iscsi class: drop session lock in iscsi_session_chkready Mike Christie
2020-12-03 21:08 ` Lee Duncan
2020-12-01 21:29 ` [PATCH 05/15] libiscsi: remove queued_cmdsn Mike Christie
2020-12-01 21:29 ` [PATCH 06/15] libiscsi: drop frwd lock for session state Mike Christie
2020-12-01 21:29 ` [PATCH 07/15] libiscsi: separate itt from task allocation Mike Christie
2020-12-01 21:29 ` [PATCH 08/15] iser, be2iscsi, qla4xxx: set scsi_host_template cmd_size Mike Christie
2020-12-01 21:29 ` [PATCH 09/15] bnx2i: " Mike Christie
2020-12-01 21:29 ` Mike Christie [this message]
2020-12-01 21:29 ` [PATCH 11/15] iscsi_tcp, libcxgbi: " Mike Christie
2020-12-01 21:29 ` [PATCH 12/15] libiscsi: use blk/scsi-ml mq cmd pre-allocator Mike Christie
2020-12-01 21:29 ` [PATCH 13/15] libiscsi: drop back_lock requirement for iscsi_put_task Mike Christie
2020-12-01 21:29 ` [PATCH 14/15] libiscsi: drop back_lock from xmit path Mike Christie
2020-12-01 21:29 ` [PATCH 15/15] libiscsi: convert ping_task to refcount handler Mike Christie
2020-12-02 1:27 ` Mike Christie
2020-12-04 12:10 ` [EXT] [RFC PATCH 00/15] libiscsi: lock clean ups Manish Rangankar
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1606858196-5421-11-git-send-email-michael.christie@oracle.com \
--to=michael.christie@oracle.com \
--cc=GR-QLogic-Storage-Upstream@marvell.com \
--cc=cleech@redhat.com \
--cc=james.bottomley@hansenpartnership.com \
--cc=jitendra.bhivare@broadcom.com \
--cc=ketan.mukadam@broadcom.com \
--cc=lduncan@suse.com \
--cc=linux-scsi@vger.kernel.org \
--cc=martin.petersen@oracle.com \
--cc=mrangankar@marvell.com \
--cc=njavali@marvell.com \
--cc=subbu.seetharaman@broadcom.com \
--cc=varun@chelsio.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).