From: Daniel Wagner <dwagner@suse.de>
To: James Smart <jsmart2021@gmail.com>
Cc: linux-scsi@vger.kernel.org, maier@linux.ibm.com,
bvanassche@acm.org, herbszt@gmx.de, natechancellor@gmail.com,
rdunlap@infradead.org, hare@suse.de,
Ram Vegesna <ram.vegesna@broadcom.com>
Subject: Re: [PATCH v3 24/31] elx: efct: LIO backend interface routines
Date: Thu, 16 Apr 2020 14:34:25 +0200 [thread overview]
Message-ID: <20200416123425.lj2clovshk5yxz6g@carbon> (raw)
In-Reply-To: <20200412033303.29574-25-jsmart2021@gmail.com>
On Sat, Apr 11, 2020 at 08:32:56PM -0700, James Smart wrote:
> This patch continues the efct driver population.
>
> This patch adds driver definitions for:
> LIO backend template registration and template functions.
>
> Signed-off-by: Ram Vegesna <ram.vegesna@broadcom.com>
> Signed-off-by: James Smart <jsmart2021@gmail.com>
>
> ---
> v3:
> Fixed as per the review comments.
> Removed vport pend list. Pending list is tracked based on the sport
> assigned to vport.
> ---
> drivers/scsi/elx/efct/efct_lio.c | 1840 ++++++++++++++++++++++++++++++++++++++
> drivers/scsi/elx/efct/efct_lio.h | 178 ++++
> 2 files changed, 2018 insertions(+)
> create mode 100644 drivers/scsi/elx/efct/efct_lio.c
> create mode 100644 drivers/scsi/elx/efct/efct_lio.h
>
> diff --git a/drivers/scsi/elx/efct/efct_lio.c b/drivers/scsi/elx/efct/efct_lio.c
> new file mode 100644
> index 000000000000..c784ef9dbbee
> --- /dev/null
> +++ b/drivers/scsi/elx/efct/efct_lio.c
> @@ -0,0 +1,1840 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (C) 2019 Broadcom. All Rights Reserved. The term
> + * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
> + */
> +
> +#include <target/target_core_base.h>
> +#include <target/target_core_fabric.h>
> +#include "efct_driver.h"
> +#include "efct_lio.h"
> +
> +static struct workqueue_struct *lio_wq;
> +
> +static int
> +efct_format_wwn(char *str, size_t len, const char *pre, u64 wwn)
> +{
> + u8 a[8];
> +
> + put_unaligned_be64(wwn, a);
> + return snprintf(str, len, "%s%8phC", pre, a);
> +}
> +
> +static int
> +efct_lio_parse_wwn(const char *name, u64 *wwp, u8 npiv)
> +{
> + int num;
> + u8 b[8];
> +
> + if (npiv) {
> + num = sscanf(name,
> + "%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx",
> + &b[0], &b[1], &b[2], &b[3], &b[4], &b[5], &b[6],
> + &b[7]);
> + } else {
> + num = sscanf(name,
> + "%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx",
> + &b[0], &b[1], &b[2], &b[3], &b[4], &b[5], &b[6],
> + &b[7]);
> + }
> +
> + if (num != 8)
> + return -EINVAL;
> +
> + *wwp = get_unaligned_be64(b);
> + return EFC_SUCCESS;
> +}
> +
> +static int
> +efct_lio_parse_npiv_wwn(const char *name, size_t size, u64 *wwpn, u64 *wwnn)
> +{
> + unsigned int cnt = size;
> + int rc;
> +
> + *wwpn = *wwnn = 0;
> + if (name[cnt - 1] == '\n' || name[cnt - 1] == 0)
> + cnt--;
> +
> + /* validate we have enough characters for WWPN */
> + if ((cnt != (16 + 1 + 16)) || (name[16] != ':'))
> + return -EINVAL;
> +
> + rc = efct_lio_parse_wwn(&name[0], wwpn, 1);
> + if (rc)
> + return rc;
> +
> + rc = efct_lio_parse_wwn(&name[17], wwnn, 1);
> + if (rc)
> + return rc;
> +
> + return EFC_SUCCESS;
> +}
> +
> +static ssize_t
> +efct_lio_tpg_enable_show(struct config_item *item, char *page)
> +{
> + struct se_portal_group *se_tpg = to_tpg(item);
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + return snprintf(page, PAGE_SIZE, "%d\n", atomic_read(&tpg->enabled));
> +}
> +
> +static ssize_t
> +efct_lio_tpg_enable_store(struct config_item *item, const char *page,
> + size_t count)
> +{
> + struct se_portal_group *se_tpg = to_tpg(item);
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> + struct efct *efct;
> + struct efc *efc;
> + unsigned long op;
> + int ret;
> +
> + if (!tpg->sport || !tpg->sport->efct) {
> + pr_err("%s: Unable to find EFCT device\n", __func__);
> + return -EINVAL;
> + }
> +
> + efct = tpg->sport->efct;
> + efc = efct->efcport;
> +
> + if (kstrtoul(page, 0, &op) < 0)
> + return -EINVAL;
> +
> + if (op == 1) {
> + atomic_set(&tpg->enabled, 1);
> + efc_log_debug(efct, "enable portal group %d\n", tpg->tpgt);
> +
> + ret = efct_xport_control(efct->xport, EFCT_XPORT_PORT_ONLINE);
> + if (ret) {
> + efct->tgt_efct.lio_sport = NULL;
> + efc_log_test(efct, "cannot bring port online\n");
> + return ret;
> + }
> + } else if (op == 0) {
> + efc_log_debug(efct, "disable portal group %d\n", tpg->tpgt);
> +
> + if (efc->domain && efc->domain->sport)
> + efct_scsi_tgt_del_sport(efc, efc->domain->sport);
> +
> + atomic_set(&tpg->enabled, 0);
> + } else {
> + return -EINVAL;
> + }
> +
> + return count;
> +}
> +
> +static ssize_t
> +efct_lio_npiv_tpg_enable_show(struct config_item *item, char *page)
> +{
> + struct se_portal_group *se_tpg = to_tpg(item);
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + return snprintf(page, PAGE_SIZE, "%d\n", atomic_read(&tpg->enabled));
> +}
> +
> +static ssize_t
> +efct_lio_npiv_tpg_enable_store(struct config_item *item, const char *page,
> + size_t count)
> +{
> + struct se_portal_group *se_tpg = to_tpg(item);
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> + struct efct_lio_vport *lio_vport = tpg->vport;
> + struct efct *efct;
> + struct efc *efc;
> + int ret;
> + unsigned long op;
> +
> + if (kstrtoul(page, 0, &op) < 0)
> + return -EINVAL;
> +
> + if (!lio_vport) {
> + pr_err("Unable to find vport\n");
> + return -EINVAL;
> + }
> +
> + efct = lio_vport->efct;
> + efc = efct->efcport;
> +
> + if (op == 1) {
> + atomic_set(&tpg->enabled, 1);
> + efc_log_debug(efct, "enable portal group %d\n", tpg->tpgt);
> +
> + if (efc->domain) {
> + ret = efc_sport_vport_new(efc->domain,
> + lio_vport->npiv_wwpn,
> + lio_vport->npiv_wwnn,
> + U32_MAX, false, true,
> + NULL, NULL);
> + if (ret != 0) {
> + efc_log_err(efct, "Failed to create Vport\n");
> + return ret;
> + }
> + return count;
> + }
> +
> + if (!(efc_vport_create_spec(efc, lio_vport->npiv_wwnn,
> + lio_vport->npiv_wwpn, U32_MAX,
> + false, true, NULL, NULL)))
> + return -ENOMEM;
> +
> + } else if (op == 0) {
> + efc_log_debug(efct, "disable portal group %d\n", tpg->tpgt);
> +
> + atomic_set(&tpg->enabled, 0);
> + /* only physical sport should exist, free lio_sport
> + * allocated in efct_lio_make_sport
> + */
> + if (efc->domain) {
> + efc_sport_vport_del(efct->efcport, efc->domain,
> + lio_vport->npiv_wwpn,
> + lio_vport->npiv_wwnn);
> + return count;
> + }
> + } else {
> + return -EINVAL;
> + }
> + return count;
> +}
> +
> +static bool efct_lio_node_is_initiator(struct efc_node *node)
Why is the function name not on new line
> +{
> + if (!node)
> + return false;
> +
> + if (node->rnode.fc_id && node->rnode.fc_id != FC_FID_FLOGI &&
> + node->rnode.fc_id != FC_FID_DIR_SERV &&
> + node->rnode.fc_id != FC_FID_FCTRL) {
> + return true;
> + }
> +
> + return false;
> +}
> +
> +static int efct_lio_tgt_session_data(struct efct *efct, u64 wwpn,
> + char *buf, int size)
same here
> +{
> + struct efc_sli_port *sport = NULL;
> + struct efc_node *node = NULL;
> + struct efc *efc = efct->efcport;
> + u16 loop_id = 0;
> + int off = 0, rc = 0;
> +
> + if (!efc->domain) {
> + efc_log_err(efct, "failed to find efct/domain\n");
> + return EFC_FAIL;
> + }
> +
> + list_for_each_entry(sport, &efc->domain->sport_list, list_entry) {
> + if (sport->wwpn != wwpn)
> + continue;
> + list_for_each_entry(node, &sport->node_list,
> + list_entry) {
> + /* Dump only remote NPORT sessions */
> + if (!efct_lio_node_is_initiator(node))
> + continue;
> +
> + rc = snprintf(buf + off, size - off,
> + "0x%016llx,0x%08x,0x%04x\n",
> + get_unaligned_be64(node->wwpn),
> + node->rnode.fc_id, loop_id);
> + if (rc < 0)
> + break;
> + off += rc;
> + }
> + }
> +
> + return EFC_SUCCESS;
> +}
> +
> +static int efct_debugfs_session_open(struct inode *inode, struct file *filp)
function name on a new line as for the rest of the file or the way around.
Just a bit more constency please.
> +{
> + struct efct_lio_sport *sport = inode->i_private;
> + int size = 17 * PAGE_SIZE; /* 34 byte per session*2048 sessions */
34 * SZ_2K
> +
> + if (!(filp->f_mode & FMODE_READ)) {
> + filp->private_data = sport;
> + return EFC_SUCCESS;
> + }
> +
> + filp->private_data = kmalloc(size, GFP_KERNEL);
> + if (!filp->private_data)
> + return -ENOMEM;
> +
> + memset(filp->private_data, 0, size);
> + efct_lio_tgt_session_data(sport->efct, sport->wwpn, filp->private_data,
> + size);
> + return EFC_SUCCESS;
> +}
> +
> +static int efct_debugfs_session_close(struct inode *inode, struct file *filp)
> +{
> + if (filp->f_mode & FMODE_READ)
> + kfree(filp->private_data);
> +
> + return EFC_SUCCESS;
> +}
> +
> +static ssize_t efct_debugfs_session_read(struct file *filp, char __user *buf,
> + size_t count, loff_t *ppos)
> +{
> + if (!(filp->f_mode & FMODE_READ))
> + return -EPERM;
New line here as in function above, or no newline above. Whatever
the prefered style is (is true for the whole file).
> + return simple_read_from_buffer(buf, count, ppos, filp->private_data,
> + strlen(filp->private_data));
> +}
> +
> +static int efct_npiv_debugfs_session_open(struct inode *inode,
> + struct file *filp)
> +{
> + struct efct_lio_vport *sport = inode->i_private;
> + int size = 17 * PAGE_SIZE; /* 34 byte per session*2048 sessions */
34 * SZ_2K
> +
> + if (!(filp->f_mode & FMODE_READ)) {
> + filp->private_data = sport;
> + return EFC_SUCCESS;
> + }
> +
> + filp->private_data = kmalloc(size, GFP_KERNEL);
> + if (!filp->private_data)
> + return -ENOMEM;
> +
> + memset(filp->private_data, 0, size);
> + efct_lio_tgt_session_data(sport->efct, sport->npiv_wwpn,
> + filp->private_data, size);
> + return EFC_SUCCESS;
> +}
> +
> +static const struct file_operations efct_debugfs_session_fops = {
> + .owner = THIS_MODULE,
> + .open = efct_debugfs_session_open,
> + .release = efct_debugfs_session_close,
> + .read = efct_debugfs_session_read,
> + .llseek = default_llseek,
> +};
> +
> +static const struct file_operations efct_npiv_debugfs_session_fops = {
> + .owner = THIS_MODULE,
> + .open = efct_npiv_debugfs_session_open,
> + .release = efct_debugfs_session_close,
> + .read = efct_debugfs_session_read,
> + .llseek = default_llseek,
> +};
> +
> +static char *efct_lio_get_fabric_wwn(struct se_portal_group *se_tpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + return tpg->sport->wwpn_str;
> +}
> +
> +static char *efct_lio_get_npiv_fabric_wwn(struct se_portal_group *se_tpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + return tpg->vport->wwpn_str;
> +}
> +
> +static u16 efct_lio_get_tag(struct se_portal_group *se_tpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + return tpg->tpgt;
> +}
> +
> +static u16 efct_lio_get_npiv_tag(struct se_portal_group *se_tpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + return tpg->tpgt;
> +}
> +
> +static int efct_lio_check_demo_mode(struct se_portal_group *se_tpg)
> +{
> + return 1;
> +}
> +
> +static int efct_lio_check_demo_mode_cache(struct se_portal_group *se_tpg)
> +{
> + return 1;
> +}
> +
> +static int efct_lio_check_demo_write_protect(struct se_portal_group *se_tpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + return tpg->tpg_attrib.demo_mode_write_protect;
> +}
> +
> +static int
> +efct_lio_npiv_check_demo_write_protect(struct se_portal_group *se_tpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + return tpg->tpg_attrib.demo_mode_write_protect;
> +}
> +
> +static int efct_lio_check_prod_write_protect(struct se_portal_group *se_tpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + return tpg->tpg_attrib.prod_mode_write_protect;
> +}
> +
> +static int
> +efct_lio_npiv_check_prod_write_protect(struct se_portal_group *se_tpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + return tpg->tpg_attrib.prod_mode_write_protect;
> +}
> +
> +static u32 efct_lio_tpg_get_inst_index(struct se_portal_group *se_tpg)
> +{
> + return EFC_SUCCESS;
> +}
> +
> +static int efct_lio_check_stop_free(struct se_cmd *se_cmd)
> +{
> + struct efct_scsi_tgt_io *ocp = container_of(se_cmd,
> + struct efct_scsi_tgt_io,
> + cmd);
> + struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
> +
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_CHK_STOP_FREE);
> + return target_put_sess_cmd(se_cmd);
> +}
> +
> +static int
> +efct_lio_abort_tgt_cb(struct efct_io *io,
> + enum efct_scsi_io_status scsi_status,
> + u32 flags, void *arg)
> +{
> + efct_lio_io_printf(io, "%s\n", __func__);
> + return EFC_SUCCESS;
> +}
> +
> +/* command has been aborted, cleanup here */
> +static void efct_lio_aborted_task(struct se_cmd *se_cmd)
> +{
> + struct efct_scsi_tgt_io *ocp = container_of(se_cmd,
> + struct efct_scsi_tgt_io,
> + cmd);
struct efct_scsi_tgt_io *ocp =
container_of(se_cmd, struct efct_scsi_tgt_io, cmd);
IMO this is better to read
> + struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
> +
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_ABORTED_TASK);
> +
> + if (!(se_cmd->transport_state & CMD_T_ABORTED) || ocp->rsp_sent)
> + return;
> +
> + ocp->aborting = true;
> + ocp->err = EFCT_SCSI_STATUS_ABORTED;
> + /* terminate the exchange */
> + efct_scsi_tgt_abort_io(io, efct_lio_abort_tgt_cb, NULL);
> +}
> +
> +/* Called when se_cmd's ref count goes to 0 */
> +static void efct_lio_release_cmd(struct se_cmd *se_cmd)
> +{
> + struct efct_scsi_tgt_io *ocp = container_of(se_cmd,
> + struct efct_scsi_tgt_io,
> + cmd);
same as above
> + struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
> + struct efct *efct = io->efct;
> +
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_RELEASE_CMD);
> + efct_scsi_io_complete(io);
> + atomic_sub_return(1, &efct->tgt_efct.ios_in_use);
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_CMPL_CMD);
> +}
> +
> +static void efct_lio_close_session(struct se_session *se_sess)
> +{
> + struct efc_node *node = se_sess->fabric_sess_ptr;
> + struct efct *efct = NULL;
> + int rc;
> +
> + pr_debug("se_sess=%p node=%p", se_sess, node);
> +
> + if (!node) {
> + pr_debug("node is NULL");
> + return;
> + }
> +
> + efct = node->efc->base;
> + rc = efct_xport_control(efct->xport,
> + EFCT_XPORT_POST_NODE_EVENT, node,
> + EFCT_XPORT_SHUTDOWN, NULL);
aligment of "EFCT_XPORT_SHUTDOWN, NULL);" is a bit off
> + if (rc != 0) {
> + efc_log_test(efct,
> + "Failed to shutdown session %p node %p\n",
> + se_sess, node);
> + return;
> + }
> +}
> +
> +static u32 efct_lio_sess_get_index(struct se_session *se_sess)
> +{
> + return EFC_SUCCESS;
> +}
> +
> +static void efct_lio_set_default_node_attrs(struct se_node_acl *nacl)
> +{
> +}
> +
> +static int efct_lio_get_cmd_state(struct se_cmd *se_cmd)
> +{
> + return EFC_SUCCESS;
> +}
> +
> +static int
> +efct_lio_sg_map(struct efct_io *io)
> +{
> + struct efct_scsi_tgt_io *ocp = &io->tgt_io;
> + struct se_cmd *cmd = &ocp->cmd;
> +
> + ocp->seg_map_cnt = pci_map_sg(io->efct->pcidev, cmd->t_data_sg,
> + cmd->t_data_nents, cmd->data_direction);
> + if (ocp->seg_map_cnt == 0)
> + return -EFAULT;
> + return EFC_SUCCESS;
> +}
> +
> +static void
> +efct_lio_sg_unmap(struct efct_io *io)
> +{
> + struct efct_scsi_tgt_io *ocp = &io->tgt_io;
> + struct se_cmd *cmd = &ocp->cmd;
> +
> + if (WARN_ON(!ocp->seg_map_cnt || !cmd->t_data_sg))
> + return;
> +
> + pci_unmap_sg(io->efct->pcidev, cmd->t_data_sg,
> + ocp->seg_map_cnt, cmd->data_direction);
> + ocp->seg_map_cnt = 0;
> +}
> +
> +static int
> +efct_lio_status_done(struct efct_io *io,
> + enum efct_scsi_io_status scsi_status,
> + u32 flags, void *arg)
> +{
> + struct efct_scsi_tgt_io *ocp = &io->tgt_io;
> +
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_RSP_DONE);
> + if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
> + efct_lio_io_printf(io, "callback completed with error=%d\n",
> + scsi_status);
> + ocp->err = scsi_status;
> + }
> + if (ocp->seg_map_cnt)
> + efct_lio_sg_unmap(io);
> +
> + efct_lio_io_printf(io, "status=%d, err=%d flags=0x%x, dir=%d\n",
> + scsi_status, ocp->err, flags, ocp->ddir);
> +
> + transport_generic_free_cmd(&io->tgt_io.cmd, 0);
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
> + return EFC_SUCCESS;
> +}
> +
> +static int
> +efct_lio_datamove_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
> + u32 flags, void *arg);
> +
> +static int
> +efct_lio_write_pending(struct se_cmd *cmd)
> +{
> + struct efct_scsi_tgt_io *ocp = container_of(cmd,
> + struct efct_scsi_tgt_io,
> + cmd);
as mentioned above, the container_of could start a new line
> + struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
> + struct efct_scsi_sgl *sgl = io->sgl;
> + struct scatterlist *sg;
> + u32 flags = 0, cnt, curcnt;
> + u64 length = 0;
> + int rc = 0;
> +
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_WRITE_PENDING);
> + efct_lio_io_printf(io, "trans_state=0x%x se_cmd_flags=0x%x\n",
> + cmd->transport_state, cmd->se_cmd_flags);
> +
> + if (ocp->seg_cnt == 0) {
> + ocp->seg_cnt = cmd->t_data_nents;
> + ocp->cur_seg = 0;
> + if (efct_lio_sg_map(io)) {
> + efct_lio_io_printf(io, "efct_lio_sg_map failed\n");
> + return -EFAULT;
> + }
> + }
> + curcnt = (ocp->seg_map_cnt - ocp->cur_seg);
> + curcnt = (curcnt < io->sgl_allocated) ? curcnt : io->sgl_allocated;
> + /* find current sg */
> + for (cnt = 0, sg = cmd->t_data_sg; cnt < ocp->cur_seg; cnt++,
> + sg = sg_next(sg))
> + ;
Please add a comment here, a single ';' is bit 'thin', e.g.
for (...)
/* do nothing */ ;
> +
> + for (cnt = 0; cnt < curcnt; cnt++, sg = sg_next(sg)) {
> + sgl[cnt].addr = sg_dma_address(sg);
> + sgl[cnt].dif_addr = 0;
> + sgl[cnt].len = sg_dma_len(sg);
> + length += sgl[cnt].len;
> + ocp->cur_seg++;
> + }
> + if (ocp->cur_seg == ocp->seg_cnt)
> + flags = EFCT_SCSI_LAST_DATAPHASE;
> + rc = efct_scsi_recv_wr_data(io, flags, sgl, curcnt, length,
> + efct_lio_datamove_done, NULL);
> + return rc;
> +}
> +
> +static int
> +efct_lio_queue_data_in(struct se_cmd *cmd)
> +{
> + struct efct_scsi_tgt_io *ocp = container_of(cmd,
> + struct efct_scsi_tgt_io,
> + cmd);
> + struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
> + struct efct_scsi_sgl *sgl = io->sgl;
> + struct scatterlist *sg = NULL;
> + uint flags = 0, cnt = 0, curcnt = 0;
> + u64 length = 0;
> + int rc = 0;
> +
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_QUEUE_DATA_IN);
> +
> + if (ocp->seg_cnt == 0) {
> + if (cmd->data_length) {
> + ocp->seg_cnt = cmd->t_data_nents;
> + ocp->cur_seg = 0;
> + if (efct_lio_sg_map(io)) {
> + efct_lio_io_printf(io,
> + "efct_lio_sg_map failed\n");
> + return -EAGAIN;
> + }
> + } else {
> + /* If command length is 0, send the response status */
> + struct efct_scsi_cmd_resp rsp;
> +
> + memset(&rsp, 0, sizeof(rsp));
> + efct_lio_io_printf(io,
> + "cmd : %p length 0, send status\n",
> + cmd);
> + return efct_scsi_send_resp(io, 0, &rsp,
> + efct_lio_status_done, NULL);
> + }
> + }
> + curcnt = min(ocp->seg_map_cnt - ocp->cur_seg, io->sgl_allocated);
> +
> + while (cnt < curcnt) {
> + sg = &cmd->t_data_sg[ocp->cur_seg];
> + sgl[cnt].addr = sg_dma_address(sg);
> + sgl[cnt].dif_addr = 0;
> + if (ocp->transferred_len + sg_dma_len(sg) >= cmd->data_length)
> + sgl[cnt].len = cmd->data_length - ocp->transferred_len;
> + else
> + sgl[cnt].len = sg_dma_len(sg);
> +
> + ocp->transferred_len += sgl[cnt].len;
> + length += sgl[cnt].len;
> + ocp->cur_seg++;
> + cnt++;
> + if (ocp->transferred_len == cmd->data_length)
> + break;
> + }
> +
> + if (ocp->transferred_len == cmd->data_length) {
> + flags = EFCT_SCSI_LAST_DATAPHASE;
> + ocp->seg_cnt = ocp->cur_seg;
> + }
> +
> + /* If there is residual, disable Auto Good Response */
> + if (cmd->residual_count)
> + flags |= EFCT_SCSI_NO_AUTO_RESPONSE;
> +
> + rc = efct_scsi_send_rd_data(io, flags, sgl, curcnt, length,
> + efct_lio_datamove_done, NULL);
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RD_DATA);
> + return rc;
> +}
> +
> +static int
> +efct_lio_datamove_done(struct efct_io *io,
> + enum efct_scsi_io_status scsi_status,
> + u32 flags, void *arg)
> +{
> + struct efct_scsi_tgt_io *ocp = &io->tgt_io;
> + struct se_cmd *cmd = &io->tgt_io.cmd;
> + int rc;
> +
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_DATA_DONE);
> + if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
> + efct_lio_io_printf(io, "callback completed with error=%d\n",
> + scsi_status);
> + ocp->err = scsi_status;
> + }
> + efct_lio_io_printf(io, "seg_map_cnt=%d\n", ocp->seg_map_cnt);
> + if (ocp->seg_map_cnt) {
> + if (ocp->err == EFCT_SCSI_STATUS_GOOD &&
> + ocp->cur_seg < ocp->seg_cnt) {
> + efct_lio_io_printf(io, "continuing cmd at segm=%d\n",
> + ocp->cur_seg);
> + if (ocp->ddir == DMA_TO_DEVICE)
> + rc = efct_lio_write_pending(&ocp->cmd);
> + else
> + rc = efct_lio_queue_data_in(&ocp->cmd);
> + if (rc == 0)
> + return EFC_SUCCESS;
> + ocp->err = EFCT_SCSI_STATUS_ERROR;
> + efct_lio_io_printf(io, "could not continue command\n");
> + }
> + efct_lio_sg_unmap(io);
> + }
> +
> + if (io->tgt_io.aborting) {
> + efct_lio_io_printf(io, "IO done aborted\n");
> + return EFC_SUCCESS;
> + }
> +
> + if (ocp->ddir == DMA_TO_DEVICE) {
> + efct_lio_io_printf(io, "Write done, trans_state=0x%x\n",
> + io->tgt_io.cmd.transport_state);
> + if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
> + transport_generic_request_failure(&io->tgt_io.cmd,
> + TCM_CHECK_CONDITION_ABORT_CMD);
> + efct_set_lio_io_state(io,
> + EFCT_LIO_STATE_TGT_GENERIC_REQ_FAILURE);
> + } else {
> + efct_set_lio_io_state(io,
> + EFCT_LIO_STATE_TGT_EXECUTE_CMD);
> + target_execute_cmd(&io->tgt_io.cmd);
> + }
> + } else {
> + if ((flags & EFCT_SCSI_IO_CMPL_RSP_SENT) == 0) {
> + struct efct_scsi_cmd_resp rsp;
> + /* send check condition if an error occurred */
> + memset(&rsp, 0, sizeof(rsp));
> + rsp.scsi_status = cmd->scsi_status;
> + rsp.sense_data = (uint8_t *)io->tgt_io.sense_buffer;
> + rsp.sense_data_length = cmd->scsi_sense_length;
> +
> + /* Check for residual underrun or overrun */
> + if (cmd->se_cmd_flags & SCF_OVERFLOW_BIT)
> + rsp.residual = -cmd->residual_count;
> + else if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT)
> + rsp.residual = cmd->residual_count;
> +
> + rc = efct_scsi_send_resp(io, 0, &rsp,
> + efct_lio_status_done, NULL);
> + efct_set_lio_io_state(io,
> + EFCT_LIO_STATE_SCSI_SEND_RSP);
> + if (rc != 0) {
> + efct_lio_io_printf(io,
> + "Read done, failed to send rsp, rc=%d\n",
> + rc);
> + transport_generic_free_cmd(&io->tgt_io.cmd, 0);
> + efct_set_lio_io_state(io,
> + EFCT_LIO_STATE_TGT_GENERIC_FREE);
> + } else {
> + ocp->rsp_sent = true;
> + }
> + } else {
> + ocp->rsp_sent = true;
> + transport_generic_free_cmd(&io->tgt_io.cmd, 0);
> + efct_set_lio_io_state(io,
> + EFCT_LIO_STATE_TGT_GENERIC_FREE);
> + }
> + }
> + return EFC_SUCCESS;
> +}
The last two function could be splitted up into few smaller function
which would help the readablilty quite a bit IMO. Many of these
function tend to be right side heavy which leads to many funky new
lines.
> +
> +static int
> +efct_lio_tmf_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
> + u32 flags, void *arg)
> +{
> + efct_lio_tmfio_printf(io, "cmd=%p status=%d, flags=0x%x\n",
> + &io->tgt_io.cmd, scsi_status, flags);
> +
> + transport_generic_free_cmd(&io->tgt_io.cmd, 0);
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
> + return EFC_SUCCESS;
> +}
> +
> +static int
> +efct_lio_null_tmf_done(struct efct_io *tmfio,
> + enum efct_scsi_io_status scsi_status,
> + u32 flags, void *arg)
> +{
> + efct_lio_tmfio_printf(tmfio, "cmd=%p status=%d, flags=0x%x\n",
> + &tmfio->tgt_io.cmd, scsi_status, flags);
> +
> + /* free struct efct_io only, no active se_cmd */
> + efct_scsi_io_complete(tmfio);
> + return EFC_SUCCESS;
> +}
> +
> +static int
> +efct_lio_queue_status(struct se_cmd *cmd)
> +{
> + struct efct_scsi_cmd_resp rsp;
> + struct efct_scsi_tgt_io *ocp = container_of(cmd,
> + struct efct_scsi_tgt_io,
> + cmd);
> + struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
> + int rc = 0;
> +
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_QUEUE_STATUS);
> + efct_lio_io_printf(io,
> + "status=0x%x trans_state=0x%x se_cmd_flags=0x%x sns_len=%d\n",
> + cmd->scsi_status, cmd->transport_state, cmd->se_cmd_flags,
> + cmd->scsi_sense_length);
> +
> + memset(&rsp, 0, sizeof(rsp));
> + rsp.scsi_status = cmd->scsi_status;
> + rsp.sense_data = (u8 *)io->tgt_io.sense_buffer;
> + rsp.sense_data_length = cmd->scsi_sense_length;
> +
> + /* Check for residual underrun or overrun, mark negitive value for
> + * underrun to recognize in HW
> + */
> + if (cmd->se_cmd_flags & SCF_OVERFLOW_BIT)
> + rsp.residual = -cmd->residual_count;
> + else if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT)
> + rsp.residual = cmd->residual_count;
> +
> + rc = efct_scsi_send_resp(io, 0, &rsp, efct_lio_status_done, NULL);
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RSP);
> + if (rc == 0)
> + ocp->rsp_sent = true;
> + return rc;
> +}
> +
> +static void efct_lio_queue_tm_rsp(struct se_cmd *cmd)
> +{
> + struct efct_scsi_tgt_io *ocp = container_of(cmd,
> + struct efct_scsi_tgt_io,
> + cmd);
> + struct efct_io *tmfio = container_of(ocp, struct efct_io, tgt_io);
> + struct se_tmr_req *se_tmr = cmd->se_tmr_req;
> + u8 rspcode;
> +
> + efct_lio_tmfio_printf(tmfio, "cmd=%p function=0x%x tmr->response=%d\n",
> + cmd, se_tmr->function, se_tmr->response);
> + switch (se_tmr->response) {
> + case TMR_FUNCTION_COMPLETE:
> + rspcode = EFCT_SCSI_TMF_FUNCTION_COMPLETE;
> + break;
> + case TMR_TASK_DOES_NOT_EXIST:
> + rspcode = EFCT_SCSI_TMF_FUNCTION_IO_NOT_FOUND;
> + break;
> + case TMR_LUN_DOES_NOT_EXIST:
> + rspcode = EFCT_SCSI_TMF_INCORRECT_LOGICAL_UNIT_NUMBER;
> + break;
> + case TMR_FUNCTION_REJECTED:
> + default:
> + rspcode = EFCT_SCSI_TMF_FUNCTION_REJECTED;
> + break;
> + }
> + efct_scsi_send_tmf_resp(tmfio, rspcode, NULL, efct_lio_tmf_done, NULL);
> +}
> +
> +static struct efct *efct_find_wwpn(u64 wwpn)
> +{
> + int efctidx;
> + struct efct *efct;
> +
> + /* Search for the HBA that has this WWPN */
> + for (efctidx = 0; efctidx < MAX_EFCT_DEVICES; efctidx++) {
> +
> + efct = efct_devices[efctidx];
> + if (!efct)
> + continue;
> +
> + if (wwpn == efct_get_wwpn(&efct->hw))
> + break;
> + }
> +
> + if (efctidx == MAX_EFCT_DEVICES)
> + return NULL;
> +
> + return efct_devices[efctidx];
> +}
> +
> +static struct dentry *
> +efct_create_dfs_session(struct efct *efct, void *data, u8 npiv)
> +{
> + char name[16];
> +
> + if (!efct->sess_debugfs_dir)
> + return NULL;
> +
> + if (!npiv) {
> + snprintf(name, sizeof(name), "efct-sessions-%d",
> + efct->instance_index);
> + return debugfs_create_file(name, 0644, efct->sess_debugfs_dir,
> + data, &efct_debugfs_session_fops);
> + }
> +
> + snprintf(name, sizeof(name), "sessions-npiv-%d", efct->instance_index);
> +
> + return debugfs_create_file(name, 0644, efct->sess_debugfs_dir, data,
> + &efct_npiv_debugfs_session_fops);
> +}
> +
> +static struct se_wwn *
> +efct_lio_make_sport(struct target_fabric_configfs *tf,
> + struct config_group *group, const char *name)
> +{
> + struct efct_lio_sport *lio_sport;
> + struct efct *efct;
> + int ret;
> + u64 wwpn;
> +
> + ret = efct_lio_parse_wwn(name, &wwpn, 0);
> + if (ret)
> + return ERR_PTR(ret);
> +
> + efct = efct_find_wwpn(wwpn);
> + if (!efct) {
> + pr_err("cannot find EFCT for base wwpn %s\n", name);
> + return ERR_PTR(-ENXIO);
> + }
> +
> + lio_sport = kzalloc(sizeof(*lio_sport), GFP_KERNEL);
> + if (!lio_sport)
> + return ERR_PTR(-ENOMEM);
> +
> + lio_sport->efct = efct;
> + lio_sport->wwpn = wwpn;
> + efct_format_wwn(lio_sport->wwpn_str, sizeof(lio_sport->wwpn_str),
> + "naa.", wwpn);
> + efct->tgt_efct.lio_sport = lio_sport;
> +
> + lio_sport->sessions = efct_create_dfs_session(efct, lio_sport, 0);
> + return &lio_sport->sport_wwn;
> +}
> +
> +static struct se_wwn *
> +efct_lio_npiv_make_sport(struct target_fabric_configfs *tf,
> + struct config_group *group, const char *name)
> +{
> + struct efct_lio_vport *lio_vport;
> + struct efct *efct;
> + int ret = -1;
> + u64 p_wwpn, npiv_wwpn, npiv_wwnn;
> + char *p, *pbuf, tmp[128];
> + struct efct_lio_vport_list_t *vport_list;
> + struct fc_vport *new_fc_vport;
> + struct fc_vport_identifiers vport_id;
> + unsigned long flags = 0;
> +
> + snprintf(tmp, sizeof(tmp), "%s", name);
> + pbuf = &tmp[0];
> +
> + p = strsep(&pbuf, "@");
> +
> + if (!p || !pbuf) {
> + pr_err("Unable to find separator operator(@)\n");
> + return ERR_PTR(-EINVAL);
> + }
> +
> + ret = efct_lio_parse_wwn(p, &p_wwpn, 0);
> + if (ret)
> + return ERR_PTR(ret);
> +
> + ret = efct_lio_parse_npiv_wwn(pbuf, strlen(pbuf), &npiv_wwpn,
> + &npiv_wwnn);
> + if (ret)
> + return ERR_PTR(ret);
> +
> + efct = efct_find_wwpn(p_wwpn);
> + if (!efct) {
> + pr_err("cannot find EFCT for base wwpn %s\n", name);
> + return ERR_PTR(-ENXIO);
> + }
> +
> + lio_vport = kzalloc(sizeof(*lio_vport), GFP_KERNEL);
> + if (!lio_vport)
> + return ERR_PTR(-ENOMEM);
> +
> + lio_vport->efct = efct;
> + lio_vport->wwpn = p_wwpn;
> + lio_vport->npiv_wwpn = npiv_wwpn;
> + lio_vport->npiv_wwnn = npiv_wwnn;
> +
> + efct_format_wwn(lio_vport->wwpn_str, sizeof(lio_vport->wwpn_str),
> + "naa.", npiv_wwpn);
> +
> + vport_list = kmalloc(sizeof(*vport_list), GFP_KERNEL);
> + if (!vport_list) {
> + kfree(lio_vport);
> + return ERR_PTR(-ENOMEM);
> + }
> +
> + memset(vport_list, 0, sizeof(struct efct_lio_vport_list_t));
> + vport_list->lio_vport = lio_vport;
> + spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
> + INIT_LIST_HEAD(&vport_list->list_entry);
> + list_add_tail(&vport_list->list_entry, &efct->tgt_efct.vport_list);
> + spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
> +
> + lio_vport->sessions = efct_create_dfs_session(efct, lio_vport, 1);
> +
> + memset(&vport_id, 0, sizeof(vport_id));
> + vport_id.port_name = npiv_wwpn;
> + vport_id.node_name = npiv_wwnn;
> + vport_id.roles = FC_PORT_ROLE_FCP_INITIATOR;
> + vport_id.vport_type = FC_PORTTYPE_NPIV;
> + vport_id.disable = false;
> +
> + new_fc_vport = fc_vport_create(efct->shost, 0, &vport_id);
> + if (!new_fc_vport) {
> + efc_log_err(efct, "fc_vport_create failed\n");
> + kfree(lio_vport);
> + kfree(vport_list);
> + return ERR_PTR(-ENOMEM);
> + }
> +
> + lio_vport->fc_vport = new_fc_vport;
> +
> + return &lio_vport->vport_wwn;
> +}
> +
> +static void
> +efct_lio_drop_sport(struct se_wwn *wwn)
> +{
> + struct efct_lio_sport *lio_sport = container_of(wwn,
> + struct efct_lio_sport, sport_wwn);
> + struct efct *efct = lio_sport->efct;
> +
> + /* only physical sport should exist, free lio_sport allocated
> + * in efct_lio_make_sport.
> + */
> +
> + debugfs_remove(lio_sport->sessions);
> + lio_sport->sessions = NULL;
> +
> + kfree(efct->tgt_efct.lio_sport);
> + efct->tgt_efct.lio_sport = NULL;
> +}
> +
> +static void
> +efct_lio_npiv_drop_sport(struct se_wwn *wwn)
> +{
> + struct efct_lio_vport *lio_vport = container_of(wwn,
> + struct efct_lio_vport, vport_wwn);
> + struct efct_lio_vport_list_t *vport, *next_vport;
> + struct efct *efct = lio_vport->efct;
> + unsigned long flags = 0;
> +
> + spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
> +
> + debugfs_remove(lio_vport->sessions);
> +
> + if (lio_vport->fc_vport)
> + fc_vport_terminate(lio_vport->fc_vport);
> +
> + lio_vport->sessions = NULL;
> +
> + list_for_each_entry_safe(vport, next_vport, &efct->tgt_efct.vport_list,
> + list_entry) {
> + if (vport->lio_vport == lio_vport) {
> + list_del(&vport->list_entry);
> + kfree(vport->lio_vport);
> + kfree(vport);
> + break;
> + }
> + }
> + spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
> +}
> +
> +static struct se_portal_group *
> +efct_lio_make_tpg(struct se_wwn *wwn, const char *name)
> +{
> + struct efct_lio_sport *lio_sport = container_of(wwn,
> + struct efct_lio_sport, sport_wwn);
> + struct efct_lio_tpg *tpg;
> + struct efct *efct;
> + unsigned long n;
> + int ret;
> +
> + if (strstr(name, "tpgt_") != name)
> + return ERR_PTR(-EINVAL);
> + if (kstrtoul(name + 5, 10, &n) || n > USHRT_MAX)
> + return ERR_PTR(-EINVAL);
> +
> + tpg = kzalloc(sizeof(*tpg), GFP_KERNEL);
> + if (!tpg)
> + return ERR_PTR(-ENOMEM);
> +
> + tpg->sport = lio_sport;
> + tpg->tpgt = n;
> + atomic_set(&tpg->enabled, 0);
> +
> + tpg->tpg_attrib.generate_node_acls = 1;
> + tpg->tpg_attrib.demo_mode_write_protect = 1;
> + tpg->tpg_attrib.cache_dynamic_acls = 1;
> + tpg->tpg_attrib.demo_mode_login_only = 1;
> + tpg->tpg_attrib.session_deletion_wait = 1;
> +
> + ret = core_tpg_register(wwn, &tpg->tpg, SCSI_PROTOCOL_FCP);
> + if (ret < 0) {
> + kfree(tpg);
> + return NULL;
> + }
> + efct = lio_sport->efct;
> + efct->tgt_efct.tpg = tpg;
> + efc_log_debug(efct, "create portal group %d\n", tpg->tpgt);
> +
> + return &tpg->tpg;
> +}
> +
> +static void
> +efct_lio_drop_tpg(struct se_portal_group *se_tpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + efc_log_debug(tpg->sport->efct, "drop portal group %d\n", tpg->tpgt);
> + tpg->sport->efct->tgt_efct.tpg = NULL;
> + core_tpg_deregister(se_tpg);
> + kfree(tpg);
> +}
> +
> +static struct se_portal_group *
> +efct_lio_npiv_make_tpg(struct se_wwn *wwn, const char *name)
> +{
> + struct efct_lio_vport *lio_vport = container_of(wwn,
> + struct efct_lio_vport, vport_wwn);
> + struct efct_lio_tpg *tpg;
> + struct efct *efct;
> + unsigned long n;
> + int ret;
> +
> + efct = lio_vport->efct;
> + if (strstr(name, "tpgt_") != name)
> + return ERR_PTR(-EINVAL);
> + if (kstrtoul(name + 5, 10, &n) || n > USHRT_MAX)
> + return ERR_PTR(-EINVAL);
> +
> + if (n != 1) {
> + efc_log_err(efct, "Invalid tpgt index: %ld provided\n", n);
> + return ERR_PTR(-EINVAL);
> + }
> +
> + tpg = kzalloc(sizeof(*tpg), GFP_KERNEL);
> + if (!tpg)
> + return ERR_PTR(-ENOMEM);
> +
> + tpg->vport = lio_vport;
> + tpg->tpgt = n;
> + atomic_set(&tpg->enabled, 0);
> +
> + tpg->tpg_attrib.generate_node_acls = 1;
> + tpg->tpg_attrib.demo_mode_write_protect = 1;
> + tpg->tpg_attrib.cache_dynamic_acls = 1;
> + tpg->tpg_attrib.demo_mode_login_only = 1;
> + tpg->tpg_attrib.session_deletion_wait = 1;
> +
> + ret = core_tpg_register(wwn, &tpg->tpg, SCSI_PROTOCOL_FCP);
> +
> + if (ret < 0) {
> + kfree(tpg);
> + return NULL;
> + }
> + lio_vport->tpg = tpg;
> + efc_log_debug(efct, "create vport portal group %d\n", tpg->tpgt);
> +
> + return &tpg->tpg;
> +}
> +
> +static void
> +efct_lio_npiv_drop_tpg(struct se_portal_group *se_tpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(se_tpg,
> + struct efct_lio_tpg, tpg);
> +
> + efc_log_debug(tpg->vport->efct, "drop npiv portal group %d\n",
> + tpg->tpgt);
> + core_tpg_deregister(se_tpg);
> + kfree(tpg);
> +}
> +
> +static int
> +efct_lio_init_nodeacl(struct se_node_acl *se_nacl, const char *name)
> +{
> + struct efct_lio_nacl *nacl;
> + u64 wwnn;
> +
> + if (efct_lio_parse_wwn(name, &wwnn, 0) < 0)
> + return -EINVAL;
> +
> + nacl = container_of(se_nacl, struct efct_lio_nacl, se_node_acl);
> + nacl->nport_wwnn = wwnn;
> +
> + efct_format_wwn(nacl->nport_name, sizeof(nacl->nport_name), "", wwnn);
> + return EFC_SUCCESS;
> +}
> +
> +static int efct_lio_check_demo_mode_login_only(struct se_portal_group *stpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(stpg, struct efct_lio_tpg, tpg);
> +
> + return tpg->tpg_attrib.demo_mode_login_only;
> +}
> +
> +static int
> +efct_lio_npiv_check_demo_mode_login_only(struct se_portal_group *stpg)
> +{
> + struct efct_lio_tpg *tpg = container_of(stpg, struct efct_lio_tpg, tpg);
> +
> + return tpg->tpg_attrib.demo_mode_login_only;
> +}
> +
> +static struct efct_lio_tpg *
> +efct_get_vport_tpg(struct efc_node *node)
> +{
> + struct efct *efct;
> + u64 wwpn = node->sport->wwpn;
> + struct efct_lio_vport_list_t *vport, *next;
> + struct efct_lio_vport *lio_vport = NULL;
> + struct efct_lio_tpg *tpg = NULL;
> + unsigned long flags = 0;
> +
> + efct = node->efc->base;
> + spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
> + list_for_each_entry_safe(vport, next, &efct->tgt_efct.vport_list,
> + list_entry) {
> + lio_vport = vport->lio_vport;
> + if (wwpn && lio_vport &&
> + lio_vport->npiv_wwpn == wwpn) {
this fits on one line
> + efc_log_test(efct, "found tpg on vport\n");
> + tpg = lio_vport->tpg;
> + break;
> + }
> + }
> + spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
> + return tpg;
> +}
> +
> +static int efct_session_cb(struct se_portal_group *se_tpg,
> + struct se_session *se_sess, void *private)
> +{
> + struct efc_node *node = private;
> + struct efct_scsi_tgt_node *tgt_node = NULL;
> +
> + tgt_node = kzalloc(sizeof(*tgt_node), GFP_KERNEL);
> + if (!tgt_node)
> + return EFC_FAIL;
> +
> + tgt_node->session = se_sess;
> + node->tgt_node = tgt_node;
> +
> + return EFC_SUCCESS;
> +}
> +
> +int efct_scsi_tgt_new_device(struct efct *efct)
> +{
> + int rc = 0;
> + u32 total_ios;
> +
> + /* Get the max settings */
> + efct->tgt_efct.max_sge = sli_get_max_sge(&efct->hw.sli);
> + efct->tgt_efct.max_sgl = sli_get_max_sgl(&efct->hw.sli);
> +
> + /* initialize IO watermark fields */
> + atomic_set(&efct->tgt_efct.ios_in_use, 0);
> + total_ios = efct->hw.config.n_io;
> + efc_log_debug(efct, "total_ios=%d\n", total_ios);
> + efct->tgt_efct.watermark_min =
> + (total_ios * EFCT_WATERMARK_LOW_PCT) / 100;
> + efct->tgt_efct.watermark_max =
> + (total_ios * EFCT_WATERMARK_HIGH_PCT) / 100;
> + atomic_set(&efct->tgt_efct.io_high_watermark,
> + efct->tgt_efct.watermark_max);
> + atomic_set(&efct->tgt_efct.watermark_hit, 0);
> + atomic_set(&efct->tgt_efct.initiator_count, 0);
> +
> + lio_wq = create_singlethread_workqueue("efct_lio_worker");
> + if (!lio_wq) {
> + efc_log_err(efct, "workqueue create failed\n");
> + return -ENOMEM;
> + }
> +
> + spin_lock_init(&efct->tgt_efct.efct_lio_lock);
> + INIT_LIST_HEAD(&efct->tgt_efct.vport_list);
> +
> + return rc;
> +}
> +
> +int efct_scsi_tgt_del_device(struct efct *efct)
> +{
> + flush_workqueue(lio_wq);
> +
> + return EFC_SUCCESS;
> +}
> +
> +int
> +efct_scsi_tgt_new_domain(struct efc *efc, struct efc_domain *domain)
> +{
> + return 0;
> +}
> +
> +void
> +efct_scsi_tgt_del_domain(struct efc *efc, struct efc_domain *domain)
> +{
> +}
> +
> +/* Called by the libefc when new sli port (sport) is discovered */
> +int
> +efct_scsi_tgt_new_sport(struct efc *efc, struct efc_sli_port *sport)
> +{
> + struct efct *efct = sport->efc->base;
> +
> + efc_log_debug(efct, "New SPORT: %s bound to %s\n", sport->display_name,
> + efct->tgt_efct.lio_sport->wwpn_str);
> +
> + return EFC_SUCCESS;
> +}
> +
> +/* Called by the libefc when a sport goes away. */
> +void
> +efct_scsi_tgt_del_sport(struct efc *efc, struct efc_sli_port *sport)
> +{
> + efc_log_debug(efc, "Del SPORT: %s\n",
> + sport->display_name);
fits on one line
> +}
> +/* Called by libefc to validate node. */
> +int
> +efct_scsi_validate_initiator(struct efc *efc, struct efc_node *node)
> +{
> + return 1;
> +}
> +
> +static void efct_lio_setup_session(struct work_struct *work)
> +{
> + struct efct_lio_wq_data *wq_data = container_of(work,
> + struct efct_lio_wq_data, work);
> + struct efct *efct = wq_data->efct;
> + struct efc_node *node = wq_data->ptr;
> + char wwpn[WWN_NAME_LEN];
> + struct efct_lio_tpg *tpg = NULL;
> + struct se_portal_group *se_tpg;
> + struct se_session *se_sess;
> + int watermark;
> + int initiator_count;
> +
> + /* Check to see if it's belongs to vport,
> + * if not get physical port
> + */
> + tpg = efct_get_vport_tpg(node);
> + if (tpg) {
> + se_tpg = &tpg->tpg;
> + } else if (efct->tgt_efct.tpg) {
> + tpg = efct->tgt_efct.tpg;
> + se_tpg = &tpg->tpg;
> + } else {
> + efc_log_err(efct, "failed to init session\n");
> + return;
> + }
> +
> + /*
> + * Format the FCP Initiator port_name into colon
> + * separated values to match the format by our explicit
> + * ConfigFS NodeACLs.
> + */
> + efct_format_wwn(wwpn, sizeof(wwpn), "",
> + efc_node_get_wwpn(node));
fits on one line
> +
> + se_sess = target_setup_session(se_tpg, 0, 0,
> + TARGET_PROT_NORMAL,
> + wwpn, node,
> + efct_session_cb);
> + if (IS_ERR(se_sess)) {
> + efc_log_err(efct, "failed to setup session\n");
> + return;
> + }
> +
> + efc_log_debug(efct, "new initiator se_sess=%p node=%p\n",
> + se_sess, node);
> +
> + /* update IO watermark: increment initiator count */
> + initiator_count =
> + atomic_add_return(1, &efct->tgt_efct.initiator_count);
> + watermark = (efct->tgt_efct.watermark_max -
> + initiator_count * EFCT_IO_WATERMARK_PER_INITIATOR);
> + watermark = (efct->tgt_efct.watermark_min > watermark) ?
> + efct->tgt_efct.watermark_min : watermark;
to many brackets in the last four lines.
> + atomic_set(&efct->tgt_efct.io_high_watermark,
> + watermark);
> +
> + kfree(wq_data);
> +}
> +
> +/* Called by the libefc when new a new remote initiator is discovered */
> +int efct_scsi_new_initiator(struct efc *efc, struct efc_node *node)
> +{
> + struct efct *efct = node->efc->base;
> + struct efct_lio_wq_data *wq_data;
> +
> + /*
> + * Since LIO only supports initiator validation at thread level,
> + * we are open minded and accept all callers.
> + */
> + wq_data = kzalloc(sizeof(*wq_data), GFP_ATOMIC);
> + if (!wq_data)
> + return -ENOMEM;
> +
> + wq_data->ptr = node;
> + wq_data->efct = efct;
> + INIT_WORK(&wq_data->work, efct_lio_setup_session);
> + queue_work(lio_wq, &wq_data->work);
> + return EFC_SUCCESS;
> +}
> +
> +static void efct_lio_remove_session(struct work_struct *work)
> +{
> + struct efct_lio_wq_data *wq_data = container_of(work,
> + struct efct_lio_wq_data, work);
> + struct efct *efct = wq_data->efct;
> + struct efc_node *node = wq_data->ptr;
> + struct efct_scsi_tgt_node *tgt_node = NULL;
> + struct se_session *se_sess;
> +
> + tgt_node = node->tgt_node;
> + se_sess = tgt_node->session;
> +
> + if (!se_sess) {
> + /* base driver has sent back-to-back requests
> + * to unreg session with no intervening
> + * register
> + */
> + efc_log_test(efct,
> + "unreg session for NULL session\n");
> + efc_scsi_del_initiator_complete(node->efc,
> + node);
> + return;
> + }
> +
> + efc_log_debug(efct, "unreg session se_sess=%p node=%p\n",
> + se_sess, node);
> +
> + /* first flag all session commands to complete */
> + target_sess_cmd_list_set_waiting(se_sess);
> +
> + /* now wait for session commands to complete */
> + target_wait_for_sess_cmds(se_sess);
> + target_remove_session(se_sess);
> +
> + kfree(node->tgt_node);
> +
> + node->tgt_node = NULL;
> + efc_scsi_del_initiator_complete(node->efc, node);
> +
> + kfree(wq_data);
> +}
> +
> +/* Called by the libefc when an initiator goes away. */
> +int efct_scsi_del_initiator(struct efc *efc, struct efc_node *node,
> + int reason)
> +{
> + struct efct *efct = node->efc->base;
> + struct efct_lio_wq_data *wq_data;
> + int watermark;
> + int initiator_count;
> +
> + if (reason == EFCT_SCSI_INITIATOR_MISSING)
> + return EFCT_SCSI_CALL_COMPLETE;
> +
> + wq_data = kmalloc(sizeof(*wq_data), GFP_ATOMIC);
> + if (!wq_data)
> + return EFCT_SCSI_CALL_COMPLETE;
> +
> + memset(wq_data, 0, sizeof(*wq_data));
> + wq_data->ptr = node;
> + wq_data->efct = efct;
> + INIT_WORK(&wq_data->work, efct_lio_remove_session);
> + queue_work(lio_wq, &wq_data->work);
> +
> + /*
> + * update IO watermark: decrement initiator count
> + */
> + initiator_count =
> + atomic_sub_return(1, &efct->tgt_efct.initiator_count);
> + watermark = (efct->tgt_efct.watermark_max -
> + initiator_count * EFCT_IO_WATERMARK_PER_INITIATOR);
> + watermark = (efct->tgt_efct.watermark_min > watermark) ?
> + efct->tgt_efct.watermark_min : watermark;
too many brackets in the last four lines.
> + atomic_set(&efct->tgt_efct.io_high_watermark, watermark);
> +
> + return EFCT_SCSI_CALL_ASYNC;
> +}
> +
> +int efct_scsi_recv_cmd(struct efct_io *io, uint64_t lun, u8 *cdb,
> + u32 cdb_len, u32 flags)
> +{
> + struct efct_scsi_tgt_io *ocp = &io->tgt_io;
> + struct efct *efct = io->efct;
> + char *ddir;
> + struct efct_scsi_tgt_node *tgt_node = NULL;
> + struct se_session *se_sess;
> + int rc = 0;
> +
> + memset(ocp, 0, sizeof(struct efct_scsi_tgt_io));
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_RECV_CMD);
> + atomic_add_return(1, &efct->tgt_efct.ios_in_use);
> +
> + /* set target timeout */
> + io->timeout = efct->target_io_timer_sec;
> +
> + if (flags & EFCT_SCSI_CMD_SIMPLE)
> + ocp->task_attr = TCM_SIMPLE_TAG;
> + else if (flags & EFCT_SCSI_CMD_HEAD_OF_QUEUE)
> + ocp->task_attr = TCM_HEAD_TAG;
> + else if (flags & EFCT_SCSI_CMD_ORDERED)
> + ocp->task_attr = TCM_ORDERED_TAG;
> + else if (flags & EFCT_SCSI_CMD_ACA)
> + ocp->task_attr = TCM_ACA_TAG;
> +
> + switch (flags & (EFCT_SCSI_CMD_DIR_IN | EFCT_SCSI_CMD_DIR_OUT)) {
> + case EFCT_SCSI_CMD_DIR_IN:
> + ddir = "FROM_INITIATOR";
> + ocp->ddir = DMA_TO_DEVICE;
> + break;
> + case EFCT_SCSI_CMD_DIR_OUT:
> + ddir = "TO_INITIATOR";
> + ocp->ddir = DMA_FROM_DEVICE;
> + break;
> + case EFCT_SCSI_CMD_DIR_IN | EFCT_SCSI_CMD_DIR_OUT:
> + ddir = "BIDIR";
> + ocp->ddir = DMA_BIDIRECTIONAL;
> + break;
> + default:
> + ddir = "NONE";
> + ocp->ddir = DMA_NONE;
> + break;
> + }
> +
> + ocp->lun = lun;
> + efct_lio_io_printf(io, "new cmd=0x%x ddir=%s dl=%u\n",
> + cdb[0], ddir, io->exp_xfer_len);
> +
> + tgt_node = io->node->tgt_node;
> + se_sess = tgt_node->session;
> + if (se_sess) {
> + efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_SUBMIT_CMD);
> + rc = target_submit_cmd(&io->tgt_io.cmd, se_sess,
> + cdb, &io->tgt_io.sense_buffer[0],
> + ocp->lun, io->exp_xfer_len,
> + ocp->task_attr, ocp->ddir,
> + TARGET_SCF_ACK_KREF);
> + if (rc) {
> + efc_log_err(efct, "failed to submit cmd se_cmd: %p\n",
> + &ocp->cmd);
> + efct_scsi_io_free(io);
> + }
> + }
> +
> + return rc;
> +}
> +
> +int
> +efct_scsi_recv_tmf(struct efct_io *tmfio, u32 lun,
> + enum efct_scsi_tmf_cmd cmd,
> + struct efct_io *io_to_abort, u32 flags)
aligment
> +{
> + unsigned char tmr_func;
> + struct efct *efct = tmfio->efct;
> + struct efct_scsi_tgt_io *ocp = &tmfio->tgt_io;
> + struct efct_scsi_tgt_node *tgt_node = NULL;
> + struct se_session *se_sess;
> + int rc;
> +
> + memset(ocp, 0, sizeof(struct efct_scsi_tgt_io));
> + efct_set_lio_io_state(tmfio, EFCT_LIO_STATE_SCSI_RECV_TMF);
> + atomic_add_return(1, &efct->tgt_efct.ios_in_use);
> + efct_lio_tmfio_printf(tmfio, "%s: new tmf %x lun=%u\n",
> + tmfio->display_name, cmd, lun);
> +
> + switch (cmd) {
> + case EFCT_SCSI_TMF_ABORT_TASK:
> + tmr_func = TMR_ABORT_TASK;
> + break;
> + case EFCT_SCSI_TMF_ABORT_TASK_SET:
> + tmr_func = TMR_ABORT_TASK_SET;
> + break;
> + case EFCT_SCSI_TMF_CLEAR_TASK_SET:
> + tmr_func = TMR_CLEAR_TASK_SET;
> + break;
> + case EFCT_SCSI_TMF_LOGICAL_UNIT_RESET:
> + tmr_func = TMR_LUN_RESET;
> + break;
> + case EFCT_SCSI_TMF_CLEAR_ACA:
> + tmr_func = TMR_CLEAR_ACA;
> + break;
> + case EFCT_SCSI_TMF_TARGET_RESET:
> + tmr_func = TMR_TARGET_WARM_RESET;
> + break;
> + case EFCT_SCSI_TMF_QUERY_ASYNCHRONOUS_EVENT:
> + case EFCT_SCSI_TMF_QUERY_TASK_SET:
> + default:
> + goto tmf_fail;
> + }
> +
> + tmfio->tgt_io.tmf = tmr_func;
> + tmfio->tgt_io.lun = lun;
> + tmfio->tgt_io.io_to_abort = io_to_abort;
> +
> + tgt_node = tmfio->node->tgt_node;
> +
> + se_sess = tgt_node->session;
> + if (!se_sess)
> + return EFC_SUCCESS;
> +
> + rc = target_submit_tmr(&ocp->cmd, se_sess, NULL, lun, ocp, tmr_func,
> + GFP_ATOMIC, tmfio->init_task_tag, TARGET_SCF_ACK_KREF);
Is GFP_ATOMIC really necessary? Are allocations done in atomic context?
> +
> + efct_set_lio_io_state(tmfio, EFCT_LIO_STATE_TGT_SUBMIT_TMR);
> + if (rc)
> + goto tmf_fail;
> +
> + return EFC_SUCCESS;
> +
> +tmf_fail:
> + efct_scsi_send_tmf_resp(tmfio, EFCT_SCSI_TMF_FUNCTION_REJECTED,
> + NULL, efct_lio_null_tmf_done, NULL);
> + return EFC_SUCCESS;
> +}
> +
> +/* Start items for efct_lio_tpg_attrib_cit */
> +
> +#define DEF_EFCT_TPG_ATTRIB(name) \
> + \
> +static ssize_t efct_lio_tpg_attrib_##name##_show( \
> + struct config_item *item, char *page) \
> +{ \
> + struct se_portal_group *se_tpg = to_tpg(item); \
> + struct efct_lio_tpg *tpg = container_of(se_tpg, \
> + struct efct_lio_tpg, tpg); \
> + \
> + return sprintf(page, "%u\n", tpg->tpg_attrib.name); \
> +} \
> + \
> +static ssize_t efct_lio_tpg_attrib_##name##_store( \
> + struct config_item *item, const char *page, size_t count) \
> +{ \
> + struct se_portal_group *se_tpg = to_tpg(item); \
> + struct efct_lio_tpg *tpg = container_of(se_tpg, \
> + struct efct_lio_tpg, tpg); \
> + struct efct_lio_tpg_attrib *a = &tpg->tpg_attrib; \
> + unsigned long val; \
> + int ret; \
> + \
> + ret = kstrtoul(page, 0, &val); \
> + if (ret < 0) { \
> + pr_err("kstrtoul() failed with ret: %d\n", ret); \
> + return ret; \
> + } \
> + \
> + if (val != 0 && val != 1) { \
> + pr_err("Illegal boolean value %lu\n", val); \
> + return -EINVAL; \
> + } \
> + \
> + a->name = val; \
> + \
> + return count; \
> +} \
> +CONFIGFS_ATTR(efct_lio_tpg_attrib_, name)
> +
> +DEF_EFCT_TPG_ATTRIB(generate_node_acls);
> +DEF_EFCT_TPG_ATTRIB(cache_dynamic_acls);
> +DEF_EFCT_TPG_ATTRIB(demo_mode_write_protect);
> +DEF_EFCT_TPG_ATTRIB(prod_mode_write_protect);
> +DEF_EFCT_TPG_ATTRIB(demo_mode_login_only);
> +DEF_EFCT_TPG_ATTRIB(session_deletion_wait);
> +
> +static struct configfs_attribute *efct_lio_tpg_attrib_attrs[] = {
> + &efct_lio_tpg_attrib_attr_generate_node_acls,
> + &efct_lio_tpg_attrib_attr_cache_dynamic_acls,
> + &efct_lio_tpg_attrib_attr_demo_mode_write_protect,
> + &efct_lio_tpg_attrib_attr_prod_mode_write_protect,
> + &efct_lio_tpg_attrib_attr_demo_mode_login_only,
> + &efct_lio_tpg_attrib_attr_session_deletion_wait,
> + NULL,
> +};
> +
> +#define DEF_EFCT_NPIV_TPG_ATTRIB(name) \
> + \
> +static ssize_t efct_lio_npiv_tpg_attrib_##name##_show( \
> + struct config_item *item, char *page) \
> +{ \
> + struct se_portal_group *se_tpg = to_tpg(item); \
> + struct efct_lio_tpg *tpg = container_of(se_tpg, \
> + struct efct_lio_tpg, tpg); \
> + \
> + return sprintf(page, "%u\n", tpg->tpg_attrib.name); \
> +} \
> + \
> +static ssize_t efct_lio_npiv_tpg_attrib_##name##_store( \
> + struct config_item *item, const char *page, size_t count) \
> +{ \
> + struct se_portal_group *se_tpg = to_tpg(item); \
> + struct efct_lio_tpg *tpg = container_of(se_tpg, \
> + struct efct_lio_tpg, tpg); \
> + struct efct_lio_tpg_attrib *a = &tpg->tpg_attrib; \
> + unsigned long val; \
> + int ret; \
> + \
> + ret = kstrtoul(page, 0, &val); \
> + if (ret < 0) { \
> + pr_err("kstrtoul() failed with ret: %d\n", ret); \
> + return ret; \
> + } \
> + \
> + if (val != 0 && val != 1) { \
> + pr_err("Illegal boolean value %lu\n", val); \
> + return -EINVAL; \
> + } \
> + \
> + a->name = val; \
> + \
> + return count; \
> +} \
> +CONFIGFS_ATTR(efct_lio_npiv_tpg_attrib_, name)
> +
> +DEF_EFCT_NPIV_TPG_ATTRIB(generate_node_acls);
> +DEF_EFCT_NPIV_TPG_ATTRIB(cache_dynamic_acls);
> +DEF_EFCT_NPIV_TPG_ATTRIB(demo_mode_write_protect);
> +DEF_EFCT_NPIV_TPG_ATTRIB(prod_mode_write_protect);
> +DEF_EFCT_NPIV_TPG_ATTRIB(demo_mode_login_only);
> +DEF_EFCT_NPIV_TPG_ATTRIB(session_deletion_wait);
> +
> +static struct configfs_attribute *efct_lio_npiv_tpg_attrib_attrs[] = {
> + &efct_lio_npiv_tpg_attrib_attr_generate_node_acls,
> + &efct_lio_npiv_tpg_attrib_attr_cache_dynamic_acls,
> + &efct_lio_npiv_tpg_attrib_attr_demo_mode_write_protect,
> + &efct_lio_npiv_tpg_attrib_attr_prod_mode_write_protect,
> + &efct_lio_npiv_tpg_attrib_attr_demo_mode_login_only,
> + &efct_lio_npiv_tpg_attrib_attr_session_deletion_wait,
> + NULL,
> +};
> +
> +CONFIGFS_ATTR(efct_lio_tpg_, enable);
> +static struct configfs_attribute *efct_lio_tpg_attrs[] = {
> + &efct_lio_tpg_attr_enable, NULL };
> +CONFIGFS_ATTR(efct_lio_npiv_tpg_, enable);
> +static struct configfs_attribute *efct_lio_npiv_tpg_attrs[] = {
> + &efct_lio_npiv_tpg_attr_enable, NULL };
> +
> +static const struct target_core_fabric_ops efct_lio_ops = {
> + .module = THIS_MODULE,
> + .fabric_name = "efct",
> + .node_acl_size = sizeof(struct efct_lio_nacl),
> + .max_data_sg_nents = 65535,
> + .tpg_get_wwn = efct_lio_get_fabric_wwn,
> + .tpg_get_tag = efct_lio_get_tag,
> + .fabric_init_nodeacl = efct_lio_init_nodeacl,
> + .tpg_check_demo_mode = efct_lio_check_demo_mode,
> + .tpg_check_demo_mode_cache = efct_lio_check_demo_mode_cache,
> + .tpg_check_demo_mode_write_protect = efct_lio_check_demo_write_protect,
> + .tpg_check_prod_mode_write_protect = efct_lio_check_prod_write_protect,
> + .tpg_get_inst_index = efct_lio_tpg_get_inst_index,
> + .check_stop_free = efct_lio_check_stop_free,
> + .aborted_task = efct_lio_aborted_task,
> + .release_cmd = efct_lio_release_cmd,
> + .close_session = efct_lio_close_session,
> + .sess_get_index = efct_lio_sess_get_index,
> + .write_pending = efct_lio_write_pending,
> + .set_default_node_attributes = efct_lio_set_default_node_attrs,
> + .get_cmd_state = efct_lio_get_cmd_state,
> + .queue_data_in = efct_lio_queue_data_in,
> + .queue_status = efct_lio_queue_status,
> + .queue_tm_rsp = efct_lio_queue_tm_rsp,
> + .fabric_make_wwn = efct_lio_make_sport,
> + .fabric_drop_wwn = efct_lio_drop_sport,
> + .fabric_make_tpg = efct_lio_make_tpg,
> + .fabric_drop_tpg = efct_lio_drop_tpg,
> + .tpg_check_demo_mode_login_only = efct_lio_check_demo_mode_login_only,
> + .tpg_check_prot_fabric_only = NULL,
> + .sess_get_initiator_sid = NULL,
> + .tfc_tpg_base_attrs = efct_lio_tpg_attrs,
> + .tfc_tpg_attrib_attrs = efct_lio_tpg_attrib_attrs,
> +};
> +
> +static const struct target_core_fabric_ops efct_lio_npiv_ops = {
> + .module = THIS_MODULE,
> + .fabric_name = "efct_npiv",
> + .node_acl_size = sizeof(struct efct_lio_nacl),
> + .max_data_sg_nents = 65535,
> + .tpg_get_wwn = efct_lio_get_npiv_fabric_wwn,
> + .tpg_get_tag = efct_lio_get_npiv_tag,
> + .fabric_init_nodeacl = efct_lio_init_nodeacl,
> + .tpg_check_demo_mode = efct_lio_check_demo_mode,
> + .tpg_check_demo_mode_cache = efct_lio_check_demo_mode_cache,
> + .tpg_check_demo_mode_write_protect =
> + efct_lio_npiv_check_demo_write_protect,
> + .tpg_check_prod_mode_write_protect =
> + efct_lio_npiv_check_prod_write_protect,
> + .tpg_get_inst_index = efct_lio_tpg_get_inst_index,
> + .check_stop_free = efct_lio_check_stop_free,
> + .aborted_task = efct_lio_aborted_task,
> + .release_cmd = efct_lio_release_cmd,
> + .close_session = efct_lio_close_session,
> + .sess_get_index = efct_lio_sess_get_index,
> + .write_pending = efct_lio_write_pending,
> + .set_default_node_attributes = efct_lio_set_default_node_attrs,
> + .get_cmd_state = efct_lio_get_cmd_state,
> + .queue_data_in = efct_lio_queue_data_in,
> + .queue_status = efct_lio_queue_status,
> + .queue_tm_rsp = efct_lio_queue_tm_rsp,
> + .fabric_make_wwn = efct_lio_npiv_make_sport,
> + .fabric_drop_wwn = efct_lio_npiv_drop_sport,
> + .fabric_make_tpg = efct_lio_npiv_make_tpg,
> + .fabric_drop_tpg = efct_lio_npiv_drop_tpg,
> + .tpg_check_demo_mode_login_only =
> + efct_lio_npiv_check_demo_mode_login_only,
> + .tpg_check_prot_fabric_only = NULL,
> + .sess_get_initiator_sid = NULL,
> + .tfc_tpg_base_attrs = efct_lio_npiv_tpg_attrs,
> + .tfc_tpg_attrib_attrs = efct_lio_npiv_tpg_attrib_attrs,
> +};
> +
> +int efct_scsi_tgt_driver_init(void)
> +{
> + int rc;
> +
> + /* Register the top level struct config_item_type with TCM core */
> + rc = target_register_template(&efct_lio_ops);
> + if (rc < 0) {
> + pr_err("target_fabric_configfs_register failed with %d\n", rc);
> + return rc;
> + }
> + rc = target_register_template(&efct_lio_npiv_ops);
> + if (rc < 0) {
> + pr_err("target_fabric_configfs_register failed with %d\n", rc);
> + target_unregister_template(&efct_lio_ops);
> + return rc;
> + }
> + return EFC_SUCCESS;
> +}
> +
> +int efct_scsi_tgt_driver_exit(void)
> +{
> + target_unregister_template(&efct_lio_ops);
> + target_unregister_template(&efct_lio_npiv_ops);
> + return EFC_SUCCESS;
> +}
> diff --git a/drivers/scsi/elx/efct/efct_lio.h b/drivers/scsi/elx/efct/efct_lio.h
> new file mode 100644
> index 000000000000..f884bcd3b240
> --- /dev/null
> +++ b/drivers/scsi/elx/efct/efct_lio.h
> @@ -0,0 +1,178 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2019 Broadcom. All Rights Reserved. The term
> + * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
> + */
> +
> +#ifndef __EFCT_LIO_H__
> +#define __EFCT_LIO_H__
> +
> +#include "efct_scsi.h"
> +#include <target/target_core_base.h>
> +
> +#define efct_lio_io_printf(io, fmt, ...) \
> + efc_log_debug(io->efct, \
> + "[%s] [%04x][i:%04x t:%04x h:%04x][c:%02x]" fmt, \
> + io->node->display_name, io->instance_index, \
> + io->init_task_tag, io->tgt_task_tag, io->hw_tag, \
> + io->tgt_io.cmd.t_task_cdb[0], ##__VA_ARGS__)
> +
> +#define efct_lio_tmfio_printf(io, fmt, ...) \
> + efc_log_debug(io->efct, \
> + "[%s] [%04x][i:%04x t:%04x h:%04x][f:%02x]" fmt, \
> + io->node->display_name, io->instance_index, \
> + io->init_task_tag, io->tgt_task_tag, io->hw_tag, \
> + io->tgt_io.tmf, ##__VA_ARGS__)
in the last file the \ were all right aligned.
> +
> +#define efct_set_lio_io_state(io, value) (io->tgt_io.state |= value)
> +
> +struct efct_lio_wq_data {
> + struct efct *efct;
> + void *ptr;
> + struct work_struct work;
> +};
> +
> +/* Target private efct structure */
> +struct efct_scsi_tgt {
> + u32 max_sge;
> + u32 max_sgl;
> +
> + /*
> + * Variables used to send task set full. We are using a high watermark
> + * method to send task set full. We will reserve a fixed number of IOs
> + * per initiator plus a fudge factor. Once we reach this number,
> + * then the target will start sending task set full/busy responses.
> + */
> + atomic_t initiator_count;
> + atomic_t ios_in_use;
> + atomic_t io_high_watermark;
> +
> + atomic_t watermark_hit;
> + int watermark_min;
> + int watermark_max;
> +
> + struct efct_lio_sport *lio_sport;
> + struct efct_lio_tpg *tpg;
> +
> + struct list_head vport_list;
> + /* Protects vport list*/
> + spinlock_t efct_lio_lock;
> +
> + u64 wwnn;
> +};
> +
> +struct efct_scsi_tgt_sport {
> + struct efct_lio_sport *lio_sport;
> +};
> +
> +struct efct_scsi_tgt_node {
> + struct se_session *session;
> +};
> +
> +#define EFCT_LIO_STATE_SCSI_RECV_CMD (1 << 0)
> +#define EFCT_LIO_STATE_TGT_SUBMIT_CMD (1 << 1)
> +#define EFCT_LIO_STATE_TFO_QUEUE_DATA_IN (1 << 2)
> +#define EFCT_LIO_STATE_TFO_WRITE_PENDING (1 << 3)
> +#define EFCT_LIO_STATE_TGT_EXECUTE_CMD (1 << 4)
> +#define EFCT_LIO_STATE_SCSI_SEND_RD_DATA (1 << 5)
> +#define EFCT_LIO_STATE_TFO_CHK_STOP_FREE (1 << 6)
> +#define EFCT_LIO_STATE_SCSI_DATA_DONE (1 << 7)
> +#define EFCT_LIO_STATE_TFO_QUEUE_STATUS (1 << 8)
> +#define EFCT_LIO_STATE_SCSI_SEND_RSP (1 << 9)
> +#define EFCT_LIO_STATE_SCSI_RSP_DONE (1 << 10)
> +#define EFCT_LIO_STATE_TGT_GENERIC_FREE (1 << 11)
> +#define EFCT_LIO_STATE_SCSI_RECV_TMF (1 << 12)
> +#define EFCT_LIO_STATE_TGT_SUBMIT_TMR (1 << 13)
> +#define EFCT_LIO_STATE_TFO_WRITE_PEND_STATUS (1 << 14)
> +#define EFCT_LIO_STATE_TGT_GENERIC_REQ_FAILURE (1 << 15)
> +
> +#define EFCT_LIO_STATE_TFO_ABORTED_TASK (1 << 29)
> +#define EFCT_LIO_STATE_TFO_RELEASE_CMD (1 << 30)
> +#define EFCT_LIO_STATE_SCSI_CMPL_CMD (1 << 31)
> +
> +struct efct_scsi_tgt_io {
> + struct se_cmd cmd;
> + unsigned char sense_buffer[TRANSPORT_SENSE_BUFFER];
> + enum dma_data_direction ddir;
> + int task_attr;
> + u64 lun;
> +
> + u32 state;
> + u8 tmf;
> + struct efct_io *io_to_abort;
> + u32 seg_map_cnt;
> + u32 seg_cnt;
> + u32 cur_seg;
> + enum efct_scsi_io_status err;
> + bool aborting;
> + bool rsp_sent;
> + uint32_t transferred_len;
> +};
> +
> +/* Handler return codes */
> +enum {
> + SCSI_HANDLER_DATAPHASE_STARTED = 1,
> + SCSI_HANDLER_RESP_STARTED,
> + SCSI_HANDLER_VALIDATED_DATAPHASE_STARTED,
> + SCSI_CMD_NOT_SUPPORTED,
> +};
> +
> +#define WWN_NAME_LEN 32
> +struct efct_lio_vport {
> + u64 wwpn;
> + u64 npiv_wwpn;
> + u64 npiv_wwnn;
> + unsigned char wwpn_str[WWN_NAME_LEN];
> + struct se_wwn vport_wwn;
> + struct efct_lio_tpg *tpg;
> + struct efct *efct;
> + struct dentry *sessions;
> + struct Scsi_Host *shost;
> + struct fc_vport *fc_vport;
> + atomic_t enable;
> +};
> +
> +struct efct_lio_sport {
> + u64 wwpn;
> + unsigned char wwpn_str[WWN_NAME_LEN];
> + struct se_wwn sport_wwn;
> + struct efct_lio_tpg *tpg;
> + struct efct *efct;
> + struct dentry *sessions;
> + atomic_t enable;
> +};
> +
> +struct efct_lio_tpg_attrib {
> + int generate_node_acls;
> + int cache_dynamic_acls;
> + int demo_mode_write_protect;
> + int prod_mode_write_protect;
> + int demo_mode_login_only;
> + bool session_deletion_wait;
> +};
> +
> +struct efct_lio_tpg {
> + struct se_portal_group tpg;
> + struct efct_lio_sport *sport;
> + struct efct_lio_vport *vport;
> + struct efct_lio_tpg_attrib tpg_attrib;
> + unsigned short tpgt;
> + atomic_t enabled;
> +};
> +
> +struct efct_lio_nacl {
> + u64 nport_wwnn;
> + char nport_name[WWN_NAME_LEN];
> + struct se_session *session;
> + struct se_node_acl se_node_acl;
> +};
> +
> +struct efct_lio_vport_list_t {
> + struct list_head list_entry;
> + struct efct_lio_vport *lio_vport;
> +};
> +
> +int efct_scsi_tgt_driver_init(void);
> +int efct_scsi_tgt_driver_exit(void);
> +
> +#endif /*__EFCT_LIO_H__ */
> --
> 2.16.4
>
Thanks,
Daniel
next prev parent reply other threads:[~2020-04-16 12:34 UTC|newest]
Thread overview: 122+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-04-12 3:32 [PATCH v3 00/31] [NEW] efct: Broadcom (Emulex) FC Target driver James Smart
2020-04-12 3:32 ` [PATCH v3 01/31] elx: libefc_sli: SLI-4 register offsets and field definitions James Smart
2020-04-14 15:23 ` Daniel Wagner
2020-04-22 4:28 ` James Smart
2020-04-15 12:06 ` Hannes Reinecke
2020-04-23 1:52 ` Roman Bolshakov
2020-04-12 3:32 ` [PATCH v3 02/31] elx: libefc_sli: SLI Descriptors and Queue entries James Smart
2020-04-14 18:02 ` Daniel Wagner
2020-04-22 4:41 ` James Smart
2020-04-15 12:14 ` Hannes Reinecke
2020-04-15 17:43 ` James Bottomley
2020-04-22 4:44 ` James Smart
2020-04-12 3:32 ` [PATCH v3 03/31] elx: libefc_sli: Data structures and defines for mbox commands James Smart
2020-04-14 19:01 ` Daniel Wagner
2020-04-15 12:22 ` Hannes Reinecke
2020-04-12 3:32 ` [PATCH v3 04/31] elx: libefc_sli: queue create/destroy/parse routines James Smart
2020-04-15 10:04 ` Daniel Wagner
2020-04-22 5:05 ` James Smart
2020-04-24 7:29 ` Daniel Wagner
2020-04-24 15:21 ` James Smart
2020-04-15 12:27 ` Hannes Reinecke
2020-04-12 3:32 ` [PATCH v3 05/31] elx: libefc_sli: Populate and post different WQEs James Smart
2020-04-15 14:34 ` Daniel Wagner
2020-04-22 5:08 ` James Smart
2020-04-12 3:32 ` [PATCH v3 06/31] elx: libefc_sli: bmbx routines and SLI config commands James Smart
2020-04-15 16:10 ` Daniel Wagner
2020-04-22 5:12 ` James Smart
2020-04-12 3:32 ` [PATCH v3 07/31] elx: libefc_sli: APIs to setup SLI library James Smart
2020-04-15 12:49 ` Hannes Reinecke
2020-04-15 17:06 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 08/31] elx: libefc: Generic state machine framework James Smart
2020-04-15 12:37 ` Hannes Reinecke
2020-04-15 17:20 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 09/31] elx: libefc: Emulex FC discovery library APIs and definitions James Smart
2020-04-15 12:41 ` Hannes Reinecke
2020-04-15 17:32 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 10/31] elx: libefc: FC Domain state machine interfaces James Smart
2020-04-15 12:50 ` Hannes Reinecke
2020-04-15 17:50 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 11/31] elx: libefc: SLI and FC PORT " James Smart
2020-04-15 15:38 ` Hannes Reinecke
2020-04-22 23:12 ` James Smart
2020-04-15 18:04 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 12/31] elx: libefc: Remote node " James Smart
2020-04-15 15:51 ` Hannes Reinecke
2020-04-23 1:35 ` James Smart
2020-04-23 8:02 ` Daniel Wagner
2020-04-23 18:24 ` James Smart
2020-04-15 18:19 ` Daniel Wagner
2020-04-23 1:32 ` James Smart
2020-04-23 7:49 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 13/31] elx: libefc: Fabric " James Smart
2020-04-15 18:51 ` Daniel Wagner
2020-04-16 6:37 ` Hannes Reinecke
2020-04-23 1:38 ` James Smart
2020-04-12 3:32 ` [PATCH v3 14/31] elx: libefc: FC node ELS and state handling James Smart
2020-04-15 18:56 ` Daniel Wagner
2020-04-23 2:50 ` James Smart
2020-04-23 8:05 ` Daniel Wagner
2020-04-23 8:12 ` Nathan Chancellor
2020-04-16 6:47 ` Hannes Reinecke
2020-04-23 2:55 ` James Smart
2020-04-12 3:32 ` [PATCH v3 15/31] elx: efct: Data structures and defines for hw operations James Smart
2020-04-16 6:51 ` Hannes Reinecke
2020-04-23 2:57 ` James Smart
2020-04-16 7:22 ` Daniel Wagner
2020-04-23 2:59 ` James Smart
2020-04-12 3:32 ` [PATCH v3 16/31] elx: efct: Driver initialization routines James Smart
2020-04-16 7:11 ` Hannes Reinecke
2020-04-23 3:09 ` James Smart
2020-04-16 8:03 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 17/31] elx: efct: Hardware queues creation and deletion James Smart
2020-04-16 7:14 ` Hannes Reinecke
2020-04-16 8:24 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 18/31] elx: efct: RQ buffer, memory pool allocation and deallocation APIs James Smart
2020-04-16 7:24 ` Hannes Reinecke
2020-04-23 3:16 ` James Smart
2020-04-16 8:41 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 19/31] elx: efct: Hardware IO and SGL initialization James Smart
2020-04-16 7:32 ` Hannes Reinecke
2020-04-16 8:47 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 20/31] elx: efct: Hardware queues processing James Smart
2020-04-16 7:37 ` Hannes Reinecke
2020-04-16 9:17 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 21/31] elx: efct: Unsolicited FC frame processing routines James Smart
2020-04-16 9:36 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 22/31] elx: efct: Extended link Service IO handling James Smart
2020-04-16 7:58 ` Hannes Reinecke
2020-04-23 3:30 ` James Smart
2020-04-16 9:49 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 23/31] elx: efct: SCSI IO handling routines James Smart
2020-04-16 11:40 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 24/31] elx: efct: LIO backend interface routines James Smart
2020-04-12 4:57 ` Bart Van Assche
2020-04-16 11:48 ` Daniel Wagner
2020-04-22 4:20 ` James Smart
2020-04-22 5:09 ` Bart Van Assche
2020-04-23 1:39 ` James Smart
2020-04-16 8:02 ` Hannes Reinecke
2020-04-16 12:34 ` Daniel Wagner [this message]
2020-04-22 4:20 ` James Smart
2020-04-12 3:32 ` [PATCH v3 25/31] elx: efct: Hardware IO submission routines James Smart
2020-04-16 8:10 ` Hannes Reinecke
2020-04-16 12:45 ` Daniel Wagner
2020-04-23 3:37 ` James Smart
2020-04-16 12:44 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 26/31] elx: efct: link statistics and SFP data James Smart
2020-04-16 12:55 ` Daniel Wagner
2020-04-12 3:32 ` [PATCH v3 27/31] elx: efct: xport and hardware teardown routines James Smart
2020-04-16 9:45 ` Hannes Reinecke
2020-04-16 13:01 ` Daniel Wagner
2020-04-12 3:33 ` [PATCH v3 28/31] elx: efct: Firmware update, async link processing James Smart
2020-04-16 10:01 ` Hannes Reinecke
2020-04-16 13:10 ` Daniel Wagner
2020-04-12 3:33 ` [PATCH v3 29/31] elx: efct: scsi_transport_fc host interface support James Smart
2020-04-12 3:33 ` [PATCH v3 30/31] elx: efct: Add Makefile and Kconfig for efct driver James Smart
2020-04-16 10:02 ` Hannes Reinecke
2020-04-16 13:15 ` Daniel Wagner
2020-04-12 3:33 ` [PATCH v3 31/31] elx: efct: Tie into kernel Kconfig and build process James Smart
2020-04-12 6:16 ` kbuild test robot
2020-04-12 7:56 ` kbuild test robot
2020-04-16 13:15 ` Daniel Wagner
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200416123425.lj2clovshk5yxz6g@carbon \
--to=dwagner@suse.de \
--cc=bvanassche@acm.org \
--cc=hare@suse.de \
--cc=herbszt@gmx.de \
--cc=jsmart2021@gmail.com \
--cc=linux-scsi@vger.kernel.org \
--cc=maier@linux.ibm.com \
--cc=natechancellor@gmail.com \
--cc=ram.vegesna@broadcom.com \
--cc=rdunlap@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox