From: Honggang LI <honli@redhat.com>
To: Bart Van Assche <bvanassche@acm.org>
Cc: Jason Gunthorpe <jgg@ziepe.ca>,
Leon Romanovsky <leonro@mellanox.com>,
Doug Ledford <dledford@redhat.com>,
linux-rdma@vger.kernel.org
Subject: Re: [PATCH] RDMA/srpt: Fix TPG creation
Date: Thu, 24 Oct 2019 21:07:46 +0800 [thread overview]
Message-ID: <20191024130746.GA21231@dhcp-128-227.nay.redhat.com> (raw)
In-Reply-To: <20191023204106.23326-1-bvanassche@acm.org>
On Wed, Oct 23, 2019 at 01:41:06PM -0700, Bart Van Assche wrote:
> Unlike the iSCSI target driver, for the SRP target driver it is sufficient
> if a single TPG can be associated with each RDMA port name. However, users
> started associating multiple TPGs with RDMA port names. Support this by
> converting the single TPG in struct srpt_port_id into a list. This patch
> fixes the following list corruption issue:
>
> list_add corruption. prev->next should be next (ffffffffc0a080c0), but was ffffa08a994ce6f0. (prev=ffffa08a994ce6f0).
> WARNING: CPU: 2 PID: 2597 at lib/list_debug.c:28 __list_add_valid+0x6a/0x70
> CPU: 2 PID: 2597 Comm: targetcli Not tainted 5.4.0-rc1.3bfa3c9602a7 #1
> RIP: 0010:__list_add_valid+0x6a/0x70
> Call Trace:
> core_tpg_register+0x116/0x200 [target_core_mod]
> srpt_make_tpg+0x3f/0x60 [ib_srpt]
> target_fabric_make_tpg+0x41/0x290 [target_core_mod]
> configfs_mkdir+0x158/0x3e0
> vfs_mkdir+0x108/0x1a0
> do_mkdirat+0x77/0xe0
> do_syscall_64+0x55/0x1d0
> entry_SYSCALL_64_after_hwframe+0x44/0xa9
>
> Cc: Honggang LI <honli@redhat.com>
> Reported-by: Honggang LI <honli@redhat.com>
> Fixes: a42d985bd5b2 ("ib_srpt: Initial SRP Target merge for v3.3-rc1")
> Signed-off-by: Bart Van Assche <bvanassche@acm.org>
> ---
> drivers/infiniband/ulp/srpt/ib_srpt.c | 77 ++++++++++++++++++---------
> drivers/infiniband/ulp/srpt/ib_srpt.h | 25 +++++++--
> 2 files changed, 73 insertions(+), 29 deletions(-)
>
> diff --git a/drivers/infiniband/ulp/srpt/ib_srpt.c b/drivers/infiniband/ulp/srpt/ib_srpt.c
> index daf811abf40a..a278e76b9e02 100644
> --- a/drivers/infiniband/ulp/srpt/ib_srpt.c
> +++ b/drivers/infiniband/ulp/srpt/ib_srpt.c
> @@ -2131,6 +2131,7 @@ static int srpt_cm_req_recv(struct srpt_device *const sdev,
> char i_port_id[36];
> u32 it_iu_len;
> int i, tag_num, tag_size, ret;
> + struct srpt_tpg *stpg;
>
> WARN_ON_ONCE(irqs_disabled());
>
> @@ -2288,19 +2289,33 @@ static int srpt_cm_req_recv(struct srpt_device *const sdev,
>
> tag_num = ch->rq_size;
> tag_size = 1; /* ib_srpt does not use se_sess->sess_cmd_map */
> - if (sport->port_guid_id.tpg.se_tpg_wwn)
> - ch->sess = target_setup_session(&sport->port_guid_id.tpg, tag_num,
> +
> + mutex_lock(&sport->port_guid_id.mutex);
> + list_for_each_entry(stpg, &sport->port_guid_id.tpg_list, entry) {
> + if (!IS_ERR_OR_NULL(ch->sess))
> + break;
> + ch->sess = target_setup_session(&stpg->tpg, tag_num,
> tag_size, TARGET_PROT_NORMAL,
> ch->sess_name, ch, NULL);
> - if (sport->port_gid_id.tpg.se_tpg_wwn && IS_ERR_OR_NULL(ch->sess))
> - ch->sess = target_setup_session(&sport->port_gid_id.tpg, tag_num,
> + }
> + mutex_unlock(&sport->port_guid_id.mutex);
> +
> + mutex_lock(&sport->port_gid_id.mutex);
> + list_for_each_entry(stpg, &sport->port_gid_id.tpg_list, entry) {
> + if (!IS_ERR_OR_NULL(ch->sess))
> + break;
> + ch->sess = target_setup_session(&stpg->tpg, tag_num,
> tag_size, TARGET_PROT_NORMAL, i_port_id,
> ch, NULL);
> - /* Retry without leading "0x" */
> - if (sport->port_gid_id.tpg.se_tpg_wwn && IS_ERR_OR_NULL(ch->sess))
> - ch->sess = target_setup_session(&sport->port_gid_id.tpg, tag_num,
> + if (!IS_ERR_OR_NULL(ch->sess))
> + break;
> + /* Retry without leading "0x" */
> + ch->sess = target_setup_session(&stpg->tpg, tag_num,
> tag_size, TARGET_PROT_NORMAL,
> i_port_id + 2, ch, NULL);
> + }
> + mutex_unlock(&sport->port_gid_id.mutex);
> +
> if (IS_ERR_OR_NULL(ch->sess)) {
> WARN_ON_ONCE(ch->sess == NULL);
> ret = PTR_ERR(ch->sess);
> @@ -3140,6 +3155,10 @@ static void srpt_add_one(struct ib_device *device)
> sport->port_attrib.srp_sq_size = DEF_SRPT_SQ_SIZE;
> sport->port_attrib.use_srq = false;
> INIT_WORK(&sport->work, srpt_refresh_port_work);
> + mutex_init(&sport->port_guid_id.mutex);
> + INIT_LIST_HEAD(&sport->port_guid_id.tpg_list);
> + mutex_init(&sport->port_gid_id.mutex);
> + INIT_LIST_HEAD(&sport->port_gid_id.tpg_list);
>
> if (srpt_refresh_port(sport)) {
> pr_err("MAD registration failed for %s-%d.\n",
> @@ -3242,18 +3261,6 @@ static struct srpt_port *srpt_tpg_to_sport(struct se_portal_group *tpg)
> return tpg->se_tpg_wwn->priv;
> }
>
> -static struct srpt_port_id *srpt_tpg_to_sport_id(struct se_portal_group *tpg)
> -{
> - struct srpt_port *sport = srpt_tpg_to_sport(tpg);
> -
> - if (tpg == &sport->port_guid_id.tpg)
> - return &sport->port_guid_id;
> - if (tpg == &sport->port_gid_id.tpg)
> - return &sport->port_gid_id;
> - WARN_ON_ONCE(true);
> - return NULL;
> -}
> -
> static struct srpt_port_id *srpt_wwn_to_sport_id(struct se_wwn *wwn)
> {
> struct srpt_port *sport = wwn->priv;
> @@ -3268,7 +3275,9 @@ static struct srpt_port_id *srpt_wwn_to_sport_id(struct se_wwn *wwn)
>
> static char *srpt_get_fabric_wwn(struct se_portal_group *tpg)
> {
> - return srpt_tpg_to_sport_id(tpg)->name;
> + struct srpt_tpg *stpg = container_of(tpg, typeof(*stpg), tpg);
> +
> + return stpg->sport_id->name;
> }
>
> static u16 srpt_get_tag(struct se_portal_group *tpg)
> @@ -3725,16 +3734,27 @@ static struct se_portal_group *srpt_make_tpg(struct se_wwn *wwn,
> const char *name)
> {
> struct srpt_port *sport = wwn->priv;
> - struct se_portal_group *tpg = &srpt_wwn_to_sport_id(wwn)->tpg;
> - int res;
> + struct srpt_port_id *sport_id = srpt_wwn_to_sport_id(wwn);
> + struct srpt_tpg *stpg;
> + int res = -ENOMEM;
>
> - res = core_tpg_register(wwn, tpg, SCSI_PROTOCOL_SRP);
> - if (res)
> + stpg = kzalloc(sizeof(*stpg), GFP_KERNEL);
> + if (!stpg)
> + return ERR_PTR(res);
> + stpg->sport_id = sport_id;
> + res = core_tpg_register(wwn, &stpg->tpg, SCSI_PROTOCOL_SRP);
> + if (res) {
> + kfree(stpg);
> return ERR_PTR(res);
> + }
> +
> + mutex_lock(&sport_id->mutex);
> + list_add_tail(&stpg->entry, &sport_id->tpg_list);
> + mutex_unlock(&sport_id->mutex);
>
> atomic_inc(&sport->refcount);
>
> - return tpg;
> + return &stpg->tpg;
> }
>
> /**
> @@ -3743,10 +3763,17 @@ static struct se_portal_group *srpt_make_tpg(struct se_wwn *wwn,
> */
> static void srpt_drop_tpg(struct se_portal_group *tpg)
> {
> + struct srpt_tpg *stpg = container_of(tpg, typeof(*stpg), tpg);
> + struct srpt_port_id *sport_id = stpg->sport_id;
> struct srpt_port *sport = srpt_tpg_to_sport(tpg);
>
> + mutex_lock(&sport_id->mutex);
> + list_del(&stpg->entry);
> + mutex_unlock(&sport_id->mutex);
> +
> sport->enabled = false;
> core_tpg_deregister(tpg);
> + kfree(stpg);
> srpt_drop_sport_ref(sport);
> }
>
> diff --git a/drivers/infiniband/ulp/srpt/ib_srpt.h b/drivers/infiniband/ulp/srpt/ib_srpt.h
> index f8bd95302ac0..27a54f777e3b 100644
> --- a/drivers/infiniband/ulp/srpt/ib_srpt.h
> +++ b/drivers/infiniband/ulp/srpt/ib_srpt.h
> @@ -363,17 +363,34 @@ struct srpt_port_attrib {
> bool use_srq;
> };
>
> +/**
> + * struct srpt_tpg - information about a single "target portal group"
> + * @entry: Entry in @sport_id->tpg_list.
> + * @sport_id: Port name this TPG is associated with.
> + * @tpg: LIO TPG data structure.
> + *
> + * Zero or more target portal groups are associated with each port name
> + * (srpt_port_id). With each TPG an ACL list is associated.
> + */
> +struct srpt_tpg {
> + struct list_head entry;
> + struct srpt_port_id *sport_id;
> + struct se_portal_group tpg;
> +};
> +
> /**
> * struct srpt_port_id - information about an RDMA port name
> - * @tpg: TPG associated with the RDMA port.
> - * @wwn: WWN associated with the RDMA port.
> - * @name: ASCII representation of the port name.
> + * @mutex: Protects @tpg_list changes.
> + * @tpg_list: TPGs associated with the RDMA port name.
> + * @wwn: WWN associated with the RDMA port name.
> + * @name: ASCII representation of the port name.
> *
> * Multiple sysfs directories can be associated with a single RDMA port. This
> * data structure represents a single (port, name) pair.
> */
> struct srpt_port_id {
> - struct se_portal_group tpg;
> + struct mutex mutex;
> + struct list_head tpg_list;
> struct se_wwn wwn;
> char name[64];
> };
Acked-by: Honggang Li <honli@redhat.com>
Thanks
next prev parent reply other threads:[~2019-10-24 13:07 UTC|newest]
Thread overview: 5+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-10-23 20:41 [PATCH] RDMA/srpt: Fix TPG creation Bart Van Assche
2019-10-24 3:37 ` Honggang LI
2019-10-24 4:01 ` Bart Van Assche
2019-10-24 13:07 ` Honggang LI [this message]
2019-10-28 16:32 ` Jason Gunthorpe
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20191024130746.GA21231@dhcp-128-227.nay.redhat.com \
--to=honli@redhat.com \
--cc=bvanassche@acm.org \
--cc=dledford@redhat.com \
--cc=jgg@ziepe.ca \
--cc=leonro@mellanox.com \
--cc=linux-rdma@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox