From: Sagi Grimberg <sagi@grimberg.me>
To: hare@kernel.org, Christoph Hellwig <hch@lst.de>
Cc: Keith Busch <kbusch@kernel.org>,
James Smart <james.smart@broadcom.com>,
linux-nvme@lists.infradead.org, linux-scsi@vger.kernel.org,
Hannes Reinecke <hare@suse.de>
Subject: Re: [PATCH RFC] nvme-fc: FPIN link integrity handling
Date: Thu, 7 Mar 2024 12:10:52 +0200 [thread overview]
Message-ID: <1dfa9a4e-e4a2-4d48-b569-85e48ce4311c@grimberg.me> (raw)
In-Reply-To: <20240219085929.31255-1-hare@kernel.org>
On 19/02/2024 10:59, hare@kernel.org wrote:
> From: Hannes Reinecke <hare@suse.de>
>
> FPIN LI (link integrity) messages are received when the attached
> fabric detects hardware errors. In response to these messages the
> affected ports should not be used for I/O, and only put back into
> service once the ports had been reset as then the hardware might
> have been replaced.
Does this mean it cannot service any type of communication over
the wire?
> This patch adds a new controller flag 'NVME_CTRL_TRANSPORT_BLOCKED'
> which will be checked during multipath path selection, causing the
> path to be skipped.
While this looks sensible to me, it also looks like this introduces a
ctrl state
outside of ctrl->state... Wouldn't it make sense to move the controller to
NVME_CTRL_DEAD ? or is it not a terminal state?
>
> Signed-off-by: Hannes Reinecke <hare@suse.de>
> ---
> drivers/nvme/host/core.c | 13 ++--
> drivers/nvme/host/fc.c | 108 +++++++++++++++++++++++++++++++++
> drivers/nvme/host/multipath.c | 2 +
> drivers/nvme/host/nvme.h | 1 +
> drivers/scsi/lpfc/lpfc_els.c | 6 +-
> drivers/scsi/qla2xxx/qla_isr.c | 1 +
> include/linux/nvme-fc-driver.h | 3 +
> 7 files changed, 129 insertions(+), 5 deletions(-)
>
> diff --git a/drivers/nvme/host/core.c b/drivers/nvme/host/core.c
> index eed3e22e24d9..5e9a0cf43636 100644
> --- a/drivers/nvme/host/core.c
> +++ b/drivers/nvme/host/core.c
> @@ -750,10 +750,14 @@ blk_status_t nvme_fail_nonready_command(struct nvme_ctrl *ctrl,
>
> if (state != NVME_CTRL_DELETING_NOIO &&
> state != NVME_CTRL_DELETING &&
> - state != NVME_CTRL_DEAD &&
> - !test_bit(NVME_CTRL_FAILFAST_EXPIRED, &ctrl->flags) &&
> - !blk_noretry_request(rq) && !(rq->cmd_flags & REQ_NVME_MPATH))
> - return BLK_STS_RESOURCE;
> + state != NVME_CTRL_DEAD) {
> + if (!test_bit(NVME_CTRL_FAILFAST_EXPIRED, &ctrl->flags) &&
> + !blk_noretry_request(rq) &&
> + !(rq->cmd_flags & REQ_NVME_MPATH))
> + return BLK_STS_RESOURCE;
> + if (test_bit(NVME_CTRL_TRANSPORT_BLOCKED, &ctrl->flags))
> + return BLK_STS_TRANSPORT;
> + }
> return nvme_host_path_error(rq);
> }
> EXPORT_SYMBOL_GPL(nvme_fail_nonready_command);
> @@ -4575,6 +4579,7 @@ int nvme_init_ctrl(struct nvme_ctrl *ctrl, struct device *dev,
> WRITE_ONCE(ctrl->state, NVME_CTRL_NEW);
> ctrl->passthru_err_log_enabled = false;
> clear_bit(NVME_CTRL_FAILFAST_EXPIRED, &ctrl->flags);
> + clear_bit(NVME_CTRL_TRANSPORT_BLOCKED, &ctrl->flags);
> spin_lock_init(&ctrl->lock);
> mutex_init(&ctrl->scan_lock);
> INIT_LIST_HEAD(&ctrl->namespaces);
> diff --git a/drivers/nvme/host/fc.c b/drivers/nvme/host/fc.c
> index 5e226728c822..fdf77f5cb944 100644
> --- a/drivers/nvme/host/fc.c
> +++ b/drivers/nvme/host/fc.c
> @@ -787,6 +787,9 @@ nvme_fc_ctrl_connectivity_loss(struct nvme_fc_ctrl *ctrl)
> "NVME-FC{%d}: controller connectivity lost. Awaiting "
> "Reconnect", ctrl->cnum);
>
> + /* clear 'transport blocked' flag as controller will be reset */
> + clear_bit(NVME_CTRL_TRANSPORT_BLOCKED, &ctrl->flags);
> +
> switch (nvme_ctrl_state(&ctrl->ctrl)) {
> case NVME_CTRL_NEW:
> case NVME_CTRL_LIVE:
> @@ -3741,6 +3744,111 @@ static struct nvmf_transport_ops nvme_fc_transport = {
> .create_ctrl = nvme_fc_create_ctrl,
> };
>
> +static struct nvme_fc_rport *nvme_fc_rport_from_wwpn(struct nvme_fc_lport *lport,
> + u64 rport_wwpn)
> +{
> + struct nvme_fc_rport *rport;
> +
> + list_for_each_entry(rport, &lport->endp_list, endp_list) {
> + if (!nvme_fc_rport_get(rport))
> + continue;
> + if (rport->remoteport.port_name == rport_wwpn &&
> + rport->remoteport.port_role & FC_PORT_ROLE_NVME_TARGET)
> + return rport;
> + nvme_fc_rport_put(rport);
> + }
> + return NULL;
> +}
> +
> +/*
> + * nvme_fc_fpin_li_lport_update - routine to update Link Integrity
> + * event statistics.
> + * @lport: local port the FPIN was received on
> + * @tlv: pointer to link integrity descriptor
> + *
> + */
> +static void
> +nvme_fc_fpin_li_lport_update(struct nvme_fc_lport *lport, struct fc_tlv_desc *tlv)
> +{
> + unsigned int i, pname_count;
> + struct nvme_fc_rport *attached_rport;
> + struct fc_fn_li_desc *li_desc = (struct fc_fn_li_desc *)tlv;
> + u64 wwpn;
> +
> + wwpn = be64_to_cpu(li_desc->attached_wwpn);
> + attached_rport = nvme_fc_rport_from_wwpn(lport, wwpn);
> + pname_count = be32_to_cpu(li_desc->pname_count);
> +
> + for (i = 0; pname_count; i++) {
> + struct nvme_fc_rport *rport;
> +
> + wwpn = be64_to_cpu(li_desc->pname_list[i]);
> + rport = nvme_fc_rport_from_wwpn(lport, wwpn);
> + if (!rport)
> + continue;
> + if (rport != attached_rport) {
> + struct nvme_fc_ctrl *ctrl;
> +
> + spin_lock_irq(&rport->lock);
> + list_for_each_entry(ctrl, &rport->ctrl_list, ctrl_list)
> + set_bit(NVME_CTRL_TRANSPORT_BLOCKED, &ctrl->ctrl.flags);
> + spin_unlock_irq(&rport->lock);
> + }
> + nvme_fc_rport_put(rport);
> + }
> + if (attached_rport) {
> + struct nvme_fc_ctrl *ctrl;
> +
> + spin_lock_irq(&attached_rport->lock);
> + list_for_each_entry(ctrl, &attached_rport->ctrl_list, ctrl_list)
> + set_bit(NVME_CTRL_TRANSPORT_BLOCKED, &ctrl->ctrl.flags);
> + spin_unlock_irq(&attached_rport->lock);
> + nvme_fc_rport_put(attached_rport);
> + }
> +}
> +
> +/**
> + * fc_host_fpin_rcv - routine to process a received FPIN.
> + * @localport: local port the FPIN was received on
> + * @fpin_len: length of FPIN payload, in bytes
> + * @fpin_buf: pointer to FPIN payload
> + * Notes:
> + * This routine assumes no locks are held on entry.
> + */
> +void
> +nvme_fc_fpin_rcv(struct nvme_fc_local_port *localport,
> + u32 fpin_len, char *fpin_buf)
> +{
> + struct nvme_fc_lport *lport;
> + struct fc_els_fpin *fpin = (struct fc_els_fpin *)fpin_buf;
> + struct fc_tlv_desc *tlv;
> + u32 bytes_remain;
> + u32 dtag;
> +
> + if (!localport)
> + return;
> + lport = localport_to_lport(localport);
> + tlv = (struct fc_tlv_desc *)&fpin->fpin_desc[0];
> + bytes_remain = fpin_len - offsetof(struct fc_els_fpin, fpin_desc);
> + bytes_remain = min_t(u32, bytes_remain, be32_to_cpu(fpin->desc_len));
> +
> + while (bytes_remain >= FC_TLV_DESC_HDR_SZ &&
> + bytes_remain >= FC_TLV_DESC_SZ_FROM_LENGTH(tlv)) {
> + dtag = be32_to_cpu(tlv->desc_tag);
> + switch (dtag) {
> + case ELS_DTAG_LNK_INTEGRITY:
> + nvme_fc_fpin_li_lport_update(lport, tlv);
> + break;
> + default:
> + break;
> + }
> +
> + bytes_remain -= FC_TLV_DESC_SZ_FROM_LENGTH(tlv);
> + tlv = fc_tlv_next_desc(tlv);
> + }
> +}
> +EXPORT_SYMBOL(nvme_fc_fpin_rcv);
> +
> /* Arbitrary successive failures max. With lots of subsystems could be high */
> #define DISCOVERY_MAX_FAIL 20
>
> diff --git a/drivers/nvme/host/multipath.c b/drivers/nvme/host/multipath.c
> index 4a90dad43303..e0dcfe9edb89 100644
> --- a/drivers/nvme/host/multipath.c
> +++ b/drivers/nvme/host/multipath.c
> @@ -235,6 +235,8 @@ static bool nvme_path_is_disabled(struct nvme_ns *ns)
> */
> if (state != NVME_CTRL_LIVE && state != NVME_CTRL_DELETING)
> return true;
> + if (test_bit(NVME_CTRL_TRANSPORT_BLOCKED, &ns->ctrl->flags))
> + return true;
> if (test_bit(NVME_NS_ANA_PENDING, &ns->flags) ||
> !test_bit(NVME_NS_READY, &ns->flags))
> return true;
> diff --git a/drivers/nvme/host/nvme.h b/drivers/nvme/host/nvme.h
> index ee30bb63e36b..6ed2ca6b35e4 100644
> --- a/drivers/nvme/host/nvme.h
> +++ b/drivers/nvme/host/nvme.h
> @@ -258,6 +258,7 @@ enum nvme_ctrl_flags {
> NVME_CTRL_SKIP_ID_CNS_CS = 4,
> NVME_CTRL_DIRTY_CAPABILITY = 5,
> NVME_CTRL_FROZEN = 6,
> + NVME_CTRL_TRANSPORT_BLOCKED = 7,
> };
>
> struct nvme_ctrl {
> diff --git a/drivers/scsi/lpfc/lpfc_els.c b/drivers/scsi/lpfc/lpfc_els.c
> index 4d723200690a..ecfe6bc8ab63 100644
> --- a/drivers/scsi/lpfc/lpfc_els.c
> +++ b/drivers/scsi/lpfc/lpfc_els.c
> @@ -33,6 +33,7 @@
> #include <scsi/scsi_transport_fc.h>
> #include <uapi/scsi/fc/fc_fs.h>
> #include <uapi/scsi/fc/fc_els.h>
> +#include <linux/nvme-fc-driver.h>
>
> #include "lpfc_hw4.h"
> #include "lpfc_hw.h"
> @@ -10343,9 +10344,12 @@ lpfc_els_rcv_fpin(struct lpfc_vport *vport, void *p, u32 fpin_length)
> fpin_length += sizeof(struct fc_els_fpin); /* the entire FPIN */
>
> /* Send every descriptor individually to the upper layer */
> - if (deliver)
> + if (deliver) {
> fc_host_fpin_rcv(lpfc_shost_from_vport(vport),
> fpin_length, (char *)fpin, 0);
> + nvme_fc_fpin_rcv(vport->localport,
> + fpin_length, (char *)fpin);
> + }
> desc_cnt++;
> }
> }
> diff --git a/drivers/scsi/qla2xxx/qla_isr.c b/drivers/scsi/qla2xxx/qla_isr.c
> index d48007e18288..b180e10053c5 100644
> --- a/drivers/scsi/qla2xxx/qla_isr.c
> +++ b/drivers/scsi/qla2xxx/qla_isr.c
> @@ -46,6 +46,7 @@ qla27xx_process_purex_fpin(struct scsi_qla_host *vha, struct purex_item *item)
> pkt, pkt_size);
>
> fc_host_fpin_rcv(vha->host, pkt_size, (char *)pkt, 0);
> + nvme_fc_fpin_rcv(vha->nvme_local_port, pkt_size, (char *)pkt);
> }
>
> const char *const port_state_str[] = {
> diff --git a/include/linux/nvme-fc-driver.h b/include/linux/nvme-fc-driver.h
> index 4109f1bd6128..994bc459e6dd 100644
> --- a/include/linux/nvme-fc-driver.h
> +++ b/include/linux/nvme-fc-driver.h
> @@ -536,6 +536,9 @@ void nvme_fc_rescan_remoteport(struct nvme_fc_remote_port *remoteport);
> int nvme_fc_set_remoteport_devloss(struct nvme_fc_remote_port *remoteport,
> u32 dev_loss_tmo);
>
> +void nvme_fc_fpin_rcv(struct nvme_fc_local_port *localport,
> + u32 fpin_len, char *fpin_buf);
> +
> /*
> * Routine called to pass a NVME-FC LS request, received by the lldd,
> * to the nvme-fc transport.
next prev parent reply other threads:[~2024-03-07 10:11 UTC|newest]
Thread overview: 8+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-02-19 8:59 [PATCH RFC] nvme-fc: FPIN link integrity handling hare
2024-03-07 10:10 ` Sagi Grimberg [this message]
2024-03-07 11:29 ` Hannes Reinecke
2024-03-07 12:01 ` Sagi Grimberg
2024-03-07 12:13 ` Hannes Reinecke
2024-03-08 10:38 ` Sagi Grimberg
2024-03-08 14:09 ` Hannes Reinecke
2024-03-08 14:19 ` Sagi Grimberg
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1dfa9a4e-e4a2-4d48-b569-85e48ce4311c@grimberg.me \
--to=sagi@grimberg.me \
--cc=hare@kernel.org \
--cc=hare@suse.de \
--cc=hch@lst.de \
--cc=james.smart@broadcom.com \
--cc=kbusch@kernel.org \
--cc=linux-nvme@lists.infradead.org \
--cc=linux-scsi@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox