From: Potnuri Bharat Teja <bharat@chelsio.com>
To: Sagi Grimberg <sagi@grimberg.me>
Cc: Samuel Jones <sjones@kalrayinc.com>, "hch@lst.de" <hch@lst.de>,
"linux-nvme@lists.infradead.org" <linux-nvme@lists.infradead.org>
Subject: Re: Request timeout seen with NVMEoF TCP
Date: Wed, 16 Dec 2020 11:21:23 +0530 [thread overview]
Message-ID: <X9mgW4AxRpQsetN1@chelsio.com> (raw)
In-Reply-To: <a204cb8e-6f3b-b5dc-9972-54651c9c75cc@grimberg.me>
On Monday, December 12/14/20, 2020 at 17:53:44 -0800, Sagi Grimberg wrote:
>
> > Hey Potnuri,
> >
> > Have you observed this further?
> >
> > I'd think that if the io_work reschedule itself when it races
> > with the direct send path this should not happen, but we may be
> > seeing a different race going on here, adding Samuel who saw
> > a similar phenomenon.
>
> I think we still have a race here with the following:
> 1. queue_rq sends h2cdata PDU (no data)
> 2. host receives r2t - prepares data PDU to send and schedules io_work
> 3. queue_rq sends another h2cdata PDU - ends up sending (2) because it was
> queued before it
> 4. io_work starts, loops but never able to acquire the send_mutex -
> eventually just ends (dosn't requeue)
> 5. (3) completes, now nothing will send (2)
>
> We can either schedule the io_work from the direct send path, but that
> is less efficient than just trying to drain the send queue in the
> direct send path and if not all was sent, the write_space callback
> will trigger it.
>
> Potnuri, does this patch solves what you are seeing?
Hi Sagi,
Below patch works fine. I have it running all night with out any issues.
Thanks.
> --
> diff --git a/drivers/nvme/host/tcp.c b/drivers/nvme/host/tcp.c
> index 1ba659927442..1b4e25624ba4 100644
> --- a/drivers/nvme/host/tcp.c
> +++ b/drivers/nvme/host/tcp.c
> @@ -262,6 +262,16 @@ static inline void nvme_tcp_advance_req(struct
> nvme_tcp_request *req,
> }
> }
>
> +static inline void nvme_tcp_send_all(struct nvme_tcp_queue *queue)
> +{
> + int ret;
> +
> + /* drain the send queue as much as we can... */
> + do {
> + ret = nvme_tcp_try_send(queue);
> + } while (ret > 0);
> +}
> +
> static inline void nvme_tcp_queue_request(struct nvme_tcp_request *req,
> bool sync, bool last)
> {
> @@ -279,7 +289,7 @@ static inline void nvme_tcp_queue_request(struct
> nvme_tcp_request *req,
> if (queue->io_cpu == smp_processor_id() &&
> sync && empty && mutex_trylock(&queue->send_mutex)) {
> queue->more_requests = !last;
> - nvme_tcp_try_send(queue);
> + nvme_tcp_send_all(queue);
> queue->more_requests = false;
> mutex_unlock(&queue->send_mutex);
> } else if (last) {
> @@ -1122,6 +1132,14 @@ static void nvme_tcp_io_work(struct work_struct *w)
> pending = true;
> else if (unlikely(result < 0))
> break;
> + } else {
> + /*
> + * submission path is sending, we need to
> + * continue or resched because the submission
> + * path direct send is not concerned with
> + * rescheduling...
> + */
> + pending = true;
> }
>
> result = nvme_tcp_try_recv(queue);
> --
_______________________________________________
Linux-nvme mailing list
Linux-nvme@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/linux-nvme
next prev parent reply other threads:[~2020-12-16 5:52 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-12-09 17:19 Request timeout seen with NVMEoF TCP Potnuri Bharat Teja
2020-12-09 18:02 ` Sagi Grimberg
2020-12-10 17:06 ` Potnuri Bharat Teja
2020-12-10 21:47 ` Sagi Grimberg
2020-12-11 7:26 ` Potnuri Bharat Teja
2020-12-14 17:51 ` Sagi Grimberg
2020-12-14 18:05 ` Potnuri Bharat Teja
2020-12-14 19:27 ` Samuel Jones
2020-12-14 20:13 ` Sagi Grimberg
2020-12-15 1:53 ` Sagi Grimberg
2020-12-15 18:30 ` Wunderlich, Mark
2020-12-19 1:04 ` Sagi Grimberg
2020-12-16 5:51 ` Potnuri Bharat Teja [this message]
2020-12-19 1:35 ` Sagi Grimberg
2020-12-21 7:50 ` Potnuri Bharat Teja
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=X9mgW4AxRpQsetN1@chelsio.com \
--to=bharat@chelsio.com \
--cc=hch@lst.de \
--cc=linux-nvme@lists.infradead.org \
--cc=sagi@grimberg.me \
--cc=sjones@kalrayinc.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).