From: Pankaj Gupta <pagupta@redhat.com>
To: "Jakub Staroń" <jstaron@google.com>
Cc: linux-nvdimm@lists.01.org, linux-kernel@vger.kernel.org,
virtualization@lists.linux-foundation.org, kvm@vger.kernel.org,
linux-fsdevel@vger.kernel.org, linux-acpi@vger.kernel.org,
qemu-devel@nongnu.org, linux-ext4@vger.kernel.org,
linux-xfs@vger.kernel.org, jack@suse.cz, mst@redhat.com,
jasowang@redhat.com, david@fromorbit.com, lcapitulino@redhat.com,
adilger kernel <adilger.kernel@dilger.ca>,
zwisler@kernel.org, aarcange@redhat.com,
dave jiang <dave.jiang@intel.com>,
darrick wong <darrick.wong@oracle.com>,
vishal l verma <vishal.l.verma@intel.com>,
david@redhat.com, willy@infradead.org, hch@infradead.org,
jmoyer@redhat.com, nilal@redhat.com, lenb@kernel.org,
kilobyte@angband.pl, riel@surriel.com,
yuval shaia <yuval.shaia@oracle.com>,
stefanha@redhat.com, pbonzini@redhat.com,
dan j williams <dan.j.williams@intel.com>,
kwolf@redhat.com, tytso@mit.edu,
xiaoguangrong eric <xiaoguangrong.eric@gmail.com>,
cohuck@redhat.com, rjw@rjwysocki.net, imammedo@redhat.com,
smbarber@google.com
Subject: Re: [Qemu-devel] [PATCH v7 2/6] virtio-pmem: Add virtio pmem driver
Date: Wed, 8 May 2019 07:12:47 -0400 (EDT) [thread overview]
Message-ID: <1555943483.27247564.1557313967518.JavaMail.zimbra@redhat.com> (raw)
In-Reply-To: <3d6479ae-6c39-d614-f1d9-aa1978e2e438@google.com>
>
> On 4/25/19 10:00 PM, Pankaj Gupta wrote:
>
> > +void host_ack(struct virtqueue *vq)
> > +{
> > + unsigned int len;
> > + unsigned long flags;
> > + struct virtio_pmem_request *req, *req_buf;
> > + struct virtio_pmem *vpmem = vq->vdev->priv;
> > +
> > + spin_lock_irqsave(&vpmem->pmem_lock, flags);
> > + while ((req = virtqueue_get_buf(vq, &len)) != NULL) {
> > + req->done = true;
> > + wake_up(&req->host_acked);
> > +
> > + if (!list_empty(&vpmem->req_list)) {
> > + req_buf = list_first_entry(&vpmem->req_list,
> > + struct virtio_pmem_request, list);
> > + list_del(&vpmem->req_list);
>
> Shouldn't it be rather `list_del(vpmem->req_list.next)`? We are trying to
> unlink
> first element of the list and `vpmem->req_list` is just the list head.
This looks correct. We are not deleting head but first entry in 'req_list'
which is device corresponding list of pending requests.
Please see below:
/**
* Retrieve the first list entry for the given list pointer.
*
* Example:
* struct foo *first;
* first = list_first_entry(&bar->list_of_foos, struct foo, list_of_foos);
*
* @param ptr The list head
* @param type Data type of the list element to retrieve
* @param member Member name of the struct list_head field in the list element.
* @return A pointer to the first list element.
*/
#define list_first_entry(ptr, type, member) \
list_entry((ptr)->next, type, member)
>
> > +int virtio_pmem_flush(struct nd_region *nd_region)
> > +{
> > + int err;
> > + unsigned long flags;
> > + struct scatterlist *sgs[2], sg, ret;
> > + struct virtio_device *vdev = nd_region->provider_data;
> > + struct virtio_pmem *vpmem = vdev->priv;
> > + struct virtio_pmem_request *req;
> > +
> > + might_sleep();
> > + req = kmalloc(sizeof(*req), GFP_KERNEL);
> > + if (!req)
> > + return -ENOMEM;
> > +
> > + req->done = req->wq_buf_avail = false;
> > + strcpy(req->name, "FLUSH");
> > + init_waitqueue_head(&req->host_acked);
> > + init_waitqueue_head(&req->wq_buf);
> > + sg_init_one(&sg, req->name, strlen(req->name));
> > + sgs[0] = &sg;
> > + sg_init_one(&ret, &req->ret, sizeof(req->ret));
> > + sgs[1] = &ret;
> > +
> > + spin_lock_irqsave(&vpmem->pmem_lock, flags);
> > + err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req, GFP_ATOMIC);
> > + if (err) {
> > + dev_err(&vdev->dev, "failed to send command to virtio pmem device\n");
> > +
> > + list_add_tail(&vpmem->req_list, &req->list);
> > + spin_unlock_irqrestore(&vpmem->pmem_lock, flags);
> > +
> > + /* When host has read buffer, this completes via host_ack */
> > + wait_event(req->wq_buf, req->wq_buf_avail);
> > + spin_lock_irqsave(&vpmem->pmem_lock, flags);
> > + }
>
> Aren't the arguments in `list_add_tail` swapped? The element we are adding
No, this is intentional. 'vpmem->req_list' maintains a list of pending requests
for entire pmem device. 'req->list'is per request list and maintains pending
request on virtio queue add failure. I think we don't need this list.
> should
> be first, the list should be second. Also, shouldn't we resubmit the request
> after
> waking up from `wait_event(req->wq_buf, req->wq_buf_avail)`?
Yes. we should. Good point.
>
> I propose rewriting it like that:
>
> diff --git a/drivers/nvdimm/virtio_pmem.c b/drivers/nvdimm/virtio_pmem.c
> index 66b582f751a3..ff0556b04e86 100644
> --- a/drivers/nvdimm/virtio_pmem.c
> +++ b/drivers/nvdimm/virtio_pmem.c
> @@ -25,7 +25,7 @@ void host_ack(struct virtqueue *vq)
> if (!list_empty(&vpmem->req_list)) {
> req_buf = list_first_entry(&vpmem->req_list,
> struct virtio_pmem_request, list);
> - list_del(&vpmem->req_list);
> + list_del(vpmem->req_list.next);
Don't think its correct.
> req_buf->wq_buf_avail = true;
> wake_up(&req_buf->wq_buf);
> }
> @@ -59,17 +59,33 @@ int virtio_pmem_flush(struct nd_region *nd_region)
> sgs[1] = &ret;
>
> spin_lock_irqsave(&vpmem->pmem_lock, flags);
> - err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req, GFP_ATOMIC);
> - if (err) {
> - dev_err(&vdev->dev, "failed to send command to virtio pmem device\n");
> + /*
> + * If virtqueue_add_sgs returns -ENOSPC then req_vq virtual queue does not
> + * have free descriptor slots. We add the request to req_list and wait
> + * for host_ack to wake us up when free slots are available.
> + */
> + while ((err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req, GFP_ATOMIC))
> == -ENOSPC) {
> + dev_err(&vdev->dev, "failed to send command to virtio pmem device, no free
> slots in the virtqueue, postponing request\n");
> + req->wq_buf_avail = false;
>
> - list_add_tail(&vpmem->req_list, &req->list);
> + list_add_tail(&req->list, &vpmem->req_list);
> spin_unlock_irqrestore(&vpmem->pmem_lock, flags);
>
> /* When host has read buffer, this completes via host_ack */
> wait_event(req->wq_buf, req->wq_buf_avail);
> spin_lock_irqsave(&vpmem->pmem_lock, flags);
> }
> +
> + /*
> + * virtqueue_add_sgs failed with error different than -ENOSPC, we can't
> + * do anything about that.
> + */
> + if (err) {
> + dev_info(&vdev->dev, "failed to send command to virtio pmem device, error
> code %d\n", err);
> + spin_unlock_irqrestore(&vpmem->pmem_lock, flags);
> + err = -EIO;
> + goto ret;
> + }
> err = virtqueue_kick(vpmem->req_vq);
> spin_unlock_irqrestore(&vpmem->pmem_lock, flags);
>
>
> Let me know if it looks reasonable to you.
Don't think this is fulfilling entire logic correctly. But thanks, I spotted a bug in my code :)
Will fix it.
>
> Thank you,
> Jakub Staron
>
>
next prev parent reply other threads:[~2019-05-08 11:12 UTC|newest]
Thread overview: 24+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-04-26 5:00 [PATCH v7 0/6] virtio pmem driver Pankaj Gupta
2019-04-26 5:00 ` [PATCH v7 1/6] libnvdimm: nd_region flush callback support Pankaj Gupta
2019-04-26 5:00 ` [PATCH v7 2/6] virtio-pmem: Add virtio pmem driver Pankaj Gupta
2019-04-30 5:53 ` [Qemu-devel] " Yuval Shaia
2019-04-30 6:06 ` Pankaj Gupta
2019-05-07 15:35 ` Dan Williams
2019-05-08 11:19 ` [Qemu-devel] " Pankaj Gupta
2019-05-10 23:33 ` Dan Williams
2019-05-11 1:26 ` Pankaj Gupta
2019-05-07 20:25 ` Jakub Staroń
2019-05-08 11:12 ` Pankaj Gupta [this message]
2019-05-08 15:23 ` Pankaj Gupta
2019-05-08 19:05 ` Jakub Staroń
2019-04-26 5:00 ` [PATCH v7 3/6] libnvdimm: add dax_dev sync flag Pankaj Gupta
2019-05-07 15:40 ` Dan Williams
2019-05-09 12:24 ` Pankaj Gupta
2019-04-26 5:00 ` [PATCH v7 4/6] dax: check synchronous mapping is supported Pankaj Gupta
2019-05-07 19:24 ` [Qemu-devel] " Jakub Staroń
2019-05-08 5:31 ` Pankaj Gupta
2019-04-26 5:00 ` [PATCH v7 5/6] ext4: disable map_sync for async flush Pankaj Gupta
2019-04-26 5:00 ` [PATCH v7 6/6] xfs: " Pankaj Gupta
2019-05-07 15:37 ` Dan Williams
2019-05-07 16:17 ` Darrick J. Wong
2019-05-08 5:49 ` [Qemu-devel] " Pankaj Gupta
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1555943483.27247564.1557313967518.JavaMail.zimbra@redhat.com \
--to=pagupta@redhat.com \
--cc=aarcange@redhat.com \
--cc=adilger.kernel@dilger.ca \
--cc=cohuck@redhat.com \
--cc=dan.j.williams@intel.com \
--cc=darrick.wong@oracle.com \
--cc=dave.jiang@intel.com \
--cc=david@fromorbit.com \
--cc=david@redhat.com \
--cc=hch@infradead.org \
--cc=imammedo@redhat.com \
--cc=jack@suse.cz \
--cc=jasowang@redhat.com \
--cc=jmoyer@redhat.com \
--cc=jstaron@google.com \
--cc=kilobyte@angband.pl \
--cc=kvm@vger.kernel.org \
--cc=kwolf@redhat.com \
--cc=lcapitulino@redhat.com \
--cc=lenb@kernel.org \
--cc=linux-acpi@vger.kernel.org \
--cc=linux-ext4@vger.kernel.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-nvdimm@lists.01.org \
--cc=linux-xfs@vger.kernel.org \
--cc=mst@redhat.com \
--cc=nilal@redhat.com \
--cc=pbonzini@redhat.com \
--cc=qemu-devel@nongnu.org \
--cc=riel@surriel.com \
--cc=rjw@rjwysocki.net \
--cc=smbarber@google.com \
--cc=stefanha@redhat.com \
--cc=tytso@mit.edu \
--cc=virtualization@lists.linux-foundation.org \
--cc=vishal.l.verma@intel.com \
--cc=willy@infradead.org \
--cc=xiaoguangrong.eric@gmail.com \
--cc=yuval.shaia@oracle.com \
--cc=zwisler@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).