From: "Michael S. Tsirkin" <mst@redhat.com>
To: Stefano Garzarella <sgarzare@redhat.com>
Cc: Jason Wang <jasowang@redhat.com>,
netdev@vger.kernel.org, Stefan Hajnoczi <stefanha@redhat.com>
Subject: Re: Question about IRQs during the .remove() of virtio-vsock driver
Date: Tue, 21 May 2019 09:56:42 -0400 [thread overview]
Message-ID: <20190521095206-mutt-send-email-mst@kernel.org> (raw)
In-Reply-To: <20190521134920.pulvy5pqnertbafd@steredhat>
On Tue, May 21, 2019 at 03:49:20PM +0200, Stefano Garzarella wrote:
> On Tue, May 21, 2019 at 06:05:31AM -0400, Michael S. Tsirkin wrote:
> > On Tue, May 21, 2019 at 11:44:07AM +0200, Stefano Garzarella wrote:
> > > Hi Micheal, Jason,
> > > as suggested by Stefan, I'm checking if we have some races in the
> > > virtio-vsock driver. We found some races in the .probe() and .remove()
> > > with the upper layer (socket) and I'll fix it.
> > >
> > > Now my attention is on the bottom layer (virtio device) and my question is:
> > > during the .remove() of virtio-vsock driver (virtio_vsock_remove), could happen
> > > that an IRQ comes and one of our callback (e.g. virtio_vsock_rx_done()) is
> > > executed, queueing new works?
> > >
> > > I tried to follow the code in both cases (device unplugged or module removed)
> > > and maybe it couldn't happen because we remove it from bus's knowledge,
> > > but I'm not sure and your advice would be very helpful.
> > >
> > > Thanks in advance,
> > > Stefano
> >
> >
> > Great question! This should be better documented: patches welcome!
>
> When I'm clear, I'll be happy to document this.
>
> >
> > Here's my understanding:
> >
> >
> > A typical removal flow works like this:
> >
> > - prevent linux from sending new kick requests to device
> > and flush such outstanding requests if any
> > (device can still send notifications to linux)
> >
> > - call
> > vi->vdev->config->reset(vi->vdev);
> > this will flush all device writes and interrupts.
> > device will not use any more buffers.
> > previously outstanding callbacks might still be active.
> >
> > - Then call
> > vdev->config->del_vqs(vdev);
> > to flush outstanding callbacks if any.
>
> Thanks for sharing these useful information.
>
> So, IIUC between step 1 (e.g. in virtio-vsock we flush all work-queues) and
> step 2, new IRQs could happen, and in the virtio-vsock driver new work
> will be queued.
>
> In order to handle this case, I'm thinking to add a new variable
> 'work_enabled' in the struct virtio_vsock, put it to false at the start
> of the .remove(), then call synchronize_rcu() before to flush all work
> queues
> and use an helper function virtio_transport_queue_work() to queue
> a new work, where the check of work_enabled and the queue_work are in the
> RCU read critical section.
>
> Here a pseudo code to explain better the idea:
>
> virtio_vsock_remove() {
> vsock->work_enabled = false;
>
> /* Wait for other CPUs to finish to queue works */
> synchronize_rcu();
>
> flush_works();
>
> vdev->config->reset(vdev);
>
> ...
>
> vdev->config->del_vqs(vdev);
> }
>
> virtio_vsock_queue_work(vsock, work) {
> rcu_read_lock();
>
> if (!vsock->work_enabled) {
> goto out;
> }
>
> queue_work(virtio_vsock_workqueue, work);
>
> out:
> rcu_read_unlock();
> }
>
>
> Do you think can work?
> Please tell me if there is a better way to handle this case.
>
> Thanks,
> Stefano
instead of rcu tricks I would just have rx_run and tx_run and check it
within the queued work - presumably under tx or rx lock.
then queueing an extra work becomes harmless,
and you flush it after del vqs which flushes everything for you.
next prev parent reply other threads:[~2019-05-21 13:56 UTC|newest]
Thread overview: 7+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-05-21 9:44 Question about IRQs during the .remove() of virtio-vsock driver Stefano Garzarella
2019-05-21 10:05 ` Michael S. Tsirkin
2019-05-21 13:49 ` Stefano Garzarella
2019-05-21 13:56 ` Michael S. Tsirkin [this message]
2019-05-21 14:12 ` Stefano Garzarella
2019-05-22 3:44 ` Jason Wang
2019-05-23 13:56 ` Stefano Garzarella
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20190521095206-mutt-send-email-mst@kernel.org \
--to=mst@redhat.com \
--cc=jasowang@redhat.com \
--cc=netdev@vger.kernel.org \
--cc=sgarzare@redhat.com \
--cc=stefanha@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).