From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1756334Ab0HAIra (ORCPT ); Sun, 1 Aug 2010 04:47:30 -0400 Received: from mx1.redhat.com ([209.132.183.28]:49824 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1756093Ab0HAIr1 (ORCPT ); Sun, 1 Aug 2010 04:47:27 -0400 Date: Sun, 1 Aug 2010 11:41:36 +0300 From: "Michael S. Tsirkin" To: Tejun Heo Cc: "David S. Miller" , Sridhar Samudrala , Jeff Dike , Juan Quintela , Rusty Russell , Takuya Yoshikawa , David Stevens , "Paul E. McKenney" , kvm@vger.kernel.org, virtualization@lists.osdl.org, netdev@vger.kernel.org, linux-kernel@vger.kernel.org Subject: Re: [PATCH] vhost: locking/rcu cleanup Message-ID: <20100801084136.GC16158@redhat.com> References: <20100729122325.GA24337@redhat.com> <4C52E692.3070405@kernel.org> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <4C52E692.3070405@kernel.org> User-Agent: Mutt/1.5.20 (2009-12-10) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Jul 30, 2010 at 04:49:54PM +0200, Tejun Heo wrote: > Hello, > > On 07/29/2010 02:23 PM, Michael S. Tsirkin wrote: > > I saw WARN_ON(!list_empty(&dev->work_list)) trigger > > so our custom flush is not as airtight as need be. > > Could be but it's also possible that something has queued something > after the last flush? > Is the problem reproducible? Well, We do requeue from the job itself. So need to be careful with what we do with indexes here. Bug seemed to happen all the time when qemu was killed under stress but now I can't reproduce anymore :( Will try again later. > > This patch switches to a simple atomic counter + srcu instead of > > the custom locked queue + flush implementation. > > > > This will slow down the setup ioctls, which should not matter - > > it's slow path anyway. We use the expedited flush to at least > > make sure it has a sane time bound. > > > > Works fine for me. I got reports that with many guests, > > work lock is highly contended, and this patch should in theory > > fix this as well - but I haven't tested this yet. > > Hmmm... vhost_poll_flush() becomes synchronize_srcu_expedited(). Can > you please explain how it works? synchronize_srcu_expedited() is an > extremely heavy operation involving scheduling the cpu_stop task on > all cpus. I'm not quite sure whether doing it from every flush is a > good idea. Is flush supposed to be a very rare operation? It is rare - happens on guest reboot typically. I guess I will switch to regular synchronize_srcu. > Having custom implementation is fine too but let's try to implement > something generic if at all possible. > > Thanks. Sure. It does seem that avoiding list lock would be pretty hard in generic code though. > -- > tejun