From: Zoltan Kiss <zoltan.kiss@citrix.com>
To: Jiri Pirko <jiri@resnulli.us>, <netdev@vger.kernel.org>
Cc: <jesse@nicira.com>, <dev@openvswitch.org>
Subject: Re: ovs inconsistent lock state
Date: Fri, 28 Feb 2014 19:21:35 +0000 [thread overview]
Message-ID: <5310E1BF.5060803@citrix.com> (raw)
In-Reply-To: <20140213171307.GC2829@minipsycho.orion>
I also had problems with these locks on current net-next, I believe the
issue is that "datapath: Fix deadlock during stats update." is not
merged from official OVS repo to kernel.
Zoli
On 13/02/14 17:13, Jiri Pirko wrote:
> Hi.
>
> On current net-next I'm getting following message once I add a dp:
>
> [ 3014.523665] =================================
> [ 3014.524118] [ INFO: inconsistent lock state ]
> [ 3014.524118] 3.14.0-rc2+ #1 Not tainted
> [ 3014.524118] ---------------------------------
> [ 3014.524118] inconsistent {SOFTIRQ-ON-W} -> {IN-SOFTIRQ-W} usage.
> [ 3014.524118] swapper/1/0 [HC0[0]:SC1[5]:HE1:SE0] takes:
> [ 3014.524118] (&(&flow->stats.stat->lock)->rlock){+.?...}, at: [<ffffffffa021b6ef>] ovs_flow_stats_update+0x4f/0xd0 [openvswitch]
> [ 3014.524118] {SOFTIRQ-ON-W} state was registered at:
> [ 3014.524118] [<ffffffff810d6843>] __lock_acquire+0x5a3/0x1c30
> [ 3014.524118] [<ffffffff810d7f80>] lock_acquire+0xb0/0x150
> [ 3014.524118] [<ffffffff81705c99>] _raw_spin_lock+0x39/0x50
> [ 3014.524118] [<ffffffffa021b8d3>] ovs_flow_stats_get+0x163/0x1b0 [openvswitch]
> [ 3014.524118] [<ffffffffa0218ce5>] ovs_flow_cmd_fill_info+0x165/0x2b0 [openvswitch]
> [ 3014.524118] [<ffffffffa021904c>] ovs_flow_cmd_build_info.constprop.27+0x6c/0xa0 [openvswitch]
> [ 3014.524118] [<ffffffffa0219519>] ovs_flow_cmd_new_or_set+0x499/0x4f0 [openvswitch]
> [ 3014.524118] [<ffffffff8161bf09>] genl_family_rcv_msg+0x199/0x390
> [ 3014.524118] [<ffffffff8161c18e>] genl_rcv_msg+0x8e/0xd0
> [ 3014.524118] [<ffffffff8161a189>] netlink_rcv_skb+0xa9/0xc0
> [ 3014.524118] [<ffffffff8161a6a8>] genl_rcv+0x28/0x40
> [ 3014.524118] [<ffffffff816197d0>] netlink_unicast+0xf0/0x1b0
> [ 3014.524118] [<ffffffff81619b8f>] netlink_sendmsg+0x2ff/0x740
> [ 3014.524118] [<ffffffff815ce86b>] sock_sendmsg+0x8b/0xc0
> [ 3014.524118] [<ffffffff815cf369>] ___sys_sendmsg+0x389/0x3a0
> [ 3014.524118] [<ffffffff815cfa62>] __sys_sendmsg+0x42/0x80
> [ 3014.524118] [<ffffffff815cfab2>] SyS_sendmsg+0x12/0x20
> [ 3014.524118] [<ffffffff8170f829>] system_call_fastpath+0x16/0x1b
> [ 3014.524118] irq event stamp: 6163288
> [ 3014.524118] hardirqs last enabled at (6163288): [<ffffffff8169f8a0>] ip6_finish_output2+0x380/0x670
> [ 3014.524118] hardirqs last disabled at (6163287): [<ffffffff8169f85b>] ip6_finish_output2+0x33b/0x670
> [ 3014.524118] softirqs last enabled at (6163266): [<ffffffff810888a2>] _local_bh_enable+0x22/0x50
> [ 3014.524118] softirqs last disabled at (6163267): [<ffffffff81089f15>] irq_exit+0xc5/0xd0
> [ 3014.524118]
> other info that might help us debug this:
> [ 3014.524118] Possible unsafe locking scenario:
>
> [ 3014.524118] CPU0
> [ 3014.524118] ----
> [ 3014.524118] lock(&(&flow->stats.stat->lock)->rlock);
> [ 3014.524118] <Interrupt>
> [ 3014.524118] lock(&(&flow->stats.stat->lock)->rlock);
> [ 3014.524118]
> *** DEADLOCK ***
>
> [ 3014.524118] 5 locks held by swapper/1/0:
> [ 3014.524118] #0: (((&idev->mc_ifc_timer))){+.-...}, at: [<ffffffff81090095>] call_timer_fn+0x5/0x180
> [ 3014.524118] #1: (rcu_read_lock){.+.+..}, at: [<ffffffff816c5755>] mld_sendpack+0x5/0x330
> [ 3014.524118] #2: (rcu_read_lock_bh){.+....}, at: [<ffffffff8169f579>] ip6_finish_output2+0x59/0x670
> [ 3014.524118] #3: (rcu_read_lock_bh){.+....}, at: [<ffffffff815eb585>] __dev_queue_xmit+0x5/0x6a0
> [ 3014.524118] #4: (rcu_read_lock){.+.+..}, at: [<ffffffffa0221995>] internal_dev_xmit+0x5/0x90 [openvswitch]
> [ 3014.524118]
> stack backtrace:
> [ 3014.524118] CPU: 1 PID: 0 Comm: swapper/1 Not tainted 3.14.0-rc2+ #1
> [ 3014.524118] Hardware name: Bochs Bochs, BIOS Bochs 01/01/2011
> [ 3014.524118] ffffffff82338940 ffff8800bb2838f0 ffffffff816fd1ac ffff8800b6b33570
> [ 3014.524118] ffff8800bb283940 ffffffff816f9144 0000000000000005 ffff880000000001
> [ 3014.524118] ffff880000000000 0000000000000006 ffff8800b6b33570 ffffffff810d4380
> [ 3014.524118] Call Trace:
> [ 3014.524118] <IRQ> [<ffffffff816fd1ac>] dump_stack+0x4d/0x66
> [ 3014.524118] [<ffffffff816f9144>] print_usage_bug+0x1f3/0x204
> [ 3014.524118] [<ffffffff810d4380>] ? check_usage_backwards+0x130/0x130
> [ 3014.524118] [<ffffffff810d4da0>] mark_lock+0x270/0x300
> [ 3014.524118] [<ffffffff810d6718>] __lock_acquire+0x478/0x1c30
> [ 3014.524118] [<ffffffff810d4ede>] ? mark_held_locks+0xae/0x130
> [ 3014.524118] [<ffffffffa0220047>] ? find_bucket.isra.1+0x67/0x70 [openvswitch]
> [ 3014.524118] [<ffffffffa0220228>] ? masked_flow_lookup+0x78/0x140 [openvswitch]
> [ 3014.524118] [<ffffffff810d669a>] ? __lock_acquire+0x3fa/0x1c30
> [ 3014.524118] [<ffffffff810d7f80>] lock_acquire+0xb0/0x150
> [ 3014.524118] [<ffffffffa021b6ef>] ? ovs_flow_stats_update+0x4f/0xd0 [openvswitch]
> [ 3014.524118] [<ffffffff81705c99>] _raw_spin_lock+0x39/0x50
> [ 3014.524118] [<ffffffffa021b6ef>] ? ovs_flow_stats_update+0x4f/0xd0 [openvswitch]
> [ 3014.524118] [<ffffffffa021b6ef>] ovs_flow_stats_update+0x4f/0xd0 [openvswitch]
> [ 3014.524118] [<ffffffffa021ac75>] ovs_dp_process_received_packet+0x75/0x100 [openvswitch]
> [ 3014.524118] [<ffffffffa02213ea>] ovs_vport_receive+0x2a/0x30 [openvswitch]
> [ 3014.524118] [<ffffffffa02219e6>] internal_dev_xmit+0x56/0x90 [openvswitch]
> [ 3014.524118] [<ffffffffa0221995>] ? internal_dev_xmit+0x5/0x90 [openvswitch]
> [ 3014.524118] [<ffffffff815eb276>] dev_hard_start_xmit+0x316/0x620
> [ 3014.524118] [<ffffffff815eb8ff>] __dev_queue_xmit+0x37f/0x6a0
> [ 3014.524118] [<ffffffff815eb585>] ? __dev_queue_xmit+0x5/0x6a0
> [ 3014.524118] [<ffffffff8169f8a0>] ? ip6_finish_output2+0x380/0x670
> [ 3014.524118] [<ffffffff815ebc30>] dev_queue_xmit+0x10/0x20
> [ 3014.524118] [<ffffffff8169f900>] ip6_finish_output2+0x3e0/0x670
> [ 3014.524118] [<ffffffff816a31aa>] ? ip6_finish_output+0x9a/0x200
> [ 3014.524118] [<ffffffff816a31aa>] ip6_finish_output+0x9a/0x200
> [ 3014.524118] [<ffffffff816a335f>] ip6_output+0x4f/0x1d0
> [ 3014.524118] [<ffffffff816c58e6>] mld_sendpack+0x196/0x330
> [ 3014.524118] [<ffffffff816c5755>] ? mld_sendpack+0x5/0x330
> [ 3014.524118] [<ffffffff816c63cd>] mld_ifc_timer_expire+0x19d/0x2d0
> [ 3014.524118] [<ffffffff816c6230>] ? mld_dad_timer_expire+0x70/0x70
> [ 3014.524118] [<ffffffff8109010a>] call_timer_fn+0x7a/0x180
> [ 3014.524118] [<ffffffff81090095>] ? call_timer_fn+0x5/0x180
> [ 3014.524118] [<ffffffff816c6230>] ? mld_dad_timer_expire+0x70/0x70
> [ 3014.524118] [<ffffffff8109054c>] run_timer_softirq+0x20c/0x2c0
> [ 3014.524118] [<ffffffff810899fd>] __do_softirq+0x12d/0x310
> [ 3014.524118] [<ffffffff81089f15>] irq_exit+0xc5/0xd0
> [ 3014.524118] [<ffffffff81711b75>] smp_apic_timer_interrupt+0x45/0x60
> [ 3014.524118] [<ffffffff817104b2>] apic_timer_interrupt+0x72/0x80
> [ 3014.524118] <EOI> [<ffffffff81052486>] ? native_safe_halt+0x6/0x10
> [ 3014.524118] [<ffffffff8101ec04>] default_idle+0x24/0xe0
> [ 3014.524118] [<ffffffff8101f54e>] arch_cpu_idle+0x2e/0x40
> [ 3014.524118] [<ffffffff810e96ae>] cpu_startup_entry+0x9e/0x280
> [ 3014.524118] [<ffffffff81044929>] start_secondary+0x1d9/0x280
> --
> To unsubscribe from this list: send the line "unsubscribe netdev" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
>
next prev parent reply other threads:[~2014-02-28 19:21 UTC|newest]
Thread overview: 9+ messages / expand[flat|nested] mbox.gz Atom feed top
2014-02-13 17:13 ovs inconsistent lock state Jiri Pirko
2014-02-13 17:42 ` Pravin Shelar
2014-02-13 18:12 ` Jiri Pirko
2014-02-13 19:21 ` Pravin Shelar
2014-02-13 19:26 ` David Miller
2014-02-13 20:07 ` Jiri Pirko
[not found] ` <20140213200753.GE2829-RDzucLLXGGI88b5SBfVpbw@public.gmane.org>
2014-02-13 20:23 ` David Miller
2014-02-28 19:21 ` Zoltan Kiss [this message]
2014-03-03 11:33 ` Jiri Pirko
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=5310E1BF.5060803@citrix.com \
--to=zoltan.kiss@citrix.com \
--cc=dev@openvswitch.org \
--cc=jesse@nicira.com \
--cc=jiri@resnulli.us \
--cc=netdev@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).