|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] 3.16-rc3: PV guest INFO: possible recursive locking detected netback_changed+0xecb/0x1100
Hi,
In one of my PV guest i got the lockdep warning below on boot.
--
Sander
[ 0.839799] =============================================
[ 0.839803] [ INFO: possible recursive locking detected ]
[ 0.839808] 3.16.0-rc3-20140630+ #1 Not tainted
[ 0.839812] ---------------------------------------------
[ 0.839817] xenwatch/24 is trying to acquire lock:
[ 0.839821] (&(&queue->rx_lock)->rlock){+.....}, at: [<ffffffff8173c72b>]
netback_changed+0xecb/0x1100
[ 0.839835]
[ 0.839835] but task is already holding lock:
[ 0.839841] (&(&queue->rx_lock)->rlock){+.....}, at: [<ffffffff8173c72b>]
netback_changed+0xecb/0x1100
[ 0.839851]
[ 0.839851] other info that might help us debug this:
[ 0.839856] Possible unsafe locking scenario:
[ 0.839856]
[ 0.839861] CPU0
[ 0.839864] ----
[ 0.839866] lock(&(&queue->rx_lock)->rlock);
[ 0.839871] lock(&(&queue->rx_lock)->rlock);
[ 0.839876]
[ 0.839876] *** DEADLOCK ***
[ 0.839876]
[ 0.839881] May be due to missing lock nesting notation
[ 0.839881]
[ 0.839886] 3 locks held by xenwatch/24:
[ 0.839889] #0: (xenwatch_mutex){+.+.+.}, at: [<ffffffff8151726b>]
xenwatch_thread+0x3b/0x160
[ 0.839901] #1: (&(&queue->rx_lock)->rlock){+.....}, at:
[<ffffffff8173c72b>] netback_changed+0xecb/0x1100
[ 0.839912] #2: (&(&queue->tx_lock)->rlock){......}, at:
[<ffffffff8173c738>] netback_changed+0xed8/0x1100
[ 0.839923]
[ 0.839923] stack backtrace:
[ 0.839928] CPU: 0 PID: 24 Comm: xenwatch Not tainted 3.16.0-rc3-20140630+ #1
[ 0.839934] ffffffff8292a4f0 ffff88002e713b78 ffffffff81b59387
ffff88002e7090c0
[ 0.839942] ffffffff8292a4f0 ffff88002e713c58 ffffffff8110284e
ffff88002e7090c0
[ 0.839950] 0000000000023700 000000000000014f 0000000000051990
ffffffff8306f7e0
[ 0.839959] Call Trace:
[ 0.843068] [<ffffffff81b59387>] dump_stack+0x46/0x58
[ 0.843068] [<ffffffff8110284e>] __lock_acquire+0x84e/0x2220
[ 0.843068] [<ffffffff8110055a>] ? mark_held_locks+0x6a/0x90
[ 0.843068] [<ffffffff81b61895>] ? __mutex_unlock_slowpath+0x1b5/0x270
[ 0.843068] [<ffffffff8173002b>] ? rtl_apply_firmware+0x16b/0x2b0
[ 0.843068] [<ffffffff8110483d>] lock_acquire+0xcd/0x110
[ 0.843068] [<ffffffff8173c72b>] ? netback_changed+0xecb/0x1100
[ 0.843068] [<ffffffff81b61a09>] ? mutex_unlock+0x9/0xb
[ 0.843068] [<ffffffff81b62f1a>] _raw_spin_lock_bh+0x3a/0x50
[ 0.843068] [<ffffffff8173c72b>] ? netback_changed+0xecb/0x1100
[ 0.843068] [<ffffffff8173c72b>] netback_changed+0xecb/0x1100
[ 0.843068] [<ffffffff81518710>] xenbus_otherend_changed+0xb0/0xc0
[ 0.843068] [<ffffffff81517230>] ? xenbus_thread+0x290/0x290
[ 0.843068] [<ffffffff8151a34e>] backend_changed+0xe/0x10
[ 0.843068] [<ffffffff815172d1>] xenwatch_thread+0xa1/0x160
[ 0.843068] [<ffffffff810fadc0>] ? __init_waitqueue_head+0x60/0x60
[ 0.843068] [<ffffffff810dfeef>] kthread+0xdf/0x100
[ 0.843068] [<ffffffff810dfe10>] ? __init_kthread_worker+0x70/0x70
[ 0.843068] [<ffffffff81b63afc>] ret_from_fork+0x7c/0xb0
[ 0.843068] [<ffffffff810dfe10>] ? __init_kthread_worker+0x70/0x70
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |