[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Prepping a bugfix push
On Thursday, 03 December 2009 at 15:22, Jeremy Fitzhardinge wrote: > On 12/03/09 11:35, Brendan Cully wrote: > >Not a patch, but I've just tried out xm save -c again with the latest > >xen changes, and while I no longer see the grant table version panic, > >the guest's devices (aside from the console) appear to be wedged on > >resume. Is anyone else seeing this? > > > >After a while on the console I see messages like this: > > > >INFO: task syslogd:2219 blocked for more than 120 seconds. > > > >which I assume is trouble with the block device. > > Is there any backtrace or other info associated with this? I take it you couldn't reproduce? Let me know what you want captured. In the meantime, here's the kernel log: Dec 4 00:16:12 p32 kernel: suspending xenstore... Dec 4 00:16:12 p32 kernel: Grant tables using version 2 layout. Here's a backtrace after the block message: INFO: task kjournald:880 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. kjournald D 00000000 0 880 2 0x00000000 cfb6be70 00000246 cf8fa000 00000000 00000002 00000000 c14972a0 00000001 cf8fa000 c152f380 c152f380 c152f380 cf8fa1fc cf850b10 cf850c9c d0808380 c1924584 00000000 cfb6be70 c1006f04 c1920284 c1924588 c1920284 cfb49854 Call Trace: [<c1006f04>] ? check_events+0x8/0xc [<c137b0a8>] io_schedule+0x1e/0x28 [<c10c2555>] sync_buffer+0x33/0x37 [<c137b3a1>] __wait_on_bit+0x45/0x62 [<c10c2522>] ? sync_buffer+0x0/0x37 [<c137b41d>] out_of_line_wait_on_bit+0x5f/0x72 [<c10c2522>] ? sync_buffer+0x0/0x37 [<c1049b7e>] ? wake_bit_function+0x0/0x47 [<c10c24e5>] __wait_on_buffer+0x23/0x25 [<c1102940>] journal_commit_transaction+0x50b/0xdeb [<c1006efb>] ? xen_restore_fl_direct_end+0x0/0x1 [<c137c4f4>] ? _spin_unlock_irqrestore+0x17/0x1d [<c103fad4>] ? try_to_del_timer_sync+0x49/0x52 [<c1105b66>] kjournald+0xab/0x1f0 [<c1049b46>] ? autoremove_wake_function+0x0/0x38 [<c1105abb>] ? kjournald+0x0/0x1f0 [<c1049a85>] kthread+0x74/0x7f [<c1049a11>] ? kthread+0x0/0x7f [<c1009b37>] kernel_thread_helper+0x7/0x10 and here's the result of xm sysrq <dom> l SysRq : Show backtrace of all active CPUs and xm sysrq <dom> w SysRq : Show Blocked State task PC stack pid father pdflush D cf40297c 0 150 2 0x00000000 cf95dec8 00000246 00000000 cf40297c c1550280 c1553680 c10067df c1550280 00000000 c152f380 c152f380 c152f380 c1550280 cf8b3130 cf8b32bc d0808380 c1553680 ffff995d cf95dec8 c1006f04 c1550280 c1553a6c c1550280 c1006efb Call Trace: [<c10067df>] ? xen_force_evtchn_callback+0xf/0x14 [<c1006f04>] ? check_events+0x8/0xc [<c1006efb>] ? xen_restore_fl_direct_end+0x0/0x1 [<c137b1cb>] schedule_timeout+0xd6/0x13e [<c10067df>] ? xen_force_evtchn_callback+0xf/0x14 [<c103fc07>] ? process_timeout+0x0/0xa [<c137a8ea>] io_schedule_timeout+0x1e/0x28 [<c108e9a8>] congestion_wait+0x51/0x66 [<c1049b46>] ? autoremove_wake_function+0x0/0x38 [<c1083ef3>] wb_kupdate+0xb3/0xfa [<c10842d5>] ? pdflush+0x0/0x205 [<c10843d5>] pdflush+0x100/0x205 [<c1006efb>] ? xen_restore_fl_direct_end+0x0/0x1 [<c137c4f4>] ? _spin_unlock_irqrestore+0x17/0x1d [<c1083e40>] ? wb_kupdate+0x0/0xfa [<c1049a85>] kthread+0x74/0x7f [<c1049a11>] ? kthread+0x0/0x7f [<c1009b37>] kernel_thread_helper+0x7/0x10 kjournald D 00000000 0 880 2 0x00000000 cfb6be70 00000246 cf8fa000 00000000 00000002 00000000 c14972a0 00000001 cf8fa000 c152f380 c152f380 c152f380 cf8fa1fc cf850b10 cf850c9c d0808380 c1924584 00000000 cfb6be70 c1006f04 c1920284 c1924588 c1920284 cfb49854 Call Trace: [<c1006f04>] ? check_events+0x8/0xc [<c137b0a8>] io_schedule+0x1e/0x28 [<c10c2555>] sync_buffer+0x33/0x37 [<c137b3a1>] __wait_on_bit+0x45/0x62 [<c10c2522>] ? sync_buffer+0x0/0x37 [<c137b41d>] out_of_line_wait_on_bit+0x5f/0x72 [<c10c2522>] ? sync_buffer+0x0/0x37 [<c1049b7e>] ? wake_bit_function+0x0/0x47 [<c10c24e5>] __wait_on_buffer+0x23/0x25 [<c1102940>] journal_commit_transaction+0x50b/0xdeb [<c1006efb>] ? xen_restore_fl_direct_end+0x0/0x1 [<c137c4f4>] ? _spin_unlock_irqrestore+0x17/0x1d [<c103fad4>] ? try_to_del_timer_sync+0x49/0x52 [<c1105b66>] kjournald+0xab/0x1f0 [<c1049b46>] ? autoremove_wake_function+0x0/0x38 [<c1105abb>] ? kjournald+0x0/0x1f0 [<c1049a85>] kthread+0x74/0x7f [<c1049a11>] ? kthread+0x0/0x7f [<c1009b37>] kernel_thread_helper+0x7/0x10 syslogd D 00000000 0 2219 1 0x00000000 cfb33e40 00000286 00000000 00000000 0000aa70 00000000 c10067df cfb7026c 00000001 c152f380 c152f380 c152f380 00000000 cfa2a130 cfa2a2bc d0808380 cfb33e64 cfb70254 c1006efb c137c4f4 c10282ff cfb33e58 cfb33e40 c1049d05 Call Trace: [<c10067df>] ? xen_force_evtchn_callback+0xf/0x14 [<c1006efb>] ? xen_restore_fl_direct_end+0x0/0x1 [<c137c4f4>] ? _spin_unlock_irqrestore+0x17/0x1d [<c10282ff>] ? __wake_up+0x3a/0x42 [<c1049d05>] ? prepare_to_wait+0x41/0x46 [<c110535f>] log_wait_commit+0x87/0xef [<c1049b46>] ? autoremove_wake_function+0x0/0x38 [<c1100ca8>] journal_stop+0x1dc/0x29f [<c110156d>] ? journal_start+0x88/0xb2 [<c11015b4>] journal_force_commit+0x1d/0x1f [<c10f2d8a>] ext3_force_commit+0x1c/0x22 [<c10ed096>] ext3_write_inode+0x2d/0x3b [<c10bd21f>] writeback_single_inode+0x1f5/0x2be [<c108326f>] ? do_writepages+0x37/0x39 [<c10bdaa1>] sync_inode+0x1f/0x37 [<c10e9409>] ext3_sync_file+0x99/0xb0 [<c10c06fc>] vfs_fsync+0x73/0xb1 [<c10c075c>] do_fsync+0x22/0x32 [<c10c078b>] sys_fsync+0xd/0xf [<c1009058>] sysenter_do_call+0x12/0x2c getty D cf402840 0 2330 1 0x00000004 cfb23cc0 00000286 00000000 cf402840 00000000 00000000 00000000 00000000 00000000 c152f380 c152f380 c152f380 00000000 cf8740f0 cf87427c d0808380 cfb23d14 c19202c8 c1006efb c137c4f4 00000002 cfb23d08 cfb23cc0 c1049d05 Call Trace: [<c1006efb>] ? xen_restore_fl_direct_end+0x0/0x1 [<c137c4f4>] ? _spin_unlock_irqrestore+0x17/0x1d [<c1049d05>] ? prepare_to_wait+0x41/0x46 [<c1101016>] do_get_write_access+0x2ab/0x3cd [<c1006efb>] ? xen_restore_fl_direct_end+0x0/0x1 [<c1049b7e>] ? wake_bit_function+0x0/0x47 [<c1101153>] journal_get_write_access+0x1b/0x2a [<c10f72c2>] __ext3_journal_get_write_access+0x19/0x3f [<c10eab31>] ext3_reserve_inode_write+0x34/0x6a [<c10eab7e>] ext3_mark_inode_dirty+0x17/0x2e [<c10eda05>] ext3_dirty_inode+0x6b/0x6d [<c107ce9c>] ? file_remove_suid+0x1b/0x59 [<c10bd91d>] __mark_inode_dirty+0x23/0xd2 [<c10b6618>] file_update_time+0x76/0xb3 [<c107e0b9>] __generic_file_aio_write_nolock+0x235/0x4b3 [<c107f0e5>] ? filemap_fault+0x76/0x3b1 [<c107e5a5>] generic_file_aio_write+0x5d/0xd1 [<c10e9267>] ext3_file_write+0x27/0xa5 [<c10a527a>] do_sync_write+0xcd/0x103 [<c1049b46>] ? autoremove_wake_function+0x0/0x38 [<c1006efb>] ? xen_restore_fl_direct_end+0x0/0x1 [<c10d2c2c>] ? locks_free_lock+0x31/0x40 [<c10d4103>] ? fcntl_setlk+0x3e/0x185 [<c10a596e>] vfs_write+0x8b/0x13f [<c137e474>] ? do_page_fault+0x1b4/0x3a0 [<c10a51ad>] ? do_sync_write+0x0/0x103 [<c10a5f1d>] sys_write+0x3d/0x64 [<c1009058>] sysenter_do_call+0x12/0x2c Sched Debug Version: v0.09, 2.6.31.6 #8 now at 194118.700060 msecs .jiffies : 4294940775 .sysctl_sched_latency : 20.000000 .sysctl_sched_min_granularity : 4.000000 .sysctl_sched_wakeup_granularity : 5.000000 .sysctl_sched_child_runs_first : 0.000001 .sysctl_sched_features : 113917 cpu#0, 2500.152 MHz .nr_running : 1 .load : 3121 .nr_switches : 17143 .nr_load_updates : 48354 .nr_uninterruptible : 4 .next_balance : 4294.940896 .curr->pid : 14 .clock : 193919.954084 .cpu_load[0] : 0 .cpu_load[1] : 0 .cpu_load[2] : 0 .cpu_load[3] : 0 .cpu_load[4] : 0 cfs_rq[0]: .exec_clock : 0.000000 .MIN_vruntime : 0.000001 .min_vruntime : 13296.631519 .max_vruntime : 0.000001 .spread : 0.000000 .spread0 : 0.000000 .nr_running : 1 .load : 3121 .nr_spread_over : 0 rt_rq[0]: .rt_nr_running : 0 .rt_throttled : 0 .rt_time : 0.000000 .rt_runtime : 950.000000 runnable tasks: task PID tree-key switches prio exec-runtime sum-exec sum-sleep ---------------------------------------------------------------------------------------------------------- R xenwatch 14 13276.942024 146 115 0 0 0.000000 0.000000 0.000000 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |