* sustained write to disk, frozen copy [not found] <1719908460.2201367046573138.JavaMail.root@shiva> @ 2013-04-27 7:16 ` LuVar 2013-04-27 21:48 ` Azat Khuzhin 2013-04-28 16:50 ` Zheng Liu 0 siblings, 2 replies; 6+ messages in thread From: LuVar @ 2013-04-27 7:16 UTC (permalink / raw) To: linux-ext4 Hi, I have my desktop about 24 hours in "deadlock". I was copying (as root in krusader from USB key (mounted as [1]) data to filesystem [2]) some files from one point to another. Now it is more than 24 hours with sustained disk write, see [3]. How can I help and "debug" this problem? I have 3.5.7 gentoo kernel ([4]). PS: I am an average user, so please by verbose to me. [1] sudo mount -o rw,uid=luvar,gid=luvar,iocharset=utf8 /dev/sdg1 /mnt/usbstick/ [2]: luvar@blacktroja ~ $ mount | grep music /dev/mapper/vg-music on /var/lib/mpd/music/local type ext4 (rw,noatime,commit=0) [3]: iotop, two first records : 17714 be/3 root 0.00 B/s 0.00 B/s 0.00 % 97.60 % [jbd2/dm-3-8] 6546 be/4 root 0.00 B/s 0.00 B/s 0.00 % 93.48 % kdeinit4: kio_file [kdeinit] file local:/tmp/ksocket-root/~-socket local:/tmp/ksocket-root/krusaderZz6431.slave-socket [4]: luvar@blacktroja ~ $ uname -a Linux blacktroja 3.5.7-gentoo #1 SMP Sun Oct 28 17:18:07 CET 2012 x86_64 Intel(R) Core(TM) i7 CPU 930 @ 2.80GHz GenuineIntel GNU/Linux Thanks, LuVar ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: sustained write to disk, frozen copy 2013-04-27 7:16 ` sustained write to disk, frozen copy LuVar @ 2013-04-27 21:48 ` Azat Khuzhin 2013-04-28 16:50 ` Zheng Liu 1 sibling, 0 replies; 6+ messages in thread From: Azat Khuzhin @ 2013-04-27 21:48 UTC (permalink / raw) To: LuVar; +Cc: linux-ext4 Hi Luvar, Could you see this thread http://comments.gmane.org/gmane.comp.file-systems.ext4/38323 Seems that you have something similar. On Sat, Apr 27, 2013 at 11:16 AM, LuVar <luvar@plaintext.sk> wrote: > Hi, > I have my desktop about 24 hours in "deadlock". I was copying (as root in krusader from USB key (mounted as [1]) data to filesystem [2]) some files from one point to another. Now it is more than 24 hours with sustained disk write, see [3]. > > How can I help and "debug" this problem? I have 3.5.7 gentoo kernel ([4]). > > PS: I am an average user, so please by verbose to me. > > [1] sudo mount -o rw,uid=luvar,gid=luvar,iocharset=utf8 /dev/sdg1 /mnt/usbstick/ > > [2]: > luvar@blacktroja ~ $ mount | grep music > /dev/mapper/vg-music on /var/lib/mpd/music/local type ext4 (rw,noatime,commit=0) > > [3]: > iotop, two first records : > 17714 be/3 root 0.00 B/s 0.00 B/s 0.00 % 97.60 % [jbd2/dm-3-8] > 6546 be/4 root 0.00 B/s 0.00 B/s 0.00 % 93.48 % kdeinit4: kio_file [kdeinit] file local:/tmp/ksocket-root/~-socket local:/tmp/ksocket-root/krusaderZz6431.slave-socket > > [4]: > luvar@blacktroja ~ $ uname -a > Linux blacktroja 3.5.7-gentoo #1 SMP Sun Oct 28 17:18:07 CET 2012 x86_64 Intel(R) Core(TM) i7 CPU 930 @ 2.80GHz GenuineIntel GNU/Linux > > Thanks, LuVar > -- > To unsubscribe from this list: send the line "unsubscribe linux-ext4" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Respectfully Azat Khuzhin Primary email a3at.mail@gmail.com ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: sustained write to disk, frozen copy 2013-04-27 7:16 ` sustained write to disk, frozen copy LuVar 2013-04-27 21:48 ` Azat Khuzhin @ 2013-04-28 16:50 ` Zheng Liu 2013-04-28 19:17 ` LuVar 1 sibling, 1 reply; 6+ messages in thread From: Zheng Liu @ 2013-04-28 16:50 UTC (permalink / raw) To: LuVar; +Cc: linux-ext4 On Sat, Apr 27, 2013 at 08:16:02AM +0100, LuVar wrote: > Hi, > I have my desktop about 24 hours in "deadlock". I was copying (as root in krusader from USB key (mounted as [1]) data to filesystem [2]) some files from one point to another. Now it is more than 24 hours with sustained disk write, see [3]. > > How can I help and "debug" this problem? I have 3.5.7 gentoo kernel ([4]). Hi LuVar, You could use 'echo w >/proc/sysrq-trigger' to look at which process has been deadlock. # echo w >/proc/sysrq-trigger (WARN: you need a root privilege) # dmesg | vim - SysRq : Show Blocked State task PC stack pid father Here is nothing because my system hasn't any deadlock. Then you could use 'echo t >/proc/sysrq-trigger' to dump current tasks and their information. That would be great if you could paste these details in mailing list. They are very useful for us to dig this problem. Thanks, - Zheng > > PS: I am an average user, so please by verbose to me. > > [1] sudo mount -o rw,uid=luvar,gid=luvar,iocharset=utf8 /dev/sdg1 /mnt/usbstick/ > > [2]: > luvar@blacktroja ~ $ mount | grep music > /dev/mapper/vg-music on /var/lib/mpd/music/local type ext4 (rw,noatime,commit=0) > > [3]: > iotop, two first records : > 17714 be/3 root 0.00 B/s 0.00 B/s 0.00 % 97.60 % [jbd2/dm-3-8] > 6546 be/4 root 0.00 B/s 0.00 B/s 0.00 % 93.48 % kdeinit4: kio_file [kdeinit] file local:/tmp/ksocket-root/~-socket local:/tmp/ksocket-root/krusaderZz6431.slave-socket > > [4]: > luvar@blacktroja ~ $ uname -a > Linux blacktroja 3.5.7-gentoo #1 SMP Sun Oct 28 17:18:07 CET 2012 x86_64 Intel(R) Core(TM) i7 CPU 930 @ 2.80GHz GenuineIntel GNU/Linux > > Thanks, LuVar > -- > To unsubscribe from this list: send the line "unsubscribe linux-ext4" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: sustained write to disk, frozen copy 2013-04-28 16:50 ` Zheng Liu @ 2013-04-28 19:17 ` LuVar 2013-04-29 9:00 ` Dmitry Monakhov 0 siblings, 1 reply; 6+ messages in thread From: LuVar @ 2013-04-28 19:17 UTC (permalink / raw) To: Zheng Liu; +Cc: linux-ext4 Fuf... Here are my deadlock things: SysRq : Show Blocked State task PC stack pid father md3_raid5 D 0000000000000001 0 16779 2 0x00000000 ffff88032f76fb70 0000000000000046 ffff88032f76e000 0000000000010c80 ffff88032f144890 0000000000010c80 ffff88032f76ffd8 0000000000004000 ffff88032f76ffd8 0000000000010c80 ffff8803330bb470 ffff88032f144890 Call Trace: [<ffffffff81355e2f>] ? __blk_run_queue+0x16/0x18 [<ffffffff81358abe>] ? blk_queue_bio+0x29a/0x2b4 [<ffffffff81356546>] ? generic_make_request+0x97/0xda [<ffffffff814ebc8c>] schedule+0x5f/0x61 [<ffffffff8143bfa5>] md_super_wait+0x68/0x80 [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 [<ffffffff8144160f>] write_page+0x1d5/0x2be [<ffffffff81441365>] bitmap_update_sb+0x115/0x117 [<ffffffff8143c27c>] md_update_sb+0x2bf/0x467 [<ffffffff814ebab1>] ? __schedule+0x6b8/0x7be [<ffffffff8143ca00>] md_check_recovery+0x26b/0x5ff [<ffffffffa04a3624>] raid5d+0x1f/0x4c8 [raid456] [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 [<ffffffff81034cee>] ? del_timer_sync+0x3c/0x48 [<ffffffff814e9fdd>] ? schedule_timeout+0x189/0x1a9 [<ffffffff8143a69c>] md_thread+0xfd/0x11b [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 [<ffffffff8143a59f>] ? md_register_thread+0xc8/0xc8 [<ffffffff8104189d>] kthread+0x84/0x8c [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d [<ffffffff814ee310>] ? gs_change+0xb/0xb jbd2/dm-3-8 D 0000000000000002 0 17714 2 0x00000000 ffff88032f647bb0 0000000000000046 ffff88032f646000 0000000000010c80 ffff880330a70440 0000000000010c80 ffff88032f647fd8 0000000000004000 ffff88032f647fd8 0000000000010c80 ffff8801a6c3e100 ffff880330a70440 Call Trace: [<ffffffff810dae71>] ? __find_get_block_slow+0x113/0x12a [<ffffffff81438706>] ? md_make_request+0xc4/0x1b9 [<ffffffff810597da>] ? ktime_get_ts+0xa9/0xb5 [<ffffffff810db7ad>] ? unmap_underlying_metadata+0x39/0x39 [<ffffffff814ebc8c>] schedule+0x5f/0x61 [<ffffffff814ebd15>] io_schedule+0x87/0xca [<ffffffff810db7b6>] sleep_on_buffer+0x9/0xd [<ffffffff814ea18f>] __wait_on_bit+0x43/0x76 [<ffffffff814ea22b>] out_of_line_wait_on_bit+0x69/0x74 [<ffffffff810db7ad>] ? unmap_underlying_metadata+0x39/0x39 [<ffffffff81041ce0>] ? autoremove_wake_function+0x34/0x34 [<ffffffff810db772>] __wait_on_buffer+0x21/0x23 [<ffffffff8118acfb>] jbd2_journal_commit_transaction+0xd19/0x1182 [<ffffffff810349e6>] ? lock_timer_base.clone.28+0x26/0x4b [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 [<ffffffff8118daf5>] kjournald2+0xc6/0x22e [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 [<ffffffff8118da2f>] ? commit_timeout+0xb/0xb [<ffffffff8104189d>] kthread+0x84/0x8c [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d [<ffffffff814ee310>] ? gs_change+0xb/0xb flush-253:3 D ffff8803314e0024 0 6471 2 0x00000000 ffff88011fedda50 0000000000000046 ffff88011fedc000 0000000000010c80 ffff880130e02b90 0000000000010c80 ffff88011feddfd8 0000000000004000 ffff88011feddfd8 0000000000010c80 ffffffff81671410 ffff880130e02b90 Call Trace: [<ffffffff8104b325>] ? try_to_wake_up+0x20a/0x21c [<ffffffff814ebc8c>] schedule+0x5f/0x61 [<ffffffff8118d7ce>] jbd2_log_wait_commit+0xc1/0x113 [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 [<ffffffff8118ebae>] jbd2_journal_force_commit_nested+0x6a/0x7c [<ffffffff8115dd54>] ext4_da_writepages+0x376/0x447 [<ffffffff81095d18>] do_writepages+0x1e/0x27 [<ffffffff810d5c56>] __writeback_single_inode.clone.24+0x3a/0xda [<ffffffff810d6132>] writeback_sb_inodes+0x1b8/0x2f7 [<ffffffff810ba680>] ? put_super+0x20/0x2b [<ffffffff810d62de>] __writeback_inodes_wb+0x6d/0xab [<ffffffff810d641f>] wb_writeback+0x103/0x194 [<ffffffff810d6b48>] wb_do_writeback+0x111/0x16d [<ffffffff810d6c29>] bdi_writeback_thread+0x85/0x14a [<ffffffff810d6ba4>] ? wb_do_writeback+0x16d/0x16d [<ffffffff810d6ba4>] ? wb_do_writeback+0x16d/0x16d [<ffffffff8104189d>] kthread+0x84/0x8c [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d [<ffffffff814ee310>] ? gs_change+0xb/0xb kio_file D ffffffffffffffff 0 6546 6443 0x00000000 ffff880270a87b98 0000000000000082 ffff880270a86000 0000000000010c80 ffff8800aa600c10 0000000000010c80 ffff880270a87fd8 0000000000004000 ffff880270a87fd8 0000000000010c80 ffff8803330bb470 ffff8800aa600c10 Call Trace: [<ffffffff810349e6>] ? lock_timer_base.clone.28+0x26/0x4b [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 [<ffffffff814ebc8c>] schedule+0x5f/0x61 [<ffffffff814e9fd5>] schedule_timeout+0x181/0x1a9 [<ffffffff8103492a>] ? run_timer_softirq+0x1ef/0x1ef [<ffffffff814ebf5b>] io_schedule_timeout+0x93/0xe4 [<ffffffff8138176e>] ? __percpu_counter_sum+0x4d/0x63 [<ffffffff8109597c>] balance_dirty_pages_ratelimited_nr+0x54d/0x615 [<ffffffff810d7fca>] generic_file_splice_write+0x11e/0x130 [<ffffffff810d7acc>] do_splice_from+0x7d/0x8a [<ffffffff810d7af4>] direct_splice_actor+0x1b/0x1d [<ffffffff810d7dfb>] splice_direct_to_actor+0xd5/0x186 [<ffffffff810d7ad9>] ? do_splice_from+0x8a/0x8a [<ffffffff810d8e01>] do_splice_direct+0x47/0x5a [<ffffffff810b8f63>] do_sendfile+0x12e/0x1c3 [<ffffffff810b9bee>] sys_sendfile64+0x54/0x92 [<ffffffff814ed062>] system_call_fastpath+0x16/0x1b sync D ffff88033089f3f0 0 6707 6477 0x00000000 ffff8801b9651d08 0000000000000086 ffff8801b9650000 0000000000010c80 ffff88033089f3f0 0000000000010c80 ffff8801b9651fd8 0000000000004000 ffff8801b9651fd8 0000000000010c80 ffff8803330ba0c0 ffff88033089f3f0 Call Trace: [<ffffffff8108dc75>] ? find_get_pages_tag+0xf3/0x12f [<ffffffff81096d23>] ? release_pages+0x19c/0x1ab [<ffffffff81096691>] ? pagevec_lookup_tag+0x20/0x29 [<ffffffff814ebc8c>] schedule+0x5f/0x61 [<ffffffff814e9e7a>] schedule_timeout+0x26/0x1a9 [<ffffffff81049017>] ? check_preempt_curr+0x3e/0x6c [<ffffffff814eb2e6>] wait_for_common+0xc8/0x13f [<ffffffff8104b337>] ? try_to_wake_up+0x21c/0x21c [<ffffffff810d9745>] ? __sync_filesystem+0x7a/0x7a [<ffffffff814eb3f7>] wait_for_completion+0x18/0x1a [<ffffffff810d657f>] writeback_inodes_sb_nr+0xb8/0xc1 [<ffffffff810d6602>] writeback_inodes_sb+0x22/0x29 [<ffffffff810d971c>] __sync_filesystem+0x51/0x7a [<ffffffff810d9756>] sync_one_sb+0x11/0x13 [<ffffffff810bb6fa>] iterate_supers+0x68/0xb8 [<ffffffff810d9695>] sync_filesystems+0x1b/0x1d [<ffffffff810d97ba>] sys_sync+0x17/0x33 [<ffffffff814ed062>] system_call_fastpath+0x16/0x1b Is there something eslse, what should I do before reboot? LuVar ----- "Zheng Liu" <gnehzuil.liu@gmail.com> wrote: > On Sat, Apr 27, 2013 at 08:16:02AM +0100, LuVar wrote: > > Hi, > > I have my desktop about 24 hours in "deadlock". I was copying (as > root in krusader from USB key (mounted as [1]) data to filesystem [2]) > some files from one point to another. Now it is more than 24 hours > with sustained disk write, see [3]. > > > > How can I help and "debug" this problem? I have 3.5.7 gentoo kernel > ([4]). > > Hi LuVar, > > You could use 'echo w >/proc/sysrq-trigger' to look at which process > has > been deadlock. > > # echo w >/proc/sysrq-trigger (WARN: you need a root privilege) > # dmesg | vim - > > SysRq : Show Blocked State > task PC stack pid father > > Here is nothing because my system hasn't any deadlock. Then you > could > use 'echo t >/proc/sysrq-trigger' to dump current tasks and their > information. That would be great if you could paste these details in > mailing list. They are very useful for us to dig this problem. > > Thanks, > - Zheng > > > > > PS: I am an average user, so please by verbose to me. > > > > [1] sudo mount -o rw,uid=luvar,gid=luvar,iocharset=utf8 /dev/sdg1 > /mnt/usbstick/ > > > > [2]: > > luvar@blacktroja ~ $ mount | grep music > > /dev/mapper/vg-music on /var/lib/mpd/music/local type ext4 > (rw,noatime,commit=0) > > > > [3]: > > iotop, two first records : > > 17714 be/3 root 0.00 B/s 0.00 B/s 0.00 % 97.60 % > [jbd2/dm-3-8] > > 6546 be/4 root 0.00 B/s 0.00 B/s 0.00 % 93.48 % > kdeinit4: kio_file [kdeinit] file local:/tmp/ksocket-root/~-socket > local:/tmp/ksocket-root/krusaderZz6431.slave-socket > > > > [4]: > > luvar@blacktroja ~ $ uname -a > > Linux blacktroja 3.5.7-gentoo #1 SMP Sun Oct 28 17:18:07 CET 2012 > x86_64 Intel(R) Core(TM) i7 CPU 930 @ 2.80GHz GenuineIntel GNU/Linux > > > > Thanks, LuVar > > -- > > To unsubscribe from this list: send the line "unsubscribe > linux-ext4" in > > the body of a message to majordomo@vger.kernel.org > > More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: sustained write to disk, frozen copy 2013-04-28 19:17 ` LuVar @ 2013-04-29 9:00 ` Dmitry Monakhov 0 siblings, 0 replies; 6+ messages in thread From: Dmitry Monakhov @ 2013-04-29 9:00 UTC (permalink / raw) To: LuVar, Zheng Liu; +Cc: linux-ext4 On Sun, 28 Apr 2013 20:17:03 +0100 (GMT+01:00), LuVar <luvar@plaintext.sk> wrote: > Fuf... Here are my deadlock things: > Strange looks md3_raid5 stuck ? Can you please post your /proc/mounts, /proc/mdstat and lvm config > SysRq : Show Blocked State > task PC stack pid father > md3_raid5 D 0000000000000001 0 16779 2 0x00000000 > ffff88032f76fb70 0000000000000046 ffff88032f76e000 0000000000010c80 > ffff88032f144890 0000000000010c80 ffff88032f76ffd8 0000000000004000 > ffff88032f76ffd8 0000000000010c80 ffff8803330bb470 ffff88032f144890 > Call Trace: > [<ffffffff81355e2f>] ? __blk_run_queue+0x16/0x18 > [<ffffffff81358abe>] ? blk_queue_bio+0x29a/0x2b4 > [<ffffffff81356546>] ? generic_make_request+0x97/0xda > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > [<ffffffff8143bfa5>] md_super_wait+0x68/0x80 > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > [<ffffffff8144160f>] write_page+0x1d5/0x2be > [<ffffffff81441365>] bitmap_update_sb+0x115/0x117 > [<ffffffff8143c27c>] md_update_sb+0x2bf/0x467 > [<ffffffff814ebab1>] ? __schedule+0x6b8/0x7be > [<ffffffff8143ca00>] md_check_recovery+0x26b/0x5ff > [<ffffffffa04a3624>] raid5d+0x1f/0x4c8 [raid456] > [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 > [<ffffffff81034cee>] ? del_timer_sync+0x3c/0x48 > [<ffffffff814e9fdd>] ? schedule_timeout+0x189/0x1a9 > [<ffffffff8143a69c>] md_thread+0xfd/0x11b > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > [<ffffffff8143a59f>] ? md_register_thread+0xc8/0xc8 > [<ffffffff8104189d>] kthread+0x84/0x8c > [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 > [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d > [<ffffffff814ee310>] ? gs_change+0xb/0xb > jbd2/dm-3-8 D 0000000000000002 0 17714 2 0x00000000 > ffff88032f647bb0 0000000000000046 ffff88032f646000 0000000000010c80 > ffff880330a70440 0000000000010c80 ffff88032f647fd8 0000000000004000 > ffff88032f647fd8 0000000000010c80 ffff8801a6c3e100 ffff880330a70440 > Call Trace: > [<ffffffff810dae71>] ? __find_get_block_slow+0x113/0x12a > [<ffffffff81438706>] ? md_make_request+0xc4/0x1b9 > [<ffffffff810597da>] ? ktime_get_ts+0xa9/0xb5 > [<ffffffff810db7ad>] ? unmap_underlying_metadata+0x39/0x39 > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > [<ffffffff814ebd15>] io_schedule+0x87/0xca > [<ffffffff810db7b6>] sleep_on_buffer+0x9/0xd > [<ffffffff814ea18f>] __wait_on_bit+0x43/0x76 > [<ffffffff814ea22b>] out_of_line_wait_on_bit+0x69/0x74 > [<ffffffff810db7ad>] ? unmap_underlying_metadata+0x39/0x39 > [<ffffffff81041ce0>] ? autoremove_wake_function+0x34/0x34 > [<ffffffff810db772>] __wait_on_buffer+0x21/0x23 > [<ffffffff8118acfb>] jbd2_journal_commit_transaction+0xd19/0x1182 > [<ffffffff810349e6>] ? lock_timer_base.clone.28+0x26/0x4b > [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 > [<ffffffff8118daf5>] kjournald2+0xc6/0x22e > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > [<ffffffff8118da2f>] ? commit_timeout+0xb/0xb > [<ffffffff8104189d>] kthread+0x84/0x8c > [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 > [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d > [<ffffffff814ee310>] ? gs_change+0xb/0xb > flush-253:3 D ffff8803314e0024 0 6471 2 0x00000000 > ffff88011fedda50 0000000000000046 ffff88011fedc000 0000000000010c80 > ffff880130e02b90 0000000000010c80 ffff88011feddfd8 0000000000004000 > ffff88011feddfd8 0000000000010c80 ffffffff81671410 ffff880130e02b90 > Call Trace: > [<ffffffff8104b325>] ? try_to_wake_up+0x20a/0x21c > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > [<ffffffff8118d7ce>] jbd2_log_wait_commit+0xc1/0x113 > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > [<ffffffff8118ebae>] jbd2_journal_force_commit_nested+0x6a/0x7c > [<ffffffff8115dd54>] ext4_da_writepages+0x376/0x447 > [<ffffffff81095d18>] do_writepages+0x1e/0x27 > [<ffffffff810d5c56>] __writeback_single_inode.clone.24+0x3a/0xda > [<ffffffff810d6132>] writeback_sb_inodes+0x1b8/0x2f7 > [<ffffffff810ba680>] ? put_super+0x20/0x2b > [<ffffffff810d62de>] __writeback_inodes_wb+0x6d/0xab > [<ffffffff810d641f>] wb_writeback+0x103/0x194 > [<ffffffff810d6b48>] wb_do_writeback+0x111/0x16d > [<ffffffff810d6c29>] bdi_writeback_thread+0x85/0x14a > [<ffffffff810d6ba4>] ? wb_do_writeback+0x16d/0x16d > [<ffffffff810d6ba4>] ? wb_do_writeback+0x16d/0x16d > [<ffffffff8104189d>] kthread+0x84/0x8c > [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 > [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d > [<ffffffff814ee310>] ? gs_change+0xb/0xb > kio_file D ffffffffffffffff 0 6546 6443 0x00000000 > ffff880270a87b98 0000000000000082 ffff880270a86000 0000000000010c80 > ffff8800aa600c10 0000000000010c80 ffff880270a87fd8 0000000000004000 > ffff880270a87fd8 0000000000010c80 ffff8803330bb470 ffff8800aa600c10 > Call Trace: > [<ffffffff810349e6>] ? lock_timer_base.clone.28+0x26/0x4b > [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > [<ffffffff814e9fd5>] schedule_timeout+0x181/0x1a9 > [<ffffffff8103492a>] ? run_timer_softirq+0x1ef/0x1ef > [<ffffffff814ebf5b>] io_schedule_timeout+0x93/0xe4 > [<ffffffff8138176e>] ? __percpu_counter_sum+0x4d/0x63 > [<ffffffff8109597c>] balance_dirty_pages_ratelimited_nr+0x54d/0x615 > [<ffffffff810d7fca>] generic_file_splice_write+0x11e/0x130 > [<ffffffff810d7acc>] do_splice_from+0x7d/0x8a > [<ffffffff810d7af4>] direct_splice_actor+0x1b/0x1d > [<ffffffff810d7dfb>] splice_direct_to_actor+0xd5/0x186 > [<ffffffff810d7ad9>] ? do_splice_from+0x8a/0x8a > [<ffffffff810d8e01>] do_splice_direct+0x47/0x5a > [<ffffffff810b8f63>] do_sendfile+0x12e/0x1c3 > [<ffffffff810b9bee>] sys_sendfile64+0x54/0x92 > [<ffffffff814ed062>] system_call_fastpath+0x16/0x1b > sync D ffff88033089f3f0 0 6707 6477 0x00000000 > ffff8801b9651d08 0000000000000086 ffff8801b9650000 0000000000010c80 > ffff88033089f3f0 0000000000010c80 ffff8801b9651fd8 0000000000004000 > ffff8801b9651fd8 0000000000010c80 ffff8803330ba0c0 ffff88033089f3f0 > Call Trace: > [<ffffffff8108dc75>] ? find_get_pages_tag+0xf3/0x12f > [<ffffffff81096d23>] ? release_pages+0x19c/0x1ab > [<ffffffff81096691>] ? pagevec_lookup_tag+0x20/0x29 > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > [<ffffffff814e9e7a>] schedule_timeout+0x26/0x1a9 > [<ffffffff81049017>] ? check_preempt_curr+0x3e/0x6c > [<ffffffff814eb2e6>] wait_for_common+0xc8/0x13f > [<ffffffff8104b337>] ? try_to_wake_up+0x21c/0x21c > [<ffffffff810d9745>] ? __sync_filesystem+0x7a/0x7a > [<ffffffff814eb3f7>] wait_for_completion+0x18/0x1a > [<ffffffff810d657f>] writeback_inodes_sb_nr+0xb8/0xc1 > [<ffffffff810d6602>] writeback_inodes_sb+0x22/0x29 > [<ffffffff810d971c>] __sync_filesystem+0x51/0x7a > [<ffffffff810d9756>] sync_one_sb+0x11/0x13 > [<ffffffff810bb6fa>] iterate_supers+0x68/0xb8 > [<ffffffff810d9695>] sync_filesystems+0x1b/0x1d > [<ffffffff810d97ba>] sys_sync+0x17/0x33 > [<ffffffff814ed062>] system_call_fastpath+0x16/0x1b > > Is there something eslse, what should I do before reboot? > > LuVar > > ----- "Zheng Liu" <gnehzuil.liu@gmail.com> wrote: > > > On Sat, Apr 27, 2013 at 08:16:02AM +0100, LuVar wrote: > > > Hi, > > > I have my desktop about 24 hours in "deadlock". I was copying (as > > root in krusader from USB key (mounted as [1]) data to filesystem [2]) > > some files from one point to another. Now it is more than 24 hours > > with sustained disk write, see [3]. > > > > > > How can I help and "debug" this problem? I have 3.5.7 gentoo kernel > > ([4]). > > > > Hi LuVar, > > > > You could use 'echo w >/proc/sysrq-trigger' to look at which process > > has > > been deadlock. > > > > # echo w >/proc/sysrq-trigger (WARN: you need a root privilege) > > # dmesg | vim - > > > > SysRq : Show Blocked State > > task PC stack pid father > > > > Here is nothing because my system hasn't any deadlock. Then you > > could > > use 'echo t >/proc/sysrq-trigger' to dump current tasks and their > > information. That would be great if you could paste these details in > > mailing list. They are very useful for us to dig this problem. > > > > Thanks, > > - Zheng > > > > > > > > PS: I am an average user, so please by verbose to me. > > > > > > [1] sudo mount -o rw,uid=luvar,gid=luvar,iocharset=utf8 /dev/sdg1 > > /mnt/usbstick/ > > > > > > [2]: > > > luvar@blacktroja ~ $ mount | grep music > > > /dev/mapper/vg-music on /var/lib/mpd/music/local type ext4 > > (rw,noatime,commit=0) > > > > > > [3]: > > > iotop, two first records : > > > 17714 be/3 root 0.00 B/s 0.00 B/s 0.00 % 97.60 % > > [jbd2/dm-3-8] > > > 6546 be/4 root 0.00 B/s 0.00 B/s 0.00 % 93.48 % > > kdeinit4: kio_file [kdeinit] file local:/tmp/ksocket-root/~-socket > > local:/tmp/ksocket-root/krusaderZz6431.slave-socket > > > > > > [4]: > > > luvar@blacktroja ~ $ uname -a > > > Linux blacktroja 3.5.7-gentoo #1 SMP Sun Oct 28 17:18:07 CET 2012 > > x86_64 Intel(R) Core(TM) i7 CPU 930 @ 2.80GHz GenuineIntel GNU/Linux > > > > > > Thanks, LuVar > > > -- > > > To unsubscribe from this list: send the line "unsubscribe > > linux-ext4" in > > > the body of a message to majordomo@vger.kernel.org > > > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- > To unsubscribe from this list: send the line "unsubscribe linux-ext4" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 6+ messages in thread
[parent not found: <1448604326.2831367229593304.JavaMail.root@shiva>]
* Re: sustained write to disk, frozen copy [not found] <1448604326.2831367229593304.JavaMail.root@shiva> @ 2013-04-29 10:13 ` luvar 0 siblings, 0 replies; 6+ messages in thread From: luvar @ 2013-04-29 10:13 UTC (permalink / raw) To: Dmitry Monakhov; +Cc: linux-ext4, Zheng Liu Hi, here are additional data: luvar@blacktroja ~ $ cat /proc/mounts rootfs / rootfs rw 0 0 proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0 sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0 udev /dev tmpfs rw,nosuid,relatime,size=10240k,mode=755 0 0 devpts /dev/pts devpts rw,relatime,gid=5,mode=620 0 0 /dev/md2 / ext3 rw,noatime,errors=continue,commit=5,barrier=1,data=ordered 0 0 tmpfs /run tmpfs rw,nosuid,nodev,relatime,mode=755 0 0 rc-svcdir /lib64/rc/init.d tmpfs rw,nosuid,nodev,noexec,relatime,size=1024k,mode=755 0 0 configfs /sys/kernel/config configfs rw,nosuid,nodev,noexec,relatime 0 0 cgroup_root /sys/fs/cgroup tmpfs rw,nosuid,nodev,noexec,relatime,size=10240k,mode=755 0 0 cpuset /sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 cpu /sys/fs/cgroup/cpu cgroup rw,nosuid,nodev,noexec,relatime,cpu 0 0 cpuacct /sys/fs/cgroup/cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpuacct 0 0 fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0 shm /dev/shm tmpfs rw,nosuid,nodev,noexec,relatime 0 0 /dev/mapper/vg-portage /usr/portage ext2 rw,noatime,errors=continue 0 0 /dev/mapper/vg-distfiles /usr/portage/distfiles ext2 rw,noatime,errors=continue 0 0 /dev/mapper/vg-home /home ext3 rw,noatime,errors=continue,commit=5,barrier=1,data=ordered 0 0 none /var/tmp/portage tmpfs rw,noatime,size=8388608k,nr_inodes=2097152 0 0 /dev/mapper/vg-svrkiNfs /mnt/nfs/svrki ext4 rw,noatime,stripe=48,data=ordered 0 0 /dev/mapper/vg-music /var/lib/mpd/music/local ext4 rw,noatime,stripe=48,data=ordered 0 0 /dev/sda2 /home/luvar/.m2 btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/eclipseWorkspace btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/eclipseWorkspaceGanz btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/eclipseWorkspaceMisc btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/eclipseWorkspaceOpen btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/eclipseWorkspaceAndroid btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/.mozilla btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/documents/trackit btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/documents/plaintext btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/programs/nexus-oss-webapp-bundle btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/programs/android-sdk-linux btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/mapper/vg-skolaDokumenty /home/luvar/documents/skola btrfs rw,noatime,thread_pool=4,compress=zlib,nospace_cache 0 0 /dev/mapper/vg-laciNilfs /mnt/nfs/laciNilfs nilfs2 rw,noatime 0 0 /dev/mapper/vg-boincPartition /var/lib/boinc ext4 rw,noatime,stripe=64 0 0 /dev/mapper/vg-postgresql91 /var/lib/postgresql/9.1 reiserfs rw,noatime 0 0 /dev/mapper/vg-yacy /home/luvar/programs/yacy/yacy_partition btrfs rw,noatime,thread_pool=1,noacl,nospace_cache 0 0 rpc_pipefs /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0 nfsd /proc/fs/nfsd nfsd rw,nosuid,nodev,noexec,relatime 0 0 /dev/mapper/vg-fotodvdbackup /mnt/fotodvdbackup ext2 rw,relatime,errors=continue 0 0 /dev/mapper/vg-pogamut /mnt/pogamut ext4 rw,noatime,stripe=48,data=ordered 0 0 /dev/mapper/vg-films /mnt/films ext4 rw,noatime,stripe=48,data=ordered 0 0 /dev/loop0 /mnt/cdrom iso9660 ro,relatime 0 0 /dev/sdg1 /mnt/usbstick vfat rw,relatime,uid=1000,gid=1000,fmask=0022,dmask=0022,codepage=cp437,iocharset=utf8,shortname=mixed,errors=remount-ro 0 0 luvar@blacktroja ~ $ cat /proc/mdstat Personalities : [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] md1 : active raid1 sdc1[0] sdf1[3] sde1[2] sdd1[1] sdb1[4] 1469824 blocks [5/5] [UUUUU] md3 : active raid5 sdc3[0] sdf3[3] sde3[2] sdd3[1] sdb3[4] 3068960768 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU] bitmap: 2/6 pages [8KB], 65536KB chunk md2 : active raid10 sdc2[3] sdf2[2] sde2[1] sdd2[0] 25398272 blocks 512K chunks 2 near-copies [4/4] [UUUU] unused devices: <none> lvm config is available on http://dawn.ynet.sk/~luvar/asdf/vgcfgbackup Thanks, PS:blacktroja luvar # uptime 12:12:58 up 8 days, 16:45, 12 users, load average: 5.95, 5.64, 5.75 Still writing... ----- "Dmitry Monakhov" <dmonakhov@openvz.org> wrote: > On Sun, 28 Apr 2013 20:17:03 +0100 (GMT+01:00), LuVar > <luvar@plaintext.sk> wrote: > > Fuf... Here are my deadlock things: > > > Strange looks md3_raid5 stuck ? > Can you please post your /proc/mounts, /proc/mdstat and lvm config > > SysRq : Show Blocked State > > task PC stack pid father > > md3_raid5 D 0000000000000001 0 16779 2 0x00000000 > > ffff88032f76fb70 0000000000000046 ffff88032f76e000 > 0000000000010c80 > > ffff88032f144890 0000000000010c80 ffff88032f76ffd8 > 0000000000004000 > > ffff88032f76ffd8 0000000000010c80 ffff8803330bb470 > ffff88032f144890 > > Call Trace: > > [<ffffffff81355e2f>] ? __blk_run_queue+0x16/0x18 > > [<ffffffff81358abe>] ? blk_queue_bio+0x29a/0x2b4 > > [<ffffffff81356546>] ? generic_make_request+0x97/0xda > > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > > [<ffffffff8143bfa5>] md_super_wait+0x68/0x80 > > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > > [<ffffffff8144160f>] write_page+0x1d5/0x2be > > [<ffffffff81441365>] bitmap_update_sb+0x115/0x117 > > [<ffffffff8143c27c>] md_update_sb+0x2bf/0x467 > > [<ffffffff814ebab1>] ? __schedule+0x6b8/0x7be > > [<ffffffff8143ca00>] md_check_recovery+0x26b/0x5ff > > [<ffffffffa04a3624>] raid5d+0x1f/0x4c8 [raid456] > > [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 > > [<ffffffff81034cee>] ? del_timer_sync+0x3c/0x48 > > [<ffffffff814e9fdd>] ? schedule_timeout+0x189/0x1a9 > > [<ffffffff8143a69c>] md_thread+0xfd/0x11b > > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > > [<ffffffff8143a59f>] ? md_register_thread+0xc8/0xc8 > > [<ffffffff8104189d>] kthread+0x84/0x8c > > [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 > > [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d > > [<ffffffff814ee310>] ? gs_change+0xb/0xb > > jbd2/dm-3-8 D 0000000000000002 0 17714 2 0x00000000 > > ffff88032f647bb0 0000000000000046 ffff88032f646000 > 0000000000010c80 > > ffff880330a70440 0000000000010c80 ffff88032f647fd8 > 0000000000004000 > > ffff88032f647fd8 0000000000010c80 ffff8801a6c3e100 > ffff880330a70440 > > Call Trace: > > [<ffffffff810dae71>] ? __find_get_block_slow+0x113/0x12a > > [<ffffffff81438706>] ? md_make_request+0xc4/0x1b9 > > [<ffffffff810597da>] ? ktime_get_ts+0xa9/0xb5 > > [<ffffffff810db7ad>] ? unmap_underlying_metadata+0x39/0x39 > > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > > [<ffffffff814ebd15>] io_schedule+0x87/0xca > > [<ffffffff810db7b6>] sleep_on_buffer+0x9/0xd > > [<ffffffff814ea18f>] __wait_on_bit+0x43/0x76 > > [<ffffffff814ea22b>] out_of_line_wait_on_bit+0x69/0x74 > > [<ffffffff810db7ad>] ? unmap_underlying_metadata+0x39/0x39 > > [<ffffffff81041ce0>] ? autoremove_wake_function+0x34/0x34 > > [<ffffffff810db772>] __wait_on_buffer+0x21/0x23 > > [<ffffffff8118acfb>] jbd2_journal_commit_transaction+0xd19/0x1182 > > [<ffffffff810349e6>] ? lock_timer_base.clone.28+0x26/0x4b > > [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 > > [<ffffffff8118daf5>] kjournald2+0xc6/0x22e > > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > > [<ffffffff8118da2f>] ? commit_timeout+0xb/0xb > > [<ffffffff8104189d>] kthread+0x84/0x8c > > [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 > > [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d > > [<ffffffff814ee310>] ? gs_change+0xb/0xb > > flush-253:3 D ffff8803314e0024 0 6471 2 0x00000000 > > ffff88011fedda50 0000000000000046 ffff88011fedc000 > 0000000000010c80 > > ffff880130e02b90 0000000000010c80 ffff88011feddfd8 > 0000000000004000 > > ffff88011feddfd8 0000000000010c80 ffffffff81671410 > ffff880130e02b90 > > Call Trace: > > [<ffffffff8104b325>] ? try_to_wake_up+0x20a/0x21c > > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > > [<ffffffff8118d7ce>] jbd2_log_wait_commit+0xc1/0x113 > > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > > [<ffffffff8118ebae>] jbd2_journal_force_commit_nested+0x6a/0x7c > > [<ffffffff8115dd54>] ext4_da_writepages+0x376/0x447 > > [<ffffffff81095d18>] do_writepages+0x1e/0x27 > > [<ffffffff810d5c56>] __writeback_single_inode.clone.24+0x3a/0xda > > [<ffffffff810d6132>] writeback_sb_inodes+0x1b8/0x2f7 > > [<ffffffff810ba680>] ? put_super+0x20/0x2b > > [<ffffffff810d62de>] __writeback_inodes_wb+0x6d/0xab > > [<ffffffff810d641f>] wb_writeback+0x103/0x194 > > [<ffffffff810d6b48>] wb_do_writeback+0x111/0x16d > > [<ffffffff810d6c29>] bdi_writeback_thread+0x85/0x14a > > [<ffffffff810d6ba4>] ? wb_do_writeback+0x16d/0x16d > > [<ffffffff810d6ba4>] ? wb_do_writeback+0x16d/0x16d > > [<ffffffff8104189d>] kthread+0x84/0x8c > > [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 > > [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d > > [<ffffffff814ee310>] ? gs_change+0xb/0xb > > kio_file D ffffffffffffffff 0 6546 6443 0x00000000 > > ffff880270a87b98 0000000000000082 ffff880270a86000 > 0000000000010c80 > > ffff8800aa600c10 0000000000010c80 ffff880270a87fd8 > 0000000000004000 > > ffff880270a87fd8 0000000000010c80 ffff8803330bb470 > ffff8800aa600c10 > > Call Trace: > > [<ffffffff810349e6>] ? lock_timer_base.clone.28+0x26/0x4b > > [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 > > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > > [<ffffffff814e9fd5>] schedule_timeout+0x181/0x1a9 > > [<ffffffff8103492a>] ? run_timer_softirq+0x1ef/0x1ef > > [<ffffffff814ebf5b>] io_schedule_timeout+0x93/0xe4 > > [<ffffffff8138176e>] ? __percpu_counter_sum+0x4d/0x63 > > [<ffffffff8109597c>] > balance_dirty_pages_ratelimited_nr+0x54d/0x615 > > [<ffffffff810d7fca>] generic_file_splice_write+0x11e/0x130 > > [<ffffffff810d7acc>] do_splice_from+0x7d/0x8a > > [<ffffffff810d7af4>] direct_splice_actor+0x1b/0x1d > > [<ffffffff810d7dfb>] splice_direct_to_actor+0xd5/0x186 > > [<ffffffff810d7ad9>] ? do_splice_from+0x8a/0x8a > > [<ffffffff810d8e01>] do_splice_direct+0x47/0x5a > > [<ffffffff810b8f63>] do_sendfile+0x12e/0x1c3 > > [<ffffffff810b9bee>] sys_sendfile64+0x54/0x92 > > [<ffffffff814ed062>] system_call_fastpath+0x16/0x1b > > sync D ffff88033089f3f0 0 6707 6477 0x00000000 > > ffff8801b9651d08 0000000000000086 ffff8801b9650000 > 0000000000010c80 > > ffff88033089f3f0 0000000000010c80 ffff8801b9651fd8 > 0000000000004000 > > ffff8801b9651fd8 0000000000010c80 ffff8803330ba0c0 > ffff88033089f3f0 > > Call Trace: > > [<ffffffff8108dc75>] ? find_get_pages_tag+0xf3/0x12f > > [<ffffffff81096d23>] ? release_pages+0x19c/0x1ab > > [<ffffffff81096691>] ? pagevec_lookup_tag+0x20/0x29 > > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > > [<ffffffff814e9e7a>] schedule_timeout+0x26/0x1a9 > > [<ffffffff81049017>] ? check_preempt_curr+0x3e/0x6c > > [<ffffffff814eb2e6>] wait_for_common+0xc8/0x13f > > [<ffffffff8104b337>] ? try_to_wake_up+0x21c/0x21c > > [<ffffffff810d9745>] ? __sync_filesystem+0x7a/0x7a > > [<ffffffff814eb3f7>] wait_for_completion+0x18/0x1a > > [<ffffffff810d657f>] writeback_inodes_sb_nr+0xb8/0xc1 > > [<ffffffff810d6602>] writeback_inodes_sb+0x22/0x29 > > [<ffffffff810d971c>] __sync_filesystem+0x51/0x7a > > [<ffffffff810d9756>] sync_one_sb+0x11/0x13 > > [<ffffffff810bb6fa>] iterate_supers+0x68/0xb8 > > [<ffffffff810d9695>] sync_filesystems+0x1b/0x1d > > [<ffffffff810d97ba>] sys_sync+0x17/0x33 > > [<ffffffff814ed062>] system_call_fastpath+0x16/0x1b > > > > Is there something eslse, what should I do before reboot? > > > > LuVar > > > > ----- "Zheng Liu" <gnehzuil.liu@gmail.com> wrote: > > > > > On Sat, Apr 27, 2013 at 08:16:02AM +0100, LuVar wrote: > > > > Hi, > > > > I have my desktop about 24 hours in "deadlock". I was copying > (as > > > root in krusader from USB key (mounted as [1]) data to filesystem > [2]) > > > some files from one point to another. Now it is more than 24 > hours > > > with sustained disk write, see [3]. > > > > > > > > How can I help and "debug" this problem? I have 3.5.7 gentoo > kernel > > > ([4]). > > > > > > Hi LuVar, > > > > > > You could use 'echo w >/proc/sysrq-trigger' to look at which > process > > > has > > > been deadlock. > > > > > > # echo w >/proc/sysrq-trigger (WARN: you need a root privilege) > > > # dmesg | vim - > > > > > > SysRq : Show Blocked State > > > task PC stack pid father > > > > > > Here is nothing because my system hasn't any deadlock. Then you > > > could > > > use 'echo t >/proc/sysrq-trigger' to dump current tasks and their > > > information. That would be great if you could paste these details > in > > > mailing list. They are very useful for us to dig this problem. > > > > > > Thanks, > > > - Zheng > > > > > > > > > > > PS: I am an average user, so please by verbose to me. > > > > > > > > [1] sudo mount -o rw,uid=luvar,gid=luvar,iocharset=utf8 > /dev/sdg1 > > > /mnt/usbstick/ > > > > > > > > [2]: > > > > luvar@blacktroja ~ $ mount | grep music > > > > /dev/mapper/vg-music on /var/lib/mpd/music/local type ext4 > > > (rw,noatime,commit=0) > > > > > > > > [3]: > > > > iotop, two first records : > > > > 17714 be/3 root 0.00 B/s 0.00 B/s 0.00 % 97.60 % > > > [jbd2/dm-3-8] > > > > 6546 be/4 root 0.00 B/s 0.00 B/s 0.00 % 93.48 % > > > kdeinit4: kio_file [kdeinit] file > local:/tmp/ksocket-root/~-socket > > > local:/tmp/ksocket-root/krusaderZz6431.slave-socket > > > > > > > > [4]: > > > > luvar@blacktroja ~ $ uname -a > > > > Linux blacktroja 3.5.7-gentoo #1 SMP Sun Oct 28 17:18:07 CET > 2012 > > > x86_64 Intel(R) Core(TM) i7 CPU 930 @ 2.80GHz GenuineIntel > GNU/Linux > > > > > > > > Thanks, LuVar > > > > -- > > > > To unsubscribe from this list: send the line "unsubscribe > > > linux-ext4" in > > > > the body of a message to majordomo@vger.kernel.org > > > > More majordomo info at > http://vger.kernel.org/majordomo-info.html > > -- > > To unsubscribe from this list: send the line "unsubscribe > linux-ext4" in > > the body of a message to majordomo@vger.kernel.org > > More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 6+ messages in thread
end of thread, other threads:[~2013-04-29 10:13 UTC | newest] Thread overview: 6+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- [not found] <1719908460.2201367046573138.JavaMail.root@shiva> 2013-04-27 7:16 ` sustained write to disk, frozen copy LuVar 2013-04-27 21:48 ` Azat Khuzhin 2013-04-28 16:50 ` Zheng Liu 2013-04-28 19:17 ` LuVar 2013-04-29 9:00 ` Dmitry Monakhov [not found] <1448604326.2831367229593304.JavaMail.root@shiva> 2013-04-29 10:13 ` luvar
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).