* raid5 crash
@ 2004-12-22 22:04 Kristian Eide
2004-12-22 22:26 ` Norbert van Nobelen
2004-12-22 23:08 ` Neil Brown
0 siblings, 2 replies; 9+ messages in thread
From: Kristian Eide @ 2004-12-22 22:04 UTC (permalink / raw)
To: linux-kernel
[-- Attachment #1: Type: text/plain, Size: 2890 bytes --]
I am running kernel 2.6.9-gentoo-r10 on an Athlon XP 2400+ computer with a SiI
3114 SATA controller hosting 4 WD2500JD-00G drives. I have combined these
drives into a raid5 array using software raid, but unfortunately the array is
not stable. I have tried several filesystems (ext3, reiserfs, xfs), but after
copying several gigabytes of data into the array (using scp) and then trying
to read them back (using rsync to compare over the network) always results in
data corruption. Here is the output from 'dmesg':
kernel BUG at drivers/md/raid5.c:813!
invalid operand: 0000 [#1]
Modules linked in: sata_sil libata sbp2 ohci1394 ieee1394 usb_storage ehci_hcd
usbcore
CPU: 0
EIP: 0060:[<c039cdd2>] Not tainted VLI
EFLAGS: 00010006 (2.6.9-gentoo-r10)
EIP is at add_stripe_bio+0x1c2/0x200
eax: 00045168 ebx: d3974b00 ecx: d3974980 edx: 00000000
esi: 00045140 edi: 00000000 ebp: e33200a4 esp: f0a05ac4
ds: 007b es: 007b ss: 0068
Process rsync (pid: 32092, threadinfo=f0a04000 task=f6c10020)
Stack: 00000000 00000296 00000140 e3320028 00045140 00000000 d3974980 c039e092
e3320028 d3974980 00000000 00000000 00000000 f0a05b1c de3e1ae0 00045158
00000000 00000003 00000004 de3e1ae0 dfe90e00 00000000 00000003 f7d85088
Call Trace:
[<c039e092>] make_request+0x122/0x200
[<c032dc6f>] generic_make_request+0x15f/0x1e0
[<c011d590>] autoremove_wake_function+0x0/0x60
[<c032dd4d>] submit_bio+0x5d/0x100
[<c0172d43>] mpage_bio_submit+0x23/0x40
[<c0173170>] do_mpage_readpage+0x2d0/0x480
[<c012367d>] __do_softirq+0x7d/0x90
[<c02b54df>] radix_tree_node_alloc+0x1f/0x60
[<c02b5762>] radix_tree_insert+0xe2/0x100
[<c0136e54>] add_to_page_cache+0x54/0x80
[<c017346b>] mpage_readpages+0x14b/0x180
[<c018f1f0>] reiserfs_get_block+0x0/0x1450
[<c013ddf4>] read_pages+0x134/0x140
[<c018f1f0>] reiserfs_get_block+0x0/0x1450
[<c013b390>] __alloc_pages+0x1d0/0x370
[<c01081c5>] do_IRQ+0xc5/0xe0
[<c013e04f>] do_page_cache_readahead+0xcf/0x130
[<c013e19f>] page_cache_readahead+0xef/0x1e0
[<c013765c>] do_generic_mapping_read+0x11c/0x4d0
[<c0137cae>] __generic_file_aio_read+0x1be/0x1f0
[<c0137a10>] file_read_actor+0x0/0xe0
[<c0137e1a>] generic_file_read+0xba/0xe0
[<c011ac24>] do_page_fault+0x194/0x591
[<c011d590>] autoremove_wake_function+0x0/0x60
[<c0126f6b>] update_wall_time+0xb/0x40
[<c012739f>] do_timer+0xdf/0xf0
[<c015270c>] vfs_read+0xbc/0x170
[<c012367d>] __do_softirq+0x7d/0x90
[<c0152a71>] sys_read+0x51/0x80
[<c010603b>] syscall_call+0x7/0xb
Code: 72 08 0f ba a8 90 00 00 00 02 83 c4 0c 5b 5e 5f 5d c3 89 cb e9 cd fe ff
ff 8b 5d 00 e9 c5 fe ff ff 77 08 39 f0 0f 86 94 fe ff ff <0f> 0b 2d 0370 92
44 c0 e9 87 fe ff ff 0f 87 a8 fe ff ff 39 f0
Any idea whether this is a kernel bug or a hardware problem?
Please CC any replies to me.
Sincerely,
--
Kristian
[-- Attachment #2: Type: application/pgp-signature, Size: 189 bytes --]
^ permalink raw reply [flat|nested] 9+ messages in thread* Re: raid5 crash 2004-12-22 22:04 raid5 crash Kristian Eide @ 2004-12-22 22:26 ` Norbert van Nobelen 2004-12-22 23:05 ` Kristian Eide 2004-12-22 23:08 ` Neil Brown 1 sibling, 1 reply; 9+ messages in thread From: Norbert van Nobelen @ 2004-12-22 22:26 UTC (permalink / raw) To: Kristian Eide; +Cc: linux-kernel The sii 3114 is a RAID controller by itself. Is in not conflicting somewhere (like running software RAID5 and at the same time hardware RAID X?) On Wednesday 22 December 2004 23:04, you wrote: > I am running kernel 2.6.9-gentoo-r10 on an Athlon XP 2400+ computer with a > SiI 3114 SATA controller hosting 4 WD2500JD-00G drives. I have combined > these drives into a raid5 array using software raid, but unfortunately the > array is not stable. I have tried several filesystems (ext3, reiserfs, > xfs), but after copying several gigabytes of data into the array (using > scp) and then trying to read them back (using rsync to compare over the > network) always results in data corruption. Here is the output from > 'dmesg': > > kernel BUG at drivers/md/raid5.c:813! > invalid operand: 0000 [#1] > Modules linked in: sata_sil libata sbp2 ohci1394 ieee1394 usb_storage > ehci_hcd usbcore > CPU: 0 > EIP: 0060:[<c039cdd2>] Not tainted VLI > EFLAGS: 00010006 (2.6.9-gentoo-r10) > EIP is at add_stripe_bio+0x1c2/0x200 > eax: 00045168 ebx: d3974b00 ecx: d3974980 edx: 00000000 > esi: 00045140 edi: 00000000 ebp: e33200a4 esp: f0a05ac4 > ds: 007b es: 007b ss: 0068 > Process rsync (pid: 32092, threadinfo=f0a04000 task=f6c10020) > Stack: 00000000 00000296 00000140 e3320028 00045140 00000000 d3974980 > c039e092 e3320028 d3974980 00000000 00000000 00000000 f0a05b1c de3e1ae0 > 00045158 00000000 00000003 00000004 de3e1ae0 dfe90e00 00000000 00000003 > f7d85088 Call Trace: > [<c039e092>] make_request+0x122/0x200 > [<c032dc6f>] generic_make_request+0x15f/0x1e0 > [<c011d590>] autoremove_wake_function+0x0/0x60 > [<c032dd4d>] submit_bio+0x5d/0x100 > [<c0172d43>] mpage_bio_submit+0x23/0x40 > [<c0173170>] do_mpage_readpage+0x2d0/0x480 > [<c012367d>] __do_softirq+0x7d/0x90 > [<c02b54df>] radix_tree_node_alloc+0x1f/0x60 > [<c02b5762>] radix_tree_insert+0xe2/0x100 > [<c0136e54>] add_to_page_cache+0x54/0x80 > [<c017346b>] mpage_readpages+0x14b/0x180 > [<c018f1f0>] reiserfs_get_block+0x0/0x1450 > [<c013ddf4>] read_pages+0x134/0x140 > [<c018f1f0>] reiserfs_get_block+0x0/0x1450 > [<c013b390>] __alloc_pages+0x1d0/0x370 > [<c01081c5>] do_IRQ+0xc5/0xe0 > [<c013e04f>] do_page_cache_readahead+0xcf/0x130 > [<c013e19f>] page_cache_readahead+0xef/0x1e0 > [<c013765c>] do_generic_mapping_read+0x11c/0x4d0 > [<c0137cae>] __generic_file_aio_read+0x1be/0x1f0 > [<c0137a10>] file_read_actor+0x0/0xe0 > [<c0137e1a>] generic_file_read+0xba/0xe0 > [<c011ac24>] do_page_fault+0x194/0x591 > [<c011d590>] autoremove_wake_function+0x0/0x60 > [<c0126f6b>] update_wall_time+0xb/0x40 > [<c012739f>] do_timer+0xdf/0xf0 > [<c015270c>] vfs_read+0xbc/0x170 > [<c012367d>] __do_softirq+0x7d/0x90 > [<c0152a71>] sys_read+0x51/0x80 > [<c010603b>] syscall_call+0x7/0xb > Code: 72 08 0f ba a8 90 00 00 00 02 83 c4 0c 5b 5e 5f 5d c3 89 cb e9 cd fe > ff ff 8b 5d 00 e9 c5 fe ff ff 77 08 39 f0 0f 86 94 fe ff ff <0f> 0b 2d 0370 > 92 44 c0 e9 87 fe ff ff 0f 87 a8 fe ff ff 39 f0 > > Any idea whether this is a kernel bug or a hardware problem? > Please CC any replies to me. > > Sincerely, ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: raid5 crash 2004-12-22 22:26 ` Norbert van Nobelen @ 2004-12-22 23:05 ` Kristian Eide 0 siblings, 0 replies; 9+ messages in thread From: Kristian Eide @ 2004-12-22 23:05 UTC (permalink / raw) To: Norbert van Nobelen; +Cc: linux-kernel [-- Attachment #1: Type: text/plain, Size: 290 bytes --] > The sii 3114 is a RAID controller by itself. > Is in not conflicting somewhere (like running software RAID5 and at the > same time hardware RAID X?) No. The SiI 3114 is only being used as an SATA controller; I have not configured any hardware raid. Sincerely, -- Kristian [-- Attachment #2: Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: raid5 crash 2004-12-22 22:04 raid5 crash Kristian Eide 2004-12-22 22:26 ` Norbert van Nobelen @ 2004-12-22 23:08 ` Neil Brown 2004-12-23 9:51 ` Prakash K. Cheemplavam ` (2 more replies) 1 sibling, 3 replies; 9+ messages in thread From: Neil Brown @ 2004-12-22 23:08 UTC (permalink / raw) To: Kristian Eide; +Cc: linux-kernel On Wednesday December 22, kreide@online.no wrote: > I am running kernel 2.6.9-gentoo-r10 on an Athlon XP 2400+ computer with a SiI > 3114 SATA controller hosting 4 WD2500JD-00G drives. I have combined these > drives into a raid5 array using software raid, but unfortunately the array is > not stable. I have tried several filesystems (ext3, reiserfs, xfs), but after > copying several gigabytes of data into the array (using scp) and then trying > to read them back (using rsync to compare over the network) always results in > data corruption. Here is the output from 'dmesg': > > kernel BUG at drivers/md/raid5.c:813! This BUG happens when there are two outstanding read (or write) requests for the same piece of storage (more accurately, two "bio"s that overlap). raid5 cannot currently handle this situation. Most filesystems would never make requests like this. I note that you are using reiserfs in this case. It is possible that reiserfs with tail-packing enabled could do this. I doubt very much that this would happen with ext3. I don't know about xfs, but I doubt it would happen their either. When using some other filesystem, what sort of data corruption are you getting? NeilBrown ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: raid5 crash 2004-12-22 23:08 ` Neil Brown @ 2004-12-23 9:51 ` Prakash K. Cheemplavam 2004-12-23 19:45 ` Kristian Eide 2005-01-13 14:58 ` raid5 crash Stephen C. Tweedie 2 siblings, 0 replies; 9+ messages in thread From: Prakash K. Cheemplavam @ 2004-12-23 9:51 UTC (permalink / raw) To: Neil Brown; +Cc: Kristian Eide, linux-kernel [-- Attachment #1: Type: text/plain, Size: 804 bytes --] Neil Brown schrieb: > On Wednesday December 22, kreide@online.no wrote: > >>I am running kernel 2.6.9-gentoo-r10 on an Athlon XP 2400+ computer with a SiI >>3114 SATA controller hosting 4 WD2500JD-00G drives. I have combined these >>drives into a raid5 array using software raid, but unfortunately the array is >>not stable. I have tried several filesystems (ext3, reiserfs, xfs), but after >>copying several gigabytes of data into the array (using scp) and then trying >>to read them back (using rsync to compare over the network) always results in >>data corruption. Here is the output from 'dmesg': >> >>kernel BUG at drivers/md/raid5.c:813! Have you a bios option called ext-p2p discard time? Try setting it higher. I posted another thread about sii3112 at lkml about this issue... Prakash [-- Attachment #2: OpenPGP digital signature --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: raid5 crash 2004-12-22 23:08 ` Neil Brown 2004-12-23 9:51 ` Prakash K. Cheemplavam @ 2004-12-23 19:45 ` Kristian Eide 2005-01-03 23:30 ` raid5 crash (possible VM problem???) Neil Brown 2005-01-13 14:58 ` raid5 crash Stephen C. Tweedie 2 siblings, 1 reply; 9+ messages in thread From: Kristian Eide @ 2004-12-23 19:45 UTC (permalink / raw) To: Neil Brown; +Cc: linux-kernel [-- Attachment #1: Type: text/plain, Size: 2741 bytes --] > I doubt very much that this would happen with ext3. I don't know > about xfs, but I doubt it would happen their either. > When using some other filesystem, what sort of data corruption are you > getting? This is with ext3: kernel BUG at drivers/md/raid5.c:813! invalid operand: 0000 [#1] Modules linked in: sata_sil libata sbp2 ohci1394 ieee1394 usb_storage ehci_hcd usbcore CPU: 0 EIP: 0060:[<c039cdd2>] Not tainted VLI EFLAGS: 00010016 (2.6.9-gentoo-r10) EIP is at add_stripe_bio+0x1c2/0x200 eax: 493c4b40 ebx: cbc5eaa0 ecx: e8257da0 edx: 00000000 esi: 493c4b18 edi: 00000000 ebp: f58958b0 esp: f5995a98 ds: 007b es: 007b ss: 0068 Process rsync (pid: 8803, threadinfo=f5994000 task=d841aaa0) Stack: c01551f7 f7ddf200 00000118 f58957c8 493c4b18 00000000 e8257da0 c039e092 f58957c8 e8257da0 00000001 00000000 00000000 f5995af0 f5fde2e0 493c4b30 00000000 00000003 00000004 f5fde2e0 dfe90e00 00000001 00000000 f7d84088 Call Trace: [<c01551f7>] __getblk+0x37/0x70 [<c039e092>] make_request+0x122/0x200 [<c032dc6f>] generic_make_request+0x15f/0x1e0 [<c032dd4d>] submit_bio+0x5d/0x100 [<c01b5ec4>] ext3_get_block+0x64/0xb0 [<c0172d43>] mpage_bio_submit+0x23/0x40 [<c0173170>] do_mpage_readpage+0x2d0/0x480 [<c0332b08>] as_next_request+0x38/0x50 [<c032a706>] elv_next_request+0x16/0x110 [<c02b4c1e>] kobject_put+0x1e/0x30 [<c02b4bf0>] kobject_release+0x0/0x10 [<c02b54df>] radix_tree_node_alloc+0x1f/0x60 [<c02b5762>] radix_tree_insert+0xe2/0x100 [<c0136e54>] add_to_page_cache+0x54/0x80 [<c017346b>] mpage_readpages+0x14b/0x180 [<c01b5e60>] ext3_get_block+0x0/0xb0 [<c013ddf4>] read_pages+0x134/0x140 [<c01b5e60>] ext3_get_block+0x0/0xb0 [<c013b390>] __alloc_pages+0x1d0/0x370 [<c013e04f>] do_page_cache_readahead+0xcf/0x130 [<c013e234>] page_cache_readahead+0x184/0x1e0 [<c013765c>] do_generic_mapping_read+0x11c/0x4d0 [<c0137cae>] __generic_file_aio_read+0x1be/0x1f0 [<c0137a10>] file_read_actor+0x0/0xe0 [<c02b4c1e>] kobject_put+0x1e/0x30 [<c0137d3a>] generic_file_aio_read+0x5a/0x80 [<c015261e>] do_sync_read+0xbe/0xf0 [<f92cdc66>] ata_qc_complete+0x46/0xe0 [libata] [<c011d590>] autoremove_wake_function+0x0/0x60 [<c036b9aa>] scsi_finish_command+0x7a/0xc0 [<c011bb9f>] recalc_task_prio+0x8f/0x190 [<c015270c>] vfs_read+0xbc/0x170 [<c040e363>] schedule+0x2a3/0x470 [<c0152a71>] sys_read+0x51/0x80 [<c010603b>] syscall_call+0x7/0xb Code: 72 08 0f ba a8 90 00 00 00 02 83 c4 0c 5b 5e 5f 5d c3 89 cb e9 cd fe ff ff 8b 5d 00 e9 c5 fe ff ff 77 08 39 f0 0f 86 94 fe ff ff <0f> 0b 2d 0370 92 44 c0 e9 87 fe ff ff 0f 87 a8 fe ff ff 39 f0 So apparently it can happen with ext3 as well. -- Kristian [-- Attachment #2: Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: raid5 crash (possible VM problem???) 2004-12-23 19:45 ` Kristian Eide @ 2005-01-03 23:30 ` Neil Brown 2005-01-16 18:33 ` Kristian Eide 0 siblings, 1 reply; 9+ messages in thread From: Neil Brown @ 2005-01-03 23:30 UTC (permalink / raw) To: Kristian Eide; +Cc: linux-kernel On Thursday December 23, kreide@online.no wrote: > > I doubt very much that this would happen with ext3. I don't know > > about xfs, but I doubt it would happen their either. > > When using some other filesystem, what sort of data corruption are you > > getting? > > This is with ext3: > > kernel BUG at drivers/md/raid5.c:813! This is very suspect... This BUG is triggered if raid5 receives a read request while there is already an outstanding read request that overlaps the new one. i.e. this->sector < that->sector and this->sector + this->size > that->sector I admit that my understanding of ext3 isn't perfect, but I would be VERY surprised if ext3 would ever submit overlapping requests. I guess it could happen if the filesystem were corrupted (and two files claimed to own the same block) but I doubt that is the case here. I suspect there is a problem somewhere else.. in the VM maybe?? You could try this patch. It might hide the real problem, or it might cause it to manifest in some other way. It changes the raid5 behaviour so that if an overlap is found, it doesn't BUG, but instead waits a little while and tries again. NeilBrown Signed-off-by: Neil Brown <neilb@cse.unsw.edu.au> ### Diffstat output ./drivers/md/raid5.c | 18 ++++++++++++++---- 1 files changed, 14 insertions(+), 4 deletions(-) diff ./drivers/md/raid5.c~current~ ./drivers/md/raid5.c --- ./drivers/md/raid5.c~current~ 2004-12-20 12:24:19.000000000 +1100 +++ ./drivers/md/raid5.c 2004-12-28 17:02:44.000000000 +1100 @@ -232,6 +232,7 @@ static struct stripe_head *__find_stripe } static void unplug_slaves(mddev_t *mddev); +static void raid5_unplug_device(request_queue_t *q); static struct stripe_head *get_active_stripe(raid5_conf_t *conf, sector_t sector, int pd_idx, int noblock) @@ -793,7 +794,7 @@ static void compute_parity(struct stripe * toread/towrite point to the first in a chain. * The bi_next chain must be in order. */ -static void add_stripe_bio (struct stripe_head *sh, struct bio *bi, int dd_idx, int forwrite) +static int add_stripe_bio (struct stripe_head *sh, struct bio *bi, int dd_idx, int forwrite) { struct bio **bip; raid5_conf_t *conf = sh->raid_conf; @@ -810,10 +811,10 @@ static void add_stripe_bio (struct strip else bip = &sh->dev[dd_idx].toread; while (*bip && (*bip)->bi_sector < bi->bi_sector) { - BUG_ON((*bip)->bi_sector + ((*bip)->bi_size >> 9) > bi->bi_sector); + if ((*bip)->bi_sector + ((*bip)->bi_size >> 9) > bi->bi_sector) + return 0; /* cannot add just now due to overlap */ bip = & (*bip)->bi_next; } -/* FIXME do I need to worry about overlapping bion */ if (*bip && bi->bi_next && (*bip) != bi->bi_next) BUG(); if (*bip) @@ -840,6 +841,7 @@ static void add_stripe_bio (struct strip if (sector >= sh->dev[dd_idx].sector + STRIPE_SECTORS) set_bit(R5_OVERWRITE, &sh->dev[dd_idx].flags); } + return 1; } @@ -1413,7 +1415,15 @@ static int make_request (request_queue_t sh = get_active_stripe(conf, new_sector, pd_idx, (bi->bi_rw&RWA_MASK)); if (sh) { - add_stripe_bio(sh, bi, dd_idx, (bi->bi_rw&RW_MASK)); + while (!add_stripe_bio(sh, bi, dd_idx, (bi->bi_rw&RW_MASK))) { + /* add failed due to overlap. Flush everything + * and wait a while + * FIXME - overlapping requests should be handled better + */ + raid5_unplug_device(mddev->queue); + set_current_state(TASK_UNINTERRUPTIBLE); + schedule_timeout(1); + } raid5_plug_device(conf); handle_stripe(sh); ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: raid5 crash (possible VM problem???) 2005-01-03 23:30 ` raid5 crash (possible VM problem???) Neil Brown @ 2005-01-16 18:33 ` Kristian Eide 0 siblings, 0 replies; 9+ messages in thread From: Kristian Eide @ 2005-01-16 18:33 UTC (permalink / raw) To: Neil Brown; +Cc: linux-kernel [-- Attachment #1: Type: text/plain, Size: 1209 bytes --] > You could try this patch. It might hide the real problem, or it might > cause it to manifest in some other way. I've applied the patch, but I now get another error; this might not be related to raid5, however, I have tested the individual SATA disks without getting any errors. This only happens when combining them into a raid5 array (other raid levels not tested). ReiserFS: md3: journal params: device md3, size 8192, journal first block 18, max trans len 1024, max batch 900, max commit age 30, max trans age 30 ReiserFS: md3: checking transaction log (md3) ReiserFS: md3: Using r5 hash to sort names attempt to access beyond end of device md3: rw=0, want=18446744063991695384, limit=1465175040 attempt to access beyond end of device md3: rw=0, want=18446744062355374704, limit=1465175040 attempt to access beyond end of device md3: rw=0, want=4913837584, limit=1465175040 attempt to access beyond end of device md3: rw=0, want=18446744071656162744, limit=1465175040 I have 4 250GB SATA disk combined into one raid5 volume (kernel 2.6.10), and this error happens after copying a few gigabytes of data into the volume and then trying to read them back. -- Kristian [-- Attachment #2: Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: raid5 crash 2004-12-22 23:08 ` Neil Brown 2004-12-23 9:51 ` Prakash K. Cheemplavam 2004-12-23 19:45 ` Kristian Eide @ 2005-01-13 14:58 ` Stephen C. Tweedie 2 siblings, 0 replies; 9+ messages in thread From: Stephen C. Tweedie @ 2005-01-13 14:58 UTC (permalink / raw) To: Neil Brown; +Cc: Stephen Tweedie, Kristian Eide, linux-kernel Hi, On Wed, 2004-12-22 at 23:08, Neil Brown wrote: > > kernel BUG at drivers/md/raid5.c:813! > > This BUG happens when there are two outstanding read (or write) > requests for the same piece of storage (more accurately, two "bio"s > that overlap). Ouch. > raid5 cannot currently handle this situation. > Most filesystems would never make requests like this. I don't think there's anything to stop me doing two O_DIRECT read()s from the same bit of a file at the same time. As far as I can see, this should be easily triggered by user-space. And if you get corruption in a filesystem such that two files share the same block, then this possibility arises again. That can happen due to corruption in an indirect block (ext2/3) or in the reiserfs tree; or more commonly due to a bitmap block getting corrupted, resulting in the same block being allocated twice. This is a situation we really need to handle. ext3 goes to great lengths to make sure that if such cases happen, the worst that results should be the filesystem taking itself readonly cleanly. It's really bad behaviour for a fault such as a bad IDE cable to be able to oops the entire kernel. > I doubt very much that this would happen with ext3. It certainly shouldn't do so for buffered IO if things are running fine, but as soon as you get corrupt data, or start using O_DIRECT, it's possible. Cheers, Stephen ^ permalink raw reply [flat|nested] 9+ messages in thread
end of thread, other threads:[~2005-01-16 18:34 UTC | newest] Thread overview: 9+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2004-12-22 22:04 raid5 crash Kristian Eide 2004-12-22 22:26 ` Norbert van Nobelen 2004-12-22 23:05 ` Kristian Eide 2004-12-22 23:08 ` Neil Brown 2004-12-23 9:51 ` Prakash K. Cheemplavam 2004-12-23 19:45 ` Kristian Eide 2005-01-03 23:30 ` raid5 crash (possible VM problem???) Neil Brown 2005-01-16 18:33 ` Kristian Eide 2005-01-13 14:58 ` raid5 crash Stephen C. Tweedie
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox