* [linux-lvm] Re: Re: Problems with dissapearing PV when mounting (Stuart D. Gathman)
@ 2009-12-07 18:47 Johan Gardell
2009-12-07 19:34 ` Stuart D. Gathman
2009-12-07 23:11 ` malahal
0 siblings, 2 replies; 4+ messages in thread
From: Johan Gardell @ 2009-12-07 18:47 UTC (permalink / raw)
To: linux-lvm
Ok, added a filter to remove /dev/fd0. But i still get
[22723.980390] device-mapper: table: 254:1: linear: dm-linear: Device
lookup failed
[22723.980395] device-mapper: ioctl: error adding target to table
[22724.001153] device-mapper: table: 254:2: linear: dm-linear: Device
lookup failed
[22724.001158] device-mapper: ioctl: error adding target to table
It doesn't mention fd0 as far as i understand?
Yes, if i remember correctly, the partition im trying to read on the
vg gardin is called root. The filesystem is ReiserFS. My current root
on the vg Dreamhack is also a ReiserFS, so i have all the modules
loaded.
mount doesn't give any messages in dmesg
lvs shows:
LV VG Attr LSize Origin Snap% Move Log Copy% Convert
dreamhacklv Dreamhack -wi-ao 1,23t
root gardin -wi-d- 928,00g
swap_1 gardin -wi-d- 2,59g
if i try to mount with:
mount -t reiserfs /dev/mapper/gardin-root /mnt/tmp
i get this in dmesg:
[23113.711247] REISERFS warning (device dm-1): sh-2006
read_super_block: bread failed (dev dm-1, block 2, size 4096)
[23113.711257] REISERFS warning (device dm-1): sh-2006
read_super_block: bread failed (dev dm-1, block 16, size 4096)
[23113.711261] REISERFS warning (device dm-1): sh-2021
reiserfs_fill_super: can not find reiserfs on dm-1
//Johan
2009/12/7 <linux-lvm-request@redhat.com>:
> Send linux-lvm mailing list submissions to
> � � � �linux-lvm@redhat.com
>
> To subscribe or unsubscribe via the World Wide Web, visit
> � � � �https://www.redhat.com/mailman/listinfo/linux-lvm
> or, via email, send a message with subject or body 'help' to
> � � � �linux-lvm-request@redhat.com
>
> You can reach the person managing the list at
> � � � �linux-lvm-owner@redhat.com
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of linux-lvm digest..."
>
>
> Today's Topics:
>
> � 1. Re: Questions regarding LVM (malahal@us.ibm.com)
> � 2. Re: Questions regarding LVM (Ray Morris)
> � 3. Re: Questions regarding LVM (Stuart D. Gathman)
> � 4. Problems with dissapearing PV when mounting (Johan Gardell)
> � 5. Re: Problems with dissapearing PV when mounting
> � � �(Stuart D. Gathman)
> � 6. LVM crash maybe due to a drbd issue (Maxence DUNNEWIND)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Thu, 3 Dec 2009 10:42:19 -0800
> From: malahal@us.ibm.com
> Subject: Re: [linux-lvm] Questions regarding LVM
> To: linux-lvm@redhat.com
> Message-ID: <20091203184219.GA29968@us.ibm.com>
> Content-Type: text/plain; charset=us-ascii
>
> Vishal Verma -X (vishaver - Embedded Resource Group at Cisco) [vishaver@cisco.com] wrote:
>> � �1. � � � Under scenario where, several hard-drives are part of LVM volume
>> � �group and if one of hard-disk gets corrupted then would whole volume group
>> � �be inaccessible ?
>
> No.
>
>> � �What would be impact on volume group's filesystem ?
>
> A volume group may have several file system images. You should have no
> problem in accessing logical volumes (or file systems on them) that
> don't include the corrupted/failed disk. Obviously, logical volumes that
> include the corrupted/failed disk will have problems unless it is
> a mirrored logical volume!
>
>> � �2. � � � From stability perspective, which version of LVM is better on
>> � �Linux kernel 2.6.x, LVM2 or LVM1 ?
>
> I would go with LVM2.
>
>
>
> ------------------------------
>
> Message: 2
> Date: Thu, 03 Dec 2009 12:43:56 -0600
> From: Ray Morris <support@bettercgi.com>
> Subject: Re: [linux-lvm] Questions regarding LVM
> To: LVM general discussion and development <linux-lvm@redhat.com>
> Message-ID: <1259865836.6713.13@raydesk1.bettercgi.com>
> Content-Type: text/plain; charset=us-ascii; DelSp=Yes; Format=Flowed
>
>> 1. � � � Under scenario where, several hard-drives are part of LVM
>> volume group and if one of hard-disk gets corrupted then would whole
>> volume group be inaccessible ?
>
>
> See the --partial option. �See also vgcfgrestore. (rtfm)
> Basically, if the drive is merely "corrupted" as you said,
> there could be some currupted data. �If the drive is missing
> or largely unuseable those extents are gone, so LVs with
> important extents on that PV would have serious problems,
> but LVs with no extents on that PV should be fine. �"Important"
> extents means, for example, if a 400GB LV which is only 10%
> full has only it's last couple on extents on the missing PV
> that may not be a major problem, if no data is stored there
> yet. �On the other hand, the first extent of the filesystem
> probably contains very important information about the
> filesystem as a whole, so if that first extent is unuseable
> you're probably reduced to greping the LV, or using an automated
> search tool that basically greps the LV - the same type of
> tools used for undelete or corrupted disks without LVM.
>
>> What would be impact on volume group's filesystem ?
>
> � �VGs don't have filesystems. �LVs do. This is the same
> question as "if I'm NOT using LVM and parts of my drive go
> bad what is the effect on the filesystem?" The ability to
> salvage files from the filesystems of affected LVs depends
> on how many extents are missing or corrupted, which extents
> those are, and what type of filesystem is used.
>
> � �So in summary, LVM doesn't change much in terms of the
> affect of a bad disk. �You should still have really solid
> backups and probably use RAID.
> --
> Ray Morris
> support@bettercgi.com
>
> Strongbox - The next generation in site security:
> http://www.bettercgi.com/strongbox/
>
> Throttlebox - Intelligent Bandwidth Control
> http://www.bettercgi.com/throttlebox/
>
> Strongbox / Throttlebox affiliate program:
> http://www.bettercgi.com/affiliates/user/register.php
>
>
> On 12/03/2009 12:22:11 PM, Vishal Verma -X (vishaver - Embedded
> Resource Group at Cisco) wrote:
>> Hello all,
>>
>>
>>
>> � � I am new to this mailing list. I have few questions regarding
>> Linux
>> LVM, would appreciate if LVM gurus could answer.
>>
>>
>>
>>
>>
>> 1. � � � Under scenario where, several hard-drives are part of LVM
>> volume group and if one of hard-disk gets corrupted then would whole
>> volume group be inaccessible ?
>>
>> What would be impact on volume group's filesystem ?
>>
>>
>>
>> 2. � � � From stability perspective, which version of LVM is better on
>> Linux kernel 2.6.x, LVM2 or LVM1 ?
>>
>>
>>
>> Regards,
>>
>> Vishal
>>
>>
>>
>>
>
> ------quoted attachment------
>> _______________________________________________
>> linux-lvm mailing list
>> linux-lvm@redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-lvm
>> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>
>
>
>
> ------------------------------
>
> Message: 3
> Date: Thu, 3 Dec 2009 13:50:52 -0500 (EST)
> From: "Stuart D. Gathman" <stuart@bmsi.com>
> Subject: Re: [linux-lvm] Questions regarding LVM
> To: LVM general discussion and development <linux-lvm@redhat.com>
> Message-ID: <Pine.LNX.4.64.0912031342400.13455@bmsred.bmsi.com>
> Content-Type: TEXT/PLAIN; charset=US-ASCII
>
> On Thu, 3 Dec 2009, Vishal Verma -X (vishaver - Embedded Resource Group at Cisco) wrote:
>
>> 1. � � � Under scenario where, several hard-drives are part of LVM
>> volume group and if one of hard-disk gets corrupted then would whole
>> volume group be inaccessible ?
>
> Short answer: it depends
>
> Long answer: For raw plain hard drives, only logical volumes using the
> affected drive are inaccessible. �Think of LVM as managing fancy run-time
> expandable partitions. �You may wish to ensure that the LVM metadata (the LVM
> equivalent of the partition table) is stored in multiple locations, or
> frequently backed up from /etc/lvm.
>
> More often, the "drives" that are part of LVM are software or hardware RAID
> drives. �In addition, Linux LVM supports mirroring (RAID1) at the LV level
> - although not yet as smoothly as other LVM systems.
>
>> What would be impact on volume group's filesystem ?
>
> Same as with any other partition that goes south.
>
>> 2. � � � From stability perspective, which version of LVM is better on
>> Linux kernel 2.6.x, LVM2 or LVM1 ?
>
> LVM2
>
> --
> � � � � � � �Stuart D. Gathman <stuart@bmsi.com>
> � �Business Management Systems Inc. �Phone: 703 591-0911 Fax: 703 591-6154
> "Confutatis maledictis, flammis acribus addictis" - background song for
> a Microsoft sponsored "Where do you want to go from here?" commercial.
>
>
>
> ------------------------------
>
> Message: 4
> Date: Sun, 6 Dec 2009 19:47:55 +0100
> From: Johan Gardell <gardin@gmail.com>
> Subject: [linux-lvm] Problems with dissapearing PV when mounting
> To: linux-lvm@redhat.com
> Message-ID:
> � � � �<691e2b620912061047i13bd740eq947dc2a67086e439@mail.gmail.com>
> Content-Type: text/plain; charset=ISO-8859-1
>
> Hi all!
>
> I am having some troubles with mounting a vg consisting of two PVs.
> After booting, the partition /dev/sdb5 does not pop up (but 1-3 does
> and 4th is a swap). If i issue partprobe -s it does show up though.
>
> partprobe shows:
>
> /dev/sda: msdos partitions 1 2 3
> /dev/sdb: msdos partitions 1 2 <5>
> /dev/sdc: msdos partitions 1 3 2
>
> I don't know what <> means but that is one of the PVs
>
> If i, after issuing partprobe, use pvscan it shows:
> �PV /dev/sdc2 � VG Dreamhack � lvm2 [1,23 TiB / 0 � �free]
> �PV /dev/sdb5 � VG gardin � � �lvm2 [465,52 GiB / 0 � �free]
> �PV /dev/sdd � �VG gardin � � �lvm2 [465,76 GiB / 704,00 MiB free]
> �Total: 3 [2,14 TiB] / in use: 3 [2,14 TiB] / in no VG: 0 [0 � ]
>
> vgscan:
> �Reading all physical volumes. �This may take a while...
> �Found volume group "Dreamhack" using metadata type lvm2
> �Found volume group "gardin" using metadata type lvm2
>
> But the problems appear when:
> vgchange -ay gardin
> �device-mapper: reload ioctl failed: Invalid argument
> �device-mapper: reload ioctl failed: Invalid argument
> �2 logical volume(s) in volume group "gardin" now active
>
> Where dmesg shows:
> [31936.135588] device-mapper: table: 254:1: linear: dm-linear: Device
> lookup failed
> [31936.135592] device-mapper: ioctl: error adding target to table
> [31936.150572] device-mapper: table: 254:2: linear: dm-linear: Device
> lookup failed
> [31936.150576] device-mapper: ioctl: error adding target to table
> [31940.024525] end_request: I/O error, dev fd0, sector 0
>
> And trying to mount the vg:
> mount /dev/mapper/gardin-root /mnt/tmp
> �mount: you must specify the filesystem type
>
> I have googled some but can't find much about this issue, does anyone
> have any ideas how i can obtain the data stored on the disk? Because i
> really need it..
>
> I am running debian squeeze with a 2.6.30-2-686 kernel, the partitions
> were originally created under debian lenny (don't know the kernel
> version back then though)
>
> lvm version shows:
> �LVM version: � � 2.02.54(1) (2009-10-26)
> �Library version: 1.02.39 (2009-10-26)
> �Driver version: �4.14.0
>
> Thanks in regards
> //Johan
>
>
>
> ------------------------------
>
> Message: 5
> Date: Sun, 6 Dec 2009 21:22:13 -0500 (EST)
> From: "Stuart D. Gathman" <stuart@bmsi.com>
> Subject: Re: [linux-lvm] Problems with dissapearing PV when mounting
> To: LVM general discussion and development <linux-lvm@redhat.com>
> Message-ID: <Pine.LNX.4.64.0912062114210.24919@bmsred.bmsi.com>
> Content-Type: TEXT/PLAIN; charset=US-ASCII
>
> On Sun, 6 Dec 2009, Johan Gardell wrote:
>
>> But the problems appear when:
>> vgchange -ay gardin
>> � device-mapper: reload ioctl failed: Invalid argument
>> � device-mapper: reload ioctl failed: Invalid argument
>> � 2 logical volume(s) in volume group "gardin" now active
>>
>> Where dmesg shows:
>> [31936.135588] device-mapper: table: 254:1: linear: dm-linear: Device
>> lookup failed
>> [31936.135592] device-mapper: ioctl: error adding target to table
>> [31936.150572] device-mapper: table: 254:2: linear: dm-linear: Device
>> lookup failed
>> [31936.150576] device-mapper: ioctl: error adding target to table
>> [31940.024525] end_request: I/O error, dev fd0, sector 0
>
> The fd0 error means you need to exclude /dev/fd0 from the vgscan in
> /etc/lvm/lvm.conf (or wherever Debian puts it). �Start to worry
> if you get I/O errors on sdb or sdc.
>
>> And trying to mount the vg:
>> mount /dev/mapper/gardin-root /mnt/tmp
>> � mount: you must specify the filesystem type
>
> You need to show us the output of 'lvs'. �(Even if it just spits out an error.)
>
> Is there an LV named "root"? �I think so, since it would say "special device
> does not exist" if not. �But what kind of filesystem did it have
> on it? �Maybe it is not an auto-detected one, or you need to load
> the filesystem driver. �Does mount report I/O error if it gets one
> trying to identify the filesystem?
>
> --
> � � � � � � �Stuart D. Gathman <stuart@bmsi.com>
> � �Business Management Systems Inc. �Phone: 703 591-0911 Fax: 703 591-6154
> "Confutatis maledictis, flammis acribus addictis" - background song for
> a Microsoft sponsored "Where do you want to go from here?" commercial.
>
>
>
> ------------------------------
>
> Message: 6
> Date: Mon, 7 Dec 2009 17:40:18 +0100
> From: Maxence DUNNEWIND <maxence@dunnewind.net>
> Subject: [linux-lvm] LVM crash maybe due to a drbd issue
> To: drbd-user@lists.linbit.com
> Message-ID: <20091207164018.GV21917@dunnewind.net>
> Content-Type: text/plain; charset="iso-8859-1"
>
> Hi,
>
> I'm using drbd with lv LVM. The layout is :
> drbd -> lv -> vg -> pv . I'm trying to do a vgchange -ay on the underlying lv,
> which is hanging and use 50% of a cpu. An echo w > /proc/sysrq-trigger gives me
> an interesting thing :
>
> Dec �7 13:35:26 z2-3 kernel: [8617743.246522] SysRq : Show Blocked State
> Dec �7 13:35:26 z2-3 kernel: [8617743.246548] � task � � � � � � � � � � � �PC
> stack � pid father
> Dec �7 13:35:26 z2-3 kernel: [8617743.246561] pdflush � � � D ffff8800280350c0
> 0 14401 � � �2
> Dec �7 13:35:26 z2-3 kernel: [8617743.246564] �ffff88012f04f8d0 0000000000000046
> ffff88000150b808 ffffffff80215d6a
> Dec �7 13:35:26 z2-3 kernel: [8617743.246566] �ffff88000150b7d0 00000000000120c0
> 000000000000e250 ffff88012f0ac770
> Dec �7 13:35:26 z2-3 kernel: [8617743.246569] �ffff88012f0aca60 0000000000000003
> 0000000000000086 ffffffff8023bc87
> Dec �7 13:35:26 z2-3 kernel: [8617743.246571] Call Trace:
> Dec �7 13:35:26 z2-3 kernel: [8617743.246578] �[<ffffffff80215d6a>] ?
> native_sched_clock+0x2e/0x5b
> Dec �7 13:35:26 z2-3 kernel: [8617743.246582] �[<ffffffff8023bc87>] ?
> try_to_wake_up+0x212/0x224
> Dec �7 13:35:26 z2-3 kernel: [8617743.246585] �[<ffffffff8024accf>] ?
> lock_timer_base+0x26/0x4c
> Dec �7 13:35:26 z2-3 kernel: [8617743.246589] �[<ffffffff804b4500>] ?
> schedule+0x9/0x1d
> Dec �7 13:35:26 z2-3 kernel: [8617743.246592] �[<ffffffff804b46eb>] ?
> schedule_timeout+0x90/0xb6
> Dec �7 13:35:26 z2-3 kernel: [8617743.246594] �[<ffffffff8024adc8>] ?
> process_timeout+0x0/0x5
> Dec �7 13:35:26 z2-3 kernel: [8617743.246596] �[<ffffffff804b46e6>] ?
> schedule_timeout+0x8b/0xb6
> Dec �7 13:35:26 z2-3 kernel: [8617743.246598] �[<ffffffff804b3ac3>] ?
> io_schedule_timeout+0x66/0xae
> Dec �7 13:35:26 z2-3 kernel: [8617743.246601] �[<ffffffff802a26f0>] ?
> congestion_wait+0x66/0x80
> Dec �7 13:35:26 z2-3 kernel: [8617743.246604] �[<ffffffff8025473e>] ?
> autoremove_wake_function+0x0/0x2e
> Dec �7 13:35:26 z2-3 kernel: [8617743.246607] �[<ffffffff802d9836>] ?
> writeback_inodes+0x9a/0xce
> Dec �7 13:35:26 z2-3 kernel: [8617743.246610] �[<ffffffff8029888c>] ?
> wb_kupdate+0xc8/0x121
> Dec �7 13:35:26 z2-3 kernel: [8617743.246613] �[<ffffffff802994d7>] ?
> pdflush+0x159/0x23b
> Dec �7 13:35:26 z2-3 kernel: [8617743.246615] �[<ffffffff802987c4>] ?
> wb_kupdate+0x0/0x121
> Dec �7 13:35:26 z2-3 kernel: [8617743.246617] �[<ffffffff8029937e>] ?
> pdflush+0x0/0x23b
> Dec �7 13:35:26 z2-3 kernel: [8617743.246620] �[<ffffffff80254382>] ?
> kthread+0x54/0x80
> Dec �7 13:35:26 z2-3 kernel: [8617743.246622] �[<ffffffff80210aca>] ?
> child_rip+0xa/0x20
> Dec �7 13:35:26 z2-3 kernel: [8617743.246638] �[<ffffffffa02df103>] ?
> handle_halt+0x0/0x12 [kvm_intel]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246641] �[<ffffffff80256fa5>] ?
> hrtimer_cancel+0xc/0x16
> Dec �7 13:35:26 z2-3 kernel: [8617743.246643] �[<ffffffff8025432e>] ?
> kthread+0x0/0x80
> Dec �7 13:35:26 z2-3 kernel: [8617743.246645] �[<ffffffff80210ac0>] ?
> child_rip+0x0/0x20
> Dec �7 13:35:26 z2-3 kernel: [8617743.246655] lvchange � � �D ffff88002804f0c0
> 0 �4969 � � �1
> Dec �7 13:35:26 z2-3 kernel: [8617743.246657] �ffff88000150b7d0 ffffffff804b4484
> 0000000000000096 ffff880084255948
> Dec �7 13:35:26 z2-3 kernel: [8617743.246659] �0000000000000001 00000000000120c0
> 000000000000e250 ffff88000145c040
> Dec �7 13:35:26 z2-3 kernel: [8617743.246662] �ffff88000145c330 0000000180254747
> ffff8800382154d0 ffffffff802348e4
> Dec �7 13:35:26 z2-3 kernel: [8617743.246664] Call Trace:
> Dec �7 13:35:26 z2-3 kernel: [8617743.246667] �[<ffffffff804b4484>] ?
> thread_return+0x3e/0xb1
> Dec �7 13:35:26 z2-3 kernel: [8617743.246670] �[<ffffffff802348e4>] ?
> __wake_up_common+0x44/0x73
> Dec �7 13:35:26 z2-3 kernel: [8617743.246672] �[<ffffffff804b4500>] ?
> schedule+0x9/0x1d
> Dec �7 13:35:26 z2-3 kernel: [8617743.246685] �[<ffffffffa0275491>] ?
> inc_ap_bio+0xde/0x12e [drbd]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246687] �[<ffffffff8025473e>] ?
> autoremove_wake_function+0x0/0x2e
> Dec �7 13:35:26 z2-3 kernel: [8617743.246698] �[<ffffffffa027768a>] ?
> drbd_make_request_26+0x36f/0x485 [drbd]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246701] �[<ffffffff8033d537>] ?
> generic_make_request+0x288/0x2d2
> Dec �7 13:35:26 z2-3 kernel: [8617743.246704] �[<ffffffff8035318f>] ?
> delay_tsc+0x26/0x57
> Dec �7 13:35:26 z2-3 kernel: [8617743.246706] �[<ffffffff8033d647>] ?
> submit_bio+0xc6/0xcd
> Dec �7 13:35:26 z2-3 kernel: [8617743.246709] �[<ffffffff802dd14b>] ?
> submit_bh+0xe3/0x103
> Dec �7 13:35:26 z2-3 kernel: [8617743.246711] �[<ffffffff802df619>] ?
> __block_write_full_page+0x1d6/0x2ac
> Dec �7 13:35:26 z2-3 kernel: [8617743.246713] �[<ffffffff802de3e1>] ?
> end_buffer_async_write+0x0/0x116
> Dec �7 13:35:26 z2-3 kernel: [8617743.246716] �[<ffffffff802e1624>] ?
> blkdev_get_block+0x0/0x57
> Dec �7 13:35:26 z2-3 kernel: [8617743.246718] �[<ffffffff80297e82>] ?
> __writepage+0xa/0x25
> Dec �7 13:35:26 z2-3 kernel: [8617743.246720] �[<ffffffff802985cc>] ?
> write_cache_pages+0x206/0x322
> Dec �7 13:35:26 z2-3 kernel: [8617743.246722] �[<ffffffff80215db6>] ?
> read_tsc+0xa/0x20
> Dec �7 13:35:26 z2-3 kernel: [8617743.246725] �[<ffffffff80297e78>] ?
> __writepage+0x0/0x25
> Dec �7 13:35:26 z2-3 kernel: [8617743.246727] �[<ffffffff802e3e9f>] ?
> __blockdev_direct_IO+0x99a/0xa41
> Dec �7 13:35:26 z2-3 kernel: [8617743.246729] �[<ffffffff802e3e9f>] ?
> __blockdev_direct_IO+0x99a/0xa41
> Dec �7 13:35:26 z2-3 kernel: [8617743.246732] �[<ffffffff80298724>] ?
> do_writepages+0x20/0x2d
> Dec �7 13:35:26 z2-3 kernel: [8617743.246734] �[<ffffffff80292a73>] ?
> __filemap_fdatawrite_range+0x4c/0x57
> Dec �7 13:35:26 z2-3 kernel: [8617743.246736] �[<ffffffff80292aa4>] ?
> filemap_write_and_wait_range+0x26/0x52
> Dec �7 13:35:26 z2-3 kernel: [8617743.246738] �[<ffffffff802932ca>] ?
> generic_file_aio_read+0xd7/0x54f
> Dec �7 13:35:26 z2-3 kernel: [8617743.246749] �[<ffffffffa027b84a>] ?
> drbd_open+0x63/0x6d [drbd]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246752] �[<ffffffff802c0b1b>] ?
> do_sync_read+0xce/0x113
> Dec �7 13:35:26 z2-3 kernel: [8617743.246754] �[<ffffffff802bf4c4>] ?
> __dentry_open+0x16f/0x260
> Dec �7 13:35:26 z2-3 kernel: [8617743.246756] �[<ffffffff802c3e26>] ?
> cp_new_stat+0xe9/0xfc
> Dec �7 13:35:26 z2-3 kernel: [8617743.246758] �[<ffffffff8025473e>] ?
> autoremove_wake_function+0x0/0x2e
> Dec �7 13:35:26 z2-3 kernel: [8617743.246761] �[<ffffffff802e1907>] ?
> block_ioctl+0x38/0x3c
> Dec �7 13:35:26 z2-3 kernel: [8617743.246763] �[<ffffffff802cc016>] ?
> vfs_ioctl+0x21/0x6c
> Dec �7 13:35:26 z2-3 kernel: [8617743.246765] �[<ffffffff802cc48c>] ?
> do_vfs_ioctl+0x42b/0x464
> Dec �7 13:35:26 z2-3 kernel: [8617743.246767] �[<ffffffff802c1585>] ?
> vfs_read+0xa6/0xff
> Dec �7 13:35:26 z2-3 kernel: [8617743.246770] �[<ffffffff802c169a>] ?
> sys_read+0x45/0x6e
> Dec �7 13:35:26 z2-3 kernel: [8617743.246773] �[<ffffffff8020fa42>] ?
> system_call_fastpath+0x16/0x1b
> Dec �7 13:35:26 z2-3 kernel: [8617743.246779] Sched Debug Version: v0.09,
> 2.6.30-1-amd64 #1
> Dec �7 13:35:26 z2-3 kernel: [8617743.246780] now at 8617743246.777559 msecs
> Dec �7 13:35:26 z2-3 kernel: [8617743.246781] � .jiffies
> : 6449328107
> Dec �7 13:35:26 z2-3 kernel: [8617743.246783] � .sysctl_sched_latency
> : 40.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246784] � .sysctl_sched_min_granularity
> : 8.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246786] � .sysctl_sched_wakeup_granularity
> : 10.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246787] � .sysctl_sched_child_runs_first
> : 0.000001
> Dec �7 13:35:26 z2-3 kernel: [8617743.246788] � .sysctl_sched_features
> : 113917
> Dec �7 13:35:26 z2-3 kernel: [8617743.246790]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246790] cpu#0, 3000.452 MHz
> Dec �7 13:35:26 z2-3 kernel: [8617743.246792] � .nr_running � � � � � � � � � �:
> 2
> Dec �7 13:35:26 z2-3 kernel: [8617743.246793] � .load � � � � � � � � � � � � �:
> 4145
> Dec �7 13:35:26 z2-3 kernel: [8617743.246794] � .nr_switches � � � � � � � � � :
> 185493369405
> Dec �7 13:35:26 z2-3 kernel: [8617743.246795] � .nr_load_updates � � � � � � � :
> 239239723
> Dec �7 13:35:26 z2-3 kernel: [8617743.246796] � .nr_uninterruptible � � � � � �:
> 55617
> Dec �7 13:35:26 z2-3 kernel: [8617743.246798] � .next_balance � � � � � � � � �:
> 6449.328222
> Dec �7 13:35:26 z2-3 kernel: [8617743.246799] � .curr->pid � � � � � � � � � � :
> 16194
> Dec �7 13:35:26 z2-3 kernel: [8617743.246800] � .clock � � � � � � � � � � � � :
> 8617743246.431307
> Dec �7 13:35:26 z2-3 kernel: [8617743.246802] � .cpu_load[0] � � � � � � � � � :
> 3121
> Dec �7 13:35:26 z2-3 kernel: [8617743.246803] � .cpu_load[1] � � � � � � � � � :
> 3121
> Dec �7 13:35:26 z2-3 kernel: [8617743.246804] � .cpu_load[2] � � � � � � � � � :
> 3121
> Dec �7 13:35:26 z2-3 kernel: [8617743.246805] � .cpu_load[3] � � � � � � � � � :
> 3121
> Dec �7 13:35:26 z2-3 kernel: [8617743.246806] � .cpu_load[4] � � � � � � � � � :
> 3121
> Dec �7 13:35:26 z2-3 kernel: [8617743.246807]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246808] cfs_rq[0]:/
> Dec �7 13:35:26 z2-3 kernel: [8617743.246809] � .exec_clock � � � � � � � � � �:
> 0.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246810] � .MIN_vruntime � � � � � � � � �:
> 352176404.012861
> Dec �7 13:35:26 z2-3 kernel: [8617743.246812] � .min_vruntime � � � � � � � � �:
> 352176404.012861
> Dec �7 13:35:26 z2-3 kernel: [8617743.246813] � .max_vruntime � � � � � � � � �:
> 352176404.012861
> Dec �7 13:35:26 z2-3 kernel: [8617743.246814] � .spread � � � � � � � � � � � �:
> 0.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246816] � .spread0 � � � � � � � � � � � :
> 0.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246817] � .nr_running � � � � � � � � � �:
> 2
> Dec �7 13:35:26 z2-3 kernel: [8617743.246818] � .load � � � � � � � � � � � � �:
> 4145
> Dec �7 13:35:26 z2-3 kernel: [8617743.246819] � .nr_spread_over � � � � � � � �:
> 0
> Dec �7 13:35:26 z2-3 kernel: [8617743.246820] � .shares � � � � � � � � � � � �:
> 0
> Dec �7 13:35:26 z2-3 kernel: [8617743.246821]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246822] rt_rq[0]:
> Dec �7 13:35:26 z2-3 kernel: [8617743.246823] � .rt_nr_running � � � � � � � � :
> 0
> Dec �7 13:35:26 z2-3 kernel: [8617743.246824] � .rt_throttled � � � � � � � � �:
> 0
> Dec �7 13:35:26 z2-3 kernel: [8617743.246825] � .rt_time � � � � � � � � � � � :
> 0.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246826] � .rt_runtime � � � � � � � � � �:
> 950.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246827]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246828] runnable tasks:
> Dec �7 13:35:26 z2-3 kernel: [8617743.246829] � � � � � � task � PID
> tree-key �switches �prio � � exec-runtime � � � � sum-exec � � � �sum-sleep
> Dec �7 13:35:26 z2-3 kernel: [8617743.246830]
> ----------------------------------------------------------------------------------------------------------
> Dec �7 13:35:26 z2-3 kernel: [8617743.246832] � � � � �kswapd0 � 219
> 352176404.012861 79827478108 � 115 � � � � � � � 0 � � � � � � � 0
> 0.000000 � � � � � � � 0.000000 � � � � � � � 0.000000 /
> Dec �7 13:35:26 z2-3 kernel: [8617743.246843] R � � � � � bash 16194
> 352176364.023477 � � �5733 � 120 � � � � � � � 0 � � � � � � � 0
> 0.000000 � � � � � � � 0.000000 � � � � � � � 0.000000 /
> Dec �7 13:35:26 z2-3 kernel: [8617743.246848]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246849] cpu#1, 3000.452 MHz
> Dec �7 13:35:26 z2-3 kernel: [8617743.246850] � .nr_running � � � � � � � � � �:
> 3
> Dec �7 13:35:26 z2-3 kernel: [8617743.246851] � .load � � � � � � � � � � � � �:
> 1024
> Dec �7 13:35:26 z2-3 kernel: [8617743.246852] � .nr_switches � � � � � � � � � :
> 227249168097
> Dec �7 13:35:26 z2-3 kernel: [8617743.246854] � .nr_load_updates � � � � � � � :
> 236708707
> Dec �7 13:35:26 z2-3 kernel: [8617743.246855] � .nr_uninterruptible � � � � � �:
> -55614
> Dec �7 13:35:26 z2-3 kernel: [8617743.246856] � .next_balance � � � � � � � � �:
> 6449.328121
> Dec �7 13:35:26 z2-3 kernel: [8617743.246857] � .curr->pid � � � � � � � � � � :
> 32275
> Dec �7 13:35:26 z2-3 kernel: [8617743.246859] � .clock � � � � � � � � � � � � :
> 8617743246.856932
> Dec �7 13:35:26 z2-3 kernel: [8617743.246860] � .cpu_load[0] � � � � � � � � � :
> 3072
> Dec �7 13:35:26 z2-3 kernel: [8617743.246861] � .cpu_load[1] � � � � � � � � � :
> 3072
> Dec �7 13:35:26 z2-3 kernel: [8617743.246862] � .cpu_load[2] � � � � � � � � � :
> 3072
> Dec �7 13:35:26 z2-3 kernel: [8617743.246863] � .cpu_load[3] � � � � � � � � � :
> 3072
> Dec �7 13:35:26 z2-3 kernel: [8617743.246864] � .cpu_load[4] � � � � � � � � � :
> 3072
> Dec �7 13:35:26 z2-3 kernel: [8617743.246865]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246866] cfs_rq[1]:/
> Dec �7 13:35:26 z2-3 kernel: [8617743.246867] � .exec_clock � � � � � � � � � �:
> 0.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246869] � .MIN_vruntime � � � � � � � � �:
> 367012701.328887
> Dec �7 13:35:26 z2-3 kernel: [8617743.246870] � .min_vruntime � � � � � � � � �:
> 367012741.328887
> Dec �7 13:35:26 z2-3 kernel: [8617743.246871] � .max_vruntime � � � � � � � � �:
> 367012701.328887
> Dec �7 13:35:26 z2-3 kernel: [8617743.246873] � .spread � � � � � � � � � � � �:
> 0.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246874] � .spread0 � � � � � � � � � � � :
> 14836337.316026
> Dec �7 13:35:26 z2-3 kernel: [8617743.246875] � .nr_running � � � � � � � � � �:
> 1
> Dec �7 13:35:26 z2-3 kernel: [8617743.246876] � .load � � � � � � � � � � � � �:
> 3072
> Dec �7 13:35:26 z2-3 kernel: [8617743.246878] � .nr_spread_over � � � � � � � �:
> 0
> Dec �7 13:35:26 z2-3 kernel: [8617743.246879] � .shares � � � � � � � � � � � �:
> 0
> Dec �7 13:35:26 z2-3 kernel: [8617743.246880]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246880] rt_rq[1]:
> Dec �7 13:35:26 z2-3 kernel: [8617743.246881] � .rt_nr_running � � � � � � � � :
> 0
> Dec �7 13:35:26 z2-3 kernel: [8617743.246882] � .rt_throttled � � � � � � � � �:
> 0
> Dec �7 13:35:26 z2-3 kernel: [8617743.246883] � .rt_time � � � � � � � � � � � :
> 0.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246885] � .rt_runtime � � � � � � � � � �:
> 950.000000
> Dec �7 13:35:26 z2-3 kernel: [8617743.246886]
> Dec �7 13:35:26 z2-3 kernel: [8617743.246886] runnable tasks:
> Dec �7 13:35:26 z2-3 kernel: [8617743.246887] � � � � � � task � PID
> tree-key �switches �prio � � exec-runtime � � � � sum-exec � � � �sum-sleep
> Dec �7 13:35:26 z2-3 kernel: [8617743.246888]
> ----------------------------------------------------------------------------------------------------------
> Dec �7 13:35:26 z2-3 kernel: [8617743.246891] � � � � � � �kvm 32275
> 367012741.340210 76292027847 � 120 � � � � � � � 0 � � � � � � � 0
> 0.000000 � � � � � � � 0.000000 � � � � � � � 0.000000 /
> Dec �7 13:35:26 z2-3 kernel: [8617743.246896] �drbd16_receiver 30799
> 367012701.341954 123640574974 � 120 � � � � � � � 0 � � � � � � � 0
> 0.000000 � � � � � � � 0.000000 � � � � � � � 0.000000 /
> Dec �7 13:35:26 z2-3 kernel: [8617743.246901]
>
>
> In the second stack trace, I can see a call to drbd_make_request_26, then to
> inc_ap_bio which seems to be blocked.
>
> Anyway, the drbddevice using this lv is down, so it shouldn't block the lvchange
> (I checked with drbdsetup /dev/drbdXX show, only the syncer part is shown).
>
> If someone has any informations about this bug ? Should I report it �(I will
> also mail it on lvm mailing list).
>
> Cheers,
>
> Maxence
>
> --
> Maxence DUNNEWIND
> Contact : maxence@dunnewind.net
> Site : http://www.dunnewind.net
> 06 32 39 39 93
> GPG : 18AE 61E4 D0B0 1C7C AAC9 �E40D 4D39 68DB 0D2E B533
> -------------- next part --------------
> A non-text attachment was scrubbed...
> Name: not available
> Type: application/pgp-signature
> Size: 197 bytes
> Desc: Digital signature
> Url : https://www.redhat.com/archives/linux-lvm/attachments/20091207/8723fd71/attachment.bin
>
> ------------------------------
>
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@redhat.com
> https://www.redhat.com/mailman/listinfo/linux-lvm
>
> End of linux-lvm Digest, Vol 70, Issue 2
> ****************************************
>
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [linux-lvm] Re: Re: Problems with dissapearing PV when mounting (Stuart D. Gathman)
2009-12-07 18:47 [linux-lvm] Re: Re: Problems with dissapearing PV when mounting (Stuart D. Gathman) Johan Gardell
@ 2009-12-07 19:34 ` Stuart D. Gathman
2009-12-07 23:11 ` malahal
1 sibling, 0 replies; 4+ messages in thread
From: Stuart D. Gathman @ 2009-12-07 19:34 UTC (permalink / raw)
To: LVM general discussion and development
On Mon, 7 Dec 2009, Johan Gardell wrote:
> Ok, added a filter to remove /dev/fd0. But i still get
> [22723.980390] device-mapper: table: 254:1: linear: dm-linear: Device
> lookup failed
> [22723.980395] device-mapper: ioctl: error adding target to table
> [22724.001153] device-mapper: table: 254:2: linear: dm-linear: Device
> lookup failed
> [22724.001158] device-mapper: ioctl: error adding target to table
Well, the 'd' in the lvs output means "device present without tables".
I googled on the error msg, and see that a bunch of Ubuntu and Debian
people had to remove evms for lvm to work properly after a certain
kernel upgrade. If that is not the problem, then I would have to start
looking at the source, but perhaps a real guru here could help.
--
Stuart D. Gathman <stuart@bmsi.com>
Business Management Systems Inc. Phone: 703 591-0911 Fax: 703 591-6154
"Confutatis maledictis, flammis acribus addictis" - background song for
a Microsoft sponsored "Where do you want to go from here?" commercial.
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [linux-lvm] Re: Re: Problems with dissapearing PV when mounting (Stuart D. Gathman)
2009-12-07 18:47 [linux-lvm] Re: Re: Problems with dissapearing PV when mounting (Stuart D. Gathman) Johan Gardell
2009-12-07 19:34 ` Stuart D. Gathman
@ 2009-12-07 23:11 ` malahal
1 sibling, 0 replies; 4+ messages in thread
From: malahal @ 2009-12-07 23:11 UTC (permalink / raw)
To: linux-lvm
Johan Gardell [gardin@gmail.com] wrote:
> Ok, added a filter to remove /dev/fd0. But i still get
> [22723.980390] device-mapper: table: 254:1: linear: dm-linear: Device
> lookup failed
> [22723.980395] device-mapper: ioctl: error adding target to table
> [22724.001153] device-mapper: table: 254:2: linear: dm-linear: Device
> lookup failed
> [22724.001158] device-mapper: ioctl: error adding target to table
There are lots of reasons why the above message shows up. Most likely
someone else using them...
> mount doesn't give any messages in dmesg
>
> lvs shows:
> LV VG Attr LSize Origin Snap% Move Log Copy% Convert
> dreamhacklv Dreamhack -wi-ao 1,23t
> root gardin -wi-d- 928,00g
> swap_1 gardin -wi-d- 2,59g
>
> if i try to mount with:
> mount -t reiserfs /dev/mapper/gardin-root /mnt/tmp
>
> i get this in dmesg:
> [23113.711247] REISERFS warning (device dm-1): sh-2006
> read_super_block: bread failed (dev dm-1, block 2, size 4096)
> [23113.711257] REISERFS warning (device dm-1): sh-2006
> read_super_block: bread failed (dev dm-1, block 16, size 4096)
> [23113.711261] REISERFS warning (device dm-1): sh-2021
> reiserfs_fill_super: can not find reiserfs on dm-1
Looks like you have some kind of LV here. What is the output of the
following two commands:
1. "dmsetup table"
1. "dmsetup ls"
Thanks, Malahal.
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: Re: [linux-lvm] Re: Re: Problems with dissapearing PV when mounting (Stuart D. Gathman)
@ 2009-12-10 16:45 Johan Gardell
2009-12-10 18:40 ` malahal
0 siblings, 1 reply; 4+ messages in thread
From: Johan Gardell @ 2009-12-10 16:45 UTC (permalink / raw)
To: linux-lvm
The output from
dmsetup table
gardin-swap_1:
gardin-root:
Dreamhack-dreamhacklv: 0 2636726272 linear 8:34 384
dmsetup ls
gardin-swap_1 (254, 2)
gardin-root (254, 1)
Dreamhack-dreamhacklv (254, 0)
Thanks!
Johan
2009/12/10 <linux-lvm-request@redhat.com>:
> Send linux-lvm mailing list submissions to
> � � � �linux-lvm@redhat.com
>
> To subscribe or unsubscribe via the World Wide Web, visit
> � � � �https://www.redhat.com/mailman/listinfo/linux-lvm
> or, via email, send a message with subject or body 'help' to
> � � � �linux-lvm-request@redhat.com
>
> You can reach the person managing the list at
> � � � �linux-lvm-owner@redhat.com
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of linux-lvm digest..."
>
>
> Today's Topics:
>
> � 1. Re: Re: Re: Problems with dissapearing PV when mounting
> � � �(Stuart D. Gathman) (Stuart D. Gathman)
> � 2. Re: Re: Re: Problems with dissapearing PV when � �mounting
> � � �(Stuart D. Gathman) (malahal@us.ibm.com)
> � 3. Re: kernel panic on lvcreate (Christopher Hawkins)
> � 4. lvm striped VG and Extend and Reallocation Question
> � � �(Vahri? Muhtaryan)
> � 5. Re: kernel panic on lvcreate (Milan Broz)
> � 6. Re: kernel panic on lvcreate (Stuart D. Gathman)
> � 7. Re: kernel panic on lvcreate (Christopher Hawkins)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Mon, 7 Dec 2009 14:34:22 -0500 (EST)
> From: "Stuart D. Gathman" <stuart@bmsi.com>
> Subject: Re: [linux-lvm] Re: Re: Problems with dissapearing PV when
> � � � �mounting � � � �(Stuart D. Gathman)
> To: LVM general discussion and development <linux-lvm@redhat.com>
> Message-ID: <Pine.LNX.4.64.0912071431030.1595@bmsred.bmsi.com>
> Content-Type: TEXT/PLAIN; charset=US-ASCII
>
> On Mon, 7 Dec 2009, Johan Gardell wrote:
>
>> Ok, added a filter to remove /dev/fd0. But i still get
>> [22723.980390] device-mapper: table: 254:1: linear: dm-linear: Device
>> lookup failed
>> [22723.980395] device-mapper: ioctl: error adding target to table
>> [22724.001153] device-mapper: table: 254:2: linear: dm-linear: Device
>> lookup failed
>> [22724.001158] device-mapper: ioctl: error adding target to table
>
> Well, the 'd' in the lvs output means "device present without tables".
> I googled on the error msg, and see that a bunch of Ubuntu and Debian
> people had to remove evms for lvm to work properly after a certain
> kernel upgrade. �If that is not the problem, then I would have to start
> looking at the source, but perhaps a real guru here could help.
>
> --
> � � � � � � �Stuart D. Gathman <stuart@bmsi.com>
> � �Business Management Systems Inc. �Phone: 703 591-0911 Fax: 703 591-6154
> "Confutatis maledictis, flammis acribus addictis" - background song for
> a Microsoft sponsored "Where do you want to go from here?" commercial.
>
>
>
> ------------------------------
>
> Message: 2
> Date: Mon, 7 Dec 2009 15:11:37 -0800
> From: malahal@us.ibm.com
> Subject: Re: [linux-lvm] Re: Re: Problems with dissapearing PV when
> � � � �mounting (Stuart D. Gathman)
> To: linux-lvm@redhat.com
> Message-ID: <20091207231136.GA31793@us.ibm.com>
> Content-Type: text/plain; charset=us-ascii
>
> Johan Gardell [gardin@gmail.com] wrote:
>> Ok, added a filter to remove /dev/fd0. But i still get
>> [22723.980390] device-mapper: table: 254:1: linear: dm-linear: Device
>> lookup failed
>> [22723.980395] device-mapper: ioctl: error adding target to table
>> [22724.001153] device-mapper: table: 254:2: linear: dm-linear: Device
>> lookup failed
>> [22724.001158] device-mapper: ioctl: error adding target to table
>
> There are lots of reasons why the above message shows up. Most likely
> someone else using them...
>
>> mount doesn't give any messages in dmesg
>>
>> lvs shows:
>> � LV � � � � �VG � � � �Attr � LSize � Origin Snap% �Move Log Copy% �Convert
>> � dreamhacklv Dreamhack -wi-ao � 1,23t
>> � root � � � �gardin � �-wi-d- 928,00g
>> � swap_1 � � �gardin � �-wi-d- � 2,59g
>>
>> if i try to mount with:
>> � mount -t reiserfs /dev/mapper/gardin-root /mnt/tmp
>>
>> i get this in dmesg:
>> � [23113.711247] REISERFS warning (device dm-1): sh-2006
>> read_super_block: bread failed (dev dm-1, block 2, size 4096)
>> � [23113.711257] REISERFS warning (device dm-1): sh-2006
>> read_super_block: bread failed (dev dm-1, block 16, size 4096)
>> � [23113.711261] REISERFS warning (device dm-1): sh-2021
>> reiserfs_fill_super: can not find reiserfs on dm-1
>
> Looks like you have some kind of LV here. What is the output of the
> following two commands:
>
> 1. "dmsetup table"
> 1. "dmsetup ls"
>
> Thanks, Malahal.
>
>
>
> ------------------------------
>
> Message: 3
> Date: Wed, 09 Dec 2009 10:00:42 -0500 (EST)
> From: Christopher Hawkins <chawkins@bplinux.com>
> Subject: Re: [linux-lvm] kernel panic on lvcreate
> To: LVM general discussion and development <linux-lvm@redhat.com>
> Message-ID: <14440243.291260370842741.JavaMail.javamailuser@localhost>
> Content-Type: text/plain; charset=utf-8
>
> Hello,
>
> After some time I revisited this issue on a freshly installed Centos 5.4 box, latest kernel (2.6.18-164.6.1.el5 ) and the panic is still reproducible. Any time I create a snapshot of the root filesystem, kernel panics. The LVM HOWTO says to post bug reports to this list. Is this the proper place?
>
> Thanks,
> Chris
>
> >From earlier post:
> OOPS message:
>
> BUG: scheduling while atomic: java/0x00000001/2959 � � � � � � � � � � � � � � � [<c061637f>] <3>BUG: scheduling while atomic: java/0x00000001/2867 � � � � � � �[<c061637f>] schedule+0x43/0xa55 � � � � � � � � � � � � � � � � � � � � � � � �[<c042c40d>] lock_timer_base+0x15/0x2f
> �[<c042c46b>] try_to_del_timer_sync+0x44/0x4a
> �[<c0437dd2>] futex_wake+0x3c/0xa5
> �[<c0434d5f>] prepare_to_wait+0x24/0x46
> �[<c0461ea7>] do_wp_page+0x1b3/0x5bb
> �[<c0438b01>] do_futex+0x239/0xb5e
> �[<c0434c13>] autoremove_wake_function+0x0/0x2d
> �[<c0463876>] __handle_mm_fault+0x9a9/0xa15
> �[<c041e727>] default_wake_function+0x0/0xc
> �[<c046548d>] unmap_region+0xe1/0xf0
> �[<c061954f>] do_page_fault+0x233/0x4e1
> �[<c061931c>] do_page_fault+0x0/0x4e1
> �[<c0405a89>] error_code+0x39/0x40
> �=======================
> schedule+0x43/0xa55
> �[<c042c40d>] <0>------------[ cut here ]------------
> kernel BUG at arch/i386/mm/highmem.c:43!
> invalid opcode: 0000 [#1]
> SMP
> last sysfs file: /devices/pci0000:00/0000:00:00.0/irq
> Modules linked in: autofs4 hidp rfcomm l2cap bluetooth lockd sunrpc ip6t_REJECTdCPU: � �3 ip6table_filter ip6_tables x_tables ipv6 xfrm_nalgo cry
> EIP: � �0060:[<c041cb08>] � �Not tainted VLI
> EFLAGS: 00010206 � (2.6.18-164.2.1.el5 #1)
> EIP is at kmap_atomic+0x5c/0x7f
> eax: c0012d6c � ebx: fff5b000 � ecx: c1fb8760 � edx: 00000180
> esi: f7be8580 � edi: f7fa7000 � ebp: 00000004 � esp: f5c54f0c
> ds: 007b � es: 007b � ss: 0068 � � � � � � � � � � � � � � � � � � � � � � � � �Process mpath_wait (pid: 3273, ti=f5c54000 task=f5c50000 task.ti=f5c54000)ne � �Stack: c073a4e0 c0462f7f f7b0eb30 f7b40780 f5c54f3c 0029c3f0 f63b5ef0 f7be8580
> � � � f7b40780 f7fa7000 00008802 c0472d75 f7b0eb30 f7c299c0 00001000 00001000
> � � � 00001000 00000101 00000001 00000000 00000000 f5c5007b 0000007b ffffffff
> Call Trace:
> �[<c0462f7f>] __handle_mm_fault+0xb2/0xa15
> �[<c0472d75>] do_filp_open+0x2b/0x31
> �[<c061954f>] do_page_fault+0x233/0x4e1
> �[<c061931c>] do_page_fault+0x0/0x4e1
> �[<c0405a89>] error_code+0x39/0x40
> �=======================
> Code: 00 89 e0 25 00 f0 ff ff 6b 50 10 1b 8d 14 13 bb 00 f0 ff ff 8d 42 44 c1 e EIP: [<c041cb08>] kmap_atomic+0x5c/0x7f SS:ESP 0068:f5c54f0c
> �<0>Kernel panic - not syncing: Fatal exception
>
> �0c 29 c3 a1 54 12 79 c0 c1 e2 02 29 d0 83 38 00 74 08 <0f> 0b 2b
>
>
> ----- "Milan Broz" <mbroz@redhat.com> wrote:
>
>> On 11/03/2009 04:07 PM, Christopher Hawkins wrote:
>> > When I create a root snapshot on a fairly typical Centos 5.3
>> server:
>> ...
>> > I get a kernel panic.
>>
>> Please try to first update kernel to version from 5.4.
>> (There were some fixes for snapshot like
>> https://bugzilla.redhat.com/show_bug.cgi?id=496100)
>>
>> If it still fails, please post the OOps trace from kernel (syslog).
>>
>> Milan
>> --
>> mbroz@redhat.com
>>
>> _______________________________________________
>> linux-lvm mailing list
>> linux-lvm@redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-lvm
>> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>
>
>
> ------------------------------
>
> Message: 4
> Date: Wed, 9 Dec 2009 22:05:59 +0200
> From: Vahri? Muhtaryan <vahric@doruk.net.tr>
> Subject: [linux-lvm] lvm striped VG and Extend and Reallocation
> � � � �Question
> To: <linux-lvm@redhat.com>
> Message-ID: <060201ca790b$066f9ea0$134edbe0$@net.tr>
> Content-Type: text/plain; charset="iso-8859-9"
>
> Hello to �All,
>
>
>
> I'm using lvm2, I will create 2 striped LV which volume group created by two
> PVs. When write happen, its will striped to two PVs step by step.
>
> I know that when need to extend stirped LV, �I have to add two PVs more �and
> extend the LV for do not get an error.
>
>
>
> Two question
>
>
>
> First; when I extend striped volume does it means I will have 2 striped 2
> linear volume group? Means chunk1 written to PV1 ,chunk2 written to PV2 and
> its over , it will pass second two PVs and chunk3 written to PV3 ,chunk4
> wirtten PV4 , right ?
>
>
>
> �f its right , when data is not big enough and chunk1 and chunk2 enough to
> store, next write request time LVM start for first pair of PVs or not ?
>
>
>
> Second;
>
>
>
> I would like to balance striped data when I add PVs to extend related VG
> because first datas are written to only olds PVs and after extend if read
> request happen still old disks will be used instead of �this and improve
> performance I would like to lay all data to all PVs after extend. �s there
> any way to �reallocation PEs ?
>
>
>
> Regards
>
> Vahric
>
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: https://www.redhat.com/archives/linux-lvm/attachments/20091209/e1641681/attachment.html
>
> ------------------------------
>
> Message: 5
> Date: Wed, 09 Dec 2009 21:18:29 +0100
> From: Milan Broz <mbroz@redhat.com>
> Subject: Re: [linux-lvm] kernel panic on lvcreate
> To: LVM general discussion and development <linux-lvm@redhat.com>
> Cc: Christopher Hawkins <chawkins@bplinux.com>
> Message-ID: <4B200615.1010702@redhat.com>
> Content-Type: text/plain; charset=UTF-8
>
> On 12/09/2009 04:00 PM, Christopher Hawkins wrote:
>>
>> After some time I revisited this issue on a freshly installed Centos 5.4 box, latest kernel (2.6.18-164.6.1.el5 )
>> and the panic is still reproducible. Any time I create a snapshot of the root filesystem, kernel panics.
>
> I guess it is already reported here https://bugzilla.redhat.com/show_bug.cgi?id=539328
> so please watch this bugzilla.
>
> Milan
> --
> mbroz@redhat.com
>
>
>
> ------------------------------
>
> Message: 6
> Date: Thu, 10 Dec 2009 10:00:07 -0500 (EST)
> From: "Stuart D. Gathman" <stuart@bmsi.com>
> Subject: Re: [linux-lvm] kernel panic on lvcreate
> To: LVM general discussion and development <linux-lvm@redhat.com>
> Message-ID: <Pine.LNX.4.64.0912100949260.8205@bmsred.bmsi.com>
> Content-Type: TEXT/PLAIN; charset=US-ASCII
>
> On Wed, 9 Dec 2009, Christopher Hawkins wrote:
>
>> After some time I revisited this issue on a freshly installed Centos 5.4 box,
>> latest kernel (2.6.18-164.6.1.el5 ) and the panic is still reproducible. Any
>> time I create a snapshot of the root filesystem, kernel panics. The LVM HOWTO
>> says to post bug reports to this list. Is this the proper place?
>
> Bummer. �I would post the bug on Centos bugzilla also. �Please post the
> bug number here if you do it (cause I'll get to it eventually).
>
> Thanks for testing this. �I have the same problem, and have a new client
> to install by next year - so not much time to work on it.
>
> Now that we know it is not yet fixed, we can form theories as to what
> is going wrong. �My guess is that the problem is caused by the fact that
> lvm is updating files in /etc/lvm on the root filesystem while taking
> the snapshot. �These updates are done by user space programs, so I would
> further speculate that *any* snapshot would crash if an update happened exactly
> when creating the snapshot - i.e. the atomic nature of snapshot creation has
> been broken. �The lvm user space probably does fsync() on files
> in /etc/lvm, which might be involved in triggering the crash.
>
> We could test the first theory by moving /etc/lvm to another volume (I
> sometimes put it on /boot - a non LVM filesystem - for easier disaster
> recovery.) Naturally, I wouldn't go moving /etc/lvm on a production server.
>
> Testing the second hypothesis is less certain, and would basically involve
> trying snapshots of LVs undergoing heavy updating.
>
> --
> � � � � � � �Stuart D. Gathman <stuart@bmsi.com>
> � �Business Management Systems Inc. �Phone: 703 591-0911 Fax: 703 591-6154
> "Confutatis maledictis, flammis acribus addictis" - background song for
> a Microsoft sponsored "Where do you want to go from here?" commercial.
>
>
>
> ------------------------------
>
> Message: 7
> Date: Thu, 10 Dec 2009 10:04:40 -0500 (EST)
> From: Christopher Hawkins <chawkins@bplinux.com>
> Subject: Re: [linux-lvm] kernel panic on lvcreate
> To: LVM general discussion and development <linux-lvm@redhat.com>
> Message-ID: <8023092.631260457480892.JavaMail.javamailuser@localhost>
> Content-Type: text/plain; charset=utf-8
>
> It is reported here:
>
> https://bugzilla.redhat.com/show_bug.cgi?id=539328
>
> That is definitely the one. And it sounds like they have a potential fix... I have already emailed the developers there asking if I can help test their patch, so hopefully soon I can post back and report status.
>
> Christopher Hawkins
>
> ----- "Stuart D. Gathman" <stuart@bmsi.com> wrote:
>
>> On Wed, 9 Dec 2009, Christopher Hawkins wrote:
>>
>> > After some time I revisited this issue on a freshly installed Centos
>> 5.4 box,
>> > latest kernel (2.6.18-164.6.1.el5 ) and the panic is still
>> reproducible. Any
>> > time I create a snapshot of the root filesystem, kernel panics. The
>> LVM HOWTO
>> > says to post bug reports to this list. Is this the proper place?
>>
>> Bummer. �I would post the bug on Centos bugzilla also. �Please post
>> the
>> bug number here if you do it (cause I'll get to it eventually).
>>
>> Thanks for testing this. �I have the same problem, and have a new
>> client
>> to install by next year - so not much time to work on it.
>>
>> Now that we know it is not yet fixed, we can form theories as to what
>> is going wrong. �My guess is that the problem is caused by the fact
>> that
>> lvm is updating files in /etc/lvm on the root filesystem while taking
>> the snapshot. �These updates are done by user space programs, so I
>> would
>> further speculate that *any* snapshot would crash if an update
>> happened exactly
>> when creating the snapshot - i.e. the atomic nature of snapshot
>> creation has
>> been broken. �The lvm user space probably does fsync() on files
>> in /etc/lvm, which might be involved in triggering the crash.
>>
>> We could test the first theory by moving /etc/lvm to another volume
>> (I
>> sometimes put it on /boot - a non LVM filesystem - for easier
>> disaster
>> recovery.) Naturally, I wouldn't go moving /etc/lvm on a production
>> server.
>>
>> Testing the second hypothesis is less certain, and would basically
>> involve
>> trying snapshots of LVs undergoing heavy updating.
>>
>> --
>> � � � � � � Stuart D. Gathman <stuart@bmsi.com>
>> � � Business Management Systems Inc. �Phone: 703 591-0911 Fax: 703
>> 591-6154
>> "Confutatis maledictis, flammis acribus addictis" - background song
>> for
>> a Microsoft sponsored "Where do you want to go from here?"
>> commercial.
>>
>> _______________________________________________
>> linux-lvm mailing list
>> linux-lvm@redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-lvm
>> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>
>
>
> ------------------------------
>
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@redhat.com
> https://www.redhat.com/mailman/listinfo/linux-lvm
>
> End of linux-lvm Digest, Vol 70, Issue 4
> ****************************************
>
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [linux-lvm] Re: Re: Problems with dissapearing PV when mounting (Stuart D. Gathman)
2009-12-10 16:45 Johan Gardell
@ 2009-12-10 18:40 ` malahal
0 siblings, 0 replies; 4+ messages in thread
From: malahal @ 2009-12-10 18:40 UTC (permalink / raw)
To: linux-lvm
Johan Gardell [gardin@gmail.com] wrote:
> The output from
> dmsetup table
> gardin-swap_1:
> gardin-root:
> Dreamhack-dreamhacklv: 0 2636726272 linear 8:34 384
The above output clearly shows that gardin-root and gardin-swap_1 have
empty tables. I remember there was a fix that cleaned up empty tables
like above but that won't help you. You need to figure out the reason
for 'device mapper lookup failure' messages.
> >> [22723.980390] device-mapper: table: 254:1: linear: dm-linear: Device
> >> lookup failed
> >> [22723.980395] device-mapper: ioctl: error adding target to table
> >> [22724.001153] device-mapper: table: 254:2: linear: dm-linear: Device
> >> lookup failed
> >> [22724.001158] device-mapper: ioctl: error adding target to table
The above messages are direct reason for not "really" creating
gardin-root or gardin-swap_1 logical volumes. If you know the
underlying PV's that make up gardin-root LV, try reading from them using
'dd'. If you can read, then the most likely problem is that someone else
has opened them. It may be hard to find though. You need to go through
usual culprits by running, 'cat /proc/mounts', 'cat /proc/swaps' etc. Is
EVMS using them by any chance..., maybe mdraid...
> >> mount doesn't give any messages in dmesg
> >>
> >> lvs shows:
> >> ? LV ? ? ? ? ?VG ? ? ? ?Attr ? LSize ? Origin Snap% ?Move Log Copy% ?Convert
> >> ? dreamhacklv Dreamhack -wi-ao ? 1,23t
> >> ? root ? ? ? ?gardin ? ?-wi-d- 928,00g
> >> ? swap_1 ? ? ?gardin ? ?-wi-d- ? 2,59g
> >>
> >> if i try to mount with:
> >> ? mount -t reiserfs /dev/mapper/gardin-root /mnt/tmp
> >>
> >> i get this in dmesg:
> >> ? [23113.711247] REISERFS warning (device dm-1): sh-2006
> >> read_super_block: bread failed (dev dm-1, block 2, size 4096)
> >> ? [23113.711257] REISERFS warning (device dm-1): sh-2006
> >> read_super_block: bread failed (dev dm-1, block 16, size 4096)
> >> ? [23113.711261] REISERFS warning (device dm-1): sh-2021
> >> reiserfs_fill_super: can not find reiserfs on dm-1
> >
> > Looks like you have some kind of LV here. What is the output of the
> > following two commands:
Your 'dmsetup table' output shows why the mount failed. You have an LV
created with no mapping.
Thanks, Malahal.
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2009-12-10 18:41 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-12-07 18:47 [linux-lvm] Re: Re: Problems with dissapearing PV when mounting (Stuart D. Gathman) Johan Gardell
2009-12-07 19:34 ` Stuart D. Gathman
2009-12-07 23:11 ` malahal
-- strict thread matches above, loose matches on Subject: below --
2009-12-10 16:45 Johan Gardell
2009-12-10 18:40 ` malahal
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).