* [linux-lvm] missing physical volumes after upgrade to rhel 5.4
@ 2009-09-24 10:06 Julie Ashworth
2009-09-24 19:38 ` [linux-lvm] " Julie Ashworth
2009-09-25 9:52 ` [linux-lvm] " Mark Round
0 siblings, 2 replies; 4+ messages in thread
From: Julie Ashworth @ 2009-09-24 10:06 UTC (permalink / raw)
To: linux-lvm
I apologize for the cross-posting (to rhelv5-list).
The lvm list is a more relevant list for my problem,
and I'm sorry I didn't realize this sooner.
After an upgrade from rhel5.3 -> rhel5.4 (and reboot)
I can no longer see PVs for 3 fibre-channel storage
devices.
The operating system still see the disk:
----------------------
# multipath -l
mpath2 (2001b4d28000064db) dm-1 JetStor,Volume Set # 00
[size=12T][features=0][hwhandler=0][rw]
\_ round-robin 0 [prio=0][active]
\_ 11:0:1:0 sdj 8:144 [active][undef]
mpath16 (1ACNCorp_FF01000113200019) dm-2 ACNCorp,R_LogVol-despo
[size=15T][features=0][hwhandler=0][rw]
\_ round-robin 0 [prio=0][active]
\_ 11:0:2:0 sdk 8:160 [active][undef]
mpath7 (32800001b4d00cf5b) dm-0 JetStor,Volume Set 416F
[size=12T][features=0][hwhandler=0][rw]
\_ round-robin 0 [prio=0][active]
\_ 11:0:0:0 sdi 8:128 [active][undef]
----------------------
There are files in /etc/lvm/backup/ that contain the
original volume information, e.g.
----------------------
jetstor642 {
id = "0e53Q3-evHX-I5f9-CWqf-NPcw-IqmC-0fVcTO"
seqno = 2
status = ["RESIZEABLE", "READ", "WRITE"]
flags = []
extent_size = 8192 # 4 Megabytes
max_lv = 0
max_pv = 0
physical_volumes {
pv0 {
id = "5wJCEA-IDC1-5GhI-jnEs-EpYF-8Uf3-sqPL4O"
device = "/dev/dm-7" # Hint only
status = ["ALLOCATABLE"]
flags = []
dev_size = 31214845952 # 14.5355 Terabytes
pe_start = 384
pe_count = 3810405 # 14.5355 Terabytes
}
}
----------------------
The devices were formatted using parted on the entire disk,
i.e. I didn't create a partition.
The partition table is "gpt" (possible label types are
"bsd", "dvh", "gpt", "loop", "mac", "msdos", "pc98"
or "sun".)
partition table information for one of the devices is below:
--------------------------
# parted /dev/sdi
GNU Parted 1.8.1
Using /dev/sdi
Welcome to GNU Parted! Type 'help' to view a list of commands.
(parted) print
Model: JetStor Volume Set 416F (scsi)
Disk /dev/sdi: 13.0TB
Sector size (logical/physical): 512B/512B
Partition Table: gpt
Number Start End Size File system Name Flags
--------------------------
output of some commands:
$ pvdisplay
returns nothing (no error)
$ lvs -a -o +devices
returns nothing (no error)
$ pvck -vvvvv /dev/sdb
#lvmcmdline.c:915 Processing: pvck -vvvvv /dev/sdb
#lvmcmdline.c:918 O_DIRECT will be used
#config/config.c:950 Setting global/locking_type to 3
#locking/locking.c:245 Cluster locking selected.
#locking/cluster_locking.c:83 connect() failed on local socket: Connection
+refused
#config/config.c:955 locking/fallback_to_local_locking not found in
+config: defaulting to 1
WARNING: Falling back to local file-based locking.
Volume Groups with the clustered attribute will be inaccessible.
#config/config.c:927 Setting global/locking_dir to /var/lock/lvm
#pvck.c:32 Scanning /dev/sdb
#device/dev-cache.c:260 /dev/sdb: Added to device cache
#device/dev-io.c:439 Opened /dev/sdb RO
#device/dev-io.c:260 /dev/sdb: size is 25395814912 sectors
#device/dev-io.c:134 /dev/sdb: block size is 4096 bytes
#filters/filter.c:124 /dev/sdb: Skipping: Partition table signature
+found
#device/dev-io.c:485 Closed /dev/sdb
#metadata/metadata.c:2337 Device /dev/sdb not found (or ignored by filtering).
-------------------------
from doing google searches, I found this gem to restore a
PV:
pvcreate --uuid "cqH4SD-VrCw-jMsN-GcwH-omCq-ThpE-dO9KmJ"
--restorefile /etc/lvm/backup/vg_04 /dev/sdd1
however, the man page says to 'use with care'. I don't want
to lose data. Can anybody comment on how safe it would be to
run this?
Thanks in advance,
Julie Ashworth
--
Julie Ashworth <julie.ashworth@berkeley.edu>
Computational Infrastructure for Research Labs, UC Berkeley
http://cirl.berkeley.edu/
PGP Key ID: 0x17F013D2
^ permalink raw reply [flat|nested] 4+ messages in thread
* [linux-lvm] Re: missing physical volumes after upgrade to rhel 5.4
2009-09-24 10:06 [linux-lvm] missing physical volumes after upgrade to rhel 5.4 Julie Ashworth
@ 2009-09-24 19:38 ` Julie Ashworth
2009-09-25 9:52 ` [linux-lvm] " Mark Round
1 sibling, 0 replies; 4+ messages in thread
From: Julie Ashworth @ 2009-09-24 19:38 UTC (permalink / raw)
To: linux-lvm
Some additional information:
----------------------
# lvm dumpconfig
devices {
dir="/dev"
scan="/dev"
preferred_names=[]
filter="a/.*/"
cache_dir="/etc/lvm/cache"
cache_file_prefix=""
write_cache_state=1
sysfs_scan=1
md_component_detection=1
ignore_suspended_devices=0
}
activation {
missing_stripe_filler="/dev/ioerror"
reserved_stack=256
reserved_memory=8192
process_priority=-18
mirror_region_size=512
readahead="auto"
mirror_log_fault_policy="allocate"
mirror_device_fault_policy="remove"
}
global {
umask=63
test=0
units="h"
activation=1
proc="/proc"
locking_type=1
fallback_to_clustered_locking=1
fallback_to_local_locking=1
locking_dir="/var/lock/lvm"
}
shell {
history_size=100
}
backup {
backup=1
backup_dir="/etc/lvm/backup"
archive=1
archive_dir="/etc/lvm/archive"
retain_min=10
retain_days=30
}
log {
verbose=0
syslog=1
overwrite=0
level=0
indent=1
command_names=0
prefix=" "
}
----------------------
(I powered off 2 of the storage devices, so only
one (12TB) device remains accessible):
# lvmdiskscan
/dev/ramdisk [ 16.00 MB]
/dev/md0 [ 148.94 MB]
/dev/ram [ 16.00 MB]
/dev/md1 [ 3.91 GB]
/dev/ram2 [ 16.00 MB]
/dev/md2 [ 105.46 GB]
/dev/dm-2 [ 11.83 TB]
/dev/ram3 [ 16.00 MB]
/dev/sda3 [ 3.91 GB]
/dev/md3 [ 3.91 GB]
/dev/ram4 [ 16.00 MB]
/dev/md4 [ 800.46 GB]
/dev/ram5 [ 16.00 MB]
/dev/ram6 [ 16.00 MB]
/dev/ram7 [ 16.00 MB]
/dev/ram8 [ 16.00 MB]
/dev/ram9 [ 16.00 MB]
/dev/ram10 [ 16.00 MB]
/dev/ram11 [ 16.00 MB]
/dev/ram12 [ 16.00 MB]
/dev/ram13 [ 16.00 MB]
/dev/ram14 [ 16.00 MB]
/dev/ram15 [ 16.00 MB]
/dev/sdb6 [ 3.91 GB]
/dev/sdb8 [ 6.05 GB]
3 disks
23 partitions
0 LVM physical volume whole disks
1 LVM physical volume
--
Julie Ashworth <julie.ashworth@berkeley.edu>
Computational Infrastructure for Research Labs, UC Berkeley
http://cirl.berkeley.edu/
PGP Key ID: 0x17F013D2
^ permalink raw reply [flat|nested] 4+ messages in thread
* RE: [linux-lvm] missing physical volumes after upgrade to rhel 5.4
2009-09-24 10:06 [linux-lvm] missing physical volumes after upgrade to rhel 5.4 Julie Ashworth
2009-09-24 19:38 ` [linux-lvm] " Julie Ashworth
@ 2009-09-25 9:52 ` Mark Round
2009-09-26 1:45 ` Julie Ashworth
1 sibling, 1 reply; 4+ messages in thread
From: Mark Round @ 2009-09-25 9:52 UTC (permalink / raw)
To: LVM general discussion and development
I just tried this myself....
1. First create a new PV on a whole disk, VG and LV
# pvcreate /dev/sdc
# vgcreate test /dev/sdc
# lvcreate -L2G -n testlv test
2. Format the LV, mount it and copy some data to it (just a random
tarball)
# mke2fs -j /dev/test/testlv
# mount /dev/test/testlv /mnt
# tar -C /mnt -xvzf ~/iscsitarget-0.4.17.tar.gz
# umount /mnt
# e2fsck /dev/test/testlv
e2fsck 1.39 (29-May-2006)
/dev/test/testlv: clean, 87/262144 files, 25554/524288 blocks
3. So the LV is OK. Now I'll make sure there's a config backup, then
wipe the PV label...
# vgcfgbackup test
Volume group "test" successfully backed up.
# vgchange -an test
0 logical volume(s) in volume group "test" now active
[root@europa ~]# pvremove -ff /dev/sdc
Really WIPE LABELS from physical volume "/dev/sdc" of volume group
"test" [y/n]? y
WARNING: Wiping physical volume label from /dev/sdc of volume group
"test"
Labels on physical volume "/dev/sdc" successfully wiped
4. Now, I'll try to recreate the PV using the backup data, and see if
the contents are intact.
# pvcreate --uuid="A0LDgs-KMlm-QEBR-sGNW-7Rlf-j3aU-x2JUKY"
--restorefile=/etc/lvm/backup/test /dev/sdc
Couldn't find device with uuid
'A0LDgs-KMlm-QEBR-sGNW-7Rlf-j3aU-x2JUKY'.
Physical volume "/dev/sdc" successfully created
# vgcfgrestore -f /etc/lvm/backup/test test
Restored volume group test
5. Check to see if we can see the previously created LV, and mount it
# lvs
LV VG Attr LSize Origin Snap% Move Log Copy% Convert
testlv test -wi--- 2.00G
# vgchange -ay test
1 logical volume(s) in volume group "test" now active
# e2fsck /dev/test/testlv
e2fsck 1.39 (29-May-2006)
/dev/test/testlv: clean, 87/262144 files, 25554/524288 blocks
# mount /dev/test/testlv /mnt
# ls /mnt
iscsitarget-0.4.17 lost+found
So, YMMV but it appears from my experiments that this operation should
be safe, and should recover your volumes. I am concerned though about
the news that the RHEL 5.3->5.4 upgrade may have caused this, as we're
looking at making the same upgrade before not too long. Do you have any
suspicion as to why this may have happened ? Have you filed a bug with
Red Hat ?
Regards,
-Mark
-----Original Message-----
From: linux-lvm-bounces@redhat.com [mailto:linux-lvm-bounces@redhat.com]
On Behalf Of Julie Ashworth
Sent: 24 September 2009 11:06
To: linux-lvm@redhat.com
Subject: [linux-lvm] missing physical volumes after upgrade to rhel 5.4
I apologize for the cross-posting (to rhelv5-list).
The lvm list is a more relevant list for my problem,
and I'm sorry I didn't realize this sooner.
After an upgrade from rhel5.3 -> rhel5.4 (and reboot)
I can no longer see PVs for 3 fibre-channel storage
devices.
The operating system still see the disk:
----------------------
# multipath -l
mpath2 (2001b4d28000064db) dm-1 JetStor,Volume Set # 00
[size=12T][features=0][hwhandler=0][rw]
\_ round-robin 0 [prio=0][active]
\_ 11:0:1:0 sdj 8:144 [active][undef]
mpath16 (1ACNCorp_FF01000113200019) dm-2 ACNCorp,R_LogVol-despo
[size=15T][features=0][hwhandler=0][rw]
\_ round-robin 0 [prio=0][active]
\_ 11:0:2:0 sdk 8:160 [active][undef]
mpath7 (32800001b4d00cf5b) dm-0 JetStor,Volume Set 416F
[size=12T][features=0][hwhandler=0][rw]
\_ round-robin 0 [prio=0][active]
\_ 11:0:0:0 sdi 8:128 [active][undef]
----------------------
There are files in /etc/lvm/backup/ that contain the
original volume information, e.g.
----------------------
jetstor642 {
id = "0e53Q3-evHX-I5f9-CWqf-NPcw-IqmC-0fVcTO"
seqno = 2
status = ["RESIZEABLE", "READ", "WRITE"]
flags = []
extent_size = 8192 # 4 Megabytes
max_lv = 0
max_pv = 0
physical_volumes {
pv0 {
id = "5wJCEA-IDC1-5GhI-jnEs-EpYF-8Uf3-sqPL4O"
device = "/dev/dm-7" # Hint only
status = ["ALLOCATABLE"]
flags = []
dev_size = 31214845952 # 14.5355 Terabytes
pe_start = 384
pe_count = 3810405 # 14.5355 Terabytes
}
}
----------------------
The devices were formatted using parted on the entire disk,
i.e. I didn't create a partition.
The partition table is "gpt" (possible label types are
"bsd", "dvh", "gpt", "loop", "mac", "msdos", "pc98"
or "sun".)
partition table information for one of the devices is below:
--------------------------
# parted /dev/sdi
GNU Parted 1.8.1
Using /dev/sdi
Welcome to GNU Parted! Type 'help' to view a list of commands.
(parted) print
Model: JetStor Volume Set 416F (scsi)
Disk /dev/sdi: 13.0TB
Sector size (logical/physical): 512B/512B
Partition Table: gpt
Number Start End Size File system Name Flags
--------------------------
output of some commands:
$ pvdisplay
returns nothing (no error)
$ lvs -a -o +devices
returns nothing (no error)
$ pvck -vvvvv /dev/sdb
#lvmcmdline.c:915 Processing: pvck -vvvvv /dev/sdb
#lvmcmdline.c:918 O_DIRECT will be used
#config/config.c:950 Setting global/locking_type to 3
#locking/locking.c:245 Cluster locking selected.
#locking/cluster_locking.c:83 connect() failed on local socket:
Connection
+refused
#config/config.c:955 locking/fallback_to_local_locking not found in
+config: defaulting to 1
WARNING: Falling back to local file-based locking.
Volume Groups with the clustered attribute will be inaccessible.
#config/config.c:927 Setting global/locking_dir to /var/lock/lvm
#pvck.c:32 Scanning /dev/sdb
#device/dev-cache.c:260 /dev/sdb: Added to device cache
#device/dev-io.c:439 Opened /dev/sdb RO
#device/dev-io.c:260 /dev/sdb: size is 25395814912 sectors
#device/dev-io.c:134 /dev/sdb: block size is 4096 bytes
#filters/filter.c:124 /dev/sdb: Skipping: Partition table
signature
+found
#device/dev-io.c:485 Closed /dev/sdb
#metadata/metadata.c:2337 Device /dev/sdb not found (or ignored by
filtering).
-------------------------
from doing google searches, I found this gem to restore a
PV:
pvcreate --uuid "cqH4SD-VrCw-jMsN-GcwH-omCq-ThpE-dO9KmJ"
--restorefile /etc/lvm/backup/vg_04 /dev/sdd1
however, the man page says to 'use with care'. I don't want
to lose data. Can anybody comment on how safe it would be to
run this?
Thanks in advance,
Julie Ashworth
--
Julie Ashworth <julie.ashworth@berkeley.edu>
Computational Infrastructure for Research Labs, UC Berkeley
http://cirl.berkeley.edu/
PGP Key ID: 0x17F013D2
_______________________________________________
linux-lvm mailing list
linux-lvm@redhat.com
https://www.redhat.com/mailman/listinfo/linux-lvm
read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [linux-lvm] missing physical volumes after upgrade to rhel 5.4
2009-09-25 9:52 ` [linux-lvm] " Mark Round
@ 2009-09-26 1:45 ` Julie Ashworth
0 siblings, 0 replies; 4+ messages in thread
From: Julie Ashworth @ 2009-09-26 1:45 UTC (permalink / raw)
To: LVM general discussion and development
hi Mark,
Thank you so much for the thorough test and
response.
I created a test volume with a gpt disk label to
stay consistent with how the failed volumes were
labelled.
The only difference between our experiences, that
I can tell, is that the pvcreate command failed
with an error similar to 'disk doesn't exist, or
is being filtered'. I believe its because of the
gpt label. I zeroed out the first 512 bytes of
the disk label, and continued with the commands
(similar to the commands you used):
dd if=/dev/zero of=/dev/sdk bs=512 count=1
pvcreate -u 5wJCEA-IDC1-5GhI-jnEs-EpYF-8Uf3-sqPL4O /dev/sdk
vgcfgrestore -f /etc/lvm/backup/jetstor642 jetstor642
vgchange -ay
I had no problems, and the volumes were
recovered. <much celebration>
Its possible that my setup is so non-standard,
that most people won't be affected (presumably
by the upgrade) as I was. Reasons my setup is
non-standard:
1) I used ext3 to format >12TB volumes (ext3
has a 8TB limit).
2) I used parted and gpt disk labels
3) I created PV on a whole disk.
Luckily, my request for a test environment was
approved (after this experience) so I can
attempt to replicate the problem and identify
the cause of the disk label corruption.
Unfortunately, the environment will certainly
arrive too late (I do work at a uni ;P) for me
to make a timely contribution.
Please let me know if I can provide any more
information.
And thanks again.
Best,
Julie
On 25-09-2009 10.52 +0100, Mark Round wrote:
> I just tried this myself....
>
> 1. First create a new PV on a whole disk, VG and LV
> # pvcreate /dev/sdc
> # vgcreate test /dev/sdc
> # lvcreate -L2G -n testlv test
>
> 2. Format the LV, mount it and copy some data to it (just a random
> tarball)
> # mke2fs -j /dev/test/testlv
> # mount /dev/test/testlv /mnt
> # tar -C /mnt -xvzf ~/iscsitarget-0.4.17.tar.gz
> # umount /mnt
> # e2fsck /dev/test/testlv
> e2fsck 1.39 (29-May-2006)
> /dev/test/testlv: clean, 87/262144 files, 25554/524288 blocks
>
> 3. So the LV is OK. Now I'll make sure there's a config backup, then
> wipe the PV label...
> # vgcfgbackup test
> Volume group "test" successfully backed up.
> # vgchange -an test
> 0 logical volume(s) in volume group "test" now active
> [root@europa ~]# pvremove -ff /dev/sdc
> Really WIPE LABELS from physical volume "/dev/sdc" of volume group
> "test" [y/n]? y
> WARNING: Wiping physical volume label from /dev/sdc of volume group
> "test"
> Labels on physical volume "/dev/sdc" successfully wiped
>
> 4. Now, I'll try to recreate the PV using the backup data, and see if
> the contents are intact.
> # pvcreate --uuid="A0LDgs-KMlm-QEBR-sGNW-7Rlf-j3aU-x2JUKY"
> --restorefile=/etc/lvm/backup/test /dev/sdc
> Couldn't find device with uuid
> 'A0LDgs-KMlm-QEBR-sGNW-7Rlf-j3aU-x2JUKY'.
> Physical volume "/dev/sdc" successfully created
> # vgcfgrestore -f /etc/lvm/backup/test test
> Restored volume group test
>
> 5. Check to see if we can see the previously created LV, and mount it
> # lvs
> LV VG Attr LSize Origin Snap% Move Log Copy% Convert
> testlv test -wi--- 2.00G
> # vgchange -ay test
> 1 logical volume(s) in volume group "test" now active
> # e2fsck /dev/test/testlv
> e2fsck 1.39 (29-May-2006)
> /dev/test/testlv: clean, 87/262144 files, 25554/524288 blocks
> # mount /dev/test/testlv /mnt
> # ls /mnt
> iscsitarget-0.4.17 lost+found
>
> So, YMMV but it appears from my experiments that this operation should
> be safe, and should recover your volumes. I am concerned though about
> the news that the RHEL 5.3->5.4 upgrade may have caused this, as we're
> looking at making the same upgrade before not too long. Do you have any
> suspicion as to why this may have happened ? Have you filed a bug with
> Red Hat ?
>
> Regards,
>
> -Mark
>
> -----Original Message-----
> From: linux-lvm-bounces@redhat.com [mailto:linux-lvm-bounces@redhat.com]
> On Behalf Of Julie Ashworth
> Sent: 24 September 2009 11:06
> To: linux-lvm@redhat.com
> Subject: [linux-lvm] missing physical volumes after upgrade to rhel 5.4
>
> I apologize for the cross-posting (to rhelv5-list).
> The lvm list is a more relevant list for my problem,
> and I'm sorry I didn't realize this sooner.
>
> After an upgrade from rhel5.3 -> rhel5.4 (and reboot)
> I can no longer see PVs for 3 fibre-channel storage
> devices.
>
> The operating system still see the disk:
> ----------------------
> # multipath -l
> mpath2 (2001b4d28000064db) dm-1 JetStor,Volume Set # 00
> [size=12T][features=0][hwhandler=0][rw]
> \_ round-robin 0 [prio=0][active]
> \_ 11:0:1:0 sdj 8:144 [active][undef]
> mpath16 (1ACNCorp_FF01000113200019) dm-2 ACNCorp,R_LogVol-despo
> [size=15T][features=0][hwhandler=0][rw]
> \_ round-robin 0 [prio=0][active]
> \_ 11:0:2:0 sdk 8:160 [active][undef]
> mpath7 (32800001b4d00cf5b) dm-0 JetStor,Volume Set 416F
> [size=12T][features=0][hwhandler=0][rw]
> \_ round-robin 0 [prio=0][active]
> \_ 11:0:0:0 sdi 8:128 [active][undef]
> ----------------------
>
>
> There are files in /etc/lvm/backup/ that contain the
> original volume information, e.g.
> ----------------------
> jetstor642 {
> id = "0e53Q3-evHX-I5f9-CWqf-NPcw-IqmC-0fVcTO"
> seqno = 2
> status = ["RESIZEABLE", "READ", "WRITE"]
> flags = []
> extent_size = 8192 # 4 Megabytes
> max_lv = 0
> max_pv = 0
>
> physical_volumes {
>
> pv0 {
> id = "5wJCEA-IDC1-5GhI-jnEs-EpYF-8Uf3-sqPL4O"
> device = "/dev/dm-7" # Hint only
>
> status = ["ALLOCATABLE"]
> flags = []
> dev_size = 31214845952 # 14.5355 Terabytes
> pe_start = 384
> pe_count = 3810405 # 14.5355 Terabytes
> }
> }
> ----------------------
>
> The devices were formatted using parted on the entire disk,
> i.e. I didn't create a partition.
> The partition table is "gpt" (possible label types are
> "bsd", "dvh", "gpt", "loop", "mac", "msdos", "pc98"
> or "sun".)
>
>
> partition table information for one of the devices is below:
> --------------------------
> # parted /dev/sdi
> GNU Parted 1.8.1
> Using /dev/sdi
> Welcome to GNU Parted! Type 'help' to view a list of commands.
> (parted) print
>
> Model: JetStor Volume Set 416F (scsi)
> Disk /dev/sdi: 13.0TB
> Sector size (logical/physical): 512B/512B
> Partition Table: gpt
>
> Number Start End Size File system Name Flags
> --------------------------
>
> output of some commands:
>
> $ pvdisplay
> returns nothing (no error)
>
> $ lvs -a -o +devices
> returns nothing (no error)
>
> $ pvck -vvvvv /dev/sdb
> #lvmcmdline.c:915 Processing: pvck -vvvvv /dev/sdb
> #lvmcmdline.c:918 O_DIRECT will be used
> #config/config.c:950 Setting global/locking_type to 3
> #locking/locking.c:245 Cluster locking selected.
> #locking/cluster_locking.c:83 connect() failed on local socket:
> Connection
> +refused
> #config/config.c:955 locking/fallback_to_local_locking not found in
> +config: defaulting to 1
> WARNING: Falling back to local file-based locking.
> Volume Groups with the clustered attribute will be inaccessible.
> #config/config.c:927 Setting global/locking_dir to /var/lock/lvm
> #pvck.c:32 Scanning /dev/sdb
> #device/dev-cache.c:260 /dev/sdb: Added to device cache
> #device/dev-io.c:439 Opened /dev/sdb RO
> #device/dev-io.c:260 /dev/sdb: size is 25395814912 sectors
> #device/dev-io.c:134 /dev/sdb: block size is 4096 bytes
> #filters/filter.c:124 /dev/sdb: Skipping: Partition table
> signature
> +found
> #device/dev-io.c:485 Closed /dev/sdb
> #metadata/metadata.c:2337 Device /dev/sdb not found (or ignored by
> filtering).
> -------------------------
>
> from doing google searches, I found this gem to restore a
> PV:
> pvcreate --uuid "cqH4SD-VrCw-jMsN-GcwH-omCq-ThpE-dO9KmJ"
> --restorefile /etc/lvm/backup/vg_04 /dev/sdd1
>
>
> however, the man page says to 'use with care'. I don't want
> to lose data. Can anybody comment on how safe it would be to
> run this?
>
> Thanks in advance,
> Julie Ashworth
>
>
> --
> Julie Ashworth <julie.ashworth@berkeley.edu>
> Computational Infrastructure for Research Labs, UC Berkeley
> http://cirl.berkeley.edu/
> PGP Key ID: 0x17F013D2
>
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@redhat.com
> https://www.redhat.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@redhat.com
> https://www.redhat.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
---end quoted text---
--
Julie Ashworth <julie.ashworth@berkeley.edu>
Computational Infrastructure for Research Labs, UC Berkeley
http://cirl.berkeley.edu/
PGP Key ID: 0x17F013D2
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2009-09-26 1:46 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-09-24 10:06 [linux-lvm] missing physical volumes after upgrade to rhel 5.4 Julie Ashworth
2009-09-24 19:38 ` [linux-lvm] " Julie Ashworth
2009-09-25 9:52 ` [linux-lvm] " Mark Round
2009-09-26 1:45 ` Julie Ashworth
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).