* [linux-lvm] LVM Volume Size Question
@ 2009-01-23 17:14 ctd
2009-01-23 17:44 ` F-D. Cami
0 siblings, 1 reply; 4+ messages in thread
From: ctd @ 2009-01-23 17:14 UTC (permalink / raw)
To: linux-lvm
Hey there,
I have most likely a simple question concerning LVM that I figured someone might be able to provide some insight into.
I just setup LVM with both /dev/sdb1 and /dev/sdc1 being assigned to my “vg” volume group. There is only one logical volume “myth” off of “vg”.
My steps:
fdisk /dev/sdc [created 1 partition to span the entire drive of type 8e]
emerge lvm2
vgscan
vgchange -a y
pvcreate /dev/sdc1
vgcreate vg /dev/sdc1
lvcreate -L900GB -nmyth vg
mkfs.jfs /dev/vg/myth
fdisk /dev/sdb [created 1 partition to span the entire drive of type 8e]
pvcreate /dev/sdb1
vgextend vg /dev/sdb1
lvextend -L+700G /dev/vg/myth
Sdb1: 700GB drive with one partition
Sdd1: 1TB drive with one partition
My question is related to the space available in /dev/vg/myth. I
would assume that I should have ~1.7TB of space on that logical
partition, but df does not seems to indicate that.
# df
Filesystem 1K-blocks Used Available Use% Mounted on
…
/dev/mapper/vg-myth 943656628 544996248 398660380 58% /mnt/store
…
mythserver michael # pvdisplay /dev/sdb1
--- Physical volume ---
PV Name /dev/sdb1
VG Name vg
PV Size 698.64 GB / not usable 2.34 MB
Allocatable yes
PE Size (KByte) 4096
Total PE 178850
Free PE 292
Allocated PE 178558
PV UUID SetyUA-DkWL-zDDo-Wm-3avR-nJH8-OnUujv
mythserver michael # pvdisplay /dev/sdc1
--- Physical volume ---
PV Name /dev/sdc1
VG Name vg
PV Size 931.51 GB / not usable 3.19 MB
Allocatable yes (but full)
PE Size (KByte) 4096
Total PE 238466
Free PE 0
Allocated PE 238466
PV UUID DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
mythserver michael # lvextend -L+700G /dev/vg/myth
Extending logical volume myth to 2.27 TB
Insufficient free space: 179200 extents needed, but only 292 available
I am guessing that I should have run these commands to extend the logical volume to its desired size:
vgextend vg /dev/sdb1
lvextend -L+700G /dev/vg/myth
before creating the filesystem with this command which I am guessing locked the size to the 900GB with I used in my setup steps
mkfs.jfs /dev/vg/myth
Does that sound like my issue?
Any thoughts on how to get out of this situation while ensuring no loss of my data that currently resides on /dev/mapper/vg-myth?
I am thinking that the following steps should work:
Copy all of my files on /dev/mapper/vg-myth to other paritions (I
assume the call to mkfs.jfs below will delete all the contents of this
partition)
"lvreduce -L-641G /dev/vg/myth" (to get the size matched up with 931GB + 698GB [ 2.27TB – 931GB – 698GB)
"mkfs.jfs /dev/vg/myth" (recreate the filesystem now that the size has been corrected)
remount /dev/vg/myth
copy back the files
Thanks in advance
Mike
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [linux-lvm] LVM Volume Size Question
2009-01-23 17:14 [linux-lvm] LVM Volume Size Question ctd
@ 2009-01-23 17:44 ` F-D. Cami
0 siblings, 0 replies; 4+ messages in thread
From: F-D. Cami @ 2009-01-23 17:44 UTC (permalink / raw)
To: ctd, LVM general discussion and development
Hi Mike,
Could you give us the output of :
vgdisplay -v vg
lvdisplay /dev/vg/myth
I think your myth LV is already 1.5TB, so you only need to run :
# mount -o remount,resize /dev/vg/myth
for the JFS filesystem to be resized.
Please backup everything before running that :)
However, since what you're doing essentially amounts to RAID0 without the
performance benefits (if you lose one drive, your data is lost), I'd run a full backup
and run the following commands to create a striped LV :
lvremove /dev/vg/myth
vgremove vg
pvremove /dev/sdb1
pvremove /dev/sdc1
dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
pvcreate /dev/sdb
pvcreate /dev/sdc
vgcreate vg /dev/sdb /dev/sdc
lcreate -i 2 -I 8 -L 1700M -n myth vg /dev/sdb /dev/sdc
(adjust 1700 to whatever your drives will take)
mkfs.jfs /dev/vg/myth
You will lose a bit of space but gain some performance ; the available VG size can
then be used for other LVs or snapshots.
Best,
Francois
On Fri, 23 Jan 2009 11:14:35 -0600
ctd@minneapolish3.com wrote:
> Hey there,
>
>
>
> I have most likely a simple question concerning LVM that I figured someone might be able to provide some insight into.
>
>
>
> I just setup LVM with both /dev/sdb1 and /dev/sdc1 being assigned to my “vg” volume group. There is only one logical volume “myth” off of “vg”.
>
>
>
> My steps:
>
> fdisk /dev/sdc [created 1 partition to span the entire drive of type 8e]
> emerge lvm2
> vgscan
> vgchange -a y
> pvcreate /dev/sdc1
> vgcreate vg /dev/sdc1
> lvcreate -L900GB -nmyth vg
> mkfs.jfs /dev/vg/myth
> fdisk /dev/sdb [created 1 partition to span the entire drive of type 8e]
> pvcreate /dev/sdb1
> vgextend vg /dev/sdb1
> lvextend -L+700G /dev/vg/myth
>
>
> Sdb1: 700GB drive with one partition
>
> Sdd1: 1TB drive with one partition
>
>
> My question is related to the space available in /dev/vg/myth. I
> would assume that I should have ~1.7TB of space on that logical
> partition, but df does not seems to indicate that.
>
>
> # df
> Filesystem 1K-blocks Used Available Use% Mounted on
> …
> /dev/mapper/vg-myth 943656628 544996248 398660380 58% /mnt/store
> …
> mythserver michael # pvdisplay /dev/sdb1
> --- Physical volume ---
> PV Name /dev/sdb1
> VG Name vg
> PV Size 698.64 GB / not usable 2.34 MB
> Allocatable yes
> PE Size (KByte) 4096
> Total PE 178850
> Free PE 292
> Allocated PE 178558
> PV UUID SetyUA-DkWL-zDDo-Wm-3avR-nJH8-OnUujv
> mythserver michael # pvdisplay /dev/sdc1
> --- Physical volume ---
> PV Name /dev/sdc1
> VG Name vg
> PV Size 931.51 GB / not usable 3.19 MB
> Allocatable yes (but full)
> PE Size (KByte) 4096
> Total PE 238466
> Free PE 0
> Allocated PE 238466
> PV UUID DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
> mythserver michael # lvextend -L+700G /dev/vg/myth
> Extending logical volume myth to 2.27 TB
> Insufficient free space: 179200 extents needed, but only 292 available
>
> I am guessing that I should have run these commands to extend the logical volume to its desired size:
> vgextend vg /dev/sdb1
> lvextend -L+700G /dev/vg/myth
>
> before creating the filesystem with this command which I am guessing locked the size to the 900GB with I used in my setup steps
> mkfs.jfs /dev/vg/myth
>
>
> Does that sound like my issue?
>
> Any thoughts on how to get out of this situation while ensuring no loss of my data that currently resides on /dev/mapper/vg-myth?
>
>
> I am thinking that the following steps should work:
> Copy all of my files on /dev/mapper/vg-myth to other paritions (I
> assume the call to mkfs.jfs below will delete all the contents of this
> partition)
> "lvreduce -L-641G /dev/vg/myth" (to get the size matched up with 931GB + 698GB [ 2.27TB – 931GB – 698GB)
> "mkfs.jfs /dev/vg/myth" (recreate the filesystem now that the size has been corrected)
> remount /dev/vg/myth
> copy back the files
>
> Thanks in advance
> Mike
>
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@redhat.com
> https://www.redhat.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [linux-lvm] LVM Volume Size Question
@ 2009-01-23 18:02 ctd
2009-01-23 18:18 ` F-D. Cami
0 siblings, 1 reply; 4+ messages in thread
From: ctd @ 2009-01-23 18:02 UTC (permalink / raw)
To: LVM general discussion and development
[-- Attachment #1: Type: text/plain, Size: 8211 bytes --]
BODY { font-family:Arial, Helvetica, sans-serif;font-size:12px;
}Francois,
Here is the output you requested:
mythserver michael # vgdisplay -v vg
Using volume group(s) on command line
Finding volume group "vg"
--- Volume group ---
VG Name vg
System ID
Format lvm2
Metadata Areas 4
Metadata Sequence No 10
VG Access read/write
VG Status resizable
MAX LV 0
Cur LV 1
Open LV 1
Max PV 0
Cur PV 2
Act PV 2
VG Size 1.59 TB
PE Size 4.00 MB
Total PE 417316
Alloc PE / Size 417024 / 1.59 TB
Free PE / Size 292 / 1.14 GB
VG UUID 2a2Vzo-3HUx-gUU0-EYk3-md1s-PgAg-MM0bQ6
--- Logical volume ---
LV Name /dev/vg/myth
VG Name vg
LV UUID 4Auu9y-47vW-6BBd-Rdd5-PP63-3sYB-lvmNmQ
LV Write Access read/write
LV Status available
# open 1
LV Size 1.59 TB
Current LE 417024
Segments 2
Allocation inherit
Read ahead sectors auto
- currently set to 256
Block device 254:0
--- Physical volumes ---
PV Name /dev/sdc1
PV UUID DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
PV Status allocatable
Total PE / Free PE 238466 / 0
PV Name /dev/sdb1
PV UUID SetyUA-DkWL-zDDo-W8Om-3avR-nJH8-OnUujv
PV Status allocatable
Total PE / Free PE 178850 / 292
mythserver michael # lvdisplay /dev/vg/myth
--- Logical volume ---
LV Name /dev/vg/myth
VG Name vg
LV UUID 4Auu9y-47vW-6BBd-Rdd5-PP63-3sYB-lvmNmQ
LV Write Access read/write
LV Status available
# open 1
LV Size 1.59 TB
Current LE 417024
Segments 2
Allocation inherit
Read ahead sectors auto
- currently set to 256
Block device 254:0
So it does appear that /dev/vg/myth is the full 1.59TB. Any reason
that it appears that the output of df does not agree with this or am I
confused?
Regarding your recommended steps, I understand that those steps will
not eliminate my reliance on the two drives not failing but the steps
will setup a stripped mapping which can improve performance, right?
In my situation, the directory mapped to /dev/vg/myth is just used
to store recorded programs from mythtv. Therefore, the data is not
critical, but the desire is to present a directory that is as big as
possible to allow mythtv to not run out of recording space (I think
1.6TB should do it!!!).
Also, I assume that the second "dd" command line you wrote should
have referenced /dev/sdc and not sdb, correct?
Thanks a ton for your insight.
Take it easy,
Mike
On Fri 09/01/23 10:44 , "F-D. Cami" fcami@winsoft.fr sent:
Hi Mike,
Could you give us the output of :
vgdisplay -v vg
lvdisplay /dev/vg/myth
I think your myth LV is already 1.5TB, so you only need to run :
# mount -o remount,resize /dev/vg/myth
for the JFS filesystem to be resized.
Please backup everything before running that :)
However, since what you're doing essentially amounts to RAID0
without the
performance benefits (if you lose one drive, your data is lost), I'd
run a full backup
and run the following commands to create a striped LV :
lvremove /dev/vg/myth
vgremove vg
pvremove /dev/sdb1
pvremove /dev/sdc1
dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
pvcreate /dev/sdb
pvcreate /dev/sdc
vgcreate vg /dev/sdb /dev/sdc
lcreate -i 2 -I 8 -L 1700M -n myth vg /dev/sdb /dev/sdc
(adjust 1700 to whatever your drives will take)
mkfs.jfs /dev/vg/myth
You will lose a bit of space but gain some performance ; the
available VG size can
then be used for other LVs or snapshots.
Best,
Francois
On Fri, 23 Jan 2009 11:14:35 -0600
ctd@minneapolish3.com [1] wrote:
> Hey there,
>
>
>
> I have most likely a simple question concerning LVM that I figured
someone might be able to provide some insight into.
>
>
>
> I just setup LVM with both /dev/sdb1 and /dev/sdc1 being assigned
to my “vg” volume group. There is only one logical volume
“myth” off of “vg”.
>
>
>
> My steps:
>
> fdisk /dev/sdc [created 1 partition to span the entire drive of
type 8e]
> emerge lvm2
> vgscan
> vgchange -a y
> pvcreate /dev/sdc1
> vgcreate vg /dev/sdc1
> lvcreate -L900GB -nmyth vg
> mkfs.jfs /dev/vg/myth
> fdisk /dev/sdb [created 1 partition to span the entire drive of
type 8e]
> pvcreate /dev/sdb1
> vgextend vg /dev/sdb1
> lvextend -L+700G /dev/vg/myth
>
>
> Sdb1: 700GB drive with one partition
>
> Sdd1: 1TB drive with one partition
>
>
> My question is related to the space available in /dev/vg/myth. I
> would assume that I should have ~1.7TB of space on that logical
> partition, but df does not seems to indicate that.
>
>
> # df
> Filesystem 1K-blocks Used Available Use% Mounted on
> …
> /dev/mapper/vg-myth 943656628 544996248 398660380 58% /mnt/store
> …
> mythserver michael # pvdisplay /dev/sdb1
> --- Physical volume ---
> PV Name /dev/sdb1
> VG Name vg
> PV Size 698.64 GB / not usable 2.34 MB
> Allocatable yes
> PE Size (KByte) 4096
> Total PE 178850
> Free PE 292
> Allocated PE 178558
> PV UUID SetyUA-DkWL-zDDo-Wm-3avR-nJH8-OnUujv
> mythserver michael # pvdisplay /dev/sdc1
> --- Physical volume ---
> PV Name /dev/sdc1
> VG Name vg
> PV Size 931.51 GB / not usable 3.19 MB
> Allocatable yes (but full)
> PE Size (KByte) 4096
> Total PE 238466
> Free PE 0
> Allocated PE 238466
> PV UUID DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
> mythserver michael # lvextend -L+700G /dev/vg/myth
> Extending logical volume myth to 2.27 TB
> Insufficient free space: 179200 extents needed, but only 292
available
>
> I am guessing that I should have run these commands to extend the
logical volume to its desired size:
> vgextend vg /dev/sdb1
> lvextend -L+700G /dev/vg/myth
>
> before creating the filesystem with this command which I am
guessing locked the size to the 900GB with I used in my setup steps
> mkfs.jfs /dev/vg/myth
>
>
> Does that sound like my issue?
>
> Any thoughts on how to get out of this situation while ensuring no
loss of my data that currently resides on /dev/mapper/vg-myth?
>
>
> I am thinking that the following steps should work:
> Copy all of my files on /dev/mapper/vg-myth to other paritions (I
> assume the call to mkfs.jfs below will delete all the contents of
this
> partition)
> "lvreduce -L-641G /dev/vg/myth" (to get the size matched up with
931GB + 698GB [ 2.27TB – 931GB – 698GB)
> "mkfs.jfs /dev/vg/myth" (recreate the filesystem now that the size
has been corrected)
> remount /dev/vg/myth
> copy back the files
>
> Thanks in advance
> Mike
>
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@redhat.com [2]
> https://www.redhat.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
_______________________________________________
linux-lvm mailing list
linux-lvm@redhat.com [3]
https://www.redhat.com/mailman/listinfo/linux-lvm
read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
Links:
------
[1] mailto:ctd@minneapolish3.com
[2] mailto:linux-lvm@redhat.com
[3] mailto:linux-lvm@redhat.com
[-- Attachment #2: Type: text/html, Size: 19114 bytes --]
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [linux-lvm] LVM Volume Size Question
2009-01-23 18:02 ctd
@ 2009-01-23 18:18 ` F-D. Cami
0 siblings, 0 replies; 4+ messages in thread
From: F-D. Cami @ 2009-01-23 18:18 UTC (permalink / raw)
To: ctd, LVM general discussion and development
Mike,
df reports filesystem sizes, not block device sizes.
You now only need to backup and grow the filesystem.
Yes, what I meant was an improvement in performance at some space
cost, if all you need is space, don't do it.
And, yes, the 2nd dd should be sdc :)
Cheers
F
On Fri, 23 Jan 2009 12:02:45 -0600
ctd@minneapolish3.com wrote:
> BODY { font-family:Arial, Helvetica, sans-serif;font-size:12px;
> }Francois,
> Here is the output you requested:
> mythserver michael # vgdisplay -v vg
> Using volume group(s) on command line
> Finding volume group "vg"
> --- Volume group ---
> VG Name vg
> System ID
> Format lvm2
> Metadata Areas 4
> Metadata Sequence No 10
> VG Access read/write
> VG Status resizable
> MAX LV 0
> Cur LV 1
> Open LV 1
> Max PV 0
> Cur PV 2
> Act PV 2
> VG Size 1.59 TB
> PE Size 4.00 MB
> Total PE 417316
> Alloc PE / Size 417024 / 1.59 TB
> Free PE / Size 292 / 1.14 GB
> VG UUID 2a2Vzo-3HUx-gUU0-EYk3-md1s-PgAg-MM0bQ6
> --- Logical volume ---
> LV Name /dev/vg/myth
> VG Name vg
> LV UUID 4Auu9y-47vW-6BBd-Rdd5-PP63-3sYB-lvmNmQ
> LV Write Access read/write
> LV Status available
> # open 1
> LV Size 1.59 TB
> Current LE 417024
> Segments 2
> Allocation inherit
> Read ahead sectors auto
> - currently set to 256
> Block device 254:0
> --- Physical volumes ---
> PV Name /dev/sdc1
> PV UUID DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
> PV Status allocatable
> Total PE / Free PE 238466 / 0
> PV Name /dev/sdb1
> PV UUID SetyUA-DkWL-zDDo-W8Om-3avR-nJH8-OnUujv
> PV Status allocatable
> Total PE / Free PE 178850 / 292
> mythserver michael # lvdisplay /dev/vg/myth
> --- Logical volume ---
> LV Name /dev/vg/myth
> VG Name vg
> LV UUID 4Auu9y-47vW-6BBd-Rdd5-PP63-3sYB-lvmNmQ
> LV Write Access read/write
> LV Status available
> # open 1
> LV Size 1.59 TB
> Current LE 417024
> Segments 2
> Allocation inherit
> Read ahead sectors auto
> - currently set to 256
> Block device 254:0
> So it does appear that /dev/vg/myth is the full 1.59TB. Any reason
> that it appears that the output of df does not agree with this or am I
> confused?
> Regarding your recommended steps, I understand that those steps will
> not eliminate my reliance on the two drives not failing but the steps
> will setup a stripped mapping which can improve performance, right?
> In my situation, the directory mapped to /dev/vg/myth is just used
> to store recorded programs from mythtv. Therefore, the data is not
> critical, but the desire is to present a directory that is as big as
> possible to allow mythtv to not run out of recording space (I think
> 1.6TB should do it!!!).
> Also, I assume that the second "dd" command line you wrote should
> have referenced /dev/sdc and not sdb, correct?
> Thanks a ton for your insight.
> Take it easy,
> Mike
> On Fri 09/01/23 10:44 , "F-D. Cami" fcami@winsoft.fr sent:
> Hi Mike,
> Could you give us the output of :
> vgdisplay -v vg
> lvdisplay /dev/vg/myth
> I think your myth LV is already 1.5TB, so you only need to run :
> # mount -o remount,resize /dev/vg/myth
> for the JFS filesystem to be resized.
> Please backup everything before running that :)
> However, since what you're doing essentially amounts to RAID0
> without the
> performance benefits (if you lose one drive, your data is lost), I'd
> run a full backup
> and run the following commands to create a striped LV :
> lvremove /dev/vg/myth
> vgremove vg
> pvremove /dev/sdb1
> pvremove /dev/sdc1
> dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
> dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
> pvcreate /dev/sdb
> pvcreate /dev/sdc
> vgcreate vg /dev/sdb /dev/sdc
> lcreate -i 2 -I 8 -L 1700M -n myth vg /dev/sdb /dev/sdc
> (adjust 1700 to whatever your drives will take)
> mkfs.jfs /dev/vg/myth
> You will lose a bit of space but gain some performance ; the
> available VG size can
> then be used for other LVs or snapshots.
> Best,
> Francois
> On Fri, 23 Jan 2009 11:14:35 -0600
> ctd@minneapolish3.com [1] wrote:
> > Hey there,
> >
> >
> >
> > I have most likely a simple question concerning LVM that I figured
> someone might be able to provide some insight into.
> >
> >
> >
> > I just setup LVM with both /dev/sdb1 and /dev/sdc1 being assigned
> to my “vg” volume group. There is only one logical volume
> “myth” off of “vg”.
> >
> >
> >
> > My steps:
> >
> > fdisk /dev/sdc [created 1 partition to span the entire drive of
> type 8e]
> > emerge lvm2
> > vgscan
> > vgchange -a y
> > pvcreate /dev/sdc1
> > vgcreate vg /dev/sdc1
> > lvcreate -L900GB -nmyth vg
> > mkfs.jfs /dev/vg/myth
> > fdisk /dev/sdb [created 1 partition to span the entire drive of
> type 8e]
> > pvcreate /dev/sdb1
> > vgextend vg /dev/sdb1
> > lvextend -L+700G /dev/vg/myth
> >
> >
> > Sdb1: 700GB drive with one partition
> >
> > Sdd1: 1TB drive with one partition
> >
> >
> > My question is related to the space available in /dev/vg/myth. I
> > would assume that I should have ~1.7TB of space on that logical
> > partition, but df does not seems to indicate that.
> >
> >
> > # df
> > Filesystem 1K-blocks Used Available Use% Mounted on
> > …
> > /dev/mapper/vg-myth 943656628 544996248 398660380 58% /mnt/store
> > …
> > mythserver michael # pvdisplay /dev/sdb1
> > --- Physical volume ---
> > PV Name /dev/sdb1
> > VG Name vg
> > PV Size 698.64 GB / not usable 2.34 MB
> > Allocatable yes
> > PE Size (KByte) 4096
> > Total PE 178850
> > Free PE 292
> > Allocated PE 178558
> > PV UUID SetyUA-DkWL-zDDo-Wm-3avR-nJH8-OnUujv
> > mythserver michael # pvdisplay /dev/sdc1
> > --- Physical volume ---
> > PV Name /dev/sdc1
> > VG Name vg
> > PV Size 931.51 GB / not usable 3.19 MB
> > Allocatable yes (but full)
> > PE Size (KByte) 4096
> > Total PE 238466
> > Free PE 0
> > Allocated PE 238466
> > PV UUID DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
> > mythserver michael # lvextend -L+700G /dev/vg/myth
> > Extending logical volume myth to 2.27 TB
> > Insufficient free space: 179200 extents needed, but only 292
> available
> >
> > I am guessing that I should have run these commands to extend the
> logical volume to its desired size:
> > vgextend vg /dev/sdb1
> > lvextend -L+700G /dev/vg/myth
> >
> > before creating the filesystem with this command which I am
> guessing locked the size to the 900GB with I used in my setup steps
> > mkfs.jfs /dev/vg/myth
> >
> >
> > Does that sound like my issue?
> >
> > Any thoughts on how to get out of this situation while ensuring no
> loss of my data that currently resides on /dev/mapper/vg-myth?
> >
> >
> > I am thinking that the following steps should work:
> > Copy all of my files on /dev/mapper/vg-myth to other paritions (I
> > assume the call to mkfs.jfs below will delete all the contents of
> this
> > partition)
> > "lvreduce -L-641G /dev/vg/myth" (to get the size matched up with
> 931GB + 698GB [ 2.27TB – 931GB – 698GB)
> > "mkfs.jfs /dev/vg/myth" (recreate the filesystem now that the size
> has been corrected)
> > remount /dev/vg/myth
> > copy back the files
> >
> > Thanks in advance
> > Mike
> >
> > _______________________________________________
> > linux-lvm mailing list
> > linux-lvm@redhat.com [2]
> > https://www.redhat.com/mailman/listinfo/linux-lvm
> > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@redhat.com [3]
> https://www.redhat.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>
>
> Links:
> ------
> [1] mailto:ctd@minneapolish3.com
> [2] mailto:linux-lvm@redhat.com
> [3] mailto:linux-lvm@redhat.com
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2009-01-23 18:18 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-01-23 17:14 [linux-lvm] LVM Volume Size Question ctd
2009-01-23 17:44 ` F-D. Cami
-- strict thread matches above, loose matches on Subject: below --
2009-01-23 18:02 ctd
2009-01-23 18:18 ` F-D. Cami
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).