linux-lvm.redhat.com archive mirror
 help / color / mirror / Atom feed
* [linux-lvm] LVM Volume Size Question
@ 2009-01-23 17:14 ctd
  2009-01-23 17:44 ` F-D. Cami
  0 siblings, 1 reply; 4+ messages in thread
From: ctd @ 2009-01-23 17:14 UTC (permalink / raw)
  To: linux-lvm

Hey there,



I have most likely a simple question concerning LVM that I figured someone might be able to provide some insight into.



I just setup LVM with both /dev/sdb1 and /dev/sdc1 being assigned to my “vg” volume group. There is only one logical volume “myth” off of “vg”.



My steps:

fdisk /dev/sdc [created 1 partition to span the entire drive of type 8e]
emerge lvm2
vgscan
vgchange -a y 
pvcreate /dev/sdc1
vgcreate vg /dev/sdc1
lvcreate -L900GB -nmyth vg
mkfs.jfs /dev/vg/myth
fdisk /dev/sdb [created 1 partition to span the entire drive of type 8e]
pvcreate /dev/sdb1
vgextend vg /dev/sdb1
lvextend -L+700G /dev/vg/myth


Sdb1: 700GB drive with one partition

Sdd1: 1TB drive with one partition


My question is related to the space available in /dev/vg/myth. I
would assume that I should have ~1.7TB of space on that logical
partition, but df does not seems to indicate that. 


 # df
Filesystem           1K-blocks      Used Available Use% Mounted on
…
/dev/mapper/vg-myth  943656628 544996248 398660380  58% /mnt/store
…
mythserver michael # pvdisplay /dev/sdb1
  --- Physical volume ---
  PV Name               /dev/sdb1
  VG Name               vg
  PV Size               698.64 GB / not usable 2.34 MB
  Allocatable           yes
  PE Size (KByte)       4096
  Total PE              178850
  Free PE               292
  Allocated PE          178558
  PV UUID               SetyUA-DkWL-zDDo-Wm-3avR-nJH8-OnUujv
mythserver michael # pvdisplay /dev/sdc1
  --- Physical volume ---
  PV Name               /dev/sdc1
  VG Name               vg
  PV Size               931.51 GB / not usable 3.19 MB
  Allocatable           yes (but full)
  PE Size (KByte)       4096
  Total PE              238466
  Free PE               0
  Allocated PE          238466
  PV UUID               DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
mythserver michael # lvextend -L+700G /dev/vg/myth
  Extending logical volume myth to 2.27 TB
  Insufficient free space: 179200 extents needed, but only 292 available

I am guessing that I should have run these commands to extend the logical volume to its desired size:
vgextend vg /dev/sdb1
lvextend -L+700G /dev/vg/myth

before creating the filesystem with this command which I am guessing locked the size to the 900GB with I used in my setup steps
mkfs.jfs /dev/vg/myth

	
Does that sound like my issue?

Any thoughts on how to get out of this situation while ensuring no loss of my data that currently resides on /dev/mapper/vg-myth?


I am thinking that the following steps should work:
Copy all of my files on /dev/mapper/vg-myth to other paritions (I
assume the call to mkfs.jfs below will delete all the contents of this
partition)
"lvreduce -L-641G /dev/vg/myth" (to get the size matched up with 931GB + 698GB [ 2.27TB – 931GB – 698GB)
"mkfs.jfs /dev/vg/myth" (recreate the filesystem now that the size has been corrected) 
remount /dev/vg/myth
copy back the files

Thanks in advance
Mike 

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [linux-lvm] LVM Volume Size Question
  2009-01-23 17:14 [linux-lvm] LVM Volume Size Question ctd
@ 2009-01-23 17:44 ` F-D. Cami
  0 siblings, 0 replies; 4+ messages in thread
From: F-D. Cami @ 2009-01-23 17:44 UTC (permalink / raw)
  To: ctd, LVM general discussion and development


Hi Mike,

Could you give us the output of :

vgdisplay -v vg
lvdisplay /dev/vg/myth

I think your myth LV is already 1.5TB, so you only need to run :
# mount -o remount,resize /dev/vg/myth
for the JFS filesystem to be resized.
Please backup everything before running that :)

However, since what you're doing essentially amounts to RAID0 without the
performance benefits (if you lose one drive, your data is lost), I'd run a full backup
and run the following commands to create a striped LV :

lvremove /dev/vg/myth
vgremove vg
pvremove /dev/sdb1
pvremove /dev/sdc1
dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
pvcreate /dev/sdb
pvcreate /dev/sdc
vgcreate vg /dev/sdb /dev/sdc
lcreate -i 2 -I 8 -L 1700M -n myth vg /dev/sdb /dev/sdc
   (adjust 1700 to whatever your drives will take)
mkfs.jfs /dev/vg/myth

You will lose a bit of space but gain some performance ; the available VG size can
then be used for other LVs or snapshots.

Best,

Francois



On Fri, 23 Jan 2009 11:14:35 -0600
ctd@minneapolish3.com wrote:

> Hey there,
> 
> 
> 
> I have most likely a simple question concerning LVM that I figured someone might be able to provide some insight into.
> 
> 
> 
> I just setup LVM with both /dev/sdb1 and /dev/sdc1 being assigned to my “vg” volume group. There is only one logical volume “myth” off of “vg”.
> 
> 
> 
> My steps:
> 
> fdisk /dev/sdc [created 1 partition to span the entire drive of type 8e]
> emerge lvm2
> vgscan
> vgchange -a y 
> pvcreate /dev/sdc1
> vgcreate vg /dev/sdc1
> lvcreate -L900GB -nmyth vg
> mkfs.jfs /dev/vg/myth
> fdisk /dev/sdb [created 1 partition to span the entire drive of type 8e]
> pvcreate /dev/sdb1
> vgextend vg /dev/sdb1
> lvextend -L+700G /dev/vg/myth
> 
> 
> Sdb1: 700GB drive with one partition
> 
> Sdd1: 1TB drive with one partition
> 
> 
> My question is related to the space available in /dev/vg/myth. I
> would assume that I should have ~1.7TB of space on that logical
> partition, but df does not seems to indicate that. 
> 
> 
>  # df
> Filesystem           1K-blocks      Used Available Use% Mounted on
> …
> /dev/mapper/vg-myth  943656628 544996248 398660380  58% /mnt/store
> …
> mythserver michael # pvdisplay /dev/sdb1
>   --- Physical volume ---
>   PV Name               /dev/sdb1
>   VG Name               vg
>   PV Size               698.64 GB / not usable 2.34 MB
>   Allocatable           yes
>   PE Size (KByte)       4096
>   Total PE              178850
>   Free PE               292
>   Allocated PE          178558
>   PV UUID               SetyUA-DkWL-zDDo-Wm-3avR-nJH8-OnUujv
> mythserver michael # pvdisplay /dev/sdc1
>   --- Physical volume ---
>   PV Name               /dev/sdc1
>   VG Name               vg
>   PV Size               931.51 GB / not usable 3.19 MB
>   Allocatable           yes (but full)
>   PE Size (KByte)       4096
>   Total PE              238466
>   Free PE               0
>   Allocated PE          238466
>   PV UUID               DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
> mythserver michael # lvextend -L+700G /dev/vg/myth
>   Extending logical volume myth to 2.27 TB
>   Insufficient free space: 179200 extents needed, but only 292 available
> 
> I am guessing that I should have run these commands to extend the logical volume to its desired size:
> vgextend vg /dev/sdb1
> lvextend -L+700G /dev/vg/myth
> 
> before creating the filesystem with this command which I am guessing locked the size to the 900GB with I used in my setup steps
> mkfs.jfs /dev/vg/myth
> 
> 	
> Does that sound like my issue?
> 
> Any thoughts on how to get out of this situation while ensuring no loss of my data that currently resides on /dev/mapper/vg-myth?
> 
> 
> I am thinking that the following steps should work:
> Copy all of my files on /dev/mapper/vg-myth to other paritions (I
> assume the call to mkfs.jfs below will delete all the contents of this
> partition)
> "lvreduce -L-641G /dev/vg/myth" (to get the size matched up with 931GB + 698GB [ 2.27TB – 931GB – 698GB)
> "mkfs.jfs /dev/vg/myth" (recreate the filesystem now that the size has been corrected) 
> remount /dev/vg/myth
> copy back the files
> 
> Thanks in advance
> Mike 
> 
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@redhat.com
> https://www.redhat.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [linux-lvm] LVM Volume Size Question
@ 2009-01-23 18:02 ctd
  2009-01-23 18:18 ` F-D. Cami
  0 siblings, 1 reply; 4+ messages in thread
From: ctd @ 2009-01-23 18:02 UTC (permalink / raw)
  To: LVM general discussion and development

[-- Attachment #1: Type: text/plain, Size: 8211 bytes --]

  BODY { font-family:Arial, Helvetica, sans-serif;font-size:12px;
}Francois,
 Here is the output you requested:
 mythserver michael # vgdisplay -v vg
     Using volume group(s) on command line
     Finding volume group "vg"
   --- Volume group ---
   VG Name               vg
   System ID
   Format                lvm2
   Metadata Areas        4
   Metadata Sequence No  10
   VG Access             read/write
   VG Status             resizable
   MAX LV                0
   Cur LV                1
   Open LV               1
   Max PV                0
   Cur PV                2
   Act PV                2
   VG Size               1.59 TB
   PE Size               4.00 MB
   Total PE              417316
   Alloc PE / Size       417024 / 1.59 TB
   Free  PE / Size       292 / 1.14 GB
   VG UUID               2a2Vzo-3HUx-gUU0-EYk3-md1s-PgAg-MM0bQ6
   --- Logical volume ---
   LV Name                /dev/vg/myth
   VG Name                vg
   LV UUID                4Auu9y-47vW-6BBd-Rdd5-PP63-3sYB-lvmNmQ
   LV Write Access        read/write
   LV Status              available
   # open                 1
   LV Size                1.59 TB
   Current LE             417024
   Segments               2
   Allocation             inherit
   Read ahead sectors     auto
   - currently set to     256
   Block device           254:0
   --- Physical volumes ---
   PV Name               /dev/sdc1
   PV UUID               DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
   PV Status             allocatable
   Total PE / Free PE    238466 / 0
   PV Name               /dev/sdb1
   PV UUID               SetyUA-DkWL-zDDo-W8Om-3avR-nJH8-OnUujv
   PV Status             allocatable
   Total PE / Free PE    178850 / 292
 mythserver michael # lvdisplay /dev/vg/myth
   --- Logical volume ---
   LV Name                /dev/vg/myth
   VG Name                vg
   LV UUID                4Auu9y-47vW-6BBd-Rdd5-PP63-3sYB-lvmNmQ
   LV Write Access        read/write
   LV Status              available
   # open                 1
   LV Size                1.59 TB
   Current LE             417024
   Segments               2
   Allocation             inherit
   Read ahead sectors     auto
   - currently set to     256
   Block device           254:0
 So it does appear that /dev/vg/myth is the full 1.59TB.  Any reason
that it appears that the output of df does not agree with this or am I
confused?
 Regarding your recommended steps, I understand that those steps will
not eliminate my reliance on the two drives not failing but the steps
will setup a stripped mapping which can improve performance, right?
 In my situation, the directory mapped to /dev/vg/myth is just used
to store recorded programs from mythtv.  Therefore, the data is not
critical, but the desire is to present a directory that is as big as
possible to allow mythtv to not run out of recording space (I think
1.6TB should do it!!!).
 Also, I assume that the second "dd" command line you wrote should
have referenced /dev/sdc and not sdb, correct?
 Thanks a ton for your insight.
 Take it easy,
 Mike
 On Fri 09/01/23 10:44 , "F-D. Cami" fcami@winsoft.fr sent:
 Hi Mike,
 Could you give us the output of :
 vgdisplay -v vg
 lvdisplay /dev/vg/myth
 I think your myth LV is already 1.5TB, so you only need to run :
 # mount -o remount,resize /dev/vg/myth
 for the JFS filesystem to be resized.
 Please backup everything before running that :)
 However, since what you're doing essentially amounts to RAID0
without the
 performance benefits (if you lose one drive, your data is lost), I'd
run a full backup
 and run the following commands to create a striped LV :
 lvremove /dev/vg/myth
 vgremove vg
 pvremove /dev/sdb1
 pvremove /dev/sdc1
 dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
 dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
 pvcreate /dev/sdb
 pvcreate /dev/sdc
 vgcreate vg /dev/sdb /dev/sdc
 lcreate -i 2 -I 8 -L 1700M -n myth vg /dev/sdb /dev/sdc
    (adjust 1700 to whatever your drives will take)
 mkfs.jfs /dev/vg/myth
 You will lose a bit of space but gain some performance ; the
available VG size can
 then be used for other LVs or snapshots.
 Best,
 Francois
 On Fri, 23 Jan 2009 11:14:35 -0600
 ctd@minneapolish3.com [1] wrote:
 > Hey there,
 > 
 > 
 > 
 > I have most likely a simple question concerning LVM that I figured
someone might be able to provide some insight into.
 > 
 > 
 > 
 > I just setup LVM with both /dev/sdb1 and /dev/sdc1 being assigned
to my “vg” volume group. There is only one logical volume
“myth” off of “vg”.
 > 
 > 
 > 
 > My steps:
 > 
 > fdisk /dev/sdc [created 1 partition to span the entire drive of
type 8e]
 > emerge lvm2
 > vgscan
 > vgchange -a y 
 > pvcreate /dev/sdc1
 > vgcreate vg /dev/sdc1
 > lvcreate -L900GB -nmyth vg
 > mkfs.jfs /dev/vg/myth
 > fdisk /dev/sdb [created 1 partition to span the entire drive of
type 8e]
 > pvcreate /dev/sdb1
 > vgextend vg /dev/sdb1
 > lvextend -L+700G /dev/vg/myth
 > 
 > 
 > Sdb1: 700GB drive with one partition
 > 
 > Sdd1: 1TB drive with one partition
 > 
 > 
 > My question is related to the space available in /dev/vg/myth. I
 > would assume that I should have ~1.7TB of space on that logical
 > partition, but df does not seems to indicate that. 
 > 
 > 
 >  # df
 > Filesystem           1K-blocks      Used Available Use% Mounted on
 > …
 > /dev/mapper/vg-myth  943656628 544996248 398660380  58% /mnt/store
 > …
 > mythserver michael # pvdisplay /dev/sdb1
 >   --- Physical volume ---
 >   PV Name               /dev/sdb1
 >   VG Name               vg
 >   PV Size               698.64 GB / not usable 2.34 MB
 >   Allocatable           yes
 >   PE Size (KByte)       4096
 >   Total PE              178850
 >   Free PE               292
 >   Allocated PE          178558
 >   PV UUID               SetyUA-DkWL-zDDo-Wm-3avR-nJH8-OnUujv
 > mythserver michael # pvdisplay /dev/sdc1
 >   --- Physical volume ---
 >   PV Name               /dev/sdc1
 >   VG Name               vg
 >   PV Size               931.51 GB / not usable 3.19 MB
 >   Allocatable           yes (but full)
 >   PE Size (KByte)       4096
 >   Total PE              238466
 >   Free PE               0
 >   Allocated PE          238466
 >   PV UUID               DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
 > mythserver michael # lvextend -L+700G /dev/vg/myth
 >   Extending logical volume myth to 2.27 TB
 >   Insufficient free space: 179200 extents needed, but only 292
available
 > 
 > I am guessing that I should have run these commands to extend the
logical volume to its desired size:
 > vgextend vg /dev/sdb1
 > lvextend -L+700G /dev/vg/myth
 > 
 > before creating the filesystem with this command which I am
guessing locked the size to the 900GB with I used in my setup steps
 > mkfs.jfs /dev/vg/myth
 > 
 >  
 > Does that sound like my issue?
 > 
 > Any thoughts on how to get out of this situation while ensuring no
loss of my data that currently resides on /dev/mapper/vg-myth?
 > 
 > 
 > I am thinking that the following steps should work:
 > Copy all of my files on /dev/mapper/vg-myth to other paritions (I
 > assume the call to mkfs.jfs below will delete all the contents of
this
 > partition)
 > "lvreduce -L-641G /dev/vg/myth" (to get the size matched up with
931GB + 698GB [ 2.27TB – 931GB – 698GB)
 > "mkfs.jfs /dev/vg/myth" (recreate the filesystem now that the size
has been corrected) 
 > remount /dev/vg/myth
 > copy back the files
 > 
 > Thanks in advance
 > Mike 
 > 
 > _______________________________________________
 > linux-lvm mailing list
 > linux-lvm@redhat.com [2]
 > https://www.redhat.com/mailman/listinfo/linux-lvm
 > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
 _______________________________________________
 linux-lvm mailing list
 linux-lvm@redhat.com [3]
 https://www.redhat.com/mailman/listinfo/linux-lvm
 read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
 

Links:
------
[1] mailto:ctd@minneapolish3.com
[2] mailto:linux-lvm@redhat.com
[3] mailto:linux-lvm@redhat.com

[-- Attachment #2: Type: text/html, Size: 19114 bytes --]

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [linux-lvm] LVM Volume Size Question
  2009-01-23 18:02 ctd
@ 2009-01-23 18:18 ` F-D. Cami
  0 siblings, 0 replies; 4+ messages in thread
From: F-D. Cami @ 2009-01-23 18:18 UTC (permalink / raw)
  To: ctd, LVM general discussion and development


Mike,

df reports filesystem sizes, not block device sizes.
You now only need to backup and grow the filesystem.
Yes, what I meant was an improvement in performance at some space
cost, if all you need is space, don't do it. 
And, yes, the 2nd dd should be sdc :)

Cheers

F


On Fri, 23 Jan 2009 12:02:45 -0600
ctd@minneapolish3.com wrote:

>   BODY { font-family:Arial, Helvetica, sans-serif;font-size:12px;
> }Francois,
>  Here is the output you requested:
>  mythserver michael # vgdisplay -v vg
>      Using volume group(s) on command line
>      Finding volume group "vg"
>    --- Volume group ---
>    VG Name               vg
>    System ID
>    Format                lvm2
>    Metadata Areas        4
>    Metadata Sequence No  10
>    VG Access             read/write
>    VG Status             resizable
>    MAX LV                0
>    Cur LV                1
>    Open LV               1
>    Max PV                0
>    Cur PV                2
>    Act PV                2
>    VG Size               1.59 TB
>    PE Size               4.00 MB
>    Total PE              417316
>    Alloc PE / Size       417024 / 1.59 TB
>    Free  PE / Size       292 / 1.14 GB
>    VG UUID               2a2Vzo-3HUx-gUU0-EYk3-md1s-PgAg-MM0bQ6
>    --- Logical volume ---
>    LV Name                /dev/vg/myth
>    VG Name                vg
>    LV UUID                4Auu9y-47vW-6BBd-Rdd5-PP63-3sYB-lvmNmQ
>    LV Write Access        read/write
>    LV Status              available
>    # open                 1
>    LV Size                1.59 TB
>    Current LE             417024
>    Segments               2
>    Allocation             inherit
>    Read ahead sectors     auto
>    - currently set to     256
>    Block device           254:0
>    --- Physical volumes ---
>    PV Name               /dev/sdc1
>    PV UUID               DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
>    PV Status             allocatable
>    Total PE / Free PE    238466 / 0
>    PV Name               /dev/sdb1
>    PV UUID               SetyUA-DkWL-zDDo-W8Om-3avR-nJH8-OnUujv
>    PV Status             allocatable
>    Total PE / Free PE    178850 / 292
>  mythserver michael # lvdisplay /dev/vg/myth
>    --- Logical volume ---
>    LV Name                /dev/vg/myth
>    VG Name                vg
>    LV UUID                4Auu9y-47vW-6BBd-Rdd5-PP63-3sYB-lvmNmQ
>    LV Write Access        read/write
>    LV Status              available
>    # open                 1
>    LV Size                1.59 TB
>    Current LE             417024
>    Segments               2
>    Allocation             inherit
>    Read ahead sectors     auto
>    - currently set to     256
>    Block device           254:0
>  So it does appear that /dev/vg/myth is the full 1.59TB.  Any reason
> that it appears that the output of df does not agree with this or am I
> confused?
>  Regarding your recommended steps, I understand that those steps will
> not eliminate my reliance on the two drives not failing but the steps
> will setup a stripped mapping which can improve performance, right?
>  In my situation, the directory mapped to /dev/vg/myth is just used
> to store recorded programs from mythtv.  Therefore, the data is not
> critical, but the desire is to present a directory that is as big as
> possible to allow mythtv to not run out of recording space (I think
> 1.6TB should do it!!!).
>  Also, I assume that the second "dd" command line you wrote should
> have referenced /dev/sdc and not sdb, correct?
>  Thanks a ton for your insight.
>  Take it easy,
>  Mike
>  On Fri 09/01/23 10:44 , "F-D. Cami" fcami@winsoft.fr sent:
>  Hi Mike,
>  Could you give us the output of :
>  vgdisplay -v vg
>  lvdisplay /dev/vg/myth
>  I think your myth LV is already 1.5TB, so you only need to run :
>  # mount -o remount,resize /dev/vg/myth
>  for the JFS filesystem to be resized.
>  Please backup everything before running that :)
>  However, since what you're doing essentially amounts to RAID0
> without the
>  performance benefits (if you lose one drive, your data is lost), I'd
> run a full backup
>  and run the following commands to create a striped LV :
>  lvremove /dev/vg/myth
>  vgremove vg
>  pvremove /dev/sdb1
>  pvremove /dev/sdc1
>  dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
>  dd if=/dev/zero of=/dev/sdb bs=4096 count=10000
>  pvcreate /dev/sdb
>  pvcreate /dev/sdc
>  vgcreate vg /dev/sdb /dev/sdc
>  lcreate -i 2 -I 8 -L 1700M -n myth vg /dev/sdb /dev/sdc
>     (adjust 1700 to whatever your drives will take)
>  mkfs.jfs /dev/vg/myth
>  You will lose a bit of space but gain some performance ; the
> available VG size can
>  then be used for other LVs or snapshots.
>  Best,
>  Francois
>  On Fri, 23 Jan 2009 11:14:35 -0600
>  ctd@minneapolish3.com [1] wrote:
>  > Hey there,
>  > 
>  > 
>  > 
>  > I have most likely a simple question concerning LVM that I figured
> someone might be able to provide some insight into.
>  > 
>  > 
>  > 
>  > I just setup LVM with both /dev/sdb1 and /dev/sdc1 being assigned
> to my “vg” volume group. There is only one logical volume
> “myth” off of “vg”.
>  > 
>  > 
>  > 
>  > My steps:
>  > 
>  > fdisk /dev/sdc [created 1 partition to span the entire drive of
> type 8e]
>  > emerge lvm2
>  > vgscan
>  > vgchange -a y 
>  > pvcreate /dev/sdc1
>  > vgcreate vg /dev/sdc1
>  > lvcreate -L900GB -nmyth vg
>  > mkfs.jfs /dev/vg/myth
>  > fdisk /dev/sdb [created 1 partition to span the entire drive of
> type 8e]
>  > pvcreate /dev/sdb1
>  > vgextend vg /dev/sdb1
>  > lvextend -L+700G /dev/vg/myth
>  > 
>  > 
>  > Sdb1: 700GB drive with one partition
>  > 
>  > Sdd1: 1TB drive with one partition
>  > 
>  > 
>  > My question is related to the space available in /dev/vg/myth. I
>  > would assume that I should have ~1.7TB of space on that logical
>  > partition, but df does not seems to indicate that. 
>  > 
>  > 
>  >  # df
>  > Filesystem           1K-blocks      Used Available Use% Mounted on
>  > …
>  > /dev/mapper/vg-myth  943656628 544996248 398660380  58% /mnt/store
>  > …
>  > mythserver michael # pvdisplay /dev/sdb1
>  >   --- Physical volume ---
>  >   PV Name               /dev/sdb1
>  >   VG Name               vg
>  >   PV Size               698.64 GB / not usable 2.34 MB
>  >   Allocatable           yes
>  >   PE Size (KByte)       4096
>  >   Total PE              178850
>  >   Free PE               292
>  >   Allocated PE          178558
>  >   PV UUID               SetyUA-DkWL-zDDo-Wm-3avR-nJH8-OnUujv
>  > mythserver michael # pvdisplay /dev/sdc1
>  >   --- Physical volume ---
>  >   PV Name               /dev/sdc1
>  >   VG Name               vg
>  >   PV Size               931.51 GB / not usable 3.19 MB
>  >   Allocatable           yes (but full)
>  >   PE Size (KByte)       4096
>  >   Total PE              238466
>  >   Free PE               0
>  >   Allocated PE          238466
>  >   PV UUID               DX11mo-r0Eh-jN5N-objS-oqo6-eVSU-MShkS2
>  > mythserver michael # lvextend -L+700G /dev/vg/myth
>  >   Extending logical volume myth to 2.27 TB
>  >   Insufficient free space: 179200 extents needed, but only 292
> available
>  > 
>  > I am guessing that I should have run these commands to extend the
> logical volume to its desired size:
>  > vgextend vg /dev/sdb1
>  > lvextend -L+700G /dev/vg/myth
>  > 
>  > before creating the filesystem with this command which I am
> guessing locked the size to the 900GB with I used in my setup steps
>  > mkfs.jfs /dev/vg/myth
>  > 
>  >  
>  > Does that sound like my issue?
>  > 
>  > Any thoughts on how to get out of this situation while ensuring no
> loss of my data that currently resides on /dev/mapper/vg-myth?
>  > 
>  > 
>  > I am thinking that the following steps should work:
>  > Copy all of my files on /dev/mapper/vg-myth to other paritions (I
>  > assume the call to mkfs.jfs below will delete all the contents of
> this
>  > partition)
>  > "lvreduce -L-641G /dev/vg/myth" (to get the size matched up with
> 931GB + 698GB [ 2.27TB – 931GB – 698GB)
>  > "mkfs.jfs /dev/vg/myth" (recreate the filesystem now that the size
> has been corrected) 
>  > remount /dev/vg/myth
>  > copy back the files
>  > 
>  > Thanks in advance
>  > Mike 
>  > 
>  > _______________________________________________
>  > linux-lvm mailing list
>  > linux-lvm@redhat.com [2]
>  > https://www.redhat.com/mailman/listinfo/linux-lvm
>  > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>  _______________________________________________
>  linux-lvm mailing list
>  linux-lvm@redhat.com [3]
>  https://www.redhat.com/mailman/listinfo/linux-lvm
>  read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>  
> 
> Links:
> ------
> [1] mailto:ctd@minneapolish3.com
> [2] mailto:linux-lvm@redhat.com
> [3] mailto:linux-lvm@redhat.com

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2009-01-23 18:18 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-01-23 17:14 [linux-lvm] LVM Volume Size Question ctd
2009-01-23 17:44 ` F-D. Cami
  -- strict thread matches above, loose matches on Subject: below --
2009-01-23 18:02 ctd
2009-01-23 18:18 ` F-D. Cami

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).