From: Lars Ellenberg <lars.ellenberg@linbit.com>
To: linux-lvm@redhat.com
Subject: Re: [linux-lvm] Allocation Policy for Cloud Computing needed
Date: Mon, 20 Feb 2012 22:59:11 +0100 [thread overview]
Message-ID: <20120220215911.GJ6331@barkeeper1-xen.linbit> (raw)
In-Reply-To: <4F3D098F.4010709@profitbricks.com>
On Thu, Feb 16, 2012 at 02:50:07PM +0100, Sebastian Riemer wrote:
> Hi LVM list,
>
> I'm experimenting with storage for many QEMU/KVM virtual machines in
> cloud computing. I've got many concurrent IO processes and 24 hard
> drives. I've tested the scalability with a single IO reader process per
> hard drive. Single drives scale best and have the best performance of
> cause, but we need mirroring and volume management. So I've created MD
> RAID-1 arrays and created on each a VG and two LVs. This gives me good
> overall performance (up to 2 GB/s, HBA limit: 2.2 GB/s).
>
> Then, I've tested to put all my RAID-1 arrays into a single VG, because
> LV size should be adjustable over all hard drives. I've tried all
> allocation policies but none does what I want to achieve here. Yeah,
> that this isn't implemented fully is in the man page, ... .
>
> I want to have an allocation which distributes the LVs equally over the
> PVs as long as space is left and LVs aren't resized. The goal is to
> minimize the number of concurrent IO processes per hard drive (striping
> is total crap in this situation).
>
> I've tested LVM2 2.02.66 and kernel 3.0.15. Is something like that
> implemented in newer releases or is something like that intended to be
> implemented in near future?
I don't know. Does not look like it, though.
> Or does someone want to implement this together with me?
I would certainly be here for discussions.
Though, as you always will be more flexible with scripts than with
pre-implemented fixed algorithms, I probably would first check if I can
solve it with some scripting.
[completely untested, but you get the idea]
#!/bin/bash
export LANG=C LC_ALL=C
name=$1 vg=$2 size_in_MiB=$3
PVS=$(vgs --nohead --unit m -o pv_name,pv_free -O -pv_free,pv_name $vg |
awk -v need=$size_in_MiB '{ print $1; sum += $2;
if (sum >= need) exit; }')
lvcreate -n $name -L ${size_in_MiB}m $vg $PVS
(similar for lvextend)
Which basically implements this allocation policy:
use the pvs with most free space available,
and no more than necessary.
If I understand you correctly, that would almost do what you asked for.
You can get pretty complex in similar scripts, if you really want to...
consider using
pvs -o vg_name,lv_name,pv_name,pvseg_start,pvseg_size,seg_pe_ranges
and explicitly listing not only the PVS, but even the PE ranges to your
lvcreate commands...
Lars
--
: Lars Ellenberg
: LINBIT | Your Way to High Availability
: DRBD/HA support and consulting http://www.linbit.com
next prev parent reply other threads:[~2012-02-20 21:59 UTC|newest]
Thread overview: 7+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-02-16 13:50 [linux-lvm] Allocation Policy for Cloud Computing needed Sebastian Riemer
2012-02-16 14:41 ` James Hawtin
2012-02-16 15:27 ` Sebastian Riemer
2012-02-20 21:59 ` Lars Ellenberg [this message]
2012-02-20 22:41 ` Ray Morris
2012-02-21 9:48 ` Sebastian Riemer
2012-02-21 9:34 ` Sebastian Riemer
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20120220215911.GJ6331@barkeeper1-xen.linbit \
--to=lars.ellenberg@linbit.com \
--cc=linux-lvm@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).