[linux-lvm] combined linear + striped
gpowiedziuk at gmail.com
Mon Sep 30 23:14:34 UTC 2013
> -----Original Message-----
> From: linux-lvm-bounces at redhat.com [mailto:linux-lvm-
> bounces at redhat.com] On Behalf Of matthew patton
> Sent: Monday, September 30, 2013 2:54 AM
> To: LVM general discussion and development
> Subject: Re: [linux-lvm] combined linear + striped
> >I have 400 disks over here. All these disks live on 4 separate raids 5
> >(ibm ds6000). So stripping across disks within a raid doesn't make any
> you don't give the SAN administrators of the US Census Bureau enough
> credit. ;-)
> >But, what I'd like to do is to stripe it across separate raids.
> >This is how I see it
> >Raid a - disks 1001-1100
> >Raid b - disks 2001-2100
> >Raid c - disks 3001-3100
> >Raid d - disks 4001-3100
> >Stripping would be across every set of 4 disks: 1001,2001,3001,4001 +
> 1002,2002,3002,4002 ... and so on.
> >And all stripped sets would be combined in to one big LV.
> >Is it even possible?
> Of course. but I still wouldn't do it en-mass like that. Unless (and even
> you have exactly one workload defined that you want all 400 disks to
> participate in, you'll be FAR better off assigning different workloads to
> different spindle sets (ie. raids). I would go farther and recommend those
> raid sets be broken into much smaller sets of spindles, say 8+P at most.
> Unless you've been handed 100 x 4 "disks" because your SAN admin thought
> it was a brilliant idea to cap the size of each LUN at some silly number
> 25GB (yes, EMC used to advocate for such foolishness) and thus they
> represent increasing offsets on the same spindle set.
Well, I am also a SAN admin here but I have a good explanation :)
It is a DS6000 + IBM Mainframe and we are limited to 3390 CKD disks which
are pretty small.
But it is not a big deal. Most of our data live on FCP disks (different
storage system) where size is not an issue.
This 'old guy' (DS6000) is going to be used for backups only. So one big
LVM makes things much easier.
> All that said, lets assume you assigned all 400 LUNs to a single Linux LVM
> Volume Group. Then when you LVCREATE just specify the list of PV (or
> extents therein) you want to use and the order. So,
> # lvcreate -L ... -n ... vgname pv1001 pv2001 pv3001 pv4001 pv1002
Perfect! That is exactly what I wanted. After doing it this way, performance
seems to be about 3.5 times better!
And I can see when I write data to this device, that at the beginning only
1001, 2001, 3001, 4001 are being used and after certain number of gigabytes
it switches to 1002,2002,3002,4002 and so on. Awesome!
> Frankly I'd also look at using the LVcreate Interleave directives to
> bit more effectively.
I am already happy here - 3,5x boost from one simple idea and few hours of
work made my day.
> >I know that I can just create a stripped volume across all 400 volumes
> >but it doesn't really make sens in this set up does it?
> Probably not unless you're doing just massive streaming reads or writes. I
> think your first stop should be the SAN admin's cube. Bring a nice, maple
> bat and practice your follow-through. Perfect form is important after
As I said, it is 'just' for backups. It's of course nice when it is fast as
much it is possible but currently, with the new striped LVM setup, the
source won't be able to push the data faster than we can write it here.
Maple clue bat you say... I am glad I do both here - sys admin and san admin
... but there are developers out there so it still may happen ;)
> linux-lvm mailing list
> linux-lvm at redhat.com
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
More information about the linux-lvm