[linux-lvm] LVM in shared parallel SCSI environment

Matthew O'Keefe okeefe at brule.borg.umn.edu
Wed Nov 22 00:42:32 UTC 2000


Hi,

On Tue, Nov 21, 2000 at 11:25:15PM +0100, Jos Visser wrote:
> Hi,
> 
> Are the plans public? Are comments invited?

Heinz and his team is working on a draft for this and will post it 
"soon":  I'll let Heinz define "soon" :-)  

Of course comments are welcome.  I think we are talking about
1.0 being released in Q1 2000, but again, Heinz and others 
should make that prediction.
Regards,
Matt

Matthew O'Keefe
Sistina Software, Inc. 
> 
> ++Jos
> 
> And thus it came to pass that Matthew O'Keefe wrote:
> (on Tue, Nov 21, 2000 at 07:44:52AM -0600 to be exact)
> 
> > 
> > Hi,
> > 
> > Heinz and his LVM team (we've hired two new LVM developers)
> > as well as the GFS team have worked
> > out a preliminary design for cluster LVM.  The plan is too
> > include it in the 1.0 release.
> > 
> > I totally agree with Jos:  a cluster volume manager is very
> > useful, and should stand alone (but also be compatible with)
> > a cluster file system like GFS.  There is a tremendous amount
> > of commercial activity in the area of volume management
> > software for shared SAN storage.  Imagine you have 2 
> > $3 million dollar EMC symmetrix disk arrays, each attached
> > to independent servers.  If one of these symmetrix fills up,
> > you have to buy another for just that server alone, even if
> > the other server's symmetrix has lots of free space.
> > 
> > If instead you share these 2 symmetrix boxen across a san,
> > then you can expand the PV for one machine into the other
> > the symmetrix with free space, and there is no need to buy
> > another array.  This is a key reason why shared SAN storage is
> > taking off.
> > 
> > 
> > 
> > Matt O'Keefe
> > Sistina Software, Inc.
> > 
> > On Wed, Nov 15, 2000 at 08:04:14AM +0100, Jos Visser wrote:
> > > Hi,
> > > 
> > > Though most has already been said in this thread, just a small followup
> > > with some notes and thoughts.
> > > 
> > > The traditional volume managers on HP-UX, Solaris (VxVM) and AIX do not
> > > usually support shared access to a volume group from two or more nodes,
> > > even if the nodes access different logical volumes. This is done
> > > explicitly to prevent the kind of problems that have been pointed out in
> > > this thread (the chance that two nodes have different in-core metadata
> > > about the VG). HP's LVM supports a read-only vgchange that allows only
> > > read-only access to the VG and its LV's, but I've never used it.
> > > 
> > > In these traditional environment, the clustering software exports and
> > > imports the VG's as necessary, and run some clusterwide resource manager
> > > that takes care of who currently "owns" the VG. Veritas has a special
> > > Cluster Volume Manager (CVM) that allows shared access to volume groups,
> > > but AFAIK it is only used with parallel databases such as Oracle
> > > Parallel Server.
> > > 
> > > For myself, I would not choose a solution like Jesse's. However, the fun
> > > and power of Unix is that everyone can handcraft his/her own optimal
> > > environment. As long as you're aware of the consequences what you're
> > > doing: please be my guest :-)
> > > 
> > > I must admit that I have not looked at what LVM 0.9 will bring to the
> > > table, but some added features in the clustering arena would be very
> > > welcome.
> > > 
> > > ++Jos
> > > 
> > > And thus it came to pass that Jesse Sipprell wrote:
> > > (on Tue, Nov 14, 2000 at 02:29:02PM -0500 to be exact)
> > > 
> > > > On Tue, Nov 14, 2000 at 04:09:47PM +0000, Paul Jakma wrote:
> > > > > On Tue, 14 Nov 2000, Jesse Sipprell wrote:
> > > > > 
> > > > > > In the mean time, I'll just have to do things the old fashioned
> > > > > > way.  I'll put a procedure in place that any LVM changes done from
> > > > > > a particular node require the bouncing of VGs on all other
> > > > > > attached nodes.  Fortunately, after initial cluster setup,
> > > > > > manipulation of LVs won't really be performed on a routine basis.
> > > > > 
> > > > > and so what do you do with these LV's? The filesystem/application you
> > > > > run on them has to be aware of the shared-access nature of the
> > > > > device.. so that rules out all but GFS - which IIRC already has some
> > > > > LVM like features.
> > > > 
> > > > Actually, it's entirely possible to run a non-shared-media-aware filesystem as
> > > > long as no more than one cluster node has a given file system mounted at a
> > > > time.
> > > > 
> > > > To illustrate:
> > > > 
> > > > |-------- VG --------|
> > > > ||====== LV0 =======||
> > > > || (ext2)           || --> Mounted on Cluster Node 1
> > > > ||==================||
> > > > ||====== LV1 =======||
> > > > || (ext2)           || --> Mounted on Cluster Node 2
> > > > ||==================||
> > > > ||====== LV2 =======||
> > > > || (ext2)           || --> Mounted on Cluster Node 3
> > > > ||==================||
> > > > ||====== LV3 =======||
> > > > || (ext2)           || --> Mounted on Cluster Node 4
> > > > ||==================||
> > > > |                    |
> > > > |  Free Space in VG  |
> > > > |                    |
> > > > |====================|
> > > > 
> > > > Because none of the cluster nodes are attempting to share access to the actual
> > > > blocks where each filesystem is stored, there are no concurrency issues.
> > > > 
> > > > One can use the benefits of LVM to unmount LV0's fs on Cluster Node 1, resize
> > > > the LV, resize the fs and remount.  Now, Cluster Node's 2, 3 and 4 need to
> > > > have their in-core LVM metadata updated in order to see the new size of LV0.
> > > > Once this is done via the vgchange bounce, everything is consistant.
> > > > 
> > > > -- 
> > > > Jesse Sipprell
> > > > Technical Operations Director
> > > > Evolution Communications, Inc.
> > > > 800.496.4736
> > > > 
> > > > * Finger jss at evcom.net for my PGP Public Key *
> > > 
> > > -- 
> > > Success and happiness can not be pursued; it must ensue as the 
> > > unintended side-effect of one's personal dedication to a course greater 
> > > than oneself.
> 
> -- 
> Success and happiness can not be pursued; it must ensue as the 
> unintended side-effect of one's personal dedication to a course greater 
> than oneself.



More information about the linux-lvm mailing list