[Linux-cluster] Distributed Storage in Cluster Suite
Gordan Bobic
gordan at bobich.net
Tue Feb 23 20:40:57 UTC 2010
Joseph L. Casale wrote:
> Hi Gordan,
> I'm under the gun to get something operational asap and have been reading like
> mad but given the severity of getting it wrong, I want some experience under my
> belt before I step off the reservation. I don't have the infrastructure available
> to export an iscsi device, it's the backend I need redundant just the way drbd
> would provide.
You can do that with iSCSI. Make a sparse file to represent the exported
storage and export that via iSCSI. Do the same on both nodes. Then from
both nodes connect to it, get CLVM on top and get it to mirror the two
exported volumes. Put GFS on top. Nowhere nearly as clean and nice as
DRBD, though, especially cometh time to recover from failure.
> I am ok with drbd, but its integration with rhcs concerns me as I don't really
> know how the init scripts get constructed. I saw an archive post about the wrapper
> for /usr/share/cluster/drbd.sh to hand it the resource name.
>
> I can assume if a node tanks, drbd goes primary on the remaining node, everything
> is good.
Or you can go active-active with GFS. :)
> Looking through drbd.sh I see case's for promote/demote, I added this (the wrapper actually)
> as a resource "script" with a child "fs" in a mock cluster as the underlying device
> was primary on that node, it just started:) How does rhcs know when adding a "script"
> to pass the required cases? How do I configure this?
I use it in active-active mode with GFS. In that case I just use the
fencing agent in DRBD's "stonith" configuration so that when
disconnection occurs, the failed node gets fenced.
Gordan
More information about the Linux-cluster
mailing list