[Linux-cluster] DR node in a cluster

Paras pradhan pradhanparas at gmail.com
Wed Jul 6 17:16:39 UTC 2011


Chris,

All the nodes are connected to a single SAN at this moment through fibre.


@steven:

--
 If you don't have enough nodes at a site to allow quorum to be
established, then when communication fails between sites you must fence
those nodes or risk data corruption when communication is
re-established,
-----

Yes true, but in this case a single node can made the cluster quorate.
(qdisk vote=3 ,node votes=3, total=6) which is not recommened I guess (?).


Steve

On Wed, Jul 6, 2011 at 11:46 AM, Jankowski, Chris <Chris.Jankowski at hp.com>wrote:

> Paras,****
>
> ** **
>
> A curiosity question:****
>
> ** **
>
> How do you make sure that your storage will survive failure of **either**
> of your site without loss of data and continuity of service?****
>
> What storage configuration are you using?****
>
> ** **
>
> Thanks and regards,****
>
>
> Chris****
>
> ** **
>
> *From:* linux-cluster-bounces at redhat.com [mailto:
> linux-cluster-bounces at redhat.com] *On Behalf Of *Paras pradhan
> *Sent:* Thursday, 7 July 2011 02:15
> *To:* linux clustering
> *Subject:* [Linux-cluster] DR node in a cluster****
>
> ** **
>
> Hi,****
>
> ** **
>
> My GFS2 linux cluster has three nodes. Two at the data center and one at
> the DR site. If the nodes at DR site break/turnoff, all the services move to
> DR node. But if the 2 nodes at the data center lost communication with the
> DR node, I am not sure how does the cluster handles the split brain. So I am
> looking for some recommendation in this kind of scenario. I am usig Qdisk
> votes (=3) in this case.****
>
> ** **
>
> --****
>
> Here is the cman_tool status output.****
>
> ** **
>
> ** **
>
> -****
>
> Version: 6.2.0****
>
> Config Version: 74****
>
> Cluster Name: vrprd****
>
> Cluster Id: 3304****
>
> Cluster Member: Yes****
>
> Cluster Generation: 1720****
>
> Membership state: Cluster-Member****
>
> Nodes: 3****
>
> Expected votes: 6****
>
> Quorum device votes: 3****
>
> Total votes: 6****
>
> Quorum: 4  ****
>
> Active subsystems: 10****
>
> Flags: Dirty ****
>
> Ports Bound: 0 11 177  ****
>
> Node name: vrprd1.hostmy.com****
>
> Node ID: 2****
>
> Multicast addresses: x.x.x.244 ****
>
> Node addresses: x.x.x.96 ****
>
> --****
>
> ** **
>
> Thanks!****
>
> Paras.****
>
> ** **
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20110706/02a2fe08/attachment.htm>


More information about the Linux-cluster mailing list