[Linux-cluster] Quorum device brain the cluster when master lose network

GouNiNi gounini.geekarea at gmail.com
Wed Aug 1 08:28:07 UTC 2012


I do this test one more time and I got same result with more precisions:

When I shutdown network on 2 nodes including the master, master stay alive while the 2 online nodes are fencing the offline non-master node. The cluster goes Inquorate after.
When fenced node came back, he joins cluster and cluster becomes quorate. New master is chose and the old master is fenced.

# cman_tool status
Version: 6.2.0
Config Version: 144
Cluster Name: cluname
Cluster Id: 57462
Cluster Member: Yes
Cluster Generation: 488
Membership state: Cluster-Member
Nodes: 4
Expected votes: 5
Quorum device votes: 1
Total votes: 5
Quorum: 3
Active subsystems: 9
Flags: Dirty
Ports Bound: 0 177
Node name: nodename
Node ID: 2
Multicast addresses: ZZ.ZZ.ZZ.ZZ
Node addresses: YY.YY.YY.YY

-- 
  .`'`.   GouNiNi
 :  ': :  
 `. ` .`  GNU/Linux
   `'`    http://www.geekarea.fr


----- Mail original -----
> De: "emmanuel segura" <emi2fast at gmail.com>
> À: "linux clustering" <linux-cluster at redhat.com>
> Envoyé: Lundi 30 Juillet 2012 17:35:39
> Objet: Re: [Linux-cluster] Quorum device brain the cluster when master	lose network
> 
> 
> can you send me the ouput from cman_tool status? when the cluster
> it's running
> 
> 
> 2012/7/30 GouNiNi < gounini.geekarea at gmail.com >
> 
> 
> 
> 
> ----- Mail original -----
> > De: "Digimer" < lists at alteeve.ca >
> > À: "linux clustering" < linux-cluster at redhat.com >
> > Cc: "GouNiNi" < gounini.geekarea at gmail.com >
> > Envoyé: Lundi 30 Juillet 2012 17:10:10
> > Objet: Re: [Linux-cluster] Quorum device brain the cluster when
> > master lose network
> > 
> > On 07/30/2012 10:43 AM, GouNiNi wrote:
> > > Hello,
> > > 
> > > I did some tests on 4 nodes cluster with quorum device and I find
> > > a
> > > bad situation with one test, so I need your knowledges to correct
> > > my configuration.
> > > 
> > > Configuation:
> > > 4 nodes, all vote for 1
> > > quorum device vote for 1 (to hold services with minimum 2 nodes
> > > up)
> > > cman expected votes 5
> > > 
> > > Situation:
> > > I shut down network on 2 nodes, one of them is master.
> > > 
> > > Observation:
> > > Fencing of one node (the master)... Quorum device Offline, Quorum
> > > disolved ! Services stopped.
> > > Fenced node reboot, cluster is quorate, 2nd offline node is
> > > fenced.
> > > Services restart.
> > > 2nd node offline reboot.
> > > 
> > > My cluster is not quorate for 8 min (very long hardware boot :-)
> > > and my services were offline.
> > > 
> > > Do you know how to prevent this situation?
> > > 
> > > Regards,
> > 
> > Please tell us the name and version of the cluster software you are
> > using, Please also share your configuration file(s).
> > 
> > --
> > Digimer
> > Papers and Projects: https://alteeve.com
> > 
> 
> Sorry, RHEL5.6 64bits
> 
> # rpm -q cman rgmanager
> cman-2.0.115-68.el5
> rgmanager-2.0.52-9.el5
> 
> 
> <?xml version="1.0"?>
> <cluster alias="cluname" config_version="144" name="cluname">
> <clusternodes>
> <clusternode name="node1" nodeid="1" votes="1">
> <fence>
> <method name="single">
> <device name="fenceIBM_307" port="12"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="node2" nodeid="2" votes="1">
> <fence>
> <method name="single">
> <device name="fenceIBM_307" port="11"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="node3" nodeid="3" votes="1">
> <fence>
> <method name="single">
> <device name="fenceIBM_308" port="6"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="node4" nodeid="4" votes="1">
> <fence>
> <method name="single">
> <device name="fenceIBM_308" port="7"/>
> </method>
> </fence>
> </clusternode>
> </clusternodes>
> <fencedevices>
> <fencedevice agent="fence_bladecenter" ipaddr="XX.XX.XX.XX"
> login="xxxx" name="fenceIBM_307" passwd="yyyy"/>
> <fencedevice agent="fence_bladecenter" ipaddr="YY.YY.YY.YY"
> login="xxxx" name="fenceIBM_308" passwd="yyyy"/>
> </fencedevices>
> <rm log_level="7">
> <failoverdomains/>
> <resources/>
> <service ...>
> <...>
> </service>
> </rm>
> <fence_daemon clean_start="0" post_fail_delay="15"
> post_join_delay="300"/>
> <cman expected_votes="5">
> <multicast addr="ZZ.ZZ.ZZ.ZZ"/>
> </cman>
> <quorumd interval="7" label="quorum" tko="12" votes="1"/>
> </cluster>
> 
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
> 
> 
> --
> esta es mi vida e me la vivo hasta que dios quiera
> 
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster




More information about the Linux-cluster mailing list