[Linux-cluster] Problems configuring fence devices

carlopmart carlopmart at gmail.com
Wed Dec 17 13:29:03 UTC 2008


Correct, but if i use "expected_votes=1" under cman section, errors doesn't 
disappears ...

Alexandre Gondim dos Santos wrote:
> Your cluster must have more then 50% votes to be quorate.
> 
> Alexandre Gondim dos Santos
> 
> carlopmart escreveu:
>> Hi all,
>>
>>  I have setup three nodes with redhat cluster suite using rhel5.2. One 
>> of this three nodes acts as a iscsi server an provides fence_gnbd to 
>> the other ones. In this node I have configured fence_manual because is 
>> the first node to start-up. But every time returns this errors until 
>> node1 or node2 or both starts ...:
>>
>>
>> Dec 17 12:42:21 node3 dlm_controld[2568]: connect to ccs error -111, 
>> check ccsd or cluster status
>> Dec 17 12:42:21 node3 ccsd[2538]: Cluster is not quorate.  Refusing 
>> connection.
>> Dec 17 12:42:21 node3 ccsd[2538]: Error while processing connect: 
>> Connection refused
>> Dec 17 12:42:21 node3 ccsd[2538]: Cluster is not quorate.  Refusing 
>> connection.
>> Dec 17 12:42:21 node3 ccsd[2538]: Error while processing connect: 
>> Connection refused
>> Dec 17 12:42:22 node3 ccsd[2538]: Cluster is not quorate.  Refusing 
>> connection.
>>
>>
>>  My cluster.conf is:
>>
>> <?xml version="1.0"?>
>> <cluster alias="RhelHomeCluster" config_version="4" 
>> name="RhelHomeCluster">
>>         <fence_daemon post_fail_delay="0" post_join_delay="3"/>
>>         <clusternodes>
>>                 <clusternode name="node1" nodeid="2" votes="1">
>>                         <fence>
>>                                 <method name="1">
>>                                         <device name="gnbd-fence01" 
>> nodename="node1"/>
>>                                 </method>
>>                         </fence>
>>                         <multicast addr="239.192.75.55" 
>> interface="eth0"/>
>>                 </clusternode>
>>                 <clusternode name="node2" nodeid="3" votes="1">
>>                         <fence>
>>                                 <method name="1">
>>                                         <device name="gnbd-fence01" 
>> nodename="node2"/>
>>                                 </method>
>>                         </fence>
>>                         <multicast addr="239.192.75.55" 
>> interface="eth0"/>
>>                 </clusternode>
>>                 <clusternode name="node3" nodeid="1" votes="1">
>>                         <fence>
>>                                 <method name="1">
>>                                         <device name="last_resort" 
>> nodename="node3"/>
>>                                 </method>
>>                         </fence>
>>                         <multicast addr="239.192.75.55" 
>> interface="eth1"/>
>>                 </clusternode>
>>         </clusternodes>
>>         <cman>
>>                 <multicast addr="239.192.75.55"/>
>>         </cman>
>>         <fencedevices>
>>                 <fencedevice agent="fence_gnbd" name="gnbd-fence01" 
>> servers="node3"/>
>>                 <fencedevice agent="fence_manual" name="last_resort"/>
>>         </fencedevices>
>>         <rm log_facility="local4" log_level="7">
>>                 <failoverdomains>
>>                         <failoverdomain name="PriCluster" ordered="1" 
>> restricted="1">
>>                                 <failoverdomainnode name="node1" 
>> priority="1"/>
>>                                 <failoverdomainnode name="node2" 
>> priority="2"/>
>>                         </failoverdomain>
>>                         <failoverdomain name="SecCluster" ordered="1" 
>> restricted="1">
>>                                 <failoverdomainnode name="node2" 
>> priority="1"/>
>>                                 <failoverdomainnode name="node1" 
>> priority="2"/>
>>                         </failoverdomain>
>>                 </failoverdomains>
>>                 <resources>
>>                         <ip address="172.25.50.18" monitor_link="1"/>
>>                 </resources>
>>                 <service autostart="1" domain="SecCluster" 
>> name="proxy-svc" recovery="relocate">
>>                         <ip ref="172.25.50.18">
>>                                 <script 
>> file="/data/configs/etc/init.d/squid" name="squid"/>
>>                         </ip>
>>                 </service>
>>         </rm>
>> </cluster>
> 
> -- 



-- 
CL Martinez
carlopmart {at} gmail {d0t} com




More information about the Linux-cluster mailing list