[Linux-cluster] Fencing issue using IPMI (nodes fencing each other ending in a loop)

Stevan Colaco stevan.colaco at gmail.com
Sat Sep 27 08:44:16 UTC 2008


Hello All,

To test i have moved network connectivity from cisco swictes to 5 port
DLINK swicth.
Cluster is workig with fencing properly.

Best Regards,
-Stevan Colaco


2008/9/24 Grisha G. <grigorygor at gmail.com>:
> In 2 node cluster you should use a quorum disk to solve the split brain
> problem.
> after you create a quorum disk change this line in you cluster.conf
> from <cman expected_votes="1" two_node="1"/>
> to  <cman expected_votes="3" two_node="0"/>
>
> Grisha
>
>
> On Tue, Sep 23, 2008 at 7:27 PM, Stevan Colaco <stevan.colaco at gmail.com>
> wrote:
>>
>> Hello
>>
>> issue: Fencing using fence_ipmilan, each node keeps fencing the other
>> node ending in a fence loop.....
>>
>> We have implemented RH Cluster on RHEL5.2 64bit.
>> Server Hardware: SUN X4150
>> Storage: SUN 6140
>> Fencing Machnism: fence_ipmilan
>>
>>  We have downloaded the IPMI fence_ipmilan and configured two node
>> cluster with ipmi fencing. But..
>>
>> when we ifdown the NIC interface, the node gets fenced but the service
>> does not relocate to the other node. at the same time when the
>> initially fenced node joins back the cluster it fences the other
>> node......
>> this keeps on ending in a loop.
>>
>> We downloaded and followed the intructions from the ipmi site
>> mentioned below
>> http://docs.sun.com/source/819-6588-13/ipmi_com.html#0_74891
>>
>> we tested with following  Cmd line method which works fine.
>> #fence_ipmilan -a "ip addr" -l root -p <Passkey> -o <on|off|reboot>
>>
>> here is my cluster.conf
>>
>> <?xml version="1.0"?>
>> <cluster alias="tibcouat" config_version="12" name="tibcouat">
>>        <fence_daemon clean_start="0" post_fail_delay="0"
>> post_join_delay="3"/>
>>        <clusternodes>
>>                <clusternode name="tibco-node1-uat.kmefic.com.kw"
>> nodeid="1" votes="1">
>>                        <fence>
>>                                <method name="1">
>>                                        <device name="tibco-node1"/>
>>                                </method>
>>                        </fence>
>>                </clusternode>
>>                <clusternode name="tibco-node2-uat.kmefic.com.kw"
>> nodeid="2" votes="1">
>>                        <fence>
>>                                <method name="1">
>>                                        <device name="tibco-node2"/>
>>                                </method>
>>                        </fence>
>>                </clusternode>
>>        </clusternodes>
>>        <cman expected_votes="1" two_node="1"/>
>>        <fencedevices>
>>                <fencedevice agent="fence_ipmilan" ipaddr="172.16.71.41"
>> login="root" name="tibco-node1" passwd="changeme"/>
>>                <fencedevice agent="fence_ipmilan" ipaddr="172.16.71.42"
>> login="root" name="tibco-node2" passwd="changeme"/>
>>        </fencedevices>
>>        <rm>
>>                <failoverdomains>
>>                        <failoverdomain name="prefer_node1" nofailback="0"
>> ordered="1"
>> restricted="1">
>>                                <failoverdomainnode
>> name="tibco-node1-uat.kmefic.com.kw" priority="1"/>
>>                                <failoverdomainnode
>> name="tibco-node2-uat.kmefic.com.kw" priority="2"/>
>>                        </failoverdomain>
>>                </failoverdomains>
>>                <resources>
>>                        <ip address="172.16.71.55" monitor_link="1"/>
>>                        <clusterfs device="/dev/vg0/gfsdata"
>> force_unmount="0" fsid="63282"
>> fstype="gfs" mountpoint="/var/www/html" name="gfsdata"
>> self_fence="0"/>
>>                        <apache config_file="conf/httpd.conf"
>> name="docroot"
>> server_root="/etc/httpd" shutdown_wait="0"/>
>>                </resources>
>>                <service autostart="1" domain="prefer_node1" exclusive="0"
>> name="webby" recovery="relocate">
>>                        <ip ref="172.16.71.55"/>
>>                </service>
>>        </rm>
>> </cluster>
>>
>>
>> Kindly investigate and provide us the solution at the earliest.
>>
>> Thanks & Best Regards,
>>
>> --
>> Linux-cluster mailing list
>> Linux-cluster at redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>




More information about the Linux-cluster mailing list