[Linux-cluster] Few queries about fence working

jayesh.shinde jayesh.shinde at netcore.co.in
Wed Jan 25 09:38:45 UTC 2012


Dear  Emmanuel Segura,

Find the config below.  Because of policy I have removed some login details.

#############

<?xml version="1.0"?>
<cluster config_version="6" name="new_cluster">
<fence_daemon post_fail_delay="0" post_join_delay="3"/>
<clusternodes>
<clusternode name="mailbox1" nodeid="1" votes="1">
<multicast addr="224.0.0.1" interface="bond0"/>
<fence>
<method name="1">
<device name="imap1drac"/>
</method>
</fence>
</clusternode>
<clusternode name="mailbox2" nodeid="2" votes="1">
<multicast addr="224.0.0.1" interface="bond0"/>
<fence>
<method name="1">
<device name="imap2drac"/>
</method>
</fence>
</clusternode>
</clusternodes>
<cman expected_votes="1" two_node="1">
<multicast addr="224.0.0.1"/>
</cman>
<fencedevices>
<fencedevice agent="fence_drac6" ipaddr="<drac IP>" login="<login name>" 
name="imap1drac" passwd="xxxxx"/>
<fencedevice agent="fence_drac6" ipaddr="<drac IP>" login="<login name>" 
name="imap2drac" passwd="xxxxx"/>
</fencedevices>
<rm>
<failoverdomains/>
<resources>
<ip address="192.168.1.1" monitor_link="1"/>
<fs device="/dev/drbd0" force_fsck="0" force_unmount="1" fsid="28418" 
fstype="ext3" mountpoint="/mount/path" name="imap1_fs" options="rw" 
self_fence="1"/>
<script file="/etc/init.d/cyrus-imapd" name="imap1_init"/>
</resources>
<service autostart="1" name="imap1" recovery="restart">
<ip ref="192.168.1.1"/>
<fs ref="imap1_fs"/>
<script ref="imap1_init"/>
</service>
</rm>
</cluster>
###################

Regards
Jayesh Shinde


On 01/25/2012 01:59 PM, emmanuel segura wrote:
> Can you show me your cluster config?
>
> 2012/1/25 jayesh.shinde <jayesh.shinde at netcore.co.in 
> <mailto:jayesh.shinde at netcore.co.in>>
>
>     Hi  all ,
>
>     I have few queries about fence working.
>
>     I am using 2 different  the 2 node cluster with Dell and IBM
>     hardware in two different IDC.
>     Recently I came across the network failure problem at different
>     time and I found my 2 nodes are power off state.
>
>     Below is  how the situation happened with my 2 different 2 node
>     cluster.
>
>     With 2 node IBM node cluster with SAN :--
>     ==============================
>     1)  Network connectivity  was failed totally for few minutes.
>     2) And as per the /var/log/messages both servers failed to  fence
>     to each other and both server was UP as it is with all services.
>     3) But the "clustat" was showing serves are not in cluster mode
>     and "regmanger" status was stop.
>     4) I simply reboot the server.
>     5) After that I found both server in power off stat.
>
>
>     with another  2 node Dell server with DRBD  :--
>     =================================
>     1) Network connectivity  was failed totally.
>     2) DRAC ip was unavailable so fence failed from both server.
>     3) after some time I fond the servers are shutdown.
>
>     In normal conditions both cluster work properly
>
>      my queries are now :--
>      ===============
>     1) What could be the reason for power off ?
>     2) Does cluster's fencing method  caused for the power off  of
>     server ( i.e because of previous failed fence ) ?
>     3) Is there any test cases mentioned on net / blog / wiki  about
>     the fence , i.e different situation under which  fence works.
>
>     Please guide.
>
>     Thanks & Regards
>     Jayesh Shinde
>
>
>
>
>     --
>     Linux-cluster mailing list
>     Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
>     https://www.redhat.com/mailman/listinfo/linux-cluster
>
>
>
>
> -- 
> esta es mi vida e me la vivo hasta que dios quiera
>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20120125/e5f5f267/attachment.htm>


More information about the Linux-cluster mailing list