[Linux-cluster] Few queries about fence working
jayesh.shinde
jayesh.shinde at netcore.co.in
Wed Jan 25 09:38:45 UTC 2012
Dear Emmanuel Segura,
Find the config below. Because of policy I have removed some login details.
#############
<?xml version="1.0"?>
<cluster config_version="6" name="new_cluster">
<fence_daemon post_fail_delay="0" post_join_delay="3"/>
<clusternodes>
<clusternode name="mailbox1" nodeid="1" votes="1">
<multicast addr="224.0.0.1" interface="bond0"/>
<fence>
<method name="1">
<device name="imap1drac"/>
</method>
</fence>
</clusternode>
<clusternode name="mailbox2" nodeid="2" votes="1">
<multicast addr="224.0.0.1" interface="bond0"/>
<fence>
<method name="1">
<device name="imap2drac"/>
</method>
</fence>
</clusternode>
</clusternodes>
<cman expected_votes="1" two_node="1">
<multicast addr="224.0.0.1"/>
</cman>
<fencedevices>
<fencedevice agent="fence_drac6" ipaddr="<drac IP>" login="<login name>"
name="imap1drac" passwd="xxxxx"/>
<fencedevice agent="fence_drac6" ipaddr="<drac IP>" login="<login name>"
name="imap2drac" passwd="xxxxx"/>
</fencedevices>
<rm>
<failoverdomains/>
<resources>
<ip address="192.168.1.1" monitor_link="1"/>
<fs device="/dev/drbd0" force_fsck="0" force_unmount="1" fsid="28418"
fstype="ext3" mountpoint="/mount/path" name="imap1_fs" options="rw"
self_fence="1"/>
<script file="/etc/init.d/cyrus-imapd" name="imap1_init"/>
</resources>
<service autostart="1" name="imap1" recovery="restart">
<ip ref="192.168.1.1"/>
<fs ref="imap1_fs"/>
<script ref="imap1_init"/>
</service>
</rm>
</cluster>
###################
Regards
Jayesh Shinde
On 01/25/2012 01:59 PM, emmanuel segura wrote:
> Can you show me your cluster config?
>
> 2012/1/25 jayesh.shinde <jayesh.shinde at netcore.co.in
> <mailto:jayesh.shinde at netcore.co.in>>
>
> Hi all ,
>
> I have few queries about fence working.
>
> I am using 2 different the 2 node cluster with Dell and IBM
> hardware in two different IDC.
> Recently I came across the network failure problem at different
> time and I found my 2 nodes are power off state.
>
> Below is how the situation happened with my 2 different 2 node
> cluster.
>
> With 2 node IBM node cluster with SAN :--
> ==============================
> 1) Network connectivity was failed totally for few minutes.
> 2) And as per the /var/log/messages both servers failed to fence
> to each other and both server was UP as it is with all services.
> 3) But the "clustat" was showing serves are not in cluster mode
> and "regmanger" status was stop.
> 4) I simply reboot the server.
> 5) After that I found both server in power off stat.
>
>
> with another 2 node Dell server with DRBD :--
> =================================
> 1) Network connectivity was failed totally.
> 2) DRAC ip was unavailable so fence failed from both server.
> 3) after some time I fond the servers are shutdown.
>
> In normal conditions both cluster work properly
>
> my queries are now :--
> ===============
> 1) What could be the reason for power off ?
> 2) Does cluster's fencing method caused for the power off of
> server ( i.e because of previous failed fence ) ?
> 3) Is there any test cases mentioned on net / blog / wiki about
> the fence , i.e different situation under which fence works.
>
> Please guide.
>
> Thanks & Regards
> Jayesh Shinde
>
>
>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
> https://www.redhat.com/mailman/listinfo/linux-cluster
>
>
>
>
> --
> esta es mi vida e me la vivo hasta que dios quiera
>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20120125/e5f5f267/attachment.htm>
More information about the Linux-cluster
mailing list