[Linux-cluster] fencing problem in 2 node cluster using apc fence device

garylua at singnet.com.sg garylua at singnet.com.sg
Thu Aug 17 05:28:46 UTC 2006


Hi, another related has been puzzling me since I'm new to redhat clustering. I understand that when configuring apc fencing device for each node, I need to fill in the port and switch of the apc device. Port refers to the power outlet which my node is connected to, right? For my case, my node 1 is connected to Outlet 13 of the first pdu. Do I fill in Port=13 or Port=Outlet 13? As for switch, I have absolutely no idea what it refers to. Does it refer to the state I want the outlet to go to during fencing, like 1=Immediate ON, 2=Immediate OFF, 3=Immediate Reboot, etc?

Thanks.

Gary


--- garylua at singnet.com.sg wrote:

> Hi Lon, thanks for the reply. My cluster.conf is as follows.
> coral1 and coral2 are my 2 nodes.
> 
> 
> <?xml version="1.0"?>
> <cluster config_version="264" name="MF_Cluster">
> 	<fence_daemon clean_start="1" post_fail_delay="0"
> post_join_delay="3"/>
> 	<clusternodes>
> 		<clusternode name="coral1" votes="1">
> 			<fence>
> 				<method name="1">
> 					<device name="pdu1" option="off" port="13" switch="1"/>
> 					<device name="pdu2" option="off" port="13" switch="2"/>
> 					<device name="pdu1" option="on" port="13" switch="1"/>
> 					<device name="pdu2" option="on" port="13" switch="2"/>
> 				</method>
> 			</fence>
> 		</clusternode>
> 		<clusternode name="coral2" votes="1">
> 			<fence>
> 				<method name="1">
> 					<device name="pdu1" option="off" port="20" switch="1"/>
> 					<device name="pdu2" option="off" port="20" switch="2"/>
> 					<device name="pdu1" option="on" port="20" switch="1"/>
> 					<device name="pdu2" option="on" port="20" switch="2"/>
> 				</method>
> 			</fence>
> 		</clusternode>
> 	</clusternodes>
> 	<fencedevices>
> 		<fencedevice agent="fence_apc" ipaddr="10.10.50.100" login="apc"
> name="pdu1" passwd="apc"/>
> 		<fencedevice agent="fence_apc" ipaddr="10.10.50.101" login="apc"
> name="pdu2" passwd="apc"/>
> 	</fencedevices>
> 	<rm>
> 		<failoverdomains>
> 			<failoverdomain name="MF_Failover" ordered="0" restricted="1">
> 				<failoverdomainnode name="coral2" priority="1"/>
> 				<failoverdomainnode name="coral1" priority="1"/>
> 			</failoverdomain>
> 		</failoverdomains>
> 		<resources>
> 			<fs device="/dev/sda1" force_fsck="0" force_unmount="1"
> fstype="ext3" mountpoint="/MF/MF_v1.1/shared" name="testmount"
> options="" self_fence="0"/>
> 			<script file="/etc/rc.d/init.d/msgfwd" name="Message
> Forwarder"/>
> 			<ip address="10.10.50.22" monitor_link="1"/>
> 			<script file="/etc/rc.d/init.d/namesvc" name="Name Service"/>
> 		</resources>
> 		<service autostart="1" domain="MF_Failover" name="msgfwd"
> recovery="relocate">
> 			<fs ref="testmount"/>
> 			<script ref="Message Forwarder"/>
> 			<ip ref="10.10.50.22"/>
> 		</service>
> 	</rm>
> 	<cman expected_votes="1" two_node="1"/>
> </cluster>
> 
> 
> 
> 
> --- Lon Hohberger <lhh at redhat.com> wrote:
> 
> > On Wed, 2006-08-16 at 22:37 +0800, Gary Lua wrote:
> > > Hi,
> > > 
> > > I'm currently configuring fencing devices for my 2 nodes on a
> > RHEL4 
> > > cluster. The problem is quite long, so please bear with me.
> > > 
> > > I have 2 nodes (let's call them stone1 and stone2) and 2 APC
> > fencing 
> > > devices (pdu1 and pdu2, both apc 7952 devices). Both stone1 and
> > stone2 
> > > has dual power supplies. Stone1's power supplies are connected
> to
> > outlet 
> > > 13 of pdu1 and pdu2. Stone2's power supplies are connected to
> > outlet 20 
> > > of both the pdus. My question is: during the fencing
> configuration
> > for 
> > > each node, i need to specify which fence device to add to the
> > fence 
> > > level of each node. Is it correct to specify for stone1 as
> follows
> > : 
> > > pdu1 -> port=13, switch=1, pdu2-> port=13, switch=2? The same
> > applies to 
> > > stone 2 : pdu1-> port=20, switch=1, pdu2-> port=20, switch=2?
> > > 
> > > After configuring as mentioned above, with both nodes on the
> > cluster 
> > > running and my application running on stone1, i pull out the
> > ethernet 
> > > cables for stone1 to simulate that the server is down. By
> right,
> > my 
> > > application should fail over to stone2 and fencing should occur
> to
> > 
> > > stone1 (ie, stone1 should be rebooted/shutdown). However, what
> > happened 
> > > is that my application is started on stone2, and stone1 is not
> > fenced. 
> > > In fact, when i reconnect by cables, my application is still
> > running on 
> > > stone1! Seems that there are 2 instances of my application
> > running, each 
> > > on stone1 and stone2.
> > 
> > Post the cluster configuration.
> > 
> > -- Lon
> > 
> > 
> > --
> > Linux-cluster mailing list
> > Linux-cluster at redhat.com
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> > 
> 
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
> 




More information about the Linux-cluster mailing list