<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
<HTML><HEAD>
<META http-equiv=Content-Type content="text/html; charset=us-ascii">
<META content="MSHTML 6.00.2900.5945" name=GENERATOR></HEAD>
<BODY>
<DIV dir=ltr align=left><FONT face=Arial color=#0000ff 
size=2></FONT> </DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2>Hello,</FONT></SPAN></DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2></FONT></SPAN> </DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2>You might want to check the syslog to see if the cluster 
has noticed the outage and what is has tried to do about it.</FONT></SPAN></DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2>You can also check the node status via 'cman nodes' 
(explanaation of states in the cman manpage).</FONT></SPAN></DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2>Does the server have another power source, by any 
chance?</FONT></SPAN></DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2>  (if not make sure you DO have dual power supplies. 
These things die Often)</FONT></SPAN></DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2></FONT></SPAN> </DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2></FONT></SPAN> </DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2>Regards,</FONT></SPAN></DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2></FONT></SPAN> </DIV>
<DIV dir=ltr align=left><SPAN class=343012313-15052010><FONT face=Arial 
color=#0000ff size=2>Kit</FONT></SPAN></DIV><BR>
<DIV class=OutlookMessageHeader lang=en-us dir=ltr align=left>
<HR tabIndex=-1>
<FONT face=Tahoma size=2><B>From:</B> linux-cluster-bounces@redhat.com 
[mailto:linux-cluster-bounces@redhat.com] <B>On Behalf Of 
</B>Dusty<BR><B>Sent:</B> vrijdag 14 mei 2010 21:45<BR><B>To:</B> 
Linux-cluster@redhat.com<BR><B>Subject:</B> [Linux-cluster] pull plug on node, 
service never relocates<BR></FONT><BR></DIV>
<DIV></DIV>Greetings,<BR><BR>Using stock "clustering" and "cluster-storage" from 
RHEL5 update 4 X86_64 ISO.<BR><BR>As an example using my below config: 
<BR><BR>Node1 is running service1, node2 is running service2, etc, etc, node5 is 
spare and available for the relocation of any failover domain / cluster 
service.<BR><BR>If I go into the APC PDU and turn off the electrical port to 
node1, node2 will fence node1 (going into the APC PDU and doing and off, on on 
node1's port), this is fine. Works well. When node1 comes back up, then it shuts 
down service1 and service1 relocates to node5.<BR><BR>Now if I go in the lab and 
literally pull the plug on node5 running service1, another node fences node5 via 
the APC - can check the APC PDU log and see that it has done an off/on on 
node5's electrical port just fine.<BR><BR>But I pulled the plug on node5 - 
resetting the power doesn't matter. I want to simulate a completely dead node, 
and have the service relocate in this case of complete node failure.<BR><BR>In 
this RHEL5.4 cluster, the service never relocates. I can similate this on any 
node for any service. What if a node's motherboard fries? <BR><BR>What can I set 
to have the remaining nodes stop waiting for the reboot of a failed node and 
just go ahead and relocate the cluster service that had been running on the now 
failed node?<BR><BR>Thank 
you!<BR><BR>versions:<BR><BR>cman-2.0.115-1.el5<BR>openais-0.80.6-8.el5<BR>modcluster-0.12.1-2.el5<BR>lvm2-cluster-2.02.46-8.el5<BR>rgmanager-2.0.52-1.el5<BR>ricci-0.12.2-6.el5<BR><BR>cluster.conf 
(sanitized, real scripts removed, all gfs2 mounts gone for clarity):<BR><?xml 
version="1.0"?><BR><cluster config_version="1" 
name="alderaanDefenseShieldRebelAllianceCluster"><BR>    
<fence_daemon clean_start="0" post_fail_delay="3" 
post_join_delay="60"/><BR>    
<clusternodes><BR>        
<clusternode name="192.168.1.1" nodeid="1" 
votes="1"><BR>            
<fence><BR>                
<method 
name="1"><BR>                    
<device name="apc_pdu" port="1" 
switch="1"/><BR>                
</method><BR>            
</fence><BR>        
</clusternode><BR>        
<clusternode name="192.168.1.2" nodeid="2" 
votes="1"><BR>            
<fence><BR>                
<method 
name="1"><BR>                    
<device name="apc_pdu" port="2" 
switch="1"/><BR>                
</method><BR>            
</fence><BR>        
</clusternode><BR>        
<clusternode name="192.168.1.3" nodeid="3" 
votes="1"><BR>            
<fence><BR>                
<method 
name="1"><BR>                    
<device name="apc_pdu" port="3" 
switch="1"/><BR>                
</method><BR>            
</fence><BR>        
</clusternode><BR>        
<clusternode name="192.168.1.4" nodeid="4" 
votes="1"><BR>            
<fence><BR>                
<method 
name="1"><BR>                    
<device name="apc_pdu" port="4" 
switch="1"/><BR>                
</method><BR>            
</fence><BR>        
</clusternode><BR>        
<clusternode name="192.168.1.5" nodeid="5" 
votes="1"><BR>            
<fence><BR>                
<method 
name="1"><BR>                    
<device name="apc_pdu" port="5" 
switch="1"/><BR>                
</method><BR>            
</fence><BR>        
</clusternode><BR>    
</clusternodes><BR>    <cman 
expected_votes="6"/><BR>    
<fencedevices><BR>        
<fencedevice agent="fence_apc" ipaddr="192.168.1.20" login="device" 
name="apc_pdu" 
passwd="wonderwomanWasAPrettyCoolSuperhero"/><BR>    
</fencedevices><BR>    
<rm><BR>        
<failoverdomains><BR>            
<failoverdomain name="fd1" nofailback="0" ordered="1" 
restricted="1"><BR>                
<failoverdomainnode name="192.168.1.1" 
priority="1"/><BR>                
<failoverdomainnode name="192.168.1.2" 
priority="2"/><BR>                
<failoverdomainnode name="192.168.1.3" 
priority="3"/><BR>                
<failoverdomainnode name="192.168.1.4" 
priority="4"/><BR>                
<failoverdomainnode name="192.168.1.5" 
priority="5"/><BR>            
</failoverdomain><BR>            
<failoverdomain name="fd2" nofailback="0" ordered="1" 
restricted="1"><BR>                
<failoverdomainnode name="192.168.1.1" 
priority="5"/><BR>                
<failoverdomainnode name="192.168.1.2" 
priority="1"/><BR>                
<failoverdomainnode name="192.168.1.3" 
priority="2"/><BR>                
<failoverdomainnode name="192.168.1.4" 
priority="3"/><BR>                
<failoverdomainnode name="192.168.1.5" 
priority="4"/><BR>            
</failoverdomain><BR>            
<failoverdomain name="fd3" nofailback="0" ordered="1" 
restricted="1"><BR>                
<failoverdomainnode name="192.168.1.1" 
priority="4"/><BR>                
<failoverdomainnode name="192.168.1.2" 
priority="5"/><BR>                
<failoverdomainnode name="192.168.1.3" 
priority="1"/><BR>                
<failoverdomainnode name="192.168.1.4" 
priority="2"/><BR>                
<failoverdomainnode name="192.168.1.5" 
priority="3"/><BR>            
</failoverdomain><BR>            
<failoverdomain name="fd4" nofailback="0" ordered="1" 
restricted="1"><BR>                
<failoverdomainnode name="192.168.1.1" 
priority="3"/><BR>                
<failoverdomainnode name="192.168.1.2" 
priority="4"/><BR>                
<failoverdomainnode name="192.168.1.3" 
priority="5"/><BR>                
<failoverdomainnode name="192.168.1.4" 
priority="1"/><BR>                
<failoverdomainnode name="192.168.1.5" 
priority="2"/><BR>            
</failoverdomain><BR>        
</failoverdomains><BR>        
<resources><BR>            
<ip address="10.1.1.1" 
monitor_link="1"/><BR>            
<ip address="10.1.1.2" 
monitor_link="1"/><BR>            
<ip address="10.1.1.3" 
monitor_link="1"/><BR>            
<ip address="10.1.1.4" 
monitor_link="1"/><BR>            
<ip address="10.1.1.5" 
monitor_link="1"/><BR>            
<script file="/usr/local/bin/service1" 
name="service1"/><BR>            
<script file="/usr/local/bin/service2" 
name="service2"/><BR>            
<script file="/usr/local/bin/service3" 
name="service3"/><BR>            
<script file="/usr/local/bin/service4" 
name="service4"/><BR>       
</resources><BR>        <service 
autostart="1" domain="fd1" exclusive="1" name="service1" 
recovery="relocate"><BR>            
<ip 
ref="10.1.1.1"/><BR>            
<script ref="service1"/><BR>        
</service><BR>        <service 
autostart="1" domain="fd2" exclusive="1" name="service2" 
recovery="relocate"><BR>            
<ip 
ref="10.1.1.2"/><BR>            
<script ref="service2"/><BR>        
</service><BR>        <service 
autostart="1" domain="fd3" exclusive="1" name="service3" 
recovery="relocate"><BR>            
<ip 
ref="10.1.1.3"/><BR>            
<script ref="service3"/><BR>        
</service><BR>        <service 
autostart="1" domain="fd4" exclusive="1" name="service4" 
recovery="relocate"><BR>            
<ip 
ref="10.1.1.4"/><BR>            
<script ref="service4"/><BR>        
</service><BR>    </rm><BR></cluster><BR><BR>
<P><FONT face=Arial size=2>No virus found in this incoming message.<BR>Checked 
by AVG - www.avg.com<BR>Version: 9.0.819 / Virus Database: 271.1.1/2874 - 
Release Date: 05/14/10 20:26:00<BR></FONT></P>
<P><FONT face=Arial size=2></FONT></P></BODY></HTML>