[Linux-cluster] error clusvcadm

Delphine Ramalingom delphine.ramalingom at univ-reunion.fr
Mon May 13 07:32:27 UTC 2013


Hi,

This is the cluster.conf :

[root at titan0 11:29:14 ~]# cat /etc/cluster/cluster.conf
<?xml version="1.0" ?>
<cluster config_version="7" name="HA_MGMT">
         <fence_daemon clean_start="1" post_fail_delay="0" 
post_join_delay="60"/>
         <clusternodes>
                 <clusternode name="titan0"  nodeid="1" votes="1">
                         <fence>
                                 <method name="1">
                                         <device name="titan0fence" 
option="reboot"/>
                                 </method>
                         </fence>
                 </clusternode>
                 <clusternode name="titan1" nodeid="2" votes="1">
                         <fence>
                                 <method name="1">
                                         <device name="titan1fence" 
option="reboot"/>
                                 </method>
                         </fence>
                 </clusternode>
         </clusternodes>
         <cman  cluster_id="0" expected_votes="1" two_node="1"/>
         <fencedevices>
                 <fencedevice agent="fence_ipmilan" 
ipaddr="172.17.0.101" login="administrator" name="titan0fence" 
passwd="administrator"/>
                 <fencedevice agent="fence_ipmilan" 
ipaddr="172.17.0.102" login="administrator" name="titan1fence" 
passwd="administrator"/>
         </fencedevices>
         <rm>
                 <failoverdomains>
                         <failoverdomain name="titan0_heuristic" 
ordered="0" restricted="1">
                                 <failoverdomainnode name="titan0" 
priority="1"/>
                         </failoverdomain>
                         <failoverdomain name="titan1_heuristic" 
ordered="0" restricted="1">
                                 <failoverdomainnode name="titan1" 
priority="1"/>
                         </failoverdomain>
                         <failoverdomain name="MgmtNodes" ordered="0" 
restricted="0">
                                 <failoverdomainnode name="titan0" 
priority="1"/>
                                 <failoverdomainnode name="titan1" 
priority="2"/>
                         </failoverdomain>
             <failoverdomain name="NFSHA" ordered="0" restricted="0">
                 <failoverdomainnode name="titan0" priority="2"/>
                 <failoverdomainnode name="titan1" priority="1"/>
             </failoverdomain>
                 </failoverdomains>
             <service domain="titan0_heuristic" name="ha_titan0_check" 
autostart="1" checkinterval="10">
                     <script file="/usr/sbin/ha_titan0_check" 
name="ha_titan0_check"/>
             </service>
             <service domain="titan1_heuristic" name="ha_titan1_check" 
autostart="1" checkinterval="10">
                     <script file="/usr/sbin/ha_titan1_check" 
name="ha_titan1_check"/>
             </service>
                 <service domain="MgmtNodes" name="HA_MGMT" 
autostart="0" recovery="relocate">
             <!-- ip addresses lines mgmt -->
                                 <ip address="172.17.0.99/16" 
monitor_link="1"/>
                                 <ip address="10.90.0.99/24" 
monitor_link="1"/>
             <!-- devices lines mgmt -->
                        <fs device="LABEL=postfix" 
mountpoint="/var/spool/postfix" force_unmount="1" fstype="ext3" 
name="mgmtha5" options=""/>
                        <fs device="LABEL=bigimage" 
mountpoint="/var/lib/systemimager" force_unmount="1" fstype="ext3" 
name="mgmtha4" options=""/>
                        <clusterfs device="LABEL=HA_MGMT:conman" 
mountpoint="/var/log/conman" force_unmount="0" fstype="gfs2" 
name="mgmtha3" options=""/>
                        <clusterfs device="LABEL=HA_MGMT:ganglia" 
mountpoint="/var/lib/ganglia/rrds" force_unmount="0" fstype="gfs2" 
name="mgmtha2" options=""/>
                        <clusterfs device="LABEL=HA_MGMT:syslog" 
mountpoint="/var/log/HOSTS" force_unmount="0" fstype="gfs2" 
name="mgmtha1" options=""/>
                        <clusterfs device="LABEL=HA_MGMT:cdb" 
mountpoint="/var/lib/pgsql/data" force_unmount="0" fstype="gfs2" 
name="mgmtha0" options=""/>
                         <script file="/usr/sbin/haservices" 
name="haservices"/>
                 </service>
         <service domain="NFSHA" name="HA_NFS" autostart="0" 
checkinterval="60">
             <!-- ip addresses lines nfs -->
                                 <ip address="10.31.0.99/16" 
monitor_link="1"/>
                                 <ip address="10.90.0.88/24" 
monitor_link="1"/>
                                 <ip address="172.17.0.88/16" 
monitor_link="1"/>
             <!-- devices lines nfs -->
                        <fs device="LABEL=PROGS" mountpoint="/programs" 
force_unmount="1" fstype="ext3" name="nfsha4" options=""/>
                        <fs device="LABEL=WRKTMP" mountpoint="/worktmp" 
force_unmount="1" fstype="ext3" name="nfsha3" options=""/>
                        <fs device="LABEL=LABOS" mountpoint="/labos" 
force_unmount="1" fstype="xfs" name="nfsha2" options="ikeep"/>
                        <fs device="LABEL=OPTINTEL" 
mountpoint="/opt/intel" force_unmount="1" fstype="ext3" name="nfsha1" 
options=""/>
                        <fs device="LABEL=HOMENFS" 
mountpoint="/home_nfs" force_unmount="1" fstype="ext3" name="nfsha0" 
options=""/>
             <script file="/etc/init.d/nfs" name="nfs_service"/>
         </service>
         </rm>
     <totem token="21000" />
</cluster>
<!-- !!!!! DON'T REMOVE OR CHANGE ANYTHING IN PARAMETERS SECTION BELOW
node_name=titan0
node_ipmi_ipaddr=172.17.0.101
node_hwmanager_login=administrator
node_hwmanager_passwd=administrator
ipaddr1_for_heuristics=172.17.0.200
node_ha_name=titan1
node_ha_ipmi_ipaddr=172.17.0.102
node_ha_hwmanager_login=administrator
node_ha_hwmanager_passwd=administrator
ipaddr2_for_heuristics=172.17.0.200
mngt_virt_ipaddr_for_heuristics=not used on this type of node
END OF SECTION !!!!! -->


The var/log/messages is too long and have some messages repeated :
May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP: 
[10.40.20.30]:39198
May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP: 
[10.40.20.30]:39198
May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP: 
[10.40.20.30]:39198
May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP: 
[10.40.20.30]:39198
May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP: 
[10.40.20.30]:39198
May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP: 
[10.40.20.30]:39198
May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP: 
[10.40.20.30]:39198
May 13 11:30:34 s_sys at titan0 snmpd[4584]: Connection from UDP: 
[10.40.20.30]:53030
May 13 11:30:34 s_sys at titan0 snmpd[4584]: Received SNMP packet(s) from 
UDP: [10.40.20.30]:53030
May 13 11:30:34 s_sys at titan0 snmpd[4584]: Connection from UDP: 
[10.40.20.30]:41083
May 13 11:30:34 s_sys at titan0 snmpd[4584]: Received SNMP packet(s) from 
UDP: [10.40.20.30]:41083

Regards
Delphine



Le 13/05/13 10:37, Rajveer Singh a écrit :
> Hi Delphine,
> It seems there is some filesystem crash. Please share your 
> /var/log/messages and /etc/cluster/cluster.conf file to help you futher.
>
> Regards,
> Rajveer Singh
>
>
> On Mon, May 13, 2013 at 11:58 AM, Delphine Ramalingom 
> <delphine.ramalingom at univ-reunion.fr 
> <mailto:delphine.ramalingom at univ-reunion.fr>> wrote:
>
>     Hello,
>
>     I have a problem and I need some help.
>
>     Our cluster linux have been stopped for maintenance in the room
>     server butr, an error was occured during the stopping procedure :
>     Local machine disabling service:HA_MGMT...Failure
>
>     The cluster was electrically stopped. But since the restart, I
>     don't succed to restart services with command clussvcadm.
>     I have this message :
>
>     clusvcadm -e HA_MGMT
>     Local machine trying to enable service:HA_MGMT...Aborted; service
>     failed
>     and
>     <err>    startFilesystem: Could not match LABEL=postfix with a
>     real device
>
>     Do you have a solution for me ?
>
>     Thanks a lot in advance.
>
>     Regards
>     Delphine
>
>     -- 
>     Linux-cluster mailing list
>     Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
>     https://www.redhat.com/mailman/listinfo/linux-cluster
>
>
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20130513/a591136e/attachment.htm>


More information about the Linux-cluster mailing list