[Linux-cluster] error clusvcadm

emmanuel segura emi2fast at gmail.com
Mon May 13 08:48:39 UTC 2013


Hello Delphine

your problem as you know is here
===============================================================
<fs device="LABEL=postfix" mountpoint="/var/spool/
postfix" force_unmount="1" fstype="ext3" name="mgmtha5" options=""/>
===============================================================

I don't know if you are using lvm or partition, but you should look for the
device corresponding to that LABEL, if you are using lvm use vgs and lvs to
see if your volume are actived

Thanks



2013/5/13 Delphine Ramalingom <delphine.ramalingom at univ-reunion.fr>

>  Hi,
>
> I used it :
>
> rg_test test /etc/cluster/cluster.conf start service HA_MGMT
> Running in test mode.
> Starting HA_MGMT...
>
> <err>    startFilesystem: Could not match LABEL=postfix with a real device
> Failed to start HA_MGMT
>
> But it gives me the same message.
>
> Regards
> Delphine
>
> Le 13/05/13 11:47, emmanuel segura a écrit :
>
>  Hello
>
>  If you would like see why your service doens't start, you should use
> "rg_test test /etc/cluster/cluster.conf start service HA_MGMT"
>
>
>
>
> 2013/5/13 Delphine Ramalingom <delphine.ramalingom at univ-reunion.fr>
>
>>  Hi,
>>
>> This is the cluster.conf :
>>
>> [root at titan0 11:29:14 ~]# cat /etc/cluster/cluster.conf
>> <?xml version="1.0" ?>
>> <cluster config_version="7" name="HA_MGMT">
>>         <fence_daemon clean_start="1" post_fail_delay="0"
>> post_join_delay="60"/>
>>         <clusternodes>
>>                 <clusternode name="titan0"  nodeid="1" votes="1">
>>                         <fence>
>>                                 <method name="1">
>>                                         <device name="titan0fence"
>> option="reboot"/>
>>                                 </method>
>>                         </fence>
>>                 </clusternode>
>>                 <clusternode name="titan1" nodeid="2" votes="1">
>>                         <fence>
>>                                 <method name="1">
>>                                         <device name="titan1fence"
>> option="reboot"/>
>>                                 </method>
>>                         </fence>
>>                 </clusternode>
>>         </clusternodes>
>>         <cman  cluster_id="0" expected_votes="1" two_node="1"/>
>>         <fencedevices>
>>                 <fencedevice agent="fence_ipmilan" ipaddr="172.17.0.101"
>> login="administrator" name="titan0fence" passwd="administrator"/>
>>                 <fencedevice agent="fence_ipmilan" ipaddr="172.17.0.102"
>> login="administrator" name="titan1fence" passwd="administrator"/>
>>         </fencedevices>
>>         <rm>
>>                 <failoverdomains>
>>                         <failoverdomain name="titan0_heuristic"
>> ordered="0" restricted="1">
>>                                 <failoverdomainnode name="titan0"
>> priority="1"/>
>>                         </failoverdomain>
>>                         <failoverdomain name="titan1_heuristic"
>> ordered="0" restricted="1">
>>                                 <failoverdomainnode name="titan1"
>> priority="1"/>
>>                         </failoverdomain>
>>                         <failoverdomain name="MgmtNodes" ordered="0"
>> restricted="0">
>>                                 <failoverdomainnode name="titan0"
>> priority="1"/>
>>                                 <failoverdomainnode name="titan1"
>> priority="2"/>
>>                         </failoverdomain>
>>             <failoverdomain name="NFSHA" ordered="0" restricted="0">
>>                 <failoverdomainnode name="titan0" priority="2"/>
>>                 <failoverdomainnode name="titan1" priority="1"/>
>>             </failoverdomain>
>>                 </failoverdomains>
>>             <service domain="titan0_heuristic" name="ha_titan0_check"
>> autostart="1" checkinterval="10">
>>                     <script file="/usr/sbin/ha_titan0_check"
>> name="ha_titan0_check"/>
>>             </service>
>>             <service domain="titan1_heuristic" name="ha_titan1_check"
>> autostart="1" checkinterval="10">
>>                     <script file="/usr/sbin/ha_titan1_check"
>> name="ha_titan1_check"/>
>>             </service>
>>                 <service domain="MgmtNodes" name="HA_MGMT" autostart="0"
>> recovery="relocate">
>>             <!-- ip addresses lines mgmt -->
>>                                 <ip address="172.17.0.99/16"
>> monitor_link="1"/>
>>                                 <ip address="10.90.0.99/24"
>> monitor_link="1"/>
>>             <!-- devices lines mgmt -->
>>                        <fs device="LABEL=postfix"
>> mountpoint="/var/spool/postfix" force_unmount="1" fstype="ext3"
>> name="mgmtha5" options=""/>
>>                        <fs device="LABEL=bigimage"
>> mountpoint="/var/lib/systemimager" force_unmount="1" fstype="ext3"
>> name="mgmtha4" options=""/>
>>                        <clusterfs device="LABEL=HA_MGMT:conman"
>> mountpoint="/var/log/conman" force_unmount="0" fstype="gfs2" name="mgmtha3"
>> options=""/>
>>                        <clusterfs device="LABEL=HA_MGMT:ganglia"
>> mountpoint="/var/lib/ganglia/rrds" force_unmount="0" fstype="gfs2"
>> name="mgmtha2" options=""/>
>>                        <clusterfs device="LABEL=HA_MGMT:syslog"
>> mountpoint="/var/log/HOSTS" force_unmount="0" fstype="gfs2" name="mgmtha1"
>> options=""/>
>>                        <clusterfs device="LABEL=HA_MGMT:cdb"
>> mountpoint="/var/lib/pgsql/data" force_unmount="0" fstype="gfs2"
>> name="mgmtha0" options=""/>
>>                         <script file="/usr/sbin/haservices"
>> name="haservices"/>
>>                 </service>
>>         <service domain="NFSHA" name="HA_NFS" autostart="0"
>> checkinterval="60">
>>             <!-- ip addresses lines nfs -->
>>                                 <ip address="10.31.0.99/16"
>> monitor_link="1"/>
>>                                 <ip address="10.90.0.88/24"
>> monitor_link="1"/>
>>                                 <ip address="172.17.0.88/16"
>> monitor_link="1"/>
>>             <!-- devices lines nfs -->
>>                        <fs device="LABEL=PROGS" mountpoint="/programs"
>> force_unmount="1" fstype="ext3" name="nfsha4" options=""/>
>>                        <fs device="LABEL=WRKTMP" mountpoint="/worktmp"
>> force_unmount="1" fstype="ext3" name="nfsha3" options=""/>
>>                        <fs device="LABEL=LABOS" mountpoint="/labos"
>> force_unmount="1" fstype="xfs" name="nfsha2" options="ikeep"/>
>>                        <fs device="LABEL=OPTINTEL"
>> mountpoint="/opt/intel" force_unmount="1" fstype="ext3" name="nfsha1"
>> options=""/>
>>                        <fs device="LABEL=HOMENFS" mountpoint="/home_nfs"
>> force_unmount="1" fstype="ext3" name="nfsha0" options=""/>
>>             <script file="/etc/init.d/nfs" name="nfs_service"/>
>>         </service>
>>         </rm>
>>     <totem token="21000" />
>> </cluster>
>> <!-- !!!!! DON'T REMOVE OR CHANGE ANYTHING IN PARAMETERS SECTION BELOW
>> node_name=titan0
>> node_ipmi_ipaddr=172.17.0.101
>> node_hwmanager_login=administrator
>> node_hwmanager_passwd=administrator
>> ipaddr1_for_heuristics=172.17.0.200
>> node_ha_name=titan1
>> node_ha_ipmi_ipaddr=172.17.0.102
>> node_ha_hwmanager_login=administrator
>> node_ha_hwmanager_passwd=administrator
>> ipaddr2_for_heuristics=172.17.0.200
>> mngt_virt_ipaddr_for_heuristics=not used on this type of node
>> END OF SECTION !!!!! -->
>>
>>
>> The var/log/messages is too long and have some messages repeated :
>> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>> [10.40.20.30]:39198
>> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>> [10.40.20.30]:39198
>> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>> [10.40.20.30]:39198
>> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>> [10.40.20.30]:39198
>> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>> [10.40.20.30]:39198
>> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>> [10.40.20.30]:39198
>> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>> [10.40.20.30]:39198
>> May 13 11:30:34 s_sys at titan0 snmpd[4584]: Connection from UDP:
>> [10.40.20.30]:53030
>> May 13 11:30:34 s_sys at titan0 snmpd[4584]: Received SNMP packet(s) from
>> UDP: [10.40.20.30]:53030
>> May 13 11:30:34 s_sys at titan0 snmpd[4584]: Connection from UDP:
>> [10.40.20.30]:41083
>> May 13 11:30:34 s_sys at titan0 snmpd[4584]: Received SNMP packet(s) from
>> UDP: [10.40.20.30]:41083
>>
>> Regards
>> Delphine
>>
>>
>>
>> Le 13/05/13 10:37, Rajveer Singh a écrit :
>>
>>  Hi Delphine,
>> It seems there is some filesystem crash. Please share your
>> /var/log/messages and /etc/cluster/cluster.conf file to help you futher.
>>
>>  Regards,
>> Rajveer Singh
>>
>>
>> On Mon, May 13, 2013 at 11:58 AM, Delphine Ramalingom <
>> delphine.ramalingom at univ-reunion.fr> wrote:
>>
>>> Hello,
>>>
>>> I have a problem and I need some help.
>>>
>>> Our cluster linux have been stopped for maintenance in the room server
>>> butr, an error was occured during the stopping procedure :
>>> Local machine disabling service:HA_MGMT...Failure
>>>
>>> The cluster was electrically stopped. But since the restart, I don't
>>> succed to restart services with command clussvcadm.
>>> I have this message :
>>>
>>> clusvcadm -e HA_MGMT
>>> Local machine trying to enable service:HA_MGMT...Aborted; service failed
>>> and
>>> <err>    startFilesystem: Could not match LABEL=postfix with a real
>>> device
>>>
>>> Do you have a solution for me ?
>>>
>>> Thanks a lot in advance.
>>>
>>> Regards
>>> Delphine
>>>
>>> --
>>> Linux-cluster mailing list
>>> Linux-cluster at redhat.com
>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>
>>
>>
>>
>>
>>
>> --
>> Linux-cluster mailing list
>> Linux-cluster at redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>
>
>
>
> --
> esta es mi vida e me la vivo hasta que dios quiera
>
>
>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>



-- 
esta es mi vida e me la vivo hasta que dios quiera
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20130513/3109c944/attachment.htm>


More information about the Linux-cluster mailing list