[Linux-cluster] Strange error messages in /var/log/messages
Virginian
virginian at blueyonder.co.uk
Wed Oct 15 15:01:45 UTC 2008
Hi all,
I am running Centos 5.2 on a two node physical cluster with Xen virtualisation and 4 domains clustered underneath. I am see the following in /var/log/messages on one of the physical nodes:
Oct 15 15:53:13 xen2 avahi-daemon[3363]: New relevant interface eth0.IPv4 for mDNS.
Oct 15 15:53:13 xen2 avahi-daemon[3363]: Joining mDNS multicast group on interface eth0.IPv4 with address 10.199.10.170.
Oct 15 15:53:14 xen2 avahi-daemon[3363]: Network interface enumeration completed.
Oct 15 15:53:14 xen2 avahi-daemon[3363]: Registering new address record for fe80::200:ff:fe00:0 on virbr0.
Oct 15 15:53:14 xen2 avahi-daemon[3363]: Registering new address record for 192.168.122.1 on virbr0.
Oct 15 15:53:14 xen2 avahi-daemon[3363]: Registering new address record for fe80::202:a5ff:fed9:ef74 on eth0.
Oct 15 15:53:14 xen2 avahi-daemon[3363]: Registering new address record for 10.199.10.170 on eth0.
Oct 15 15:53:14 xen2 avahi-daemon[3363]: Registering HINFO record with values 'I686'/'LINUX'.
Oct 15 15:53:15 xen2 avahi-daemon[3363]: Server startup complete. Host name is xen2.local. Local service cookie is 3231388299.
Oct 15 15:53:16 xen2 avahi-daemon[3363]: Service "SFTP File Transfer on xen2" (/services/sftp-ssh.service) successfully established.
Oct 15 15:53:23 xen2 xenstored: Checking store ...
Oct 15 15:53:23 xen2 xenstored: Checking store complete.
Oct 15 15:53:24 xen2 ccsd[2806]: Error: unable to evaluate xpath query "/cluster/fence_xvmd/@(null) "
Oct 15 15:53:24 xen2 ccsd[2806]: Error while processing get: Invalid argument
Oct 15 15:53:24 xen2 ccsd[2806]: Error: unable to evaluate xpath query "/cluster/fence_xvmd/@(null) "
Oct 15 15:53:24 xen2 ccsd[2806]: Error while processing get: Invalid argument
Oct 15 15:53:24 xen2 ccsd[2806]: Error: unable to evaluate xpath query "/cluster/fence_xvmd/@(null) "
Oct 15 15:53:24 xen2 ccsd[2806]: Error while processing get: Invalid argument
Oct 15 15:53:24 xen2 ccsd[2806]: Error: unable to evaluate xpath query "/cluster/fence_xvmd/@(null) "
Oct 15 15:53:24 xen2 ccsd[2806]: Error while processing get: Invalid argument
Oct 15 15:53:24 xen2 ccsd[2806]: Error: unable to evaluate xpath query "/cluster/fence_xvmd/@(null) "
Oct 15 15:53:24 xen2 ccsd[2806]: Error while processing get: Invalid argument
Oct 15 15:53:24 xen2 ccsd[2806]: Error: unable to evaluate xpath query "/cluster/fence_xvmd/@(null) "
Oct 15 15:53:24 xen2 ccsd[2806]: Error while processing get: Invalid argument
Oct 15 15:53:24 xen2 modclusterd: startup succeeded
Oct 15 15:53:24 xen2 clurgmgrd[3531]: <notice> Resource Group Manager Starting
Oct 15 15:53:25 xen2 oddjobd: oddjobd startup succeeded
Oct 15 15:53:26 xen2 saslauthd[3885]: detach_tty : master pid is: 3885
Oct 15 15:53:26 xen2 saslauthd[3885]: ipc_init : listening on socket: /var/run/saslauthd/mux
Oct 15 15:53:26 xen2 ricci: startup succeeded
Oct 15 15:53:39 xen2 clurgmgrd[3531]: <notice> Starting stopped service vm:hermes
Oct 15 15:53:39 xen2 clurgmgrd[3531]: <notice> Starting stopped service vm:hestia
Oct 15 15:53:43 xen2 kernel: tap tap-1-51712: 2 getting info
Oct 15 15:53:44 xen2 kernel: tap tap-1-51728: 2 getting info
Oct 15 15:53:45 xen2 kernel: device vif1.0 entered promiscuous mode
Oct 15 15:53:45 xen2 kernel: ADDRCONF(NETDEV_UP): vif1.0: link is not ready
Oct 15 15:53:47 xen2 kernel: tap tap-2-51712: 2 getting info
Oct 15 15:53:48 xen2 kernel: tap tap-2-51728: 2 getting info
Oct 15 15:53:48 xen2 kernel: device vif2.0 entered promiscuous mode
Oct 15 15:53:48 xen2 kernel: ADDRCONF(NETDEV_UP): vif2.0: link is not ready
Oct 15 15:53:49 xen2 clurgmgrd[3531]: <notice> Service vm:hestia started
Oct 15 15:53:49 xen2 clurgmgrd[3531]: <notice> Service vm:hermes started
Oct 15 15:53:53 xen2 kernel: blktap: ring-ref 8, event-channel 11, protocol 1 (x86_32-abi)
Oct 15 15:53:53 xen2 kernel: blktap: ring-ref 9, event-channel 12, protocol 1 (x86_32-abi)
Oct 15 15:53:53 xen2 kernel: blktap: ring-ref 8, event-channel 11, protocol 1 (x86_32-abi)
Oct 15 15:53:53 xen2 kernel: blktap: ring-ref 9, event-channel 12, protocol 1 (x86_32-abi)
Oct 15 15:54:23 xen2 kernel: ADDRCONF(NETDEV_CHANGE): vif2.0: link becomes ready
Oct 15 15:54:23 xen2 kernel: xenbr0: topology change detected, propagating
Oct 15 15:54:23 xen2 kernel: xenbr0: port 4(vif2.0) entering forwarding state
Oct 15 15:54:27 xen2 kernel: ADDRCONF(NETDEV_CHANGE): vif1.0: link becomes ready
Oct 15 15:54:27 xen2 kernel: xenbr0: topology change detected, propagating
Oct 15 15:54:27 xen2 kernel: xenbr0: port 3(vif1.0) entering forwarding state
Oct 15 15:56:15 xen2 clurgmgrd[3531]: <notice> Resource Groups Locked
My cluster.conf is as follows:
cluster.conf 100% 1734 1.7KB/s 00:00
[root at xen1 cluster]# cat /etc/cluster/cluster.conf.15102008
<?xml version="1.0"?>
<cluster alias="XENCluster1" config_version="42" name="XENCluster1">
<cman expected_votes="1" two_node="1"/>
<clusternodes>
<clusternode name="xen1" nodeid="1" votes="1">
<fence>
<method name="1">
<device name="xen1-ilo"/>
</method>
</fence>
</clusternode>
<clusternode name="xen2" nodeid="2" votes="1">
<fence>
<method name="1">
<device name="xen2-ilo"/>
</method>
</fence>
</clusternode>
</clusternodes>
<fencedevices>
<fencedevice agent="fence_ilo" hostname="xen1-ilo" login="root" name="xen1-ilo" passwd="deckard1"/>
<fencedevice agent="fence_ilo" hostname="xen2-ilo" login="root" name="xen2-ilo" passwd="deckard1"/>
</fencedevices>
<rm>
<failoverdomains>
<failoverdomain name="xen1-failover" nofailback="0" ordered="1" restricted="0">
<failoverdomainnode name="xen1" priority="1"/>
</failoverdomain>
<failoverdomain name="xen2-failover" nofailback="0" ordered="1" restricted="0">
<failoverdomainnode name="xen2" priority="2"/>
</failoverdomain>
</failoverdomains>
<resources/>
<vm autostart="1" domain="xen2-failover" exclusive="0" migrate="live" name="hermes" path="/guests" recovery="relocate"/>
<vm autostart="1" domain="xen2-failover" exclusive="0" migrate="live" name="hestia" path="/guests" recovery="relocate"/>
<vm autostart="1" domain="xen1-failover" exclusive="0" migrate="live" name="aether" path="/guests" recovery="relocate"/>
<vm autostart="1" domain="xen1-failover" exclusive="0" migrate="live" name="athena" path="/guests" recovery="relocate"/>
</rm>
<fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/>
<totem token="21000"/>
<fence_xvmd/>
</cluster>
Does anybody know what these messages mean?
My domain cluster.conf is as follows:
<?xml version="1.0"?>
<cluster alias="XENCluster2" config_version="13" name="XENCluster2">
<fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/>
<clusternodes>
<clusternode name="athena.private.lan" nodeid="1" votes="1">
<fence>
<method name="1">
<device domain="athena" name="virtual_fence"/>
</method>
</fence>
</clusternode>
<clusternode name="aether.private.lan" nodeid="2" votes="1">
<fence>
<method name="1">
<device domain="aether" name="virtual_fence"/>
</method>
</fence>
</clusternode>
<clusternode name="hermes.private.lan" nodeid="3" votes="1">
<fence>
<method name="1">
<device domain="hermes" name="virtual_fence"/>
</method>
</fence>
</clusternode>
<clusternode name="hestia.private.lan" nodeid="4" votes="1">
<fence>
<method name="1">
<device domain="hestia" name="virtual_fence"/>
</method>
</fence>
</clusternode>
</clusternodes>
<cman/>
<fencedevices>
<fencedevice agent="fence_xvm" name="virtual_fence"/>
</fencedevices>
<rm>
<failoverdomains/>
<resources/>
</rm>
<fence_xvmd/>
</cluster>
Thanks
John
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20081015/b367da8a/attachment.htm>
More information about the Linux-cluster
mailing list