[Linux-cluster] Problems using fence_virt as a fence agent for two kvm guests
C. L. Martinez
carlopmart at gmail.com
Thu Jul 12 09:27:49 UTC 2012
Hi all,
I have installed two kvm guests (CentOS 6.3) to do some tests using
RHCS under a CentOS 6.3 kvm host. As a fence device I am trying to use
fence_virt, but it doesn't works for me.
fence_virt.conf in kvm host is:
fence_virtd {
listener = "multicast";
backend = "libvirt";
}
listeners {
multicast {
key_file = "/etc/fence_virt.key";
interface = "siemif";
address = "225.0.0.12";
family = "ipv4";
}
}
backends {
libvirt {
uri = "qemu:///system";
}
}
fence_virt.key is located under /etc directory:
-r-------- 1 root root 18 Jul 12 09:48 /etc/fence_virt.key
cluster.conf on both kvm guest nodes is:
<?xml version="1.0"?>
<cluster config_version="1" name="TestCluster">
<cman expected_votes="1" two_node="1"/>
<clusternodes>
<clusternode name="cosnode01.domain.local" nodeid="1">
<fence>
<method name="kvm">
<device action="reboot" port="cosnode01" name="kvm_cosnode01"/>
</method>
</fence>
</clusternode>
<clusternode name="cosnode02.domain.local" nodeid="2">
<fence>
<method name="kvm">
<device action="reboot" port="cosnode02" name="kvm_cosnode02"/>
</method>
</fence>
</clusternode>
</clusternodes>
<fencedevices>
<fencedevice agent="fence_virt" ip_family="ipv4"
multicast_address="225.0.0.12" key_file="/etc/cluster/fence_virt.key"
name="kvm_cosnode01"/>
<fencedevice agent="fence_virt" ip_family="ipv4"
multicast_address="225.0.0.12" key_file="/etc/cluster/fence_virt.key"
name="kvm_cosnode02"/>
</fencedevices>
<fence_daemon post_join_delay="30"/>
<totem rrp_mode="none" secauth="off"/>
<rm log_level="5">
<failoverdomains>
<failoverdomain name="only_node01" nofailback="1" ordered="0" restricted="1">
<failoverdomainnode name="cosnode01.domain.local"/>
</failoverdomain>
<failoverdomain name="only_node02" nofailback="1" ordered="0" restricted="1">
<failoverdomainnode name="cosnode02.domain.local"/>
</failoverdomain>
<failoverdomain name="primary_clu01" nofailback="1" ordered="1"
restricted="1">
<failoverdomainnode name="cosnode01.domain.local" priority="1"/>
<failoverdomainnode name="cosnode02.domain.local" priority="2"/>
</failoverdomain>
<failoverdomain name="primary_clu02" nofailback="1" ordered="1"
restricted="1">
<failoverdomainnode name="cosnode01.domain.local" priority="2"/>
<failoverdomainnode name="cosnode02.domain.local" priority="1"/>
</failoverdomain>
</failoverdomains>
</rm>
</cluster>
of course, fence_virt.key is copied under /etc/cluster dir in both nodes.
In cosnode01 I see this error:
fenced[4074]: fence cosnode02.domain.local dev 0.0 agent fence_virt
result: error from agent
fenced[4074]: fence cosnode02.domain.loca failed
What am I doing wrong?? Do I need to modify libvirtd.conf to listen in
siemif interface??
Thanks.
More information about the Linux-cluster
mailing list