[Linux-cluster] Problems using fence_virt as a fence agent for two kvm guests
C. L. Martinez
carlopmart at gmail.com
Fri Jul 13 12:26:35 UTC 2012
On Thu, Jul 12, 2012 at 11:27 AM, C. L. Martinez <carlopmart at gmail.com> wrote:
> Hi all,
>
> I have installed two kvm guests (CentOS 6.3) to do some tests using
> RHCS under a CentOS 6.3 kvm host. As a fence device I am trying to use
> fence_virt, but it doesn't works for me.
>
> fence_virt.conf in kvm host is:
>
> fence_virtd {
> listener = "multicast";
> backend = "libvirt";
> }
>
> listeners {
> multicast {
> key_file = "/etc/fence_virt.key";
> interface = "siemif";
> address = "225.0.0.12";
> family = "ipv4";
> }
> }
>
> backends {
> libvirt {
> uri = "qemu:///system";
> }
> }
>
> fence_virt.key is located under /etc directory:
>
> -r-------- 1 root root 18 Jul 12 09:48 /etc/fence_virt.key
>
> cluster.conf on both kvm guest nodes is:
>
> <?xml version="1.0"?>
> <cluster config_version="1" name="TestCluster">
> <cman expected_votes="1" two_node="1"/>
> <clusternodes>
> <clusternode name="cosnode01.domain.local" nodeid="1">
> <fence>
> <method name="kvm">
> <device action="reboot" port="cosnode01" name="kvm_cosnode01"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="cosnode02.domain.local" nodeid="2">
> <fence>
> <method name="kvm">
> <device action="reboot" port="cosnode02" name="kvm_cosnode02"/>
> </method>
> </fence>
> </clusternode>
> </clusternodes>
> <fencedevices>
> <fencedevice agent="fence_virt" ip_family="ipv4"
> multicast_address="225.0.0.12" key_file="/etc/cluster/fence_virt.key"
> name="kvm_cosnode01"/>
> <fencedevice agent="fence_virt" ip_family="ipv4"
> multicast_address="225.0.0.12" key_file="/etc/cluster/fence_virt.key"
> name="kvm_cosnode02"/>
> </fencedevices>
> <fence_daemon post_join_delay="30"/>
> <totem rrp_mode="none" secauth="off"/>
> <rm log_level="5">
> <failoverdomains>
> <failoverdomain name="only_node01" nofailback="1" ordered="0" restricted="1">
> <failoverdomainnode name="cosnode01.domain.local"/>
> </failoverdomain>
> <failoverdomain name="only_node02" nofailback="1" ordered="0" restricted="1">
> <failoverdomainnode name="cosnode02.domain.local"/>
> </failoverdomain>
> <failoverdomain name="primary_clu01" nofailback="1" ordered="1"
> restricted="1">
> <failoverdomainnode name="cosnode01.domain.local" priority="1"/>
> <failoverdomainnode name="cosnode02.domain.local" priority="2"/>
> </failoverdomain>
> <failoverdomain name="primary_clu02" nofailback="1" ordered="1"
> restricted="1">
> <failoverdomainnode name="cosnode01.domain.local" priority="2"/>
> <failoverdomainnode name="cosnode02.domain.local" priority="1"/>
> </failoverdomain>
> </failoverdomains>
> </rm>
> </cluster>
>
> of course, fence_virt.key is copied under /etc/cluster dir in both nodes.
>
> In cosnode01 I see this error:
>
> fenced[4074]: fence cosnode02.domain.local dev 0.0 agent fence_virt
> result: error from agent
> fenced[4074]: fence cosnode02.domain.loca failed
>
> What am I doing wrong?? Do I need to modify libvirtd.conf to listen in
> siemif interface??
>
> Thanks.
Please, any help??
More information about the Linux-cluster
mailing list