[Linux-cluster] Some problems using fence_vmware_ng under ESXi 4
Jan Friesse
jfriesse at redhat.com
Wed Oct 14 15:45:34 UTC 2009
Carlopmart,
it's not problem of fence_vmware_ng and definitively not malfunction.
Cluster is trying to keep all possible nodes going. So what will happend:
- You will boot node01
- This will start fencing, and finds out, that node02 isn't running (is
falled, ...) -> fence node2
- Node2 is fenced and starts
Regards,
Honza
carlopmart wrote:
> Hi all,
>
> I have installed two rhel5.4 nodes virtual guests, el5prodnode01 and
> el5prodnode02, under esxi 4 host and I need to use fence_vmware_ng as a
> fence device. All works ok except when ESXi starts or is rebooted. I
> have configured under ESXi host to start automatically el5prodnode01
> only when host is rebooted or starts, but when el5prodnode01 guest
> automatically starts tries to launch el5prodnode02 every time. Why? Is
> this the normal procedure for fence_vmware_ng device?? How can I stop
> this feature or malfunction??
>
> My cluster.conf is:
>
> <?xml version="1.0"?>
> <cluster alias="VirtualRHELCluster" config_version="4" name="VirtCluster">
> <fence_daemon post_fail_delay="0" post_join_delay="3"/>
> <quorumd interval="1" tko="10" votes="1" label="prodqdisk"
> log_level="4">
> <heuristic program="ping 172.25.50.11 -c1 -t1" score="1"
> interval="2" tko="3"/>
> </quorumd>
> <clusternodes>
> <clusternode name="node01.hpulabs.org" nodeid="1"
> votes="1">
> <multicast addr="239.192.25.11" interface="eth2"/>
> <fence>
> <method name="1">
> <device name="vmware_fence_node1"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="node02.hpulabs.org" nodeid="2"
> votes="1">
> <multicast addr="239.192.25.11" interface="eth2"/>
> <fence>
> <method name="1">
> <device name="vmware_fence_node2"/>
> </method>
> </fence>
> </clusternode>
> </clusternodes>
> <cman expected_votes="3" two_node="0">
> <multicast addr="239.192.25.11"/>
> </cman>
> <fencedevices>
> <fencedevice agent="fence_vmware_ng"
> name="vmware_fence_node1" ipaddr="172.25.50.11" login="root"
> passwd="rootpass" port="el5prodnode01"/>
> <fencedevice agent="fence_vmware_ng"
> name="vmware_fence_node2" ipaddr="172.25.50.11" login="root"
> passwd="rootpass" port="el5prodnode02"/>
> <fencedevice agent="fence_manual" name="manual-fence"/>
> </fencedevices>
> <rm log_facility="local4" log_level="7">
> <failoverdomains>
> <failoverdomain name="PriCluster1" ordered="1"
> restricted="1">
> <failoverdomainnode
> name="node01.hpulabs.org" priority="1"/>
> <failoverdomainnode
> name="node02.hpulabs.org" priority="2"/>
> </failoverdomain>
> <failoverdomain name="PriCluster2" ordered="1"
> restricted="1">
> <failoverdomainnode
> name="node02.hpulabs.org" priority="1"/>
> <failoverdomainnode
> name="node01.hpulabs.org" priority="2"/>
> </failoverdomain>
> <failoverdomain name="FirstNode" restricted="1">
> <failoverdomainnode
> name="node01.hpulabs.org" priority="1"/>
> </failoverdomain>
> <failoverdomain name="SecondNode" restricted="1">
> <failoverdomainnode
> name="node02.hpulabs.org" priority="1"/>
> </failoverdomain>
> </failoverdomains>
> <resources>
> <fs device="/dev/clustervol/infravol"
> force_fsck="0" force_unmount="1" fstype="ext3"
> mountpoint="/data/services/infra" name="infradata" options="rw"/>
> <fs device="/dev/clustervol/mirrorvol"
> force_fsck="0" force_unmount="1" fstype="ext3"
> mountpoint="/data/services/www" name="mirrordata" options="rw"/>
> <script
> file="/data/config/etc/init.d/postfix-cluster" name="postfix-cluster"/>
> </resources>
> </rm>
> </cluster>
>
> Many thanks for your help.
>
More information about the Linux-cluster
mailing list