[Linux-cluster] openais issue

Paras pradhan pradhanparas at gmail.com
Wed Sep 30 14:49:06 UTC 2009


All of the nodes are synced with ntp server. So this is not the case with me.

Thanks
Paras.

On Tue, Sep 29, 2009 at 6:29 PM, Johannes Rußek
<johannes.russek at io-consulting.net> wrote:
> make sure the time on the nodes is in sync, apparently when a node has too
> much offset, you won't see rgmanager (even though the process is running).
> this happened today and setting the time fixed it for me. afaicr there was
> no sign of this in the logs though.
> johannes
>
> Paras pradhan schrieb:
>>
>> I don't see rgmanager .
>>
>> Here is the o/p from clustat
>>
>> [root at cvtst1 cluster]# clustat
>> Cluster Status for test @ Tue Sep 29 15:53:33 2009
>> Member Status: Quorate
>>
>>  Member Name                                                     ID
>> Status
>>  ------ ----                                                     ----
>> ------
>>  cvtst2                                                    1 Online
>>  cvtst1                                                     2 Online,
>> Local
>>  cvtst3                                                     3 Online
>>
>>
>> Thanks
>> Paras.
>>
>> On Tue, Sep 29, 2009 at 3:44 PM, brem belguebli
>> <brem.belguebli at gmail.com> wrote:
>>
>>>
>>> It looks correct, rgmanager seems to start on all nodes
>>>
>>> what gives you clustat ?
>>>
>>> If rgmanager doesn't show, check out the logs something may have gone
>>> wrong.
>>>
>>>
>>> 2009/9/29 Paras pradhan <pradhanparas at gmail.com>:
>>>
>>>>
>>>> Change to 7 and i got this log
>>>>
>>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting down
>>>> Cluster Service Manager...
>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown complete,
>>>> exiting
>>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster Service
>>>> Manager is stopped.
>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group
>>>> Manager Starting
>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service Data
>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading Resource Rules
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building Resource Trees
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources defined
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading Failover
>>>> Domains
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains defined
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events defined
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing Services
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services Initialized
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port Opened
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: Local UP
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst2 UP
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst3 UP
>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1) Processed
>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1) Processed
>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1) Processed
>>>> Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events processed
>>>>
>>>>
>>>> Anything unusual here?
>>>>
>>>> Paras.
>>>>
>>>> On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli
>>>> <brem.belguebli at gmail.com> wrote:
>>>>
>>>>>
>>>>> I use log_level=7 to have more debugging info.
>>>>>
>>>>> It seems 4 is not enough.
>>>>>
>>>>> Brem
>>>>>
>>>>>
>>>>> 2009/9/29, Paras pradhan <pradhanparas at gmail.com>:
>>>>>
>>>>>>
>>>>>> Withe log_level of 3 I got only this
>>>>>>
>>>>>> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down
>>>>>> Cluster Service Manager...
>>>>>> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down
>>>>>> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete,
>>>>>> exiting
>>>>>> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service
>>>>>> Manager is stopped.
>>>>>> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group
>>>>>> Manager Starting
>>>>>> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down
>>>>>> Cluster Service Manager...
>>>>>> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service
>>>>>> Manager is stopped.
>>>>>> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group
>>>>>> Manager Starting
>>>>>> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting
>>>>>> down
>>>>>>
>>>>>> I do not know what the last line means.
>>>>>>
>>>>>> rgmanager version I am running is:
>>>>>> rgmanager-2.0.52-1.el5.centos
>>>>>>
>>>>>> I don't what has gone wrong.
>>>>>>
>>>>>> Thanks
>>>>>> Paras.
>>>>>>
>>>>>>
>>>>>> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli
>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>
>>>>>>>
>>>>>>> you mean it stopped successfully on all the nodes but it is failing
>>>>>>> to
>>>>>>> start only on node cvtst1 ?
>>>>>>>
>>>>>>> look at the following page  to make rgmanager more verbose. It 'll
>>>>>>> help debug....
>>>>>>>
>>>>>>> http://sources.redhat.com/cluster/wiki/RGManager
>>>>>>>
>>>>>>> at Logging Configuration section
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> 2009/9/29 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>
>>>>>>>>
>>>>>>>> Brem,
>>>>>>>>
>>>>>>>> When I try to restart rgmanager on all the nodes, this time i do not
>>>>>>>> see rgmanager running on the first node. But I do see on other 2
>>>>>>>> nodes.
>>>>>>>>
>>>>>>>> Log on the first node:
>>>>>>>>
>>>>>>>> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group
>>>>>>>> Manager Starting
>>>>>>>> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down
>>>>>>>> Cluster Service Manager...
>>>>>>>> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down
>>>>>>>> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete,
>>>>>>>> exiting
>>>>>>>> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service
>>>>>>>> Manager is stopped.
>>>>>>>> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group
>>>>>>>> Manager Starting
>>>>>>>>
>>>>>>>> -
>>>>>>>> It seems service is running ,  but I do not see rgmanger running
>>>>>>>> using clustat
>>>>>>>>
>>>>>>>>
>>>>>>>> Don't know what is going on.
>>>>>>>>
>>>>>>>> Thanks
>>>>>>>> Paras.
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli
>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>
>>>>>>>>>
>>>>>>>>> Paras,
>>>>>>>>>
>>>>>>>>> Another thing, it would have been more interesting to have a start
>>>>>>>>> DEBUG not a stop.
>>>>>>>>>
>>>>>>>>> That's why I was asking you to first stop the vm manually on all
>>>>>>>>> your
>>>>>>>>> nodes, stop eventually rgmanager on all the nodes to reset the
>>>>>>>>> potential wrong states you may have, restart rgmanager.
>>>>>>>>>
>>>>>>>>> If your VM is configured to autostart, this will make it start.
>>>>>>>>>
>>>>>>>>> It should normally fail (as it does now). Send out your newly
>>>>>>>>> created
>>>>>>>>> DEBUG file.
>>>>>>>>>
>>>>>>>>> 2009/9/29 brem belguebli <brem.belguebli at gmail.com>:
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Hi Paras,
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> I don't know the xen/cluster combination well, but if I do
>>>>>>>>>> remember
>>>>>>>>>> well, I think I've read somewhere that when using xen you have to
>>>>>>>>>> declare the use_virsh=0 key in the VM definition in the
>>>>>>>>>> cluster.conf.
>>>>>>>>>>
>>>>>>>>>> This would make rgmanager use xm commands instead of virsh
>>>>>>>>>> The DEBUG output shows clearly that you are using virsh to manage
>>>>>>>>>> your
>>>>>>>>>> VM instead of xm commands.
>>>>>>>>>> Check out the RH docs about virtualization
>>>>>>>>>>
>>>>>>>>>> I'm not a 100% sure about that, I may be completely wrong.
>>>>>>>>>>
>>>>>>>>>> Brem
>>>>>>>>>>
>>>>>>>>>> 2009/9/28 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> The only thing I noticed is the message after stopping the vm
>>>>>>>>>>> using xm
>>>>>>>>>>> in all nodes and starting using clusvcadm is
>>>>>>>>>>>
>>>>>>>>>>> "Virtual machine guest1 is blocked"
>>>>>>>>>>>
>>>>>>>>>>> The whole DEBUG file is attached.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Thanks
>>>>>>>>>>> Paras.
>>>>>>>>>>>
>>>>>>>>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli
>>>>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> There's a problem with the script that is called by rgmanager to
>>>>>>>>>>>> start
>>>>>>>>>>>> the VM, I don't know what causes it
>>>>>>>>>>>>
>>>>>>>>>>>> May be you should try something like :
>>>>>>>>>>>>
>>>>>>>>>>>> 1) stop the VM on all nodes with xm commands
>>>>>>>>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the
>>>>>>>>>>>> following
>>>>>>>>>>>> lines (after the #!/bin/bash ):
>>>>>>>>>>>>  exec >/tmp/DEBUG 2>&1
>>>>>>>>>>>>  set -x
>>>>>>>>>>>> 3) start the VM with clusvcadm -e vm:guest1
>>>>>>>>>>>>
>>>>>>>>>>>> It should fail as it did before.
>>>>>>>>>>>>
>>>>>>>>>>>> edit the the /tmp/DEBUG file and you will be able to see where
>>>>>>>>>>>> it
>>>>>>>>>>>> fails (it may generate a lot of debug)
>>>>>>>>>>>>
>>>>>>>>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh
>>>>>>>>>>>>
>>>>>>>>>>>> Post the DEBUG file if you're not able to see where it fails.
>>>>>>>>>>>>
>>>>>>>>>>>> Brem
>>>>>>>>>>>>
>>>>>>>>>>>> 2009/9/26 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> No I am not manually starting not using automatic init scripts.
>>>>>>>>>>>>>
>>>>>>>>>>>>> I started the vm using: clusvcadm -e vm:guest1
>>>>>>>>>>>>>
>>>>>>>>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few
>>>>>>>>>>>>> seconds it
>>>>>>>>>>>>> says guest1 started . But after a while I can see the guest1 on
>>>>>>>>>>>>> all
>>>>>>>>>>>>> three nodes.
>>>>>>>>>>>>>
>>>>>>>>>>>>> clustat says:
>>>>>>>>>>>>>
>>>>>>>>>>>>>  Service Name                                            Owner
>>>>>>>>>>>>> (Last)
>>>>>>>>>>>>>                                         State
>>>>>>>>>>>>>  ------- ----                                            -----
>>>>>>>>>>>>> ------
>>>>>>>>>>>>>                                         -----
>>>>>>>>>>>>>  vm:guest1                                               (none)
>>>>>>>>>>>>>                                         stopped
>>>>>>>>>>>>>
>>>>>>>>>>>>> But I can see the vm from xm li.
>>>>>>>>>>>>>
>>>>>>>>>>>>> This is what I can see from the log:
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm
>>>>>>>>>>>>> "guest1"
>>>>>>>>>>>>> returned 1 (generic error)
>>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed
>>>>>>>>>>>>> to start
>>>>>>>>>>>>> vm:guest1; return value: 1
>>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping
>>>>>>>>>>>>> service vm:guest1
>>>>>>>>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service
>>>>>>>>>>>>> vm:guest1 is
>>>>>>>>>>>>> recovering
>>>>>>>>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering
>>>>>>>>>>>>> failed
>>>>>>>>>>>>> service vm:guest1
>>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm
>>>>>>>>>>>>> "guest1"
>>>>>>>>>>>>> returned 1 (generic error)
>>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed
>>>>>>>>>>>>> to start
>>>>>>>>>>>>> vm:guest1; return value: 1
>>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping
>>>>>>>>>>>>> service vm:guest1
>>>>>>>>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service
>>>>>>>>>>>>> vm:guest1 is
>>>>>>>>>>>>> recovering
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli
>>>>>>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Have you started  your VM via rgmanager (clusvcadm -e
>>>>>>>>>>>>>> vm:guest1) or
>>>>>>>>>>>>>> using xm commands out of cluster control  (or maybe a thru an
>>>>>>>>>>>>>> automatic init script ?)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> When clustered, you should never be starting services
>>>>>>>>>>>>>> (manually or
>>>>>>>>>>>>>> thru automatic init script) out of cluster control
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> The thing would be to stop your vm on all the nodes with the
>>>>>>>>>>>>>> adequate
>>>>>>>>>>>>>> xm command (not using xen myself) and try to start it with
>>>>>>>>>>>>>> clusvcadm.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Then see if it is started on all nodes (send clustat output)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Ok. Please see below. my vm is running on all nodes though
>>>>>>>>>>>>>>> clustat
>>>>>>>>>>>>>>> says it is stopped.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> [root at cvtst1 ~]# clustat
>>>>>>>>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009
>>>>>>>>>>>>>>> Member Status: Quorate
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>  Member Name
>>>>>>>>>>>>>>>     ID   Status
>>>>>>>>>>>>>>>  ------ ----
>>>>>>>>>>>>>>>     ---- ------
>>>>>>>>>>>>>>>  cvtst2                                                    1
>>>>>>>>>>>>>>> Online, rgmanager
>>>>>>>>>>>>>>>  cvtst1                                                     2
>>>>>>>>>>>>>>> Online,
>>>>>>>>>>>>>>> Local, rgmanager
>>>>>>>>>>>>>>>  cvtst3                                                     3
>>>>>>>>>>>>>>> Online, rgmanager
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>  Service Name
>>>>>>>>>>>>>>>  Owner (Last)
>>>>>>>>>>>>>>>                                         State
>>>>>>>>>>>>>>>  ------- ----
>>>>>>>>>>>>>>>  ----- ------
>>>>>>>>>>>>>>>                                         -----
>>>>>>>>>>>>>>>  vm:guest1
>>>>>>>>>>>>>>> (none)
>>>>>>>>>>>>>>>                                         stopped
>>>>>>>>>>>>>>> [root at cvtst1 ~]#
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> ---
>>>>>>>>>>>>>>> o/p of xm li on cvtst1
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> [root at cvtst1 ~]# xm li
>>>>>>>>>>>>>>> Name                                      ID Mem(MiB) VCPUs
>>>>>>>>>>>>>>> State   Time(s)
>>>>>>>>>>>>>>> Domain-0                                   0     3470     2
>>>>>>>>>>>>>>> r-----  28939.4
>>>>>>>>>>>>>>> guest1                                     7      511     1
>>>>>>>>>>>>>>> -b----   7727.8
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> o/p of xm li on cvtst2
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> [root at cvtst2 ~]# xm li
>>>>>>>>>>>>>>> Name                                      ID Mem(MiB) VCPUs
>>>>>>>>>>>>>>> State   Time(s)
>>>>>>>>>>>>>>> Domain-0                                   0     3470     2
>>>>>>>>>>>>>>> r-----  31558.9
>>>>>>>>>>>>>>> guest1                                    21      511     1
>>>>>>>>>>>>>>> -b----   7558.2
>>>>>>>>>>>>>>> ---
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli
>>>>>>>>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> It looks like no.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> can you send an output of clustat  of when the VM is running
>>>>>>>>>>>>>>>> on
>>>>>>>>>>>>>>>> multiple nodes at the same time?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> And by the way, another one after having stopped (clusvcadm
>>>>>>>>>>>>>>>> -s vm:guest1) ?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not
>>>>>>>>>>>>>>>>> being
>>>>>>>>>>>>>>>>> properly handled by the cluster.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan
>>>>>>>>>>>>>>>>> <pradhanparas at gmail.com> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Ok.. here is my cluster.conf file
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>> [root at cvtst1 cluster]# more cluster.conf
>>>>>>>>>>>>>>>>>> <?xml version="1.0"?>
>>>>>>>>>>>>>>>>>> <cluster alias="test" config_version="9" name="test">
>>>>>>>>>>>>>>>>>>       <fence_daemon clean_start="0" post_fail_delay="0"
>>>>>>>>>>>>>>>>>> post_join_delay="3"/>
>>>>>>>>>>>>>>>>>>       <clusternodes>
>>>>>>>>>>>>>>>>>>               <clusternode name="cvtst2" nodeid="1"
>>>>>>>>>>>>>>>>>> votes="1">
>>>>>>>>>>>>>>>>>>                       <fence/>
>>>>>>>>>>>>>>>>>>               </clusternode>
>>>>>>>>>>>>>>>>>>               <clusternode name="cvtst1" nodeid="2"
>>>>>>>>>>>>>>>>>> votes="1">
>>>>>>>>>>>>>>>>>>                       <fence/>
>>>>>>>>>>>>>>>>>>               </clusternode>
>>>>>>>>>>>>>>>>>>               <clusternode name="cvtst3" nodeid="3"
>>>>>>>>>>>>>>>>>> votes="1">
>>>>>>>>>>>>>>>>>>                       <fence/>
>>>>>>>>>>>>>>>>>>               </clusternode>
>>>>>>>>>>>>>>>>>>       </clusternodes>
>>>>>>>>>>>>>>>>>>       <cman/>
>>>>>>>>>>>>>>>>>>       <fencedevices/>
>>>>>>>>>>>>>>>>>>       <rm>
>>>>>>>>>>>>>>>>>>               <failoverdomains>
>>>>>>>>>>>>>>>>>>                       <failoverdomain name="myfd1"
>>>>>>>>>>>>>>>>>> nofailback="0" ordered="1" restricted="0">
>>>>>>>>>>>>>>>>>>                               <failoverdomainnode
>>>>>>>>>>>>>>>>>> name="cvtst2" priority="3"/>
>>>>>>>>>>>>>>>>>>                               <failoverdomainnode
>>>>>>>>>>>>>>>>>> name="cvtst1" priority="1"/>
>>>>>>>>>>>>>>>>>>                               <failoverdomainnode
>>>>>>>>>>>>>>>>>> name="cvtst3" priority="2"/>
>>>>>>>>>>>>>>>>>>                       </failoverdomain>
>>>>>>>>>>>>>>>>>>               </failoverdomains>
>>>>>>>>>>>>>>>>>>               <resources/>
>>>>>>>>>>>>>>>>>>               <vm autostart="1" domain="myfd1"
>>>>>>>>>>>>>>>>>> exclusive="0" max_restarts="0"
>>>>>>>>>>>>>>>>>> name="guest1" path="/vms" recovery="r
>>>>>>>>>>>>>>>>>> estart" restart_expire_time="0"/>
>>>>>>>>>>>>>>>>>>       </rm>
>>>>>>>>>>>>>>>>>> </cluster>
>>>>>>>>>>>>>>>>>> [root at cvtst1 cluster]#
>>>>>>>>>>>>>>>>>> ------
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer
>>>>>>>>>>>>>>>>>> <volker at ixolution.de> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500,
>>>>>>>>>>>>>>>>>>> Paras pradhan <pradhanparas at gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines.
>>>>>>>>>>>>>>>>>>>> Now I am
>>>>>>>>>>>>>>>>>>>> having another problem. When I start the my xen vm in
>>>>>>>>>>>>>>>>>>>> one node, it
>>>>>>>>>>>>>>>>>>>> also starts on other nodes. Which daemon controls  this?
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the
>>>>>>>>>>>>>>>>>>> rgmanager
>>>>>>>>>>>>>>>>>>> package). To me, this sounds like a configuration
>>>>>>>>>>>>>>>>>>> problem. Maybe,
>>>>>>>>>>>>>>>>>>> you can post your cluster.conf?
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Regards,
>>>>>>>>>>>>>>>>>>> Volker
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> --
>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Linux-cluster mailing list
>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Linux-cluster mailing list
>>>>>>>> Linux-cluster at redhat.com
>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Linux-cluster mailing list
>>>>>>> Linux-cluster at redhat.com
>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>> Linux-cluster mailing list
>>>>>> Linux-cluster at redhat.com
>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>> Linux-cluster mailing list
>>>>> Linux-cluster at redhat.com
>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>
>>>>>
>>>>
>>>> --
>>>> Linux-cluster mailing list
>>>> Linux-cluster at redhat.com
>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>
>>>>
>>>
>>> --
>>> Linux-cluster mailing list
>>> Linux-cluster at redhat.com
>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>
>>>
>>
>> --
>> Linux-cluster mailing list
>> Linux-cluster at redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>




More information about the Linux-cluster mailing list