[Linux-cluster] openais issue

Daniela Anzellotti daniela.anzellotti at roma1.infn.it
Tue Oct 6 17:59:48 UTC 2009


Hi Paras,

did you reboot all the cluster nodes? I needed a complete reboot 
(actually I was angry enough to switch everything off and on again): a 
restart of the cluster suite was not enough.

As far as I understood, since the cluster was not able to bring a VM in 
a good running state, it decided that the VM has to migrate to another 
node... and at the end all the VMs was trying to migrate from one node 
to another with the result that I had all VMs starting on all the 
cluster nodes.
Restarting the cluster suite didn't kill a lot of processes that was 
stubbornly trying to migrate virtual machines...

Daniela

Paras pradhan wrote:
> Adding use_virsh=0 works great. Now I do not have my vm starting at
> all the nodes. This is a good fix. Thanks ..
> 
> The only problem left is I do not see rgmanager running on my node1
> and the clustat of node2 and node3 is reporting vm as migrating
> 
> o/p
> 
> Service Name                                          Owner (Last)
>                                      State
>  ------- ----                                          ----- ------
>                                       -----
>  vm:guest1                                             cvtst1
>                              migrating
> [root at cvtst3 vms]#
> 
> 
> Thanks
> Paras.
> 
> On Tue, Oct 6, 2009 at 10:48 AM, Daniela Anzellotti
> <daniela.anzellotti at roma1.infn.it> wrote:
>> Hi Paras,
>>
>> yes. At least it looks so...
>>
>> We have a cluster of two nodes + a quorum disk (it's not configured as a
>> "two-node cluster")
>>
>> They are running Scientific Linux 5.x, kernel 2.6.18-128.7.1.el5xen and
>>
>>  openais-0.80.6-8.el5.x86_64
>>  cman-2.0.115-1.el5.x86_64
>>  rgmanager-2.0.52-1.el5.x86_64
>>
>> The XEN VMs access the disk as simple block devices.
>> Disks are on a SAN, configured with Clustered LVM.
>>
>>  xen-3.0.3-94.el5_4.1.x86_64
>>  xen-libs-3.0.3-94.el5_4.1.x86_64
>>
>> VM configuration files are as the following
>>
>>  name = "www1"
>>  uuid = "3bd3e910-23c0-97ee-55ab-086260ef1e53"
>>  memory = 1024
>>  maxmem = 1024
>>  vcpus = 1
>>  bootloader = "/usr/bin/pygrub"
>>  vfb = [ "type=vnc,vncunused=1,keymap=en-us" ]
>>  disk = [ "phy:/dev/vg_cluster/www1.disk,xvda,w", \
>>  "phy:/dev/vg_cluster/www1.swap,xvdb,w" ]
>>  vif = [ "mac=00:16:3e:da:00:07,bridge=xenbr1" ]
>>  on_poweroff = "destroy"
>>  on_reboot = "restart"
>>  on_crash = "restart"
>>  extra = "xencons=tty0 console=tty0"
>>
>>
>> I changed in /etc/cluster/cluster.conf all the VM directive from
>>
>>  <vm autostart="1" domain="rhcs1_dom" exclusive="0" \
>>  migrate="live" name="www1" path="/etc/xen" recovery="restart"/>
>>
>> to
>>
>>  <vm autostart="1" use_virsh="0" domain="rhcs1_dom" exclusive="0" \
>>  migrate="live" name="www1" path="/etc/xen" recovery="restart"/>
>>
>>
>> Rebooted the cluster nodes and it started working again...
>>
>> As i said I hope I'll not have any other bad surprise (I tested a VM
>> migration and it is working too), but at least cluster it's working now (it
>> was not able to start a VM, before)!
>>
>> Ciao
>> Daniela
>>
>>
>> Paras pradhan wrote:
>>> So you mean your cluster is running fine with the CMAN
>>> cman-2.0.115-1.el5.x86_64 ?
>>>
>>> Which version of openais are you running?
>>>
>>> Thanks
>>> Paras.
>>>
>>>
>>> On Mon, Oct 5, 2009 at 7:19 AM, Daniela Anzellotti
>>> <daniela.anzellotti at roma1.infn.it> wrote:
>>>> Hi all,
>>>>
>>>> I had a problem similar to Paras's one today: yum updated the following
>>>> rpms
>>>> last week and today (I had to restart the cluster) the cluster was not
>>>> able
>>>> to start vm: services.
>>>>
>>>> Oct 02 05:31:05 Updated: openais-0.80.6-8.el5.x86_64
>>>> Oct 02 05:31:07 Updated: cman-2.0.115-1.el5.x86_64
>>>> Oct 02 05:31:10 Updated: rgmanager-2.0.52-1.el5.x86_64
>>>>
>>>> Oct 03 04:03:12 Updated: xen-libs-3.0.3-94.el5_4.1.x86_64
>>>> Oct 03 04:03:12 Updated: xen-libs-3.0.3-94.el5_4.1.i386
>>>> Oct 03 04:03:16 Updated: xen-3.0.3-94.el5_4.1.x86_64
>>>>
>>>>
>>>> So, after checked the vm.sh script, I add the declaration use_virsh="0"
>>>> in
>>>> the VM definition in the cluster.conf (as suggested by Brem, thanks!) and
>>>> everything is now working again.
>>>>
>>>>
>>>> BTW I didn't understand if the problem was caused by the new XEN version
>>>> or
>>>> the new openais one, thus I disabled automatic updates for both.
>>>>
>>>> I hope I'll not have any other bad surprise...
>>>>
>>>> Thank you,
>>>> cheers,
>>>> Daniela
>>>>
>>>>
>>>> Paras pradhan wrote:
>>>>> Yes this is very strange. I don't know what to do now. May be re
>>>>> create the cluster? But not a good solution actually.
>>>>>
>>>>> Packages :
>>>>>
>>>>> Kernel: kernel-xen-2.6.18-164.el5
>>>>> OS: Full updated of CentOS 5.3 except CMAN downgraded to
>>>>> cman-2.0.98-1.el5
>>>>>
>>>>> Other packages related to cluster suite:
>>>>>
>>>>> rgmanager-2.0.52-1.el5.centos
>>>>> cman-2.0.98-1.el5
>>>>> xen-3.0.3-80.el5_3.3
>>>>> xen-libs-3.0.3-80.el5_3.3
>>>>> kmod-gfs-xen-0.1.31-3.el5_3.1
>>>>> kmod-gfs-xen-0.1.31-3.el5_3.1
>>>>> kmod-gfs-0.1.31-3.el5_3.1
>>>>> gfs-utils-0.1.18-1.el5
>>>>> gfs2-utils-0.1.62-1.el5
>>>>> lvm2-2.02.40-6.el5
>>>>> lvm2-cluster-2.02.40-7.el5
>>>>> openais-0.80.3-22.el5_3.9
>>>>>
>>>>> Thanks!
>>>>> Paras.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Sep 30, 2009 at 10:02 AM, brem belguebli
>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>> Hi Paras,
>>>>>>
>>>>>> Your cluster.conf file seems correct. If it is not a ntp issue, I
>>>>>> don't see anything except a bug that causes this, or some prerequisite
>>>>>> that is not respected.
>>>>>>
>>>>>> May be you could post the versions (os, kernel, packages etc...) you
>>>>>> are using, someone may have hit the same issue with your versions.
>>>>>>
>>>>>> Brem
>>>>>>
>>>>>> 2009/9/30, Paras pradhan <pradhanparas at gmail.com>:
>>>>>>> All of the nodes are synced with ntp server. So this is not the case
>>>>>>> with me.
>>>>>>>
>>>>>>> Thanks
>>>>>>> Paras.
>>>>>>>
>>>>>>> On Tue, Sep 29, 2009 at 6:29 PM, Johannes Rußek
>>>>>>> <johannes.russek at io-consulting.net> wrote:
>>>>>>>> make sure the time on the nodes is in sync, apparently when a node
>>>>>>>> has
>>>>>>>> too
>>>>>>>> much offset, you won't see rgmanager (even though the process is
>>>>>>>> running).
>>>>>>>> this happened today and setting the time fixed it for me. afaicr
>>>>>>>> there
>>>>>>>> was
>>>>>>>> no sign of this in the logs though.
>>>>>>>> johannes
>>>>>>>>
>>>>>>>> Paras pradhan schrieb:
>>>>>>>>> I don't see rgmanager .
>>>>>>>>>
>>>>>>>>> Here is the o/p from clustat
>>>>>>>>>
>>>>>>>>> [root at cvtst1 cluster]# clustat
>>>>>>>>> Cluster Status for test @ Tue Sep 29 15:53:33 2009
>>>>>>>>> Member Status: Quorate
>>>>>>>>>
>>>>>>>>>  Member Name                                                     ID
>>>>>>>>> Status
>>>>>>>>>  ------ ----
>>>>>>>>> ----
>>>>>>>>> ------
>>>>>>>>>  cvtst2                                                    1 Online
>>>>>>>>>  cvtst1                                                     2
>>>>>>>>> Online,
>>>>>>>>> Local
>>>>>>>>>  cvtst3                                                     3 Online
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Thanks
>>>>>>>>> Paras.
>>>>>>>>>
>>>>>>>>> On Tue, Sep 29, 2009 at 3:44 PM, brem belguebli
>>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> It looks correct, rgmanager seems to start on all nodes
>>>>>>>>>>
>>>>>>>>>> what gives you clustat ?
>>>>>>>>>>
>>>>>>>>>> If rgmanager doesn't show, check out the logs something may have
>>>>>>>>>> gone
>>>>>>>>>> wrong.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> 2009/9/29 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>
>>>>>>>>>>> Change to 7 and i got this log
>>>>>>>>>>>
>>>>>>>>>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting down
>>>>>>>>>>> Cluster Service Manager...
>>>>>>>>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
>>>>>>>>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
>>>>>>>>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown
>>>>>>>>>>> complete,
>>>>>>>>>>> exiting
>>>>>>>>>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster
>>>>>>>>>>> Service
>>>>>>>>>>> Manager is stopped.
>>>>>>>>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group
>>>>>>>>>>> Manager Starting
>>>>>>>>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service
>>>>>>>>>>> Data
>>>>>>>>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading Resource
>>>>>>>>>>> Rules
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building Resource
>>>>>>>>>>> Trees
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources
>>>>>>>>>>> defined
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading Failover
>>>>>>>>>>> Domains
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains defined
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events defined
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing
>>>>>>>>>>> Services
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services
>>>>>>>>>>> Initialized
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port
>>>>>>>>>>> Opened
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change:
>>>>>>>>>>> Local
>>>>>>>>>>> UP
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change:
>>>>>>>>>>> cvtst2
>>>>>>>>>>> UP
>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change:
>>>>>>>>>>> cvtst3
>>>>>>>>>>> UP
>>>>>>>>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1)
>>>>>>>>>>> Processed
>>>>>>>>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1)
>>>>>>>>>>> Processed
>>>>>>>>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1)
>>>>>>>>>>> Processed
>>>>>>>>>>> Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events
>>>>>>>>>>> processed
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Anything unusual here?
>>>>>>>>>>>
>>>>>>>>>>> Paras.
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli
>>>>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> I use log_level=7 to have more debugging info.
>>>>>>>>>>>>
>>>>>>>>>>>> It seems 4 is not enough.
>>>>>>>>>>>>
>>>>>>>>>>>> Brem
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> 2009/9/29, Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>>>
>>>>>>>>>>>>> Withe log_level of 3 I got only this
>>>>>>>>>>>>>
>>>>>>>>>>>>> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down
>>>>>>>>>>>>> Cluster Service Manager...
>>>>>>>>>>>>> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down
>>>>>>>>>>>>> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown
>>>>>>>>>>>>> complete,
>>>>>>>>>>>>> exiting
>>>>>>>>>>>>> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster
>>>>>>>>>>>>> Service
>>>>>>>>>>>>> Manager is stopped.
>>>>>>>>>>>>> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group
>>>>>>>>>>>>> Manager Starting
>>>>>>>>>>>>> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting
>>>>>>>>>>>>> down
>>>>>>>>>>>>> Cluster Service Manager...
>>>>>>>>>>>>> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster
>>>>>>>>>>>>> Service
>>>>>>>>>>>>> Manager is stopped.
>>>>>>>>>>>>> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group
>>>>>>>>>>>>> Manager Starting
>>>>>>>>>>>>> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1
>>>>>>>>>>>>> shutting
>>>>>>>>>>>>> down
>>>>>>>>>>>>>
>>>>>>>>>>>>> I do not know what the last line means.
>>>>>>>>>>>>>
>>>>>>>>>>>>> rgmanager version I am running is:
>>>>>>>>>>>>> rgmanager-2.0.52-1.el5.centos
>>>>>>>>>>>>>
>>>>>>>>>>>>> I don't what has gone wrong.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli
>>>>>>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> you mean it stopped successfully on all the nodes but it is
>>>>>>>>>>>>>> failing
>>>>>>>>>>>>>> to
>>>>>>>>>>>>>> start only on node cvtst1 ?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> look at the following page  to make rgmanager more verbose. It
>>>>>>>>>>>>>> 'll
>>>>>>>>>>>>>> help debug....
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> http://sources.redhat.com/cluster/wiki/RGManager
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> at Logging Configuration section
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 2009/9/29 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Brem,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> When I try to restart rgmanager on all the nodes, this time i
>>>>>>>>>>>>>>> do
>>>>>>>>>>>>>>> not
>>>>>>>>>>>>>>> see rgmanager running on the first node. But I do see on other
>>>>>>>>>>>>>>> 2
>>>>>>>>>>>>>>> nodes.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Log on the first node:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource
>>>>>>>>>>>>>>> Group
>>>>>>>>>>>>>>> Manager Starting
>>>>>>>>>>>>>>> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting
>>>>>>>>>>>>>>> down
>>>>>>>>>>>>>>> Cluster Service Manager...
>>>>>>>>>>>>>>> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting
>>>>>>>>>>>>>>> down
>>>>>>>>>>>>>>> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown
>>>>>>>>>>>>>>> complete,
>>>>>>>>>>>>>>> exiting
>>>>>>>>>>>>>>> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster
>>>>>>>>>>>>>>> Service
>>>>>>>>>>>>>>> Manager is stopped.
>>>>>>>>>>>>>>> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource
>>>>>>>>>>>>>>> Group
>>>>>>>>>>>>>>> Manager Starting
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> -
>>>>>>>>>>>>>>> It seems service is running ,  but I do not see rgmanger
>>>>>>>>>>>>>>> running
>>>>>>>>>>>>>>> using clustat
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Don't know what is going on.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli
>>>>>>>>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Paras,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Another thing, it would have been more interesting to have a
>>>>>>>>>>>>>>>> start
>>>>>>>>>>>>>>>> DEBUG not a stop.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> That's why I was asking you to first stop the vm manually on
>>>>>>>>>>>>>>>> all
>>>>>>>>>>>>>>>> your
>>>>>>>>>>>>>>>> nodes, stop eventually rgmanager on all the nodes to reset
>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>> potential wrong states you may have, restart rgmanager.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> If your VM is configured to autostart, this will make it
>>>>>>>>>>>>>>>> start.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> It should normally fail (as it does now). Send out your newly
>>>>>>>>>>>>>>>> created
>>>>>>>>>>>>>>>> DEBUG file.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> 2009/9/29 brem belguebli <brem.belguebli at gmail.com>:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hi Paras,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> I don't know the xen/cluster combination well, but if I do
>>>>>>>>>>>>>>>>> remember
>>>>>>>>>>>>>>>>> well, I think I've read somewhere that when using xen you
>>>>>>>>>>>>>>>>> have
>>>>>>>>>>>>>>>>> to
>>>>>>>>>>>>>>>>> declare the use_virsh=0 key in the VM definition in the
>>>>>>>>>>>>>>>>> cluster.conf.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> This would make rgmanager use xm commands instead of virsh
>>>>>>>>>>>>>>>>> The DEBUG output shows clearly that you are using virsh to
>>>>>>>>>>>>>>>>> manage
>>>>>>>>>>>>>>>>> your
>>>>>>>>>>>>>>>>> VM instead of xm commands.
>>>>>>>>>>>>>>>>> Check out the RH docs about virtualization
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> I'm not a 100% sure about that, I may be completely wrong.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Brem
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> 2009/9/28 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> The only thing I noticed is the message after stopping the
>>>>>>>>>>>>>>>>>> vm
>>>>>>>>>>>>>>>>>> using xm
>>>>>>>>>>>>>>>>>> in all nodes and starting using clusvcadm is
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> "Virtual machine guest1 is blocked"
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> The whole DEBUG file is attached.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli
>>>>>>>>>>>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> There's a problem with the script that is called by
>>>>>>>>>>>>>>>>>>> rgmanager to
>>>>>>>>>>>>>>>>>>> start
>>>>>>>>>>>>>>>>>>> the VM, I don't know what causes it
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> May be you should try something like :
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> 1) stop the VM on all nodes with xm commands
>>>>>>>>>>>>>>>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the
>>>>>>>>>>>>>>>>>>> following
>>>>>>>>>>>>>>>>>>> lines (after the #!/bin/bash ):
>>>>>>>>>>>>>>>>>>>  exec >/tmp/DEBUG 2>&1
>>>>>>>>>>>>>>>>>>>  set -x
>>>>>>>>>>>>>>>>>>> 3) start the VM with clusvcadm -e vm:guest1
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> It should fail as it did before.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> edit the the /tmp/DEBUG file and you will be able to see
>>>>>>>>>>>>>>>>>>> where
>>>>>>>>>>>>>>>>>>> it
>>>>>>>>>>>>>>>>>>> fails (it may generate a lot of debug)
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Post the DEBUG file if you're not able to see where it
>>>>>>>>>>>>>>>>>>> fails.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Brem
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> 2009/9/26 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> No I am not manually starting not using automatic init
>>>>>>>>>>>>>>>>>>>> scripts.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> I started the vm using: clusvcadm -e vm:guest1
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few
>>>>>>>>>>>>>>>>>>>> seconds it
>>>>>>>>>>>>>>>>>>>> says guest1 started . But after a while I can see the
>>>>>>>>>>>>>>>>>>>> guest1 on
>>>>>>>>>>>>>>>>>>>> all
>>>>>>>>>>>>>>>>>>>> three nodes.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> clustat says:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>  Service Name
>>>>>>>>>>>>>>>>>>>>  Owner
>>>>>>>>>>>>>>>>>>>> (Last)
>>>>>>>>>>>>>>>>>>>>                                       State
>>>>>>>>>>>>>>>>>>>>  ------- ----
>>>>>>>>>>>>>>>>>>>>  -----
>>>>>>>>>>>>>>>>>>>> ------
>>>>>>>>>>>>>>>>>>>>                                       -----
>>>>>>>>>>>>>>>>>>>>  vm:guest1
>>>>>>>>>>>>>>>>>>>> (none)
>>>>>>>>>>>>>>>>>>>>                                       stopped
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> But I can see the vm from xm li.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> This is what I can see from the log:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on
>>>>>>>>>>>>>>>>>>>> vm
>>>>>>>>>>>>>>>>>>>> "guest1"
>>>>>>>>>>>>>>>>>>>> returned 1 (generic error)
>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68:
>>>>>>>>>>>>>>>>>>>> Failed
>>>>>>>>>>>>>>>>>>>> to start
>>>>>>>>>>>>>>>>>>>> vm:guest1; return value: 1
>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping
>>>>>>>>>>>>>>>>>>>> service vm:guest1
>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service
>>>>>>>>>>>>>>>>>>>> vm:guest1 is
>>>>>>>>>>>>>>>>>>>> recovering
>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice>
>>>>>>>>>>>>>>>>>>>> Recovering
>>>>>>>>>>>>>>>>>>>> failed
>>>>>>>>>>>>>>>>>>>> service vm:guest1
>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on
>>>>>>>>>>>>>>>>>>>> vm
>>>>>>>>>>>>>>>>>>>> "guest1"
>>>>>>>>>>>>>>>>>>>> returned 1 (generic error)
>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68:
>>>>>>>>>>>>>>>>>>>> Failed
>>>>>>>>>>>>>>>>>>>> to start
>>>>>>>>>>>>>>>>>>>> vm:guest1; return value: 1
>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping
>>>>>>>>>>>>>>>>>>>> service vm:guest1
>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service
>>>>>>>>>>>>>>>>>>>> vm:guest1 is
>>>>>>>>>>>>>>>>>>>> recovering
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli
>>>>>>>>>>>>>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Have you started  your VM via rgmanager (clusvcadm -e
>>>>>>>>>>>>>>>>>>>>> vm:guest1) or
>>>>>>>>>>>>>>>>>>>>> using xm commands out of cluster control  (or maybe a
>>>>>>>>>>>>>>>>>>>>> thru
>>>>>>>>>>>>>>>>>>>>> an
>>>>>>>>>>>>>>>>>>>>> automatic init script ?)
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> When clustered, you should never be starting services
>>>>>>>>>>>>>>>>>>>>> (manually or
>>>>>>>>>>>>>>>>>>>>> thru automatic init script) out of cluster control
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> The thing would be to stop your vm on all the nodes with
>>>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>> adequate
>>>>>>>>>>>>>>>>>>>>> xm command (not using xen myself) and try to start it
>>>>>>>>>>>>>>>>>>>>> with
>>>>>>>>>>>>>>>>>>>>> clusvcadm.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Then see if it is started on all nodes (send clustat
>>>>>>>>>>>>>>>>>>>>> output)
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Ok. Please see below. my vm is running on all nodes
>>>>>>>>>>>>>>>>>>>>>> though
>>>>>>>>>>>>>>>>>>>>>> clustat
>>>>>>>>>>>>>>>>>>>>>> says it is stopped.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>> [root at cvtst1 ~]# clustat
>>>>>>>>>>>>>>>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009
>>>>>>>>>>>>>>>>>>>>>> Member Status: Quorate
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>  Member Name
>>>>>>>>>>>>>>>>>>>>>>   ID   Status
>>>>>>>>>>>>>>>>>>>>>>  ------ ----
>>>>>>>>>>>>>>>>>>>>>>   ---- ------
>>>>>>>>>>>>>>>>>>>>>>  cvtst2
>>>>>>>>>>>>>>>>>>>>>>  1
>>>>>>>>>>>>>>>>>>>>>> Online, rgmanager
>>>>>>>>>>>>>>>>>>>>>>  cvtst1
>>>>>>>>>>>>>>>>>>>>>>  2
>>>>>>>>>>>>>>>>>>>>>> Online,
>>>>>>>>>>>>>>>>>>>>>> Local, rgmanager
>>>>>>>>>>>>>>>>>>>>>>  cvtst3
>>>>>>>>>>>>>>>>>>>>>>  3
>>>>>>>>>>>>>>>>>>>>>> Online, rgmanager
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>  Service Name
>>>>>>>>>>>>>>>>>>>>>>  Owner (Last)
>>>>>>>>>>>>>>>>>>>>>>                                       State
>>>>>>>>>>>>>>>>>>>>>>  ------- ----
>>>>>>>>>>>>>>>>>>>>>>  ----- ------
>>>>>>>>>>>>>>>>>>>>>>                                       -----
>>>>>>>>>>>>>>>>>>>>>>  vm:guest1
>>>>>>>>>>>>>>>>>>>>>> (none)
>>>>>>>>>>>>>>>>>>>>>>                                       stopped
>>>>>>>>>>>>>>>>>>>>>> [root at cvtst1 ~]#
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> ---
>>>>>>>>>>>>>>>>>>>>>> o/p of xm li on cvtst1
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>> [root at cvtst1 ~]# xm li
>>>>>>>>>>>>>>>>>>>>>> Name                                      ID Mem(MiB)
>>>>>>>>>>>>>>>>>>>>>> VCPUs
>>>>>>>>>>>>>>>>>>>>>> State   Time(s)
>>>>>>>>>>>>>>>>>>>>>> Domain-0                                   0     3470
>>>>>>>>>>>>>>>>>>>>>> 2
>>>>>>>>>>>>>>>>>>>>>> r-----  28939.4
>>>>>>>>>>>>>>>>>>>>>> guest1                                     7      511
>>>>>>>>>>>>>>>>>>>>>> 1
>>>>>>>>>>>>>>>>>>>>>> -b----   7727.8
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> o/p of xm li on cvtst2
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>> [root at cvtst2 ~]# xm li
>>>>>>>>>>>>>>>>>>>>>> Name                                      ID Mem(MiB)
>>>>>>>>>>>>>>>>>>>>>> VCPUs
>>>>>>>>>>>>>>>>>>>>>> State   Time(s)
>>>>>>>>>>>>>>>>>>>>>> Domain-0                                   0     3470
>>>>>>>>>>>>>>>>>>>>>> 2
>>>>>>>>>>>>>>>>>>>>>> r-----  31558.9
>>>>>>>>>>>>>>>>>>>>>> guest1                                    21      511
>>>>>>>>>>>>>>>>>>>>>> 1
>>>>>>>>>>>>>>>>>>>>>> -b----   7558.2
>>>>>>>>>>>>>>>>>>>>>> ---
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli
>>>>>>>>>>>>>>>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> It looks like no.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> can you send an output of clustat  of when the VM is
>>>>>>>>>>>>>>>>>>>>>>> running
>>>>>>>>>>>>>>>>>>>>>>> on
>>>>>>>>>>>>>>>>>>>>>>> multiple nodes at the same time?
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> And by the way, another one after having stopped
>>>>>>>>>>>>>>>>>>>>>>> (clusvcadm
>>>>>>>>>>>>>>>>>>>>>>> -s vm:guest1) ?
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas at gmail.com>:
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Anyone having issue as mine? Virtual machine service
>>>>>>>>>>>>>>>>>>>>>>>> is
>>>>>>>>>>>>>>>>>>>>>>>> not
>>>>>>>>>>>>>>>>>>>>>>>> being
>>>>>>>>>>>>>>>>>>>>>>>> properly handled by the cluster.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan
>>>>>>>>>>>>>>>>>>>>>>>> <pradhanparas at gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> Ok.. here is my cluster.conf file
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>>>> [root at cvtst1 cluster]# more cluster.conf
>>>>>>>>>>>>>>>>>>>>>>>>> <?xml version="1.0"?>
>>>>>>>>>>>>>>>>>>>>>>>>> <cluster alias="test" config_version="9"
>>>>>>>>>>>>>>>>>>>>>>>>> name="test">
>>>>>>>>>>>>>>>>>>>>>>>>>     <fence_daemon clean_start="0"
>>>>>>>>>>>>>>>>>>>>>>>>> post_fail_delay="0"
>>>>>>>>>>>>>>>>>>>>>>>>> post_join_delay="3"/>
>>>>>>>>>>>>>>>>>>>>>>>>>     <clusternodes>
>>>>>>>>>>>>>>>>>>>>>>>>>             <clusternode name="cvtst2" nodeid="1"
>>>>>>>>>>>>>>>>>>>>>>>>> votes="1">
>>>>>>>>>>>>>>>>>>>>>>>>>                     <fence/>
>>>>>>>>>>>>>>>>>>>>>>>>>             </clusternode>
>>>>>>>>>>>>>>>>>>>>>>>>>             <clusternode name="cvtst1" nodeid="2"
>>>>>>>>>>>>>>>>>>>>>>>>> votes="1">
>>>>>>>>>>>>>>>>>>>>>>>>>                     <fence/>
>>>>>>>>>>>>>>>>>>>>>>>>>             </clusternode>
>>>>>>>>>>>>>>>>>>>>>>>>>             <clusternode name="cvtst3" nodeid="3"
>>>>>>>>>>>>>>>>>>>>>>>>> votes="1">
>>>>>>>>>>>>>>>>>>>>>>>>>                     <fence/>
>>>>>>>>>>>>>>>>>>>>>>>>>             </clusternode>
>>>>>>>>>>>>>>>>>>>>>>>>>     </clusternodes>
>>>>>>>>>>>>>>>>>>>>>>>>>     <cman/>
>>>>>>>>>>>>>>>>>>>>>>>>>     <fencedevices/>
>>>>>>>>>>>>>>>>>>>>>>>>>     <rm>
>>>>>>>>>>>>>>>>>>>>>>>>>             <failoverdomains>
>>>>>>>>>>>>>>>>>>>>>>>>>                     <failoverdomain name="myfd1"
>>>>>>>>>>>>>>>>>>>>>>>>> nofailback="0" ordered="1" restricted="0">
>>>>>>>>>>>>>>>>>>>>>>>>>                             <failoverdomainnode
>>>>>>>>>>>>>>>>>>>>>>>>> name="cvtst2" priority="3"/>
>>>>>>>>>>>>>>>>>>>>>>>>>                             <failoverdomainnode
>>>>>>>>>>>>>>>>>>>>>>>>> name="cvtst1" priority="1"/>
>>>>>>>>>>>>>>>>>>>>>>>>>                             <failoverdomainnode
>>>>>>>>>>>>>>>>>>>>>>>>> name="cvtst3" priority="2"/>
>>>>>>>>>>>>>>>>>>>>>>>>>                     </failoverdomain>
>>>>>>>>>>>>>>>>>>>>>>>>>             </failoverdomains>
>>>>>>>>>>>>>>>>>>>>>>>>>             <resources/>
>>>>>>>>>>>>>>>>>>>>>>>>>             <vm autostart="1" domain="myfd1"
>>>>>>>>>>>>>>>>>>>>>>>>> exclusive="0" max_restarts="0"
>>>>>>>>>>>>>>>>>>>>>>>>> name="guest1" path="/vms" recovery="r
>>>>>>>>>>>>>>>>>>>>>>>>> estart" restart_expire_time="0"/>
>>>>>>>>>>>>>>>>>>>>>>>>>     </rm>
>>>>>>>>>>>>>>>>>>>>>>>>> </cluster>
>>>>>>>>>>>>>>>>>>>>>>>>> [root at cvtst1 cluster]#
>>>>>>>>>>>>>>>>>>>>>>>>> ------
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer
>>>>>>>>>>>>>>>>>>>>>>>>> <volker at ixolution.de> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500,
>>>>>>>>>>>>>>>>>>>>>>>>>> Paras pradhan <pradhanparas at gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual
>>>>>>>>>>>>>>>>>>>>>>>>>>> machines.
>>>>>>>>>>>>>>>>>>>>>>>>>>> Now I am
>>>>>>>>>>>>>>>>>>>>>>>>>>> having another problem. When I start the my xen vm
>>>>>>>>>>>>>>>>>>>>>>>>>>> in
>>>>>>>>>>>>>>>>>>>>>>>>>>> one node, it
>>>>>>>>>>>>>>>>>>>>>>>>>>> also starts on other nodes. Which daemon controls
>>>>>>>>>>>>>>>>>>>>>>>>>>>  this?
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part
>>>>>>>>>>>>>>>>>>>>>>>>>> of
>>>>>>>>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>>>>> rgmanager
>>>>>>>>>>>>>>>>>>>>>>>>>> package). To me, this sounds like a configuration
>>>>>>>>>>>>>>>>>>>>>>>>>> problem. Maybe,
>>>>>>>>>>>>>>>>>>>>>>>>>> you can post your cluster.conf?
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> Regards,
>>>>>>>>>>>>>>>>>>>>>>>>>> Volker
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>> --
>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Linux-cluster mailing list
>>>>>>>>> Linux-cluster at redhat.com
>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>
>>>>>>>> --
>>>>>>>> Linux-cluster mailing list
>>>>>>>> Linux-cluster at redhat.com
>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>
>>>>>>> --
>>>>>>> Linux-cluster mailing list
>>>>>>> Linux-cluster at redhat.com
>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>
>>>>>> --
>>>>>> Linux-cluster mailing list
>>>>>> Linux-cluster at redhat.com
>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>
>>>>> --
>>>>> Linux-cluster mailing list
>>>>> Linux-cluster at redhat.com
>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>
>>>> --
>>>> - Daniela Anzellotti ------------------------------------
>>>>  INFN Roma - tel.: +39.06.49914282 - fax: +39.06.490354
>>>>  e-mail: daniela.anzellotti at roma1.infn.it
>>>> ---------------------------------------------------------
>>>>
>>>> --
>>>> Linux-cluster mailing list
>>>> Linux-cluster at redhat.com
>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>
>>> --
>>> Linux-cluster mailing list
>>> Linux-cluster at redhat.com
>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>
>> --
>> - Daniela Anzellotti ------------------------------------
>>  INFN Roma - tel.: +39.06.49914282 - fax: +39.06.490354
>>  e-mail: daniela.anzellotti at roma1.infn.it
>> ---------------------------------------------------------
>>
>> --
>> Linux-cluster mailing list
>> Linux-cluster at redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>
> 
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
> 

-- 
- Daniela Anzellotti ------------------------------------
  INFN Roma - tel.: +39.06.49914282 - fax: +39.06.490354
  e-mail: daniela.anzellotti at roma1.infn.it
---------------------------------------------------------




More information about the Linux-cluster mailing list