[Linux-cluster] openais issue

brem belguebli brem.belguebli at gmail.com
Tue Sep 29 20:44:39 UTC 2009


It looks correct, rgmanager seems to start on all nodes

what gives you clustat ?

If rgmanager doesn't show, check out the logs something may have gone wrong.


2009/9/29 Paras pradhan <pradhanparas at gmail.com>:
> Change to 7 and i got this log
>
> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting down
> Cluster Service Manager...
> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown complete, exiting
> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster Service
> Manager is stopped.
> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group
> Manager Starting
> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service Data
> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading Resource Rules
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building Resource Trees
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources defined
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading Failover Domains
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains defined
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events defined
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing Services
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services Initialized
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port Opened
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: Local UP
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst2 UP
> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst3 UP
> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1) Processed
> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1) Processed
> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1) Processed
> Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events processed
>
>
> Anything unusual here?
>
> Paras.
>
> On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli
> <brem.belguebli at gmail.com> wrote:
>> I use log_level=7 to have more debugging info.
>>
>> It seems 4 is not enough.
>>
>> Brem
>>
>>
>> 2009/9/29, Paras pradhan <pradhanparas at gmail.com>:
>>> Withe log_level of 3 I got only this
>>>
>>> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down
>>> Cluster Service Manager...
>>> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down
>>> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete, exiting
>>> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service
>>> Manager is stopped.
>>> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group
>>> Manager Starting
>>> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down
>>> Cluster Service Manager...
>>> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service
>>> Manager is stopped.
>>> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group
>>> Manager Starting
>>> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting down
>>>
>>> I do not know what the last line means.
>>>
>>> rgmanager version I am running is:
>>> rgmanager-2.0.52-1.el5.centos
>>>
>>> I don't what has gone wrong.
>>>
>>> Thanks
>>> Paras.
>>>
>>>
>>> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli
>>> <brem.belguebli at gmail.com> wrote:
>>> > you mean it stopped successfully on all the nodes but it is failing to
>>> > start only on node cvtst1 ?
>>> >
>>> > look at the following page  to make rgmanager more verbose. It 'll
>>> > help debug....
>>> >
>>> > http://sources.redhat.com/cluster/wiki/RGManager
>>> >
>>> > at Logging Configuration section
>>> >
>>> >
>>> >
>>> >
>>> > 2009/9/29 Paras pradhan <pradhanparas at gmail.com>:
>>> >> Brem,
>>> >>
>>> >> When I try to restart rgmanager on all the nodes, this time i do not
>>> >> see rgmanager running on the first node. But I do see on other 2
>>> >> nodes.
>>> >>
>>> >> Log on the first node:
>>> >>
>>> >> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group
>>> >> Manager Starting
>>> >> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down
>>> >> Cluster Service Manager...
>>> >> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down
>>> >> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete, exiting
>>> >> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service
>>> >> Manager is stopped.
>>> >> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group
>>> >> Manager Starting
>>> >>
>>> >> -
>>> >> It seems service is running ,  but I do not see rgmanger running using clustat
>>> >>
>>> >>
>>> >> Don't know what is going on.
>>> >>
>>> >> Thanks
>>> >> Paras.
>>> >>
>>> >>
>>> >> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli
>>> >> <brem.belguebli at gmail.com> wrote:
>>> >>> Paras,
>>> >>>
>>> >>> Another thing, it would have been more interesting to have a start
>>> >>> DEBUG not a stop.
>>> >>>
>>> >>> That's why I was asking you to first stop the vm manually on all your
>>> >>> nodes, stop eventually rgmanager on all the nodes to reset the
>>> >>> potential wrong states you may have, restart rgmanager.
>>> >>>
>>> >>> If your VM is configured to autostart, this will make it start.
>>> >>>
>>> >>> It should normally fail (as it does now). Send out your newly created
>>> >>> DEBUG file.
>>> >>>
>>> >>> 2009/9/29 brem belguebli <brem.belguebli at gmail.com>:
>>> >>>> Hi Paras,
>>> >>>>
>>> >>>>
>>> >>>> I don't know the xen/cluster combination well, but if I do remember
>>> >>>> well, I think I've read somewhere that when using xen you have to
>>> >>>> declare the use_virsh=0 key in the VM definition in the cluster.conf.
>>> >>>>
>>> >>>> This would make rgmanager use xm commands instead of virsh
>>> >>>> The DEBUG output shows clearly that you are using virsh to manage your
>>> >>>> VM instead of xm commands.
>>> >>>> Check out the RH docs about virtualization
>>> >>>>
>>> >>>> I'm not a 100% sure about that, I may be completely wrong.
>>> >>>>
>>> >>>> Brem
>>> >>>>
>>> >>>> 2009/9/28 Paras pradhan <pradhanparas at gmail.com>:
>>> >>>>> The only thing I noticed is the message after stopping the vm using xm
>>> >>>>> in all nodes and starting using clusvcadm is
>>> >>>>>
>>> >>>>> "Virtual machine guest1 is blocked"
>>> >>>>>
>>> >>>>> The whole DEBUG file is attached.
>>> >>>>>
>>> >>>>>
>>> >>>>> Thanks
>>> >>>>> Paras.
>>> >>>>>
>>> >>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli
>>> >>>>> <brem.belguebli at gmail.com> wrote:
>>> >>>>>> There's a problem with the script that is called by rgmanager to start
>>> >>>>>> the VM, I don't know what causes it
>>> >>>>>>
>>> >>>>>> May be you should try something like :
>>> >>>>>>
>>> >>>>>> 1) stop the VM on all nodes with xm commands
>>> >>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the following
>>> >>>>>> lines (after the #!/bin/bash ):
>>> >>>>>>   exec >/tmp/DEBUG 2>&1
>>> >>>>>>   set -x
>>> >>>>>> 3) start the VM with clusvcadm -e vm:guest1
>>> >>>>>>
>>> >>>>>> It should fail as it did before.
>>> >>>>>>
>>> >>>>>> edit the the /tmp/DEBUG file and you will be able to see where it
>>> >>>>>> fails (it may generate a lot of debug)
>>> >>>>>>
>>> >>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh
>>> >>>>>>
>>> >>>>>> Post the DEBUG file if you're not able to see where it fails.
>>> >>>>>>
>>> >>>>>> Brem
>>> >>>>>>
>>> >>>>>> 2009/9/26 Paras pradhan <pradhanparas at gmail.com>:
>>> >>>>>>> No I am not manually starting not using automatic init scripts.
>>> >>>>>>>
>>> >>>>>>> I started the vm using: clusvcadm -e vm:guest1
>>> >>>>>>>
>>> >>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few seconds it
>>> >>>>>>> says guest1 started . But after a while I can see the guest1 on all
>>> >>>>>>> three nodes.
>>> >>>>>>>
>>> >>>>>>> clustat says:
>>> >>>>>>>
>>> >>>>>>>  Service Name                                            Owner (Last)
>>> >>>>>>>                                          State
>>> >>>>>>>  ------- ----                                            ----- ------
>>> >>>>>>>                                          -----
>>> >>>>>>>  vm:guest1                                               (none)
>>> >>>>>>>                                          stopped
>>> >>>>>>>
>>> >>>>>>> But I can see the vm from xm li.
>>> >>>>>>>
>>> >>>>>>> This is what I can see from the log:
>>> >>>>>>>
>>> >>>>>>>
>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1"
>>> >>>>>>> returned 1 (generic error)
>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start
>>> >>>>>>> vm:guest1; return value: 1
>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1
>>> >>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is
>>> >>>>>>> recovering
>>> >>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering failed
>>> >>>>>>> service vm:guest1
>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1"
>>> >>>>>>> returned 1 (generic error)
>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start
>>> >>>>>>> vm:guest1; return value: 1
>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1
>>> >>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is
>>> >>>>>>> recovering
>>> >>>>>>>
>>> >>>>>>>
>>> >>>>>>> Paras.
>>> >>>>>>>
>>> >>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli
>>> >>>>>>> <brem.belguebli at gmail.com> wrote:
>>> >>>>>>>> Have you started  your VM via rgmanager (clusvcadm -e vm:guest1) or
>>> >>>>>>>> using xm commands out of cluster control  (or maybe a thru an
>>> >>>>>>>> automatic init script ?)
>>> >>>>>>>>
>>> >>>>>>>> When clustered, you should never be starting services (manually or
>>> >>>>>>>> thru automatic init script) out of cluster control
>>> >>>>>>>>
>>> >>>>>>>> The thing would be to stop your vm on all the nodes with the adequate
>>> >>>>>>>> xm command (not using xen myself) and try to start it with clusvcadm.
>>> >>>>>>>>
>>> >>>>>>>> Then see if it is started on all nodes (send clustat output)
>>> >>>>>>>>
>>> >>>>>>>>
>>> >>>>>>>>
>>> >>>>>>>> 2009/9/25 Paras pradhan <pradhanparas at gmail.com>:
>>> >>>>>>>>> Ok. Please see below. my vm is running on all nodes though clustat
>>> >>>>>>>>> says it is stopped.
>>> >>>>>>>>>
>>> >>>>>>>>> --
>>> >>>>>>>>> [root at cvtst1 ~]# clustat
>>> >>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009
>>> >>>>>>>>> Member Status: Quorate
>>> >>>>>>>>>
>>> >>>>>>>>>  Member Name                                                     ID   Status
>>> >>>>>>>>>  ------ ----                                                     ---- ------
>>> >>>>>>>>>  cvtst2                                                    1 Online, rgmanager
>>> >>>>>>>>>  cvtst1                                                     2 Online,
>>> >>>>>>>>> Local, rgmanager
>>> >>>>>>>>>  cvtst3                                                     3 Online, rgmanager
>>> >>>>>>>>>
>>> >>>>>>>>>  Service Name                                            Owner (Last)
>>> >>>>>>>>>                                          State
>>> >>>>>>>>>  ------- ----                                            ----- ------
>>> >>>>>>>>>                                          -----
>>> >>>>>>>>>  vm:guest1                                               (none)
>>> >>>>>>>>>                                          stopped
>>> >>>>>>>>> [root at cvtst1 ~]#
>>> >>>>>>>>>
>>> >>>>>>>>>
>>> >>>>>>>>> ---
>>> >>>>>>>>> o/p of xm li on cvtst1
>>> >>>>>>>>>
>>> >>>>>>>>> --
>>> >>>>>>>>> [root at cvtst1 ~]# xm li
>>> >>>>>>>>> Name                                      ID Mem(MiB) VCPUs State   Time(s)
>>> >>>>>>>>> Domain-0                                   0     3470     2 r-----  28939.4
>>> >>>>>>>>> guest1                                     7      511     1 -b----   7727.8
>>> >>>>>>>>>
>>> >>>>>>>>> o/p of xm li on cvtst2
>>> >>>>>>>>>
>>> >>>>>>>>> --
>>> >>>>>>>>> [root at cvtst2 ~]# xm li
>>> >>>>>>>>> Name                                      ID Mem(MiB) VCPUs State   Time(s)
>>> >>>>>>>>> Domain-0                                   0     3470     2 r-----  31558.9
>>> >>>>>>>>> guest1                                    21      511     1 -b----   7558.2
>>> >>>>>>>>> ---
>>> >>>>>>>>>
>>> >>>>>>>>> Thanks
>>> >>>>>>>>> Paras.
>>> >>>>>>>>>
>>> >>>>>>>>>
>>> >>>>>>>>>
>>> >>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli
>>> >>>>>>>>> <brem.belguebli at gmail.com> wrote:
>>> >>>>>>>>>> It looks like no.
>>> >>>>>>>>>>
>>> >>>>>>>>>> can you send an output of clustat  of when the VM is running on
>>> >>>>>>>>>> multiple nodes at the same time?
>>> >>>>>>>>>>
>>> >>>>>>>>>> And by the way, another one after having stopped (clusvcadm -s vm:guest1) ?
>>> >>>>>>>>>>
>>> >>>>>>>>>>
>>> >>>>>>>>>>
>>> >>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas at gmail.com>:
>>> >>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not being
>>> >>>>>>>>>>> properly handled by the cluster.
>>> >>>>>>>>>>>
>>> >>>>>>>>>>>
>>> >>>>>>>>>>> Thanks
>>> >>>>>>>>>>> Paras.
>>> >>>>>>>>>>>
>>> >>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan <pradhanparas at gmail.com> wrote:
>>> >>>>>>>>>>>> Ok.. here is my cluster.conf file
>>> >>>>>>>>>>>>
>>> >>>>>>>>>>>> --
>>> >>>>>>>>>>>> [root at cvtst1 cluster]# more cluster.conf
>>> >>>>>>>>>>>> <?xml version="1.0"?>
>>> >>>>>>>>>>>> <cluster alias="test" config_version="9" name="test">
>>> >>>>>>>>>>>>        <fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/>
>>> >>>>>>>>>>>>        <clusternodes>
>>> >>>>>>>>>>>>                <clusternode name="cvtst2" nodeid="1" votes="1">
>>> >>>>>>>>>>>>                        <fence/>
>>> >>>>>>>>>>>>                </clusternode>
>>> >>>>>>>>>>>>                <clusternode name="cvtst1" nodeid="2" votes="1">
>>> >>>>>>>>>>>>                        <fence/>
>>> >>>>>>>>>>>>                </clusternode>
>>> >>>>>>>>>>>>                <clusternode name="cvtst3" nodeid="3" votes="1">
>>> >>>>>>>>>>>>                        <fence/>
>>> >>>>>>>>>>>>                </clusternode>
>>> >>>>>>>>>>>>        </clusternodes>
>>> >>>>>>>>>>>>        <cman/>
>>> >>>>>>>>>>>>        <fencedevices/>
>>> >>>>>>>>>>>>        <rm>
>>> >>>>>>>>>>>>                <failoverdomains>
>>> >>>>>>>>>>>>                        <failoverdomain name="myfd1" nofailback="0" ordered="1" restricted="0">
>>> >>>>>>>>>>>>                                <failoverdomainnode name="cvtst2" priority="3"/>
>>> >>>>>>>>>>>>                                <failoverdomainnode name="cvtst1" priority="1"/>
>>> >>>>>>>>>>>>                                <failoverdomainnode name="cvtst3" priority="2"/>
>>> >>>>>>>>>>>>                        </failoverdomain>
>>> >>>>>>>>>>>>                </failoverdomains>
>>> >>>>>>>>>>>>                <resources/>
>>> >>>>>>>>>>>>                <vm autostart="1" domain="myfd1" exclusive="0" max_restarts="0"
>>> >>>>>>>>>>>> name="guest1" path="/vms" recovery="r
>>> >>>>>>>>>>>> estart" restart_expire_time="0"/>
>>> >>>>>>>>>>>>        </rm>
>>> >>>>>>>>>>>> </cluster>
>>> >>>>>>>>>>>> [root at cvtst1 cluster]#
>>> >>>>>>>>>>>> ------
>>> >>>>>>>>>>>>
>>> >>>>>>>>>>>> Thanks!
>>> >>>>>>>>>>>> Paras.
>>> >>>>>>>>>>>>
>>> >>>>>>>>>>>>
>>> >>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer <volker at ixolution.de> wrote:
>>> >>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500,
>>> >>>>>>>>>>>>> Paras pradhan <pradhanparas at gmail.com> wrote:
>>> >>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines. Now I am
>>> >>>>>>>>>>>>>> having another problem. When I start the my xen vm in one node, it
>>> >>>>>>>>>>>>>> also starts on other nodes. Which daemon controls  this?
>>> >>>>>>>>>>>>>
>>> >>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the rgmanager
>>> >>>>>>>>>>>>> package). To me, this sounds like a configuration problem. Maybe,
>>> >>>>>>>>>>>>> you can post your cluster.conf?
>>> >>>>>>>>>>>>>
>>> >>>>>>>>>>>>> Regards,
>>> >>>>>>>>>>>>> Volker
>>> >>>>>>>>>>>>>
>>> >>>>>>>>>>>>> --
>>> >>>>>>>>>>>>> Linux-cluster mailing list
>>> >>>>>>>>>>>>> Linux-cluster at redhat.com
>>> >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >>>>>>>>>>>>>
>>> >>>>>>>>>>>>
>>> >>>>>>>>>>>
>>> >>>>>>>>>>> --
>>> >>>>>>>>>>> Linux-cluster mailing list
>>> >>>>>>>>>>> Linux-cluster at redhat.com
>>> >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >>>>>>>>>>>
>>> >>>>>>>>>>
>>> >>>>>>>>>> --
>>> >>>>>>>>>> Linux-cluster mailing list
>>> >>>>>>>>>> Linux-cluster at redhat.com
>>> >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >>>>>>>>>>
>>> >>>>>>>>>
>>> >>>>>>>>> --
>>> >>>>>>>>> Linux-cluster mailing list
>>> >>>>>>>>> Linux-cluster at redhat.com
>>> >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >>>>>>>>>
>>> >>>>>>>>
>>> >>>>>>>> --
>>> >>>>>>>> Linux-cluster mailing list
>>> >>>>>>>> Linux-cluster at redhat.com
>>> >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >>>>>>>>
>>> >>>>>>>
>>> >>>>>>> --
>>> >>>>>>> Linux-cluster mailing list
>>> >>>>>>> Linux-cluster at redhat.com
>>> >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >>>>>>>
>>> >>>>>>
>>> >>>>>> --
>>> >>>>>> Linux-cluster mailing list
>>> >>>>>> Linux-cluster at redhat.com
>>> >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >>>>>>
>>> >>>>>
>>> >>>>> --
>>> >>>>> Linux-cluster mailing list
>>> >>>>> Linux-cluster at redhat.com
>>> >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >>>>>
>>> >>>>
>>> >>>
>>> >>> --
>>> >>> Linux-cluster mailing list
>>> >>> Linux-cluster at redhat.com
>>> >>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >>>
>>> >>
>>> >> --
>>> >> Linux-cluster mailing list
>>> >> Linux-cluster at redhat.com
>>> >> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >>
>>> >
>>> > --
>>> > Linux-cluster mailing list
>>> > Linux-cluster at redhat.com
>>> > https://www.redhat.com/mailman/listinfo/linux-cluster
>>> >
>>>
>>> --
>>> Linux-cluster mailing list
>>> Linux-cluster at redhat.com
>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>
>>
>> --
>> Linux-cluster mailing list
>> Linux-cluster at redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>




More information about the Linux-cluster mailing list