[Linux-cluster] clusvcadm : Could not connect to resource group manager

PARAM KRISH mkparam at gmail.com
Sat Aug 25 04:51:25 UTC 2012


I was able to download RHEL5 32bit ISO evaluation copy so i went with it
rather than downloading RHEL6 just for my PoC on two node cluster.

I was not blinding following RHEL4 doc which is why i did not look for
installing those packages once after i did "Cluster" and "ClusterStorage"
group install assuming everything must be there.
I started having a doubt about any missing packages because of those dlm
errors reported in the messages i sent earlier.

Aug 23 05:02:59 server1 dlm_controld[27892]: group_init error (nil) 111
Aug 23 05:02:59 server1 fenced[27886]: group_init error (nil) 111

Searching on some forums always gives me old replies about RHEL4 and error
which talks about some dlm rpm's, thats why.

-Param


On Sat, Aug 25, 2012 at 10:07 AM, Digimer <lists at alteeve.ca> wrote:

> I haven't used the rhcs stable 2 (version in RHEL 5) in some time, and I
> don't remember for sure. I don't think I've seen magma though.
>
> Why are you following the docs for RHEL 4, installing on RHEL 5 when
> RHEL 6 is the current version?
>
> On 08/25/2012 12:14 AM, PARAM KRISH wrote:
> > Digimer,
> >
> > I just want to confirm if i am missing any rpm's in my setup.
> >
> > I referred this site
> > http://www.centos.org/docs/4/html/rh-cs-en-4/ap-rhcs-sw-inst-cust.html
> > to install the rpm's in red hat 5.6 but these packages *magma**, *dlm**
> > are missing in the CD, does it mean these are really important ? I
> > installed using 'yum group install' for "Cluster" and "ClusterStorage"
> >
> > This is what the yum.log says from server1. Please confirm if there is
> > something missing.
> >
> > -Param
> >
> > On Fri, Aug 24, 2012 at 9:06 PM, PARAM KRISH <mkparam at gmail.com
> > <mailto:mkparam at gmail.com>> wrote:
> >
> >     Please find below the details.
> >
> >     1. cluster.conf : Please find attached herewith.
> >
> >     2. Is fencing in any form mandatory for a setup as simple as this ?
> >     I am just using two redhat5 VM's within VMware fusion in my Mac.
> >     What kind of fencing is applicable to this kind of setup, that can
> >     also benefit me ? All that i want to see from this PoC is to add
> >     some services like Apache, MySQL to see how quick and reliable the
> >     cluster to pick the service failures etc.,
> >
> >     3. Please find attached the messages from server1 and server2 when i
> >     did "service cman stop and start" on both nodes one after the other.
> >     In server1 both stop and start went fine but clustat showed "Could
> >     not connect to CMAN: connection refused". If i do "cman_tool join"
> >     on both nodes one after the other, things look green
> >
> >     Also, Am i missing any rpm's that are most important ? I used yum
> >     group install "Clustering" and "ClusterStorage" to install all the
> >     packages.
> >
> >     -Param
> >
> >     On Fri, Aug 24, 2012 at 7:28 PM, Digimer <lists at alteeve.ca
> >     <mailto:lists at alteeve.ca>> wrote:
> >
> >         A few things;
> >
> >         1. Please repost your cluster.conf file with line wraps in plain
> >         text.
> >
> >         2. Manual fencing is not supported in any way, please use real
> >         fencing,
> >         like IPMI, iLO, etc.
> >
> >         3. Please stop the cluster entirely, start 'tail -f -n 0
> >         /var/log/messages' on both nodes, then start cman, then start
> >         rgmanager.
> >         Please share the output from the logs.
> >
> >         Digimer
> >
> >         On 08/24/2012 06:43 AM, PARAM KRISH wrote:
> >         > Hi, Thanks for the help. I hope we are nearing to the problem.
> >         >
> >         > I enabled logging , this is how my cluster.conf looks like
> >         >
> >         > <?xml version="1.0"?>
> >         > <cluster alias="newCluster" config_version="16"
> name="newCluster">
> >         > <logging debug="on"/>
> >         > <cman expected_votes="1" two_node="1"/>
> >         > <clusternodes>
> >         > <clusternode name="server1" nodeid="1" votes="1">
> >         > <fence><method name="single"><device
> >         > name="human"/></method></fence></clusternode><clusternode
> >         name="server2"
> >         > nodeid="2" votes="1"><fence><method name="single"><device
> >         >
> >
> name="human"/></method></fence></clusternode></clusternodes><fencedevices>
> >         >
> >         >         </fencedevices><rm><failoverdomains><failoverdomain
> >         > name="failOver" nofailback="0" ordered="1"
> >         > restricted="0"><failoverdomainnode name="server1"
> >         > priority="1"/><failoverdomainnode name="server2"
> >         > priority="2"/></failoverdomain></failoverdomains><resources><ip
> >         > address="192.168.61.130" monitor_link="1"/><apache
> >         > config_file="conf/httpd.conf" name="httpd"
> >         server_root="/etc/httpd"
> >         > shutdown_wait="0"/></resources><service autostart="1"
> >         domain="failOver"
> >         > exclusive="1" name="Apache" recovery="relocate"><ip
> >         > address="192.168.61.130" monitor_link="1"><apache
> >         > config_file="conf/httpd.conf" name="Apache"
> >         server_root="/etc/httpd"
> >         > shutdown_wait="0"/></ip></service><service autostart="1"
> >         > domain="failOver" exclusive="1" name="website"
> >         recovery="relocate"><ip
> >         > ref="192.168.61.130"><apache
> >         > ref="httpd"/></ip></service></rm><fence_daemon clean_start="1"
> >         > post_fail_delay="0" post_join_delay="3"/><logging
> >         debug="on"/></cluster>
> >         >
> >         > There is no logging happening in /var/run/cluster/
> >         >
> >         > [root at server1 ~]# ls /var/run/cluster/
> >         > apache  ccsd.pid  ccsd.sock  rgmanager.sk
> >         <http://rgmanager.sk> <http://rgmanager.sk>
> >         >
> >         > I started resource manager in foreground and it says like ..
> >         >
> >         > failed acquiring lockspace: No such device
> >         > Locks not working!
> >         >
> >         > What next i could do ?
> >         >
> >         > -Param
> >         >
> >         > On Fri, Aug 24, 2012 at 3:18 PM, emmanuel segura
> >         <emi2fast at gmail.com <mailto:emi2fast at gmail.com>
> >         > <mailto:emi2fast at gmail.com <mailto:emi2fast at gmail.com>>>
> wrote:
> >         >
> >         >     /etc/init.d/rgmanager start or service rgmanager start
> >         >
> >         >
> >         >     2012/8/24 Heiko Nardmann <heiko.nardmann at itechnical.de
> >         <mailto:heiko.nardmann at itechnical.de>
> >         >     <mailto:heiko.nardmann at itechnical.de
> >         <mailto:heiko.nardmann at itechnical.de>>>
> >         >
> >         >         It is strange that strace shows that
> >         >         /var/run/cluster/rgmanager.sk <http://rgmanager.sk>
> >         <http://rgmanager.sk> is missing.
> >         >
> >         >         Normally it is helpful to see the complete
> >         cluster.conf. Could
> >         >         you provide that one?
> >         >
> >         >         Also of interest is /var/log/cluster/rgmanager.log -
> >         do you have
> >         >         debug enabled inside cluster.conf?
> >         >
> >         >         Maybe it is possible to start rgmanager in the
> >         foreground (-f)
> >         >         with strace? That might also be a way to show why the
> >         >         rgmanager.sk <http://rgmanager.sk>
> >         <http://rgmanager.sk> is missing ...
> >         >
> >         >         Just some ideas ...
> >         >
> >         >
> >         >         Kind regards,
> >         >
> >         >             Heiko
> >         >
> >         >         Am 24.08.2012 11 <tel:24.08.2012%2011>
> >         <tel:24.08.2012%2011>:04, schrieb PARAM KRISH:
> >         >
> >         >             All,
> >         >
> >         >             I am trying to setup a simple two node cluster in
> >         my laptop
> >         >             using two RHEL VM's.
> >         >
> >         >             Everything looks just fine to me but i am unable
> >         to enable a
> >         >             apache service though it works beautifully when
> >         tried with
> >         >             "rg_test test" on both the nodes.
> >         >
> >         >             What could be the problem ? Please help. I am a
> >         novice in
> >         >             red hat cluster but learnt a bit of it in the last
> >         few days
> >         >             while trying to fix all the problems encountered.
> >         >
> >         >             Here are the details.
> >         >
> >         >             [root at server1 ~]# clustat
> >         >             Cluster Status for newCluster @ Thu Aug 23
> >         00:29:32 2012
> >         >             Member Status: Quorate
> >         >
> >         >              Member Name                 ID   Status
> >         >              ------ ----                 ---- ------
> >         >              server1                     1 Online, Local
> >         >              server2                     2 Online
> >         >
> >         >             [root at server1 ~]# clustat -x
> >         >             <?xml version="1.0"?>
> >         >             <clustat version="4.1.1">
> >         >               <cluster name="newCluster" id="43188"
> >         generation="250536"/>
> >         >               <quorum quorate="1" groupmember="0"/>
> >         >               <nodes>
> >         >                 <node name="server1" state="1" local="1"
> >         estranged="0"
> >         >             rgmanager="0" rgmanager_master="0" qdisk="0"
> >         >             nodeid="0x00000001"/>
> >         >                 <node name="server2" state="1" local="0"
> >         estranged="0"
> >         >             rgmanager="0" rgmanager_master="0" qdisk="0"
> >         >             nodeid="0x00000002"/>  </nodes>
> >         >             </clustat>
> >         >
> >         >             [root at server2 ~]# clustat
> >         >             Cluster Status for newCluster @ Thu Aug 23
> >         03:13:34 2012
> >         >             Member Status: Quorate
> >         >
> >         >              Member Name                 ID   Status
> >         >              ------ ----                 ---- ------
> >         >              server1                     1 Online
> >         >              server2                     2 Online, Local
> >         >
> >         >             [root at server2 ~]# clustat -x
> >         >             <?xml version="1.0"?>
> >         >             <clustat version="4.1.1">
> >         >               <cluster name="newCluster" id="43188"
> >         generation="250536"/>
> >         >               <quorum quorate="1" groupmember="0"/>
> >         >               <nodes>
> >         >                 <node name="server1" state="1" local="0"
> >         estranged="0"
> >         >             rgmanager="0" rgmanager_master="0" qdisk="0"
> >         >             nodeid="0x00000001"/>
> >         >                 <node name="server2" state="1" local="1"
> >         estranged="0"
> >         >             rgmanager="0" rgmanager_master="0" qdisk="0"
> >         >             nodeid="0x00000002"/>
> >         >               </nodes>
> >         >             </clustat>
> >         >
> >         >
> >         >             [root at server2 ~]# clusvcadm -e Apache
> >         >             Local machine trying to enable
> >         service:Apache...Could not
> >         >             connect to resource group manager
> >         >
> >         >             strace cluvcsadm -e Apache
> >         >             ...
> >         >             stat64(1, {st_mode=S_IFCHR|0620,
> >         st_rdev=makedev(136, 4),
> >         >             ...}) = 0
> >         >             mmap2(NULL, 4096, PROT_READ|PROT_WRITE,
> >         >             MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0xb7fb5000
> >         >             write(1, "Local machine trying to enable s"...,
> >         48Local
> >         >             machine trying to enable service:Apache...) = 48
> >         >             socket(PF_FILE, SOCK_STREAM, 0)         = 5
> >         >             connect(5, {sa_family=AF_FILE,
> >         >             path="/var/run/cluster/rgmanag__er.sk
> >         <http://rgmanag__er.sk> <http://rgmanager.sk>
> >         >             <http://rgmanager.sk>"...}, 110) = -1 ENOENT (No
> >         such file
> >         >             or directory)
> >         >
> >         >             close(5)                                = 0
> >         >             write(1, "Could not connect to resource gr"...,
> >         44Could not
> >         >             connect to resource group manager
> >         >             ) = 44
> >         >             exit_group(1)                           = ?
> >         >
> >         >
> >         >             [root at server1 ~]# hostname
> >         >             server1.localdomain
> >         >
> >         >             [root at server1 ~]# cat /etc/hosts
> >         >             # Do not remove the following line, or various
> >         programs
> >         >             # that require network functionality will fail.
> >         >             #127.0.0.1              server1.localdomain server1
> >         >             localhost.localdomain localhost
> >         >             192.168.61.132 server1.localdomain server1
> >         >             192.168.61.133 server2.localdomain server2
> >         >             ::1             localhost6.localdomain6 localhost6
> >         >
> >         >
> >         >             Package versions :
> >         >             luci-0.12.2-24.el5
> >         >             ricci-0.12.2-24.el5
> >         >             rgmanager-2.0.52-9.el5
> >         >             modcluster-0.12.1-2.el5
> >         >             cluster-cim-0.12.1-2.el5
> >         >             system-config-cluster-1.0.57-7
> >         >             lvm2-cluster-2.02.74-3.el5
> >         >             cluster-snmp-0.12.1-2.el5
> >         >
> >         >             [root at server1 log]# cman_tool status
> >         >             Version: 6.2.0
> >         >             Config Version: 15
> >         >             Cluster Name: newCluster
> >         >             Cluster Id: 43188
> >         >             Cluster Member: Yes
> >         >             Cluster Generation: 250536
> >         >             Membership state: Cluster-Member
> >         >             Nodes: 2
> >         >             Expected votes: 1
> >         >             Total votes: 2
> >         >             Quorum: 1
> >         >             Active subsystems: 2
> >         >             Flags: 2node
> >         >             Ports Bound: 0
> >         >             Node name: server1
> >         >             Node ID: 1
> >         >             Multicast addresses: 239.192.168.93
> >         >             Node addresses: 192.168.61.132
> >         >
> >         >             Redhat :Red Hat Enterprise Linux Server release 5.6
> >         >             (Tikanga)2.6.18-238.el5xen
> >         >
> >         >             [root at server1 log]# service rgmanager status
> >         >             clurgmgrd (pid  9775) is running...
> >         >
> >         >             [root at server1 log]# netstat -na | grep 11111
> >         >             tcp        0      0 0.0.0.0:11111
> >         <http://0.0.0.0:11111> <http://0.0.0.0:11111>
> >         >             <http://0.0.0.0:11111>         0.0.0.0:*
> >         >             LISTEN
> >         >
> >         >
> >         >             Please let me know if you can help. One thing i
> >         noticed was
> >         >             that in the "clustat" it does not show "rgmanager"
> >         against
> >         >             both the nodes but i see the service is just
> >         running fine.
> >         >
> >         >             *Note : No iptables, no SELinux enabled.*
> >         >             *
> >         >
> >         >             *
> >         >             Hope i have given all the details required to help
> me
> >         >             quickly. Thanks.
> >         >
> >         >             -Param
> >         >
> >         >
> >         >
> >         >         --
> >         >         Linux-cluster mailing list
> >         >         Linux-cluster at redhat.com
> >         <mailto:Linux-cluster at redhat.com>
> >         <mailto:Linux-cluster at redhat.com <mailto:
> Linux-cluster at redhat.com>>
> >         >
> https://www.redhat.com/__mailman/listinfo/linux-cluster
> >         >         <https://www.redhat.com/mailman/listinfo/linux-cluster
> >
> >         >
> >         >
> >         >
> >         >
> >         >     --
> >         >     esta es mi vida e me la vivo hasta que dios quiera
> >         >
> >         >     --
> >         >     Linux-cluster mailing list
> >         >     Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
> >         <mailto:Linux-cluster at redhat.com <mailto:
> Linux-cluster at redhat.com>>
> >         >     https://www.redhat.com/mailman/listinfo/linux-cluster
> >         >
> >         >
> >         >
> >         >
> >         > --
> >         > Linux-cluster mailing list
> >         > Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
> >         > https://www.redhat.com/mailman/listinfo/linux-cluster
> >         >
> >
> >
> >         --
> >         Digimer
> >         Papers and Projects: https://alteeve.ca
> >
> >
> >
>
>
> --
> Digimer
> Papers and Projects: https://alteeve.ca
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20120825/cf47206a/attachment.htm>


More information about the Linux-cluster mailing list