[Linux-cluster] clusvcadm : Could not connect to resource group manager
PARAM KRISH
mkparam at gmail.com
Sat Aug 25 04:51:25 UTC 2012
I was able to download RHEL5 32bit ISO evaluation copy so i went with it
rather than downloading RHEL6 just for my PoC on two node cluster.
I was not blinding following RHEL4 doc which is why i did not look for
installing those packages once after i did "Cluster" and "ClusterStorage"
group install assuming everything must be there.
I started having a doubt about any missing packages because of those dlm
errors reported in the messages i sent earlier.
Aug 23 05:02:59 server1 dlm_controld[27892]: group_init error (nil) 111
Aug 23 05:02:59 server1 fenced[27886]: group_init error (nil) 111
Searching on some forums always gives me old replies about RHEL4 and error
which talks about some dlm rpm's, thats why.
-Param
On Sat, Aug 25, 2012 at 10:07 AM, Digimer <lists at alteeve.ca> wrote:
> I haven't used the rhcs stable 2 (version in RHEL 5) in some time, and I
> don't remember for sure. I don't think I've seen magma though.
>
> Why are you following the docs for RHEL 4, installing on RHEL 5 when
> RHEL 6 is the current version?
>
> On 08/25/2012 12:14 AM, PARAM KRISH wrote:
> > Digimer,
> >
> > I just want to confirm if i am missing any rpm's in my setup.
> >
> > I referred this site
> > http://www.centos.org/docs/4/html/rh-cs-en-4/ap-rhcs-sw-inst-cust.html
> > to install the rpm's in red hat 5.6 but these packages *magma**, *dlm**
> > are missing in the CD, does it mean these are really important ? I
> > installed using 'yum group install' for "Cluster" and "ClusterStorage"
> >
> > This is what the yum.log says from server1. Please confirm if there is
> > something missing.
> >
> > -Param
> >
> > On Fri, Aug 24, 2012 at 9:06 PM, PARAM KRISH <mkparam at gmail.com
> > <mailto:mkparam at gmail.com>> wrote:
> >
> > Please find below the details.
> >
> > 1. cluster.conf : Please find attached herewith.
> >
> > 2. Is fencing in any form mandatory for a setup as simple as this ?
> > I am just using two redhat5 VM's within VMware fusion in my Mac.
> > What kind of fencing is applicable to this kind of setup, that can
> > also benefit me ? All that i want to see from this PoC is to add
> > some services like Apache, MySQL to see how quick and reliable the
> > cluster to pick the service failures etc.,
> >
> > 3. Please find attached the messages from server1 and server2 when i
> > did "service cman stop and start" on both nodes one after the other.
> > In server1 both stop and start went fine but clustat showed "Could
> > not connect to CMAN: connection refused". If i do "cman_tool join"
> > on both nodes one after the other, things look green
> >
> > Also, Am i missing any rpm's that are most important ? I used yum
> > group install "Clustering" and "ClusterStorage" to install all the
> > packages.
> >
> > -Param
> >
> > On Fri, Aug 24, 2012 at 7:28 PM, Digimer <lists at alteeve.ca
> > <mailto:lists at alteeve.ca>> wrote:
> >
> > A few things;
> >
> > 1. Please repost your cluster.conf file with line wraps in plain
> > text.
> >
> > 2. Manual fencing is not supported in any way, please use real
> > fencing,
> > like IPMI, iLO, etc.
> >
> > 3. Please stop the cluster entirely, start 'tail -f -n 0
> > /var/log/messages' on both nodes, then start cman, then start
> > rgmanager.
> > Please share the output from the logs.
> >
> > Digimer
> >
> > On 08/24/2012 06:43 AM, PARAM KRISH wrote:
> > > Hi, Thanks for the help. I hope we are nearing to the problem.
> > >
> > > I enabled logging , this is how my cluster.conf looks like
> > >
> > > <?xml version="1.0"?>
> > > <cluster alias="newCluster" config_version="16"
> name="newCluster">
> > > <logging debug="on"/>
> > > <cman expected_votes="1" two_node="1"/>
> > > <clusternodes>
> > > <clusternode name="server1" nodeid="1" votes="1">
> > > <fence><method name="single"><device
> > > name="human"/></method></fence></clusternode><clusternode
> > name="server2"
> > > nodeid="2" votes="1"><fence><method name="single"><device
> > >
> >
> name="human"/></method></fence></clusternode></clusternodes><fencedevices>
> > >
> > > </fencedevices><rm><failoverdomains><failoverdomain
> > > name="failOver" nofailback="0" ordered="1"
> > > restricted="0"><failoverdomainnode name="server1"
> > > priority="1"/><failoverdomainnode name="server2"
> > > priority="2"/></failoverdomain></failoverdomains><resources><ip
> > > address="192.168.61.130" monitor_link="1"/><apache
> > > config_file="conf/httpd.conf" name="httpd"
> > server_root="/etc/httpd"
> > > shutdown_wait="0"/></resources><service autostart="1"
> > domain="failOver"
> > > exclusive="1" name="Apache" recovery="relocate"><ip
> > > address="192.168.61.130" monitor_link="1"><apache
> > > config_file="conf/httpd.conf" name="Apache"
> > server_root="/etc/httpd"
> > > shutdown_wait="0"/></ip></service><service autostart="1"
> > > domain="failOver" exclusive="1" name="website"
> > recovery="relocate"><ip
> > > ref="192.168.61.130"><apache
> > > ref="httpd"/></ip></service></rm><fence_daemon clean_start="1"
> > > post_fail_delay="0" post_join_delay="3"/><logging
> > debug="on"/></cluster>
> > >
> > > There is no logging happening in /var/run/cluster/
> > >
> > > [root at server1 ~]# ls /var/run/cluster/
> > > apache ccsd.pid ccsd.sock rgmanager.sk
> > <http://rgmanager.sk> <http://rgmanager.sk>
> > >
> > > I started resource manager in foreground and it says like ..
> > >
> > > failed acquiring lockspace: No such device
> > > Locks not working!
> > >
> > > What next i could do ?
> > >
> > > -Param
> > >
> > > On Fri, Aug 24, 2012 at 3:18 PM, emmanuel segura
> > <emi2fast at gmail.com <mailto:emi2fast at gmail.com>
> > > <mailto:emi2fast at gmail.com <mailto:emi2fast at gmail.com>>>
> wrote:
> > >
> > > /etc/init.d/rgmanager start or service rgmanager start
> > >
> > >
> > > 2012/8/24 Heiko Nardmann <heiko.nardmann at itechnical.de
> > <mailto:heiko.nardmann at itechnical.de>
> > > <mailto:heiko.nardmann at itechnical.de
> > <mailto:heiko.nardmann at itechnical.de>>>
> > >
> > > It is strange that strace shows that
> > > /var/run/cluster/rgmanager.sk <http://rgmanager.sk>
> > <http://rgmanager.sk> is missing.
> > >
> > > Normally it is helpful to see the complete
> > cluster.conf. Could
> > > you provide that one?
> > >
> > > Also of interest is /var/log/cluster/rgmanager.log -
> > do you have
> > > debug enabled inside cluster.conf?
> > >
> > > Maybe it is possible to start rgmanager in the
> > foreground (-f)
> > > with strace? That might also be a way to show why the
> > > rgmanager.sk <http://rgmanager.sk>
> > <http://rgmanager.sk> is missing ...
> > >
> > > Just some ideas ...
> > >
> > >
> > > Kind regards,
> > >
> > > Heiko
> > >
> > > Am 24.08.2012 11 <tel:24.08.2012%2011>
> > <tel:24.08.2012%2011>:04, schrieb PARAM KRISH:
> > >
> > > All,
> > >
> > > I am trying to setup a simple two node cluster in
> > my laptop
> > > using two RHEL VM's.
> > >
> > > Everything looks just fine to me but i am unable
> > to enable a
> > > apache service though it works beautifully when
> > tried with
> > > "rg_test test" on both the nodes.
> > >
> > > What could be the problem ? Please help. I am a
> > novice in
> > > red hat cluster but learnt a bit of it in the last
> > few days
> > > while trying to fix all the problems encountered.
> > >
> > > Here are the details.
> > >
> > > [root at server1 ~]# clustat
> > > Cluster Status for newCluster @ Thu Aug 23
> > 00:29:32 2012
> > > Member Status: Quorate
> > >
> > > Member Name ID Status
> > > ------ ---- ---- ------
> > > server1 1 Online, Local
> > > server2 2 Online
> > >
> > > [root at server1 ~]# clustat -x
> > > <?xml version="1.0"?>
> > > <clustat version="4.1.1">
> > > <cluster name="newCluster" id="43188"
> > generation="250536"/>
> > > <quorum quorate="1" groupmember="0"/>
> > > <nodes>
> > > <node name="server1" state="1" local="1"
> > estranged="0"
> > > rgmanager="0" rgmanager_master="0" qdisk="0"
> > > nodeid="0x00000001"/>
> > > <node name="server2" state="1" local="0"
> > estranged="0"
> > > rgmanager="0" rgmanager_master="0" qdisk="0"
> > > nodeid="0x00000002"/> </nodes>
> > > </clustat>
> > >
> > > [root at server2 ~]# clustat
> > > Cluster Status for newCluster @ Thu Aug 23
> > 03:13:34 2012
> > > Member Status: Quorate
> > >
> > > Member Name ID Status
> > > ------ ---- ---- ------
> > > server1 1 Online
> > > server2 2 Online, Local
> > >
> > > [root at server2 ~]# clustat -x
> > > <?xml version="1.0"?>
> > > <clustat version="4.1.1">
> > > <cluster name="newCluster" id="43188"
> > generation="250536"/>
> > > <quorum quorate="1" groupmember="0"/>
> > > <nodes>
> > > <node name="server1" state="1" local="0"
> > estranged="0"
> > > rgmanager="0" rgmanager_master="0" qdisk="0"
> > > nodeid="0x00000001"/>
> > > <node name="server2" state="1" local="1"
> > estranged="0"
> > > rgmanager="0" rgmanager_master="0" qdisk="0"
> > > nodeid="0x00000002"/>
> > > </nodes>
> > > </clustat>
> > >
> > >
> > > [root at server2 ~]# clusvcadm -e Apache
> > > Local machine trying to enable
> > service:Apache...Could not
> > > connect to resource group manager
> > >
> > > strace cluvcsadm -e Apache
> > > ...
> > > stat64(1, {st_mode=S_IFCHR|0620,
> > st_rdev=makedev(136, 4),
> > > ...}) = 0
> > > mmap2(NULL, 4096, PROT_READ|PROT_WRITE,
> > > MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0xb7fb5000
> > > write(1, "Local machine trying to enable s"...,
> > 48Local
> > > machine trying to enable service:Apache...) = 48
> > > socket(PF_FILE, SOCK_STREAM, 0) = 5
> > > connect(5, {sa_family=AF_FILE,
> > > path="/var/run/cluster/rgmanag__er.sk
> > <http://rgmanag__er.sk> <http://rgmanager.sk>
> > > <http://rgmanager.sk>"...}, 110) = -1 ENOENT (No
> > such file
> > > or directory)
> > >
> > > close(5) = 0
> > > write(1, "Could not connect to resource gr"...,
> > 44Could not
> > > connect to resource group manager
> > > ) = 44
> > > exit_group(1) = ?
> > >
> > >
> > > [root at server1 ~]# hostname
> > > server1.localdomain
> > >
> > > [root at server1 ~]# cat /etc/hosts
> > > # Do not remove the following line, or various
> > programs
> > > # that require network functionality will fail.
> > > #127.0.0.1 server1.localdomain server1
> > > localhost.localdomain localhost
> > > 192.168.61.132 server1.localdomain server1
> > > 192.168.61.133 server2.localdomain server2
> > > ::1 localhost6.localdomain6 localhost6
> > >
> > >
> > > Package versions :
> > > luci-0.12.2-24.el5
> > > ricci-0.12.2-24.el5
> > > rgmanager-2.0.52-9.el5
> > > modcluster-0.12.1-2.el5
> > > cluster-cim-0.12.1-2.el5
> > > system-config-cluster-1.0.57-7
> > > lvm2-cluster-2.02.74-3.el5
> > > cluster-snmp-0.12.1-2.el5
> > >
> > > [root at server1 log]# cman_tool status
> > > Version: 6.2.0
> > > Config Version: 15
> > > Cluster Name: newCluster
> > > Cluster Id: 43188
> > > Cluster Member: Yes
> > > Cluster Generation: 250536
> > > Membership state: Cluster-Member
> > > Nodes: 2
> > > Expected votes: 1
> > > Total votes: 2
> > > Quorum: 1
> > > Active subsystems: 2
> > > Flags: 2node
> > > Ports Bound: 0
> > > Node name: server1
> > > Node ID: 1
> > > Multicast addresses: 239.192.168.93
> > > Node addresses: 192.168.61.132
> > >
> > > Redhat :Red Hat Enterprise Linux Server release 5.6
> > > (Tikanga)2.6.18-238.el5xen
> > >
> > > [root at server1 log]# service rgmanager status
> > > clurgmgrd (pid 9775) is running...
> > >
> > > [root at server1 log]# netstat -na | grep 11111
> > > tcp 0 0 0.0.0.0:11111
> > <http://0.0.0.0:11111> <http://0.0.0.0:11111>
> > > <http://0.0.0.0:11111> 0.0.0.0:*
> > > LISTEN
> > >
> > >
> > > Please let me know if you can help. One thing i
> > noticed was
> > > that in the "clustat" it does not show "rgmanager"
> > against
> > > both the nodes but i see the service is just
> > running fine.
> > >
> > > *Note : No iptables, no SELinux enabled.*
> > > *
> > >
> > > *
> > > Hope i have given all the details required to help
> me
> > > quickly. Thanks.
> > >
> > > -Param
> > >
> > >
> > >
> > > --
> > > Linux-cluster mailing list
> > > Linux-cluster at redhat.com
> > <mailto:Linux-cluster at redhat.com>
> > <mailto:Linux-cluster at redhat.com <mailto:
> Linux-cluster at redhat.com>>
> > >
> https://www.redhat.com/__mailman/listinfo/linux-cluster
> > > <https://www.redhat.com/mailman/listinfo/linux-cluster
> >
> > >
> > >
> > >
> > >
> > > --
> > > esta es mi vida e me la vivo hasta que dios quiera
> > >
> > > --
> > > Linux-cluster mailing list
> > > Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
> > <mailto:Linux-cluster at redhat.com <mailto:
> Linux-cluster at redhat.com>>
> > > https://www.redhat.com/mailman/listinfo/linux-cluster
> > >
> > >
> > >
> > >
> > > --
> > > Linux-cluster mailing list
> > > Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
> > > https://www.redhat.com/mailman/listinfo/linux-cluster
> > >
> >
> >
> > --
> > Digimer
> > Papers and Projects: https://alteeve.ca
> >
> >
> >
>
>
> --
> Digimer
> Papers and Projects: https://alteeve.ca
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20120825/cf47206a/attachment.htm>
More information about the Linux-cluster
mailing list