From pradhanparas at gmail.com Wed May 7 15:14:20 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 10:14:20 -0500 Subject: [Linux-cluster] clusvcadm Message-ID: Hi, We had some network problem the other day and today I noticed the clusvcadm commands are not working. For example: it does not want to stop the service, migrate a vm etc etc. On one of the nodes clustat does not show any running services. I should restart the rgmanager? This is RHEL 5. Thanks Paras. -------------- next part -------------- An HTML attachment was scrubbed... URL: From emi2fast at gmail.com Wed May 7 16:36:21 2014 From: emi2fast at gmail.com (emmanuel segura) Date: Wed, 7 May 2014 18:36:21 +0200 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: sorry, but not enough information for help you 2014-05-07 17:14 GMT+02:00 Paras pradhan : > Hi, > We had some network problem the other day and today I noticed the > clusvcadm commands are not working. For example: it does not want to stop > the service, migrate a vm etc etc. On one of the nodes clustat does not > show any running services. I should restart the rgmanager? > > This is RHEL 5. > > Thanks > Paras. > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera -------------- next part -------------- An HTML attachment was scrubbed... URL: From jmd_singhsaini at yahoo.com Wed May 7 16:41:31 2014 From: jmd_singhsaini at yahoo.com (jmd_singhsaini at yahoo.com) Date: Thu, 8 May 2014 00:41:31 +0800 (SGT) Subject: [Linux-cluster] clusvcadm In-Reply-To: Message-ID: <1399480891.16076.YahooMailAndroidMobile@web193503.mail.sg3.yahoo.com> L Sent from Yahoo Mail on Android -------------- next part -------------- An HTML attachment was scrubbed... URL: From pradhanparas at gmail.com Wed May 7 17:20:48 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 12:20:48 -0500 Subject: [Linux-cluster] clusvcadm In-Reply-To: <1399480891.16076.YahooMailAndroidMobile@web193503.mail.sg3.yahoo.com> References: <1399480891.16076.YahooMailAndroidMobile@web193503.mail.sg3.yahoo.com> Message-ID: Emmanuel, When I do : clusvcadm -s vm:vm.domain Local machine stopping vm:vm.domain... This process never ends and on one of the nodes the services are not shown #clustat Service states unavailable: Temporary failure; try again Cluster Status for vprd @ Wed May 7 12:04:37 2014 Member Status: Quorate Member Name ID Status ------ ---- ---- ------ vprd2.domain 1 Online, Local vprd1.domain 2 Online vprd3.domain 3 Online /dev/dm-3 0 Online, Quorum Disk [root at cvprd2 log]# ---- Thanks Paras. On Wed, May 7, 2014 at 11:41 AM, jmd_singhsaini at yahoo.com < jmd_singhsaini at yahoo.com> wrote: > L > > Sent from Yahoo Mail on Android > > ------------------------------ > * From: * Paras pradhan ; > * To: * linux clustering ; > * Subject: * [Linux-cluster] clusvcadm > * Sent: * Wed, May 7, 2014 3:14:20 PM > > Hi, > We had some network problem the other day and today I noticed the > clusvcadm commands are not working. For example: it does not want to stop > the service, migrate a vm etc etc. On one of the nodes clustat does not > show any running services. I should restart the rgmanager? > > This is RHEL 5. > > Thanks > Paras. > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -------------- next part -------------- An HTML attachment was scrubbed... URL: From haldunce at gmail.com Wed May 7 17:28:52 2014 From: haldunce at gmail.com (hugo aldunce) Date: Wed, 7 May 2014 13:28:52 -0400 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: ohh good luck! 2014-05-07 11:14 GMT-04:00 Paras pradhan : > Hi, > We had some network problem the other day and today I noticed the > clusvcadm commands are not working. For example: it does not want to stop > the service, migrate a vm etc etc. On one of the nodes clustat does not > show any running services. I should restart the rgmanager? > > This is RHEL 5. > > Thanks > Paras. > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- --------------------------------------------------------------------------------------------------------------------- Hugo Aldunce E Tel. 09 82121045 mail: haldunce at gmail.com --------------------------------------------------------------------------------------------------------------------- -------------- next part -------------- An HTML attachment was scrubbed... URL: From emi2fast at gmail.com Wed May 7 17:46:56 2014 From: emi2fast at gmail.com (emmanuel segura) Date: Wed, 7 May 2014 19:46:56 +0200 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: cman_tool services? 2014-05-07 19:28 GMT+02:00 hugo aldunce : > ohh good luck! > > > 2014-05-07 11:14 GMT-04:00 Paras pradhan : > >> Hi, >> We had some network problem the other day and today I noticed the >> clusvcadm commands are not working. For example: it does not want to stop >> the service, migrate a vm etc etc. On one of the nodes clustat does not >> show any running services. I should restart the rgmanager? >> >> This is RHEL 5. >> >> Thanks >> Paras. >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > > --------------------------------------------------------------------------------------------------------------------- > Hugo Aldunce E > Tel. 09 82121045 > mail: haldunce at gmail.com > > --------------------------------------------------------------------------------------------------------------------- > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera -------------- next part -------------- An HTML attachment was scrubbed... URL: From pradhanparas at gmail.com Wed May 7 17:52:44 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 12:52:44 -0500 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: Thats looks good. #cman_tool services type level name id state fence 0 default 00010001 none [1 2 3] dlm 1 clvmd 00020001 none [1 2 3] dlm 1 guest_comp_vms1 00020003 none [1 2 3] dlm 1 guest_comp_vms2 00040003 none [1 2 3] dlm 1 guest_comp_vms3 00060003 none [1 2 3] dlm 1 rgmanager 00030001 none [1 2 3] gfs 2 guest_comp_vms1 00010003 none [1 2 3] gfs 2 guest_comp_vms2 00030003 none [1 2 3] gfs 2 guest_comp_vms3 00050003 none [1 2 3] On Wed, May 7, 2014 at 12:46 PM, emmanuel segura wrote: > cman_tool services? > > > 2014-05-07 19:28 GMT+02:00 hugo aldunce : > > ohh good luck! >> >> >> 2014-05-07 11:14 GMT-04:00 Paras pradhan : >> >>> Hi, >>> We had some network problem the other day and today I noticed the >>> clusvcadm commands are not working. For example: it does not want to stop >>> the service, migrate a vm etc etc. On one of the nodes clustat does not >>> show any running services. I should restart the rgmanager? >>> >>> This is RHEL 5. >>> >>> Thanks >>> Paras. >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> >> -- >> >> --------------------------------------------------------------------------------------------------------------------- >> Hugo Aldunce E >> Tel. 09 82121045 >> mail: haldunce at gmail.com >> >> --------------------------------------------------------------------------------------------------------------------- >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -------------- next part -------------- An HTML attachment was scrubbed... URL: From emi2fast at gmail.com Wed May 7 18:07:42 2014 From: emi2fast at gmail.com (emmanuel segura) Date: Wed, 7 May 2014 20:07:42 +0200 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: i saw your rgmanager lockspace is there, you see any error in your msg? can show your cluster config? 2014-05-07 19:52 GMT+02:00 Paras pradhan : > Thats looks good. > > #cman_tool services > > type level name id state > > fence 0 default 00010001 none > > [1 2 3] > > dlm 1 clvmd 00020001 none > > [1 2 3] > > dlm 1 guest_comp_vms1 00020003 none > > [1 2 3] > > dlm 1 guest_comp_vms2 00040003 none > > [1 2 3] > > dlm 1 guest_comp_vms3 00060003 none > > [1 2 3] > > dlm 1 rgmanager 00030001 none > > [1 2 3] > > gfs 2 guest_comp_vms1 00010003 none > > [1 2 3] > > gfs 2 guest_comp_vms2 00030003 none > > [1 2 3] > > gfs 2 guest_comp_vms3 00050003 none > > [1 2 3] > > > On Wed, May 7, 2014 at 12:46 PM, emmanuel segura wrote: > >> cman_tool services? >> >> >> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >> >> ohh good luck! >>> >>> >>> 2014-05-07 11:14 GMT-04:00 Paras pradhan : >>> >>>> Hi, >>>> We had some network problem the other day and today I noticed the >>>> clusvcadm commands are not working. For example: it does not want to stop >>>> the service, migrate a vm etc etc. On one of the nodes clustat does not >>>> show any running services. I should restart the rgmanager? >>>> >>>> This is RHEL 5. >>>> >>>> Thanks >>>> Paras. >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> >>> -- >>> >>> --------------------------------------------------------------------------------------------------------------------- >>> Hugo Aldunce E >>> Tel. 09 82121045 >>> mail: haldunce at gmail.com >>> >>> --------------------------------------------------------------------------------------------------------------------- >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera -------------- next part -------------- An HTML attachment was scrubbed... URL: From pradhanparas at gmail.com Wed May 7 18:24:07 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 13:24:07 -0500 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: Oh. How did you see that? Here is the cluster.conf http://pastebin.com/DveLMGXT Thanks! -Paras. On Wed, May 7, 2014 at 1:07 PM, emmanuel segura wrote: > i saw your rgmanager lockspace is there, you see any error in your msg? > can show your cluster config? > > > 2014-05-07 19:52 GMT+02:00 Paras pradhan : > > Thats looks good. >> >> #cman_tool services >> >> type level name id state >> >> fence 0 default 00010001 none >> >> [1 2 3] >> >> dlm 1 clvmd 00020001 none >> >> [1 2 3] >> >> dlm 1 guest_comp_vms1 00020003 none >> >> [1 2 3] >> >> dlm 1 guest_comp_vms2 00040003 none >> >> [1 2 3] >> >> dlm 1 guest_comp_vms3 00060003 none >> >> [1 2 3] >> >> dlm 1 rgmanager 00030001 none >> >> [1 2 3] >> >> gfs 2 guest_comp_vms1 00010003 none >> >> [1 2 3] >> >> gfs 2 guest_comp_vms2 00030003 none >> >> [1 2 3] >> >> gfs 2 guest_comp_vms3 00050003 none >> >> [1 2 3] >> >> >> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura wrote: >> >>> cman_tool services? >>> >>> >>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>> >>> ohh good luck! >>>> >>>> >>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan : >>>> >>>>> Hi, >>>>> We had some network problem the other day and today I noticed the >>>>> clusvcadm commands are not working. For example: it does not want to stop >>>>> the service, migrate a vm etc etc. On one of the nodes clustat does not >>>>> show any running services. I should restart the rgmanager? >>>>> >>>>> This is RHEL 5. >>>>> >>>>> Thanks >>>>> Paras. >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster at redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> >>>> -- >>>> >>>> --------------------------------------------------------------------------------------------------------------------- >>>> Hugo Aldunce E >>>> Tel. 09 82121045 >>>> mail: haldunce at gmail.com >>>> >>>> --------------------------------------------------------------------------------------------------------------------- >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -------------- next part -------------- An HTML attachment was scrubbed... URL: From emi2fast at gmail.com Wed May 7 18:38:16 2014 From: emi2fast at gmail.com (emmanuel segura) Date: Wed, 7 May 2014 20:38:16 +0200 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: from your previous outpout of cman_tool services [1 2 3] dlm 1 rgmanager 00030001 none 2014-05-07 20:24 GMT+02:00 Paras pradhan : > Oh. How did you see that? > > Here is the cluster.conf http://pastebin.com/DveLMGXT > > Thanks! > -Paras. > > > On Wed, May 7, 2014 at 1:07 PM, emmanuel segura wrote: > >> i saw your rgmanager lockspace is there, you see any error in your msg? >> can show your cluster config? >> >> >> 2014-05-07 19:52 GMT+02:00 Paras pradhan : >> >> Thats looks good. >>> >>> #cman_tool services >>> >>> type level name id state >>> >>> fence 0 default 00010001 none >>> >>> [1 2 3] >>> >>> dlm 1 clvmd 00020001 none >>> >>> [1 2 3] >>> >>> dlm 1 guest_comp_vms1 00020003 none >>> >>> [1 2 3] >>> >>> dlm 1 guest_comp_vms2 00040003 none >>> >>> [1 2 3] >>> >>> dlm 1 guest_comp_vms3 00060003 none >>> >>> [1 2 3] >>> >>> dlm 1 rgmanager 00030001 none >>> >>> [1 2 3] >>> >>> gfs 2 guest_comp_vms1 00010003 none >>> >>> [1 2 3] >>> >>> gfs 2 guest_comp_vms2 00030003 none >>> >>> [1 2 3] >>> >>> gfs 2 guest_comp_vms3 00050003 none >>> >>> [1 2 3] >>> >>> >>> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura wrote: >>> >>>> cman_tool services? >>>> >>>> >>>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>>> >>>> ohh good luck! >>>>> >>>>> >>>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan : >>>>> >>>>>> Hi, >>>>>> We had some network problem the other day and today I noticed the >>>>>> clusvcadm commands are not working. For example: it does not want to stop >>>>>> the service, migrate a vm etc etc. On one of the nodes clustat does not >>>>>> show any running services. I should restart the rgmanager? >>>>>> >>>>>> This is RHEL 5. >>>>>> >>>>>> Thanks >>>>>> Paras. >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster at redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> >>>>> --------------------------------------------------------------------------------------------------------------------- >>>>> Hugo Aldunce E >>>>> Tel. 09 82121045 >>>>> mail: haldunce at gmail.com >>>>> >>>>> --------------------------------------------------------------------------------------------------------------------- >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster at redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> >>>> -- >>>> esta es mi vida e me la vivo hasta que dios quiera >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera -------------- next part -------------- An HTML attachment was scrubbed... URL: From emi2fast at gmail.com Wed May 7 18:51:25 2014 From: emi2fast at gmail.com (emmanuel segura) Date: Wed, 7 May 2014 20:51:25 +0200 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: where is your log? I don't think this is the problem, but anyway from your config i saw : > from your previous outpout of cman_tool services > > [1 2 3] > dlm 1 rgmanager 00030001 none > > > 2014-05-07 20:24 GMT+02:00 Paras pradhan : > > Oh. How did you see that? >> >> Here is the cluster.conf http://pastebin.com/DveLMGXT >> >> Thanks! >> -Paras. >> >> >> On Wed, May 7, 2014 at 1:07 PM, emmanuel segura wrote: >> >>> i saw your rgmanager lockspace is there, you see any error in your msg? >>> can show your cluster config? >>> >>> >>> 2014-05-07 19:52 GMT+02:00 Paras pradhan : >>> >>> Thats looks good. >>>> >>>> #cman_tool services >>>> >>>> type level name id state >>>> >>>> fence 0 default 00010001 none >>>> >>>> [1 2 3] >>>> >>>> dlm 1 clvmd 00020001 none >>>> >>>> [1 2 3] >>>> >>>> dlm 1 guest_comp_vms1 00020003 none >>>> >>>> [1 2 3] >>>> >>>> dlm 1 guest_comp_vms2 00040003 none >>>> >>>> [1 2 3] >>>> >>>> dlm 1 guest_comp_vms3 00060003 none >>>> >>>> [1 2 3] >>>> >>>> dlm 1 rgmanager 00030001 none >>>> >>>> [1 2 3] >>>> >>>> gfs 2 guest_comp_vms1 00010003 none >>>> >>>> [1 2 3] >>>> >>>> gfs 2 guest_comp_vms2 00030003 none >>>> >>>> [1 2 3] >>>> >>>> gfs 2 guest_comp_vms3 00050003 none >>>> >>>> [1 2 3] >>>> >>>> >>>> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura wrote: >>>> >>>>> cman_tool services? >>>>> >>>>> >>>>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>>>> >>>>> ohh good luck! >>>>>> >>>>>> >>>>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan : >>>>>> >>>>>>> Hi, >>>>>>> We had some network problem the other day and today I noticed the >>>>>>> clusvcadm commands are not working. For example: it does not want to stop >>>>>>> the service, migrate a vm etc etc. On one of the nodes clustat does not >>>>>>> show any running services. I should restart the rgmanager? >>>>>>> >>>>>>> This is RHEL 5. >>>>>>> >>>>>>> Thanks >>>>>>> Paras. >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster at redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>> Hugo Aldunce E >>>>>> Tel. 09 82121045 >>>>>> mail: haldunce at gmail.com >>>>>> >>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster at redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster at redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > -- esta es mi vida e me la vivo hasta que dios quiera -------------- next part -------------- An HTML attachment was scrubbed... URL: From pradhanparas at gmail.com Wed May 7 19:05:29 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 14:05:29 -0500 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: Well I have a qdisk with vote 3 . Thats why it is 6. Here is the log. I see some GFS hung but no issue with GFS mounts at this time. http://pastebin.com/MP4BF86c I am seeing this at clumond.log not sure if this is related and what is it. Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue 23340->11670 Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue 23340->11670 Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue 23340->11670 Thanks Paras On Wed, May 7, 2014 at 1:51 PM, emmanuel segura wrote: > where is your log? > > I don't think this is the problem, but anyway from your config i saw expected_votes="6"...... > > from man cman > > Expected votes > The expected votes value is used by cman to determine quorum. > The cluster is quorate if the sum of votes of existing members is > over half of the expected votes value. By default, cman sets the > expected votes value to be the sum of votes of all nodes listed in > cluster.conf. This can be overriden by setting an explicit > expected_votes value as follows: > > If you remove this expected_votes="6", the cluster will set this parameter > to 3 > > > > 2014-05-07 20:38 GMT+02:00 emmanuel segura : > > from your previous outpout of cman_tool services >> >> [1 2 3] >> dlm 1 rgmanager 00030001 none >> >> >> 2014-05-07 20:24 GMT+02:00 Paras pradhan : >> >> Oh. How did you see that? >>> >>> Here is the cluster.conf http://pastebin.com/DveLMGXT >>> >>> Thanks! >>> -Paras. >>> >>> >>> On Wed, May 7, 2014 at 1:07 PM, emmanuel segura wrote: >>> >>>> i saw your rgmanager lockspace is there, you see any error in your msg? >>>> can show your cluster config? >>>> >>>> >>>> 2014-05-07 19:52 GMT+02:00 Paras pradhan : >>>> >>>> Thats looks good. >>>>> >>>>> #cman_tool services >>>>> >>>>> type level name id state >>>>> >>>>> fence 0 default 00010001 none >>>>> >>>>> [1 2 3] >>>>> >>>>> dlm 1 clvmd 00020001 none >>>>> >>>>> [1 2 3] >>>>> >>>>> dlm 1 guest_comp_vms1 00020003 none >>>>> >>>>> [1 2 3] >>>>> >>>>> dlm 1 guest_comp_vms2 00040003 none >>>>> >>>>> [1 2 3] >>>>> >>>>> dlm 1 guest_comp_vms3 00060003 none >>>>> >>>>> [1 2 3] >>>>> >>>>> dlm 1 rgmanager 00030001 none >>>>> >>>>> [1 2 3] >>>>> >>>>> gfs 2 guest_comp_vms1 00010003 none >>>>> >>>>> [1 2 3] >>>>> >>>>> gfs 2 guest_comp_vms2 00030003 none >>>>> >>>>> [1 2 3] >>>>> >>>>> gfs 2 guest_comp_vms3 00050003 none >>>>> >>>>> [1 2 3] >>>>> >>>>> >>>>> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura wrote: >>>>> >>>>>> cman_tool services? >>>>>> >>>>>> >>>>>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>>>>> >>>>>> ohh good luck! >>>>>>> >>>>>>> >>>>>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan : >>>>>>> >>>>>>>> Hi, >>>>>>>> We had some network problem the other day and today I noticed the >>>>>>>> clusvcadm commands are not working. For example: it does not want to stop >>>>>>>> the service, migrate a vm etc etc. On one of the nodes clustat does not >>>>>>>> show any running services. I should restart the rgmanager? >>>>>>>> >>>>>>>> This is RHEL 5. >>>>>>>> >>>>>>>> Thanks >>>>>>>> Paras. >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster at redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>> Hugo Aldunce E >>>>>>> Tel. 09 82121045 >>>>>>> mail: haldunce at gmail.com >>>>>>> >>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster at redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster at redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster at redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> >>>> -- >>>> esta es mi vida e me la vivo hasta que dios quiera >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -------------- next part -------------- An HTML attachment was scrubbed... URL: From emi2fast at gmail.com Wed May 7 20:40:57 2014 From: emi2fast at gmail.com (emmanuel segura) Date: Wed, 7 May 2014 22:40:57 +0200 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: dlm_tool ls ? 2014-05-07 21:05 GMT+02:00 Paras pradhan : > Well I have a qdisk with vote 3 . Thats why it is 6. > > Here is the log. I see some GFS hung but no issue with GFS mounts at this > time. > > http://pastebin.com/MP4BF86c > > I am seeing this at clumond.log not sure if this is related and what is it. > > Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue > 23340->11670 > > Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue > 23340->11670 > > Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue > 23340->11670 > > Thanks > Paras > > > On Wed, May 7, 2014 at 1:51 PM, emmanuel segura wrote: > >> where is your log? >> >> I don't think this is the problem, but anyway from your config i saw >> > >> from man cman >> >> Expected votes >> The expected votes value is used by cman to determine quorum. >> The cluster is quorate if the sum of votes of existing members is >> over half of the expected votes value. By default, cman sets the >> expected votes value to be the sum of votes of all nodes listed in >> cluster.conf. This can be overriden by setting an explicit >> expected_votes value as follows: >> >> If you remove this expected_votes="6", the cluster will set this >> parameter to 3 >> >> >> >> 2014-05-07 20:38 GMT+02:00 emmanuel segura : >> >> from your previous outpout of cman_tool services >>> >>> [1 2 3] >>> dlm 1 rgmanager 00030001 none >>> >>> >>> 2014-05-07 20:24 GMT+02:00 Paras pradhan : >>> >>> Oh. How did you see that? >>>> >>>> Here is the cluster.conf http://pastebin.com/DveLMGXT >>>> >>>> Thanks! >>>> -Paras. >>>> >>>> >>>> On Wed, May 7, 2014 at 1:07 PM, emmanuel segura wrote: >>>> >>>>> i saw your rgmanager lockspace is there, you see any error in your >>>>> msg? can show your cluster config? >>>>> >>>>> >>>>> 2014-05-07 19:52 GMT+02:00 Paras pradhan : >>>>> >>>>> Thats looks good. >>>>>> >>>>>> #cman_tool services >>>>>> >>>>>> type level name id state >>>>>> >>>>>> fence 0 default 00010001 none >>>>>> >>>>>> [1 2 3] >>>>>> >>>>>> dlm 1 clvmd 00020001 none >>>>>> >>>>>> [1 2 3] >>>>>> >>>>>> dlm 1 guest_comp_vms1 00020003 none >>>>>> >>>>>> [1 2 3] >>>>>> >>>>>> dlm 1 guest_comp_vms2 00040003 none >>>>>> >>>>>> [1 2 3] >>>>>> >>>>>> dlm 1 guest_comp_vms3 00060003 none >>>>>> >>>>>> [1 2 3] >>>>>> >>>>>> dlm 1 rgmanager 00030001 none >>>>>> >>>>>> [1 2 3] >>>>>> >>>>>> gfs 2 guest_comp_vms1 00010003 none >>>>>> >>>>>> [1 2 3] >>>>>> >>>>>> gfs 2 guest_comp_vms2 00030003 none >>>>>> >>>>>> [1 2 3] >>>>>> >>>>>> gfs 2 guest_comp_vms3 00050003 none >>>>>> >>>>>> [1 2 3] >>>>>> >>>>>> >>>>>> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura wrote: >>>>>> >>>>>>> cman_tool services? >>>>>>> >>>>>>> >>>>>>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>>>>>> >>>>>>> ohh good luck! >>>>>>>> >>>>>>>> >>>>>>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan : >>>>>>>> >>>>>>>>> Hi, >>>>>>>>> We had some network problem the other day and today I noticed the >>>>>>>>> clusvcadm commands are not working. For example: it does not want to stop >>>>>>>>> the service, migrate a vm etc etc. On one of the nodes clustat does not >>>>>>>>> show any running services. I should restart the rgmanager? >>>>>>>>> >>>>>>>>> This is RHEL 5. >>>>>>>>> >>>>>>>>> Thanks >>>>>>>>> Paras. >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster at redhat.com >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>> Hugo Aldunce E >>>>>>>> Tel. 09 82121045 >>>>>>>> mail: haldunce at gmail.com >>>>>>>> >>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster at redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster at redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster at redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster at redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera -------------- next part -------------- An HTML attachment was scrubbed... URL: From pradhanparas at gmail.com Wed May 7 21:01:33 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 16:01:33 -0500 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: "dlm_tools ls lockdebug" you mean? "dlm_tool ls" returns -- Usage: dlm_tool [options] [join|leave|lockdump|lockdebug] Options: -v Verbose output -d Resource directory off/on (0/1), default 0 -m Permission mode for lockspace device (octal), default 0600 -M Print MSTCPY locks in lockdump (remote locks, locally mastered) -h Print this help, then exit -V Print program version information, then exit - On Wed, May 7, 2014 at 3:40 PM, emmanuel segura wrote: > dlm_tool ls ? > > > 2014-05-07 21:05 GMT+02:00 Paras pradhan : > >> Well I have a qdisk with vote 3 . Thats why it is 6. >> >> Here is the log. I see some GFS hung but no issue with GFS mounts at this >> time. >> >> http://pastebin.com/MP4BF86c >> >> I am seeing this at clumond.log not sure if this is related and what is >> it. >> >> Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue >> 23340->11670 >> >> Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue >> 23340->11670 >> >> Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue >> 23340->11670 >> >> Thanks >> Paras >> >> >> On Wed, May 7, 2014 at 1:51 PM, emmanuel segura wrote: >> >>> where is your log? >>> >>> I don't think this is the problem, but anyway from your config i saw >>> >> >>> from man cman >>> >>> Expected votes >>> The expected votes value is used by cman to determine >>> quorum. The cluster is quorate if the sum of votes of existing members is >>> over half of the expected votes value. By default, cman sets the >>> expected votes value to be the sum of votes of all nodes listed in >>> cluster.conf. This can be overriden by setting an explicit >>> expected_votes value as follows: >>> >>> If you remove this expected_votes="6", the cluster will set this >>> parameter to 3 >>> >>> >>> >>> 2014-05-07 20:38 GMT+02:00 emmanuel segura : >>> >>> from your previous outpout of cman_tool services >>>> >>>> [1 2 3] >>>> dlm 1 rgmanager 00030001 none >>>> >>>> >>>> 2014-05-07 20:24 GMT+02:00 Paras pradhan : >>>> >>>> Oh. How did you see that? >>>>> >>>>> Here is the cluster.conf http://pastebin.com/DveLMGXT >>>>> >>>>> Thanks! >>>>> -Paras. >>>>> >>>>> >>>>> On Wed, May 7, 2014 at 1:07 PM, emmanuel segura wrote: >>>>> >>>>>> i saw your rgmanager lockspace is there, you see any error in your >>>>>> msg? can show your cluster config? >>>>>> >>>>>> >>>>>> 2014-05-07 19:52 GMT+02:00 Paras pradhan : >>>>>> >>>>>> Thats looks good. >>>>>>> >>>>>>> #cman_tool services >>>>>>> >>>>>>> type level name id state >>>>>>> >>>>>>> fence 0 default 00010001 none >>>>>>> >>>>>>> [1 2 3] >>>>>>> >>>>>>> dlm 1 clvmd 00020001 none >>>>>>> >>>>>>> [1 2 3] >>>>>>> >>>>>>> dlm 1 guest_comp_vms1 00020003 none >>>>>>> >>>>>>> [1 2 3] >>>>>>> >>>>>>> dlm 1 guest_comp_vms2 00040003 none >>>>>>> >>>>>>> [1 2 3] >>>>>>> >>>>>>> dlm 1 guest_comp_vms3 00060003 none >>>>>>> >>>>>>> [1 2 3] >>>>>>> >>>>>>> dlm 1 rgmanager 00030001 none >>>>>>> >>>>>>> [1 2 3] >>>>>>> >>>>>>> gfs 2 guest_comp_vms1 00010003 none >>>>>>> >>>>>>> [1 2 3] >>>>>>> >>>>>>> gfs 2 guest_comp_vms2 00030003 none >>>>>>> >>>>>>> [1 2 3] >>>>>>> >>>>>>> gfs 2 guest_comp_vms3 00050003 none >>>>>>> >>>>>>> [1 2 3] >>>>>>> >>>>>>> >>>>>>> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura >>>>>> > wrote: >>>>>>> >>>>>>>> cman_tool services? >>>>>>>> >>>>>>>> >>>>>>>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>>>>>>> >>>>>>>> ohh good luck! >>>>>>>>> >>>>>>>>> >>>>>>>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan : >>>>>>>>> >>>>>>>>>> Hi, >>>>>>>>>> We had some network problem the other day and today I noticed the >>>>>>>>>> clusvcadm commands are not working. For example: it does not want to stop >>>>>>>>>> the service, migrate a vm etc etc. On one of the nodes clustat does not >>>>>>>>>> show any running services. I should restart the rgmanager? >>>>>>>>>> >>>>>>>>>> This is RHEL 5. >>>>>>>>>> >>>>>>>>>> Thanks >>>>>>>>>> Paras. >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> >>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>> Hugo Aldunce E >>>>>>>>> Tel. 09 82121045 >>>>>>>>> mail: haldunce at gmail.com >>>>>>>>> >>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster at redhat.com >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster at redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster at redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster at redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster at redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> >>>> -- >>>> esta es mi vida e me la vivo hasta que dios quiera >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -------------- next part -------------- An HTML attachment was scrubbed... URL: From emi2fast at gmail.com Wed May 7 21:34:37 2014 From: emi2fast at gmail.com (emmanuel segura) Date: Wed, 7 May 2014 23:34:37 +0200 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: dlm_tool lockdebug rgmanager or dlm_tool lockdump rgmanager, anyway you can tell me when this problem started to happen? are you sure your fencing is working ok? 2014-05-07 23:01 GMT+02:00 Paras pradhan : > "dlm_tools ls lockdebug" you mean? > > "dlm_tool ls" returns > > -- > > Usage: > > > dlm_tool [options] [join|leave|lockdump|lockdebug] > > > Options: > > -v Verbose output > > -d Resource directory off/on (0/1), default 0 > > -m Permission mode for lockspace device (octal), default > 0600 > > -M Print MSTCPY locks in lockdump (remote locks, locally > mastered) > > -h Print this help, then exit > > -V Print program version information, then exit > > - > > > > > > On Wed, May 7, 2014 at 3:40 PM, emmanuel segura wrote: > >> dlm_tool ls ? >> >> >> 2014-05-07 21:05 GMT+02:00 Paras pradhan : >> >>> Well I have a qdisk with vote 3 . Thats why it is 6. >>> >>> Here is the log. I see some GFS hung but no issue with GFS mounts at >>> this time. >>> >>> http://pastebin.com/MP4BF86c >>> >>> I am seeing this at clumond.log not sure if this is related and what is >>> it. >>> >>> Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue >>> 23340->11670 >>> >>> Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue >>> 23340->11670 >>> >>> Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue >>> 23340->11670 >>> >>> Thanks >>> Paras >>> >>> >>> On Wed, May 7, 2014 at 1:51 PM, emmanuel segura wrote: >>> >>>> where is your log? >>>> >>>> I don't think this is the problem, but anyway from your config i saw >>>> >>> >>>> from man cman >>>> >>>> Expected votes >>>> The expected votes value is used by cman to determine >>>> quorum. The cluster is quorate if the sum of votes of existing members is >>>> over half of the expected votes value. By default, cman sets >>>> the expected votes value to be the sum of votes of all nodes listed in >>>> cluster.conf. This can be overriden by setting an explicit >>>> expected_votes value as follows: >>>> >>>> If you remove this expected_votes="6", the cluster will set this >>>> parameter to 3 >>>> >>>> >>>> >>>> 2014-05-07 20:38 GMT+02:00 emmanuel segura : >>>> >>>> from your previous outpout of cman_tool services >>>>> >>>>> [1 2 3] >>>>> dlm 1 rgmanager 00030001 none >>>>> >>>>> >>>>> 2014-05-07 20:24 GMT+02:00 Paras pradhan : >>>>> >>>>> Oh. How did you see that? >>>>>> >>>>>> Here is the cluster.conf http://pastebin.com/DveLMGXT >>>>>> >>>>>> Thanks! >>>>>> -Paras. >>>>>> >>>>>> >>>>>> On Wed, May 7, 2014 at 1:07 PM, emmanuel segura wrote: >>>>>> >>>>>>> i saw your rgmanager lockspace is there, you see any error in your >>>>>>> msg? can show your cluster config? >>>>>>> >>>>>>> >>>>>>> 2014-05-07 19:52 GMT+02:00 Paras pradhan : >>>>>>> >>>>>>> Thats looks good. >>>>>>>> >>>>>>>> #cman_tool services >>>>>>>> >>>>>>>> type level name id state >>>>>>>> >>>>>>>> fence 0 default 00010001 none >>>>>>>> >>>>>>>> [1 2 3] >>>>>>>> >>>>>>>> dlm 1 clvmd 00020001 none >>>>>>>> >>>>>>>> [1 2 3] >>>>>>>> >>>>>>>> dlm 1 guest_comp_vms1 00020003 none >>>>>>>> >>>>>>>> [1 2 3] >>>>>>>> >>>>>>>> dlm 1 guest_comp_vms2 00040003 none >>>>>>>> >>>>>>>> [1 2 3] >>>>>>>> >>>>>>>> dlm 1 guest_comp_vms3 00060003 none >>>>>>>> >>>>>>>> [1 2 3] >>>>>>>> >>>>>>>> dlm 1 rgmanager 00030001 none >>>>>>>> >>>>>>>> [1 2 3] >>>>>>>> >>>>>>>> gfs 2 guest_comp_vms1 00010003 none >>>>>>>> >>>>>>>> [1 2 3] >>>>>>>> >>>>>>>> gfs 2 guest_comp_vms2 00030003 none >>>>>>>> >>>>>>>> [1 2 3] >>>>>>>> >>>>>>>> gfs 2 guest_comp_vms3 00050003 none >>>>>>>> >>>>>>>> [1 2 3] >>>>>>>> >>>>>>>> >>>>>>>> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura < >>>>>>>> emi2fast at gmail.com> wrote: >>>>>>>> >>>>>>>>> cman_tool services? >>>>>>>>> >>>>>>>>> >>>>>>>>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>>>>>>>> >>>>>>>>> ohh good luck! >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan >>>>>>>>>> : >>>>>>>>>> >>>>>>>>>>> Hi, >>>>>>>>>>> We had some network problem the other day and today I noticed >>>>>>>>>>> the clusvcadm commands are not working. For example: it does not want to >>>>>>>>>>> stop the service, migrate a vm etc etc. On one of the nodes clustat does >>>>>>>>>>> not show any running services. I should restart the rgmanager? >>>>>>>>>>> >>>>>>>>>>> This is RHEL 5. >>>>>>>>>>> >>>>>>>>>>> Thanks >>>>>>>>>>> Paras. >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>> Hugo Aldunce E >>>>>>>>>> Tel. 09 82121045 >>>>>>>>>> mail: haldunce at gmail.com >>>>>>>>>> >>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster at redhat.com >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster at redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster at redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster at redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>> >>>> >>>> >>>> >>>> -- >>>> esta es mi vida e me la vivo hasta que dios quiera >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera -------------- next part -------------- An HTML attachment was scrubbed... URL: From pradhanparas at gmail.com Wed May 7 21:45:33 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 16:45:33 -0500 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: Yeah they work fine . This started when we had a network problem. I see this: dlm_tool lockdebug rgmanager can't open /sys/kernel/debug/dlm/rgmanager_locks: No such file or directory On Wed, May 7, 2014 at 4:34 PM, emmanuel segura wrote: > dlm_tool lockdebug rgmanager or dlm_tool lockdump rgmanager, anyway you > can tell me when this problem started to happen? are you sure your fencing > is working ok? > > > 2014-05-07 23:01 GMT+02:00 Paras pradhan : > > "dlm_tools ls lockdebug" you mean? >> >> "dlm_tool ls" returns >> >> -- >> >> Usage: >> >> >> dlm_tool [options] [join|leave|lockdump|lockdebug] >> >> >> Options: >> >> -v Verbose output >> >> -d Resource directory off/on (0/1), default 0 >> >> -m Permission mode for lockspace device (octal), default >> 0600 >> >> -M Print MSTCPY locks in lockdump (remote locks, locally >> mastered) >> >> -h Print this help, then exit >> >> -V Print program version information, then exit >> >> - >> >> >> >> >> >> On Wed, May 7, 2014 at 3:40 PM, emmanuel segura wrote: >> >>> dlm_tool ls ? >>> >>> >>> 2014-05-07 21:05 GMT+02:00 Paras pradhan : >>> >>>> Well I have a qdisk with vote 3 . Thats why it is 6. >>>> >>>> Here is the log. I see some GFS hung but no issue with GFS mounts at >>>> this time. >>>> >>>> http://pastebin.com/MP4BF86c >>>> >>>> I am seeing this at clumond.log not sure if this is related and what is >>>> it. >>>> >>>> Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue >>>> 23340->11670 >>>> >>>> Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue >>>> 23340->11670 >>>> >>>> Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue >>>> 23340->11670 >>>> >>>> Thanks >>>> Paras >>>> >>>> >>>> On Wed, May 7, 2014 at 1:51 PM, emmanuel segura wrote: >>>> >>>>> where is your log? >>>>> >>>>> I don't think this is the problem, but anyway from your config i saw >>>>> >>>> >>>>> from man cman >>>>> >>>>> Expected votes >>>>> The expected votes value is used by cman to determine >>>>> quorum. The cluster is quorate if the sum of votes of existing members is >>>>> over half of the expected votes value. By default, cman sets >>>>> the expected votes value to be the sum of votes of all nodes listed in >>>>> cluster.conf. This can be overriden by setting an explicit >>>>> expected_votes value as follows: >>>>> >>>>> If you remove this expected_votes="6", the cluster will set this >>>>> parameter to 3 >>>>> >>>>> >>>>> >>>>> 2014-05-07 20:38 GMT+02:00 emmanuel segura : >>>>> >>>>> from your previous outpout of cman_tool services >>>>>> >>>>>> [1 2 3] >>>>>> dlm 1 rgmanager 00030001 none >>>>>> >>>>>> >>>>>> 2014-05-07 20:24 GMT+02:00 Paras pradhan : >>>>>> >>>>>> Oh. How did you see that? >>>>>>> >>>>>>> Here is the cluster.conf http://pastebin.com/DveLMGXT >>>>>>> >>>>>>> Thanks! >>>>>>> -Paras. >>>>>>> >>>>>>> >>>>>>> On Wed, May 7, 2014 at 1:07 PM, emmanuel segura wrote: >>>>>>> >>>>>>>> i saw your rgmanager lockspace is there, you see any error in your >>>>>>>> msg? can show your cluster config? >>>>>>>> >>>>>>>> >>>>>>>> 2014-05-07 19:52 GMT+02:00 Paras pradhan : >>>>>>>> >>>>>>>> Thats looks good. >>>>>>>>> >>>>>>>>> #cman_tool services >>>>>>>>> >>>>>>>>> type level name id state >>>>>>>>> >>>>>>>>> fence 0 default 00010001 none >>>>>>>>> >>>>>>>>> [1 2 3] >>>>>>>>> >>>>>>>>> dlm 1 clvmd 00020001 none >>>>>>>>> >>>>>>>>> [1 2 3] >>>>>>>>> >>>>>>>>> dlm 1 guest_comp_vms1 00020003 none >>>>>>>>> >>>>>>>>> [1 2 3] >>>>>>>>> >>>>>>>>> dlm 1 guest_comp_vms2 00040003 none >>>>>>>>> >>>>>>>>> [1 2 3] >>>>>>>>> >>>>>>>>> dlm 1 guest_comp_vms3 00060003 none >>>>>>>>> >>>>>>>>> [1 2 3] >>>>>>>>> >>>>>>>>> dlm 1 rgmanager 00030001 none >>>>>>>>> >>>>>>>>> [1 2 3] >>>>>>>>> >>>>>>>>> gfs 2 guest_comp_vms1 00010003 none >>>>>>>>> >>>>>>>>> [1 2 3] >>>>>>>>> >>>>>>>>> gfs 2 guest_comp_vms2 00030003 none >>>>>>>>> >>>>>>>>> [1 2 3] >>>>>>>>> >>>>>>>>> gfs 2 guest_comp_vms3 00050003 none >>>>>>>>> >>>>>>>>> [1 2 3] >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura < >>>>>>>>> emi2fast at gmail.com> wrote: >>>>>>>>> >>>>>>>>>> cman_tool services? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>>>>>>>>> >>>>>>>>>> ohh good luck! >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan >>>>>>>>>> >: >>>>>>>>>>> >>>>>>>>>>>> Hi, >>>>>>>>>>>> We had some network problem the other day and today I noticed >>>>>>>>>>>> the clusvcadm commands are not working. For example: it does not want to >>>>>>>>>>>> stop the service, migrate a vm etc etc. On one of the nodes clustat does >>>>>>>>>>>> not show any running services. I should restart the rgmanager? >>>>>>>>>>>> >>>>>>>>>>>> This is RHEL 5. >>>>>>>>>>>> >>>>>>>>>>>> Thanks >>>>>>>>>>>> Paras. >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> >>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>> Hugo Aldunce E >>>>>>>>>>> Tel. 09 82121045 >>>>>>>>>>> mail: haldunce at gmail.com >>>>>>>>>>> >>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster at redhat.com >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster at redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster at redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster at redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -------------- next part -------------- An HTML attachment was scrubbed... URL: From emi2fast at gmail.com Wed May 7 22:01:07 2014 From: emi2fast at gmail.com (emmanuel segura) Date: Thu, 8 May 2014 00:01:07 +0200 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: mount -t debugfs none /sys/kernel/debug/, i now this happen when a fencing calls had problem 2014-05-07 23:45 GMT+02:00 Paras pradhan : > Yeah they work fine . This started when we had a network problem. > > I see this: > > dlm_tool lockdebug rgmanager > > can't open /sys/kernel/debug/dlm/rgmanager_locks: No such file or directory > > > > > On Wed, May 7, 2014 at 4:34 PM, emmanuel segura wrote: > >> dlm_tool lockdebug rgmanager or dlm_tool lockdump rgmanager, anyway you >> can tell me when this problem started to happen? are you sure your fencing >> is working ok? >> >> >> 2014-05-07 23:01 GMT+02:00 Paras pradhan : >> >> "dlm_tools ls lockdebug" you mean? >>> >>> "dlm_tool ls" returns >>> >>> -- >>> >>> Usage: >>> >>> >>> dlm_tool [options] [join|leave|lockdump|lockdebug] >>> >>> >>> Options: >>> >>> -v Verbose output >>> >>> -d Resource directory off/on (0/1), default 0 >>> >>> -m Permission mode for lockspace device (octal), default >>> 0600 >>> >>> -M Print MSTCPY locks in lockdump (remote locks, locally >>> mastered) >>> >>> -h Print this help, then exit >>> >>> -V Print program version information, then exit >>> >>> - >>> >>> >>> >>> >>> >>> On Wed, May 7, 2014 at 3:40 PM, emmanuel segura wrote: >>> >>>> dlm_tool ls ? >>>> >>>> >>>> 2014-05-07 21:05 GMT+02:00 Paras pradhan : >>>> >>>>> Well I have a qdisk with vote 3 . Thats why it is 6. >>>>> >>>>> Here is the log. I see some GFS hung but no issue with GFS mounts at >>>>> this time. >>>>> >>>>> http://pastebin.com/MP4BF86c >>>>> >>>>> I am seeing this at clumond.log not sure if this is related and what >>>>> is it. >>>>> >>>>> Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue >>>>> 23340->11670 >>>>> >>>>> Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue >>>>> 23340->11670 >>>>> >>>>> Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue >>>>> 23340->11670 >>>>> >>>>> Thanks >>>>> Paras >>>>> >>>>> >>>>> On Wed, May 7, 2014 at 1:51 PM, emmanuel segura wrote: >>>>> >>>>>> where is your log? >>>>>> >>>>>> I don't think this is the problem, but anyway from your config i saw >>>>>> >>>>> >>>>>> from man cman >>>>>> >>>>>> Expected votes >>>>>> The expected votes value is used by cman to determine >>>>>> quorum. The cluster is quorate if the sum of votes of existing members is >>>>>> over half of the expected votes value. By default, cman sets >>>>>> the expected votes value to be the sum of votes of all nodes listed in >>>>>> cluster.conf. This can be overriden by setting an explicit >>>>>> expected_votes value as follows: >>>>>> >>>>>> If you remove this expected_votes="6", the cluster will set this >>>>>> parameter to 3 >>>>>> >>>>>> >>>>>> >>>>>> 2014-05-07 20:38 GMT+02:00 emmanuel segura : >>>>>> >>>>>> from your previous outpout of cman_tool services >>>>>>> >>>>>>> [1 2 3] >>>>>>> dlm 1 rgmanager 00030001 none >>>>>>> >>>>>>> >>>>>>> 2014-05-07 20:24 GMT+02:00 Paras pradhan : >>>>>>> >>>>>>> Oh. How did you see that? >>>>>>>> >>>>>>>> Here is the cluster.conf http://pastebin.com/DveLMGXT >>>>>>>> >>>>>>>> Thanks! >>>>>>>> -Paras. >>>>>>>> >>>>>>>> >>>>>>>> On Wed, May 7, 2014 at 1:07 PM, emmanuel segura >>>>>>> > wrote: >>>>>>>> >>>>>>>>> i saw your rgmanager lockspace is there, you see any error in your >>>>>>>>> msg? can show your cluster config? >>>>>>>>> >>>>>>>>> >>>>>>>>> 2014-05-07 19:52 GMT+02:00 Paras pradhan : >>>>>>>>> >>>>>>>>> Thats looks good. >>>>>>>>>> >>>>>>>>>> #cman_tool services >>>>>>>>>> >>>>>>>>>> type level name id state >>>>>>>>>> >>>>>>>>>> fence 0 default 00010001 none >>>>>>>>>> >>>>>>>>>> [1 2 3] >>>>>>>>>> >>>>>>>>>> dlm 1 clvmd 00020001 none >>>>>>>>>> >>>>>>>>>> [1 2 3] >>>>>>>>>> >>>>>>>>>> dlm 1 guest_comp_vms1 00020003 none >>>>>>>>>> >>>>>>>>>> [1 2 3] >>>>>>>>>> >>>>>>>>>> dlm 1 guest_comp_vms2 00040003 none >>>>>>>>>> >>>>>>>>>> [1 2 3] >>>>>>>>>> >>>>>>>>>> dlm 1 guest_comp_vms3 00060003 none >>>>>>>>>> >>>>>>>>>> [1 2 3] >>>>>>>>>> >>>>>>>>>> dlm 1 rgmanager 00030001 none >>>>>>>>>> >>>>>>>>>> [1 2 3] >>>>>>>>>> >>>>>>>>>> gfs 2 guest_comp_vms1 00010003 none >>>>>>>>>> >>>>>>>>>> [1 2 3] >>>>>>>>>> >>>>>>>>>> gfs 2 guest_comp_vms2 00030003 none >>>>>>>>>> >>>>>>>>>> [1 2 3] >>>>>>>>>> >>>>>>>>>> gfs 2 guest_comp_vms3 00050003 none >>>>>>>>>> >>>>>>>>>> [1 2 3] >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura < >>>>>>>>>> emi2fast at gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> cman_tool services? >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>>>>>>>>>> >>>>>>>>>>> ohh good luck! >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan < >>>>>>>>>>>> pradhanparas at gmail.com>: >>>>>>>>>>>> >>>>>>>>>>>>> Hi, >>>>>>>>>>>>> We had some network problem the other day and today I noticed >>>>>>>>>>>>> the clusvcadm commands are not working. For example: it does not want to >>>>>>>>>>>>> stop the service, migrate a vm etc etc. On one of the nodes clustat does >>>>>>>>>>>>> not show any running services. I should restart the rgmanager? >>>>>>>>>>>>> >>>>>>>>>>>>> This is RHEL 5. >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks >>>>>>>>>>>>> Paras. >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> >>>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> Hugo Aldunce E >>>>>>>>>>>> Tel. 09 82121045 >>>>>>>>>>>> mail: haldunce at gmail.com >>>>>>>>>>>> >>>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster at redhat.com >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster at redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster at redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster at redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> >>>> -- >>>> esta es mi vida e me la vivo hasta que dios quiera >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera -------------- next part -------------- An HTML attachment was scrubbed... URL: From pradhanparas at gmail.com Wed May 7 22:17:16 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 17:17:16 -0500 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: Ok i got this when running dlm_tool lockdebug rgmanager # dlm_tool lockdebug rgmanager Resource ffff880082e131c0 Name (len=22) "rg="vm:wadev.domain"" Master Copy Granted Queue 00c1054d NL Remote: 3 00c0003f 02790500 NL Remote: 2 022d9a84 03830554 EX Conversion Queue Waiting Queue Resource ffff880082e132c0 Name (len=8) "usrm::vf" Local Copy, Master is node 2 Granted Queue Conversion Queue Waiting Queue Thanks! Paras. On Wed, May 7, 2014 at 5:01 PM, emmanuel segura wrote: > mount -t debugfs none /sys/kernel/debug/, i now this happen when a fencing > calls had problem > > > 2014-05-07 23:45 GMT+02:00 Paras pradhan : > > Yeah they work fine . This started when we had a network problem. >> >> I see this: >> >> dlm_tool lockdebug rgmanager >> >> can't open /sys/kernel/debug/dlm/rgmanager_locks: No such file or >> directory >> >> >> >> >> On Wed, May 7, 2014 at 4:34 PM, emmanuel segura wrote: >> >>> dlm_tool lockdebug rgmanager or dlm_tool lockdump rgmanager, anyway you >>> can tell me when this problem started to happen? are you sure your fencing >>> is working ok? >>> >>> >>> 2014-05-07 23:01 GMT+02:00 Paras pradhan : >>> >>> "dlm_tools ls lockdebug" you mean? >>>> >>>> "dlm_tool ls" returns >>>> >>>> -- >>>> >>>> Usage: >>>> >>>> >>>> dlm_tool [options] [join|leave|lockdump|lockdebug] >>>> >>>> >>>> Options: >>>> >>>> -v Verbose output >>>> >>>> -d Resource directory off/on (0/1), default 0 >>>> >>>> -m Permission mode for lockspace device (octal), >>>> default 0600 >>>> >>>> -M Print MSTCPY locks in lockdump (remote locks, >>>> locally mastered) >>>> >>>> -h Print this help, then exit >>>> >>>> -V Print program version information, then exit >>>> >>>> - >>>> >>>> >>>> >>>> >>>> >>>> On Wed, May 7, 2014 at 3:40 PM, emmanuel segura wrote: >>>> >>>>> dlm_tool ls ? >>>>> >>>>> >>>>> 2014-05-07 21:05 GMT+02:00 Paras pradhan : >>>>> >>>>>> Well I have a qdisk with vote 3 . Thats why it is 6. >>>>>> >>>>>> Here is the log. I see some GFS hung but no issue with GFS mounts at >>>>>> this time. >>>>>> >>>>>> http://pastebin.com/MP4BF86c >>>>>> >>>>>> I am seeing this at clumond.log not sure if this is related and what >>>>>> is it. >>>>>> >>>>>> Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue >>>>>> 23340->11670 >>>>>> >>>>>> Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue >>>>>> 23340->11670 >>>>>> >>>>>> Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue >>>>>> 23340->11670 >>>>>> >>>>>> Thanks >>>>>> Paras >>>>>> >>>>>> >>>>>> On Wed, May 7, 2014 at 1:51 PM, emmanuel segura wrote: >>>>>> >>>>>>> where is your log? >>>>>>> >>>>>>> I don't think this is the problem, but anyway from your config i saw >>>>>>> >>>>>> >>>>>>> from man cman >>>>>>> >>>>>>> Expected votes >>>>>>> The expected votes value is used by cman to determine >>>>>>> quorum. The cluster is quorate if the sum of votes of existing members is >>>>>>> over half of the expected votes value. By default, cman sets >>>>>>> the expected votes value to be the sum of votes of all nodes listed in >>>>>>> cluster.conf. This can be overriden by setting an explicit >>>>>>> expected_votes value as follows: >>>>>>> >>>>>>> If you remove this expected_votes="6", the cluster will set this >>>>>>> parameter to 3 >>>>>>> >>>>>>> >>>>>>> >>>>>>> 2014-05-07 20:38 GMT+02:00 emmanuel segura : >>>>>>> >>>>>>> from your previous outpout of cman_tool services >>>>>>>> >>>>>>>> [1 2 3] >>>>>>>> dlm 1 rgmanager 00030001 none >>>>>>>> >>>>>>>> >>>>>>>> 2014-05-07 20:24 GMT+02:00 Paras pradhan : >>>>>>>> >>>>>>>> Oh. How did you see that? >>>>>>>>> >>>>>>>>> Here is the cluster.conf http://pastebin.com/DveLMGXT >>>>>>>>> >>>>>>>>> Thanks! >>>>>>>>> -Paras. >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, May 7, 2014 at 1:07 PM, emmanuel segura < >>>>>>>>> emi2fast at gmail.com> wrote: >>>>>>>>> >>>>>>>>>> i saw your rgmanager lockspace is there, you see any error in >>>>>>>>>> your msg? can show your cluster config? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2014-05-07 19:52 GMT+02:00 Paras pradhan >>>>>>>>>> : >>>>>>>>>> >>>>>>>>>> Thats looks good. >>>>>>>>>>> >>>>>>>>>>> #cman_tool services >>>>>>>>>>> >>>>>>>>>>> type level name id state >>>>>>>>>>> >>>>>>>>>>> fence 0 default 00010001 none >>>>>>>>>>> >>>>>>>>>>> [1 2 3] >>>>>>>>>>> >>>>>>>>>>> dlm 1 clvmd 00020001 none >>>>>>>>>>> >>>>>>>>>>> [1 2 3] >>>>>>>>>>> >>>>>>>>>>> dlm 1 guest_comp_vms1 00020003 none >>>>>>>>>>> >>>>>>>>>>> [1 2 3] >>>>>>>>>>> >>>>>>>>>>> dlm 1 guest_comp_vms2 00040003 none >>>>>>>>>>> >>>>>>>>>>> [1 2 3] >>>>>>>>>>> >>>>>>>>>>> dlm 1 guest_comp_vms3 00060003 none >>>>>>>>>>> >>>>>>>>>>> [1 2 3] >>>>>>>>>>> >>>>>>>>>>> dlm 1 rgmanager 00030001 none >>>>>>>>>>> >>>>>>>>>>> [1 2 3] >>>>>>>>>>> >>>>>>>>>>> gfs 2 guest_comp_vms1 00010003 none >>>>>>>>>>> >>>>>>>>>>> [1 2 3] >>>>>>>>>>> >>>>>>>>>>> gfs 2 guest_comp_vms2 00030003 none >>>>>>>>>>> >>>>>>>>>>> [1 2 3] >>>>>>>>>>> >>>>>>>>>>> gfs 2 guest_comp_vms3 00050003 none >>>>>>>>>>> >>>>>>>>>>> [1 2 3] >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura < >>>>>>>>>>> emi2fast at gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> cman_tool services? >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>>>>>>>>>>> >>>>>>>>>>>> ohh good luck! >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan < >>>>>>>>>>>>> pradhanparas at gmail.com>: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>> We had some network problem the other day and today I noticed >>>>>>>>>>>>>> the clusvcadm commands are not working. For example: it does not want to >>>>>>>>>>>>>> stop the service, migrate a vm etc etc. On one of the nodes clustat does >>>>>>>>>>>>>> not show any running services. I should restart the rgmanager? >>>>>>>>>>>>>> >>>>>>>>>>>>>> This is RHEL 5. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks >>>>>>>>>>>>>> Paras. >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> >>>>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>>> Hugo Aldunce E >>>>>>>>>>>>> Tel. 09 82121045 >>>>>>>>>>>>> mail: haldunce at gmail.com >>>>>>>>>>>>> >>>>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster at redhat.com >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster at redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster at redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster at redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -------------- next part -------------- An HTML attachment was scrubbed... URL: From pradhanparas at gmail.com Wed May 7 22:18:55 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 17:18:55 -0500 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: and this dlm_tool lockdump rgmanager id 00c0003f gr NL rq IV pid 12633 master 1 "rg=" Thanks Paras On Wed, May 7, 2014 at 5:17 PM, Paras pradhan wrote: > Ok i got this when running dlm_tool lockdebug rgmanager > > # dlm_tool lockdebug rgmanager > > > Resource ffff880082e131c0 Name (len=22) "rg="vm:wadev.domain"" > > Master Copy > > Granted Queue > > 00c1054d NL Remote: 3 00c0003f > > 02790500 NL Remote: 2 022d9a84 > > 03830554 EX > > Conversion Queue > > Waiting Queue > > > Resource ffff880082e132c0 Name (len=8) "usrm::vf" > > Local Copy, Master is node 2 > > Granted Queue > > Conversion Queue > > Waiting Queue > > > Thanks! > > Paras. > > > On Wed, May 7, 2014 at 5:01 PM, emmanuel segura wrote: > >> mount -t debugfs none /sys/kernel/debug/, i now this happen when a >> fencing calls had problem >> >> >> 2014-05-07 23:45 GMT+02:00 Paras pradhan : >> >> Yeah they work fine . This started when we had a network problem. >>> >>> I see this: >>> >>> dlm_tool lockdebug rgmanager >>> >>> can't open /sys/kernel/debug/dlm/rgmanager_locks: No such file or >>> directory >>> >>> >>> >>> >>> On Wed, May 7, 2014 at 4:34 PM, emmanuel segura wrote: >>> >>>> dlm_tool lockdebug rgmanager or dlm_tool lockdump rgmanager, anyway you >>>> can tell me when this problem started to happen? are you sure your fencing >>>> is working ok? >>>> >>>> >>>> 2014-05-07 23:01 GMT+02:00 Paras pradhan : >>>> >>>> "dlm_tools ls lockdebug" you mean? >>>>> >>>>> "dlm_tool ls" returns >>>>> >>>>> -- >>>>> >>>>> Usage: >>>>> >>>>> >>>>> dlm_tool [options] [join|leave|lockdump|lockdebug] >>>>> >>>>> >>>>> Options: >>>>> >>>>> -v Verbose output >>>>> >>>>> -d Resource directory off/on (0/1), default 0 >>>>> >>>>> -m Permission mode for lockspace device (octal), >>>>> default 0600 >>>>> >>>>> -M Print MSTCPY locks in lockdump (remote locks, >>>>> locally mastered) >>>>> >>>>> -h Print this help, then exit >>>>> >>>>> -V Print program version information, then exit >>>>> >>>>> - >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> On Wed, May 7, 2014 at 3:40 PM, emmanuel segura wrote: >>>>> >>>>>> dlm_tool ls ? >>>>>> >>>>>> >>>>>> 2014-05-07 21:05 GMT+02:00 Paras pradhan : >>>>>> >>>>>>> Well I have a qdisk with vote 3 . Thats why it is 6. >>>>>>> >>>>>>> Here is the log. I see some GFS hung but no issue with GFS mounts at >>>>>>> this time. >>>>>>> >>>>>>> http://pastebin.com/MP4BF86c >>>>>>> >>>>>>> I am seeing this at clumond.log not sure if this is related and what >>>>>>> is it. >>>>>>> >>>>>>> Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue >>>>>>> 23340->11670 >>>>>>> >>>>>>> Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue >>>>>>> 23340->11670 >>>>>>> >>>>>>> Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue >>>>>>> 23340->11670 >>>>>>> >>>>>>> Thanks >>>>>>> Paras >>>>>>> >>>>>>> >>>>>>> On Wed, May 7, 2014 at 1:51 PM, emmanuel segura wrote: >>>>>>> >>>>>>>> where is your log? >>>>>>>> >>>>>>>> I don't think this is the problem, but anyway from your config i >>>>>>>> saw >>>>>>> >>>>>>>> from man cman >>>>>>>> >>>>>>>> Expected votes >>>>>>>> The expected votes value is used by cman to determine >>>>>>>> quorum. The cluster is quorate if the sum of votes of existing members is >>>>>>>> over half of the expected votes value. By default, cman >>>>>>>> sets the expected votes value to be the sum of votes of all nodes listed in >>>>>>>> cluster.conf. This can be overriden by setting an explicit >>>>>>>> expected_votes value as follows: >>>>>>>> >>>>>>>> If you remove this expected_votes="6", the cluster will set this >>>>>>>> parameter to 3 >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> 2014-05-07 20:38 GMT+02:00 emmanuel segura : >>>>>>>> >>>>>>>> from your previous outpout of cman_tool services >>>>>>>>> >>>>>>>>> [1 2 3] >>>>>>>>> dlm 1 rgmanager 00030001 none >>>>>>>>> >>>>>>>>> >>>>>>>>> 2014-05-07 20:24 GMT+02:00 Paras pradhan : >>>>>>>>> >>>>>>>>> Oh. How did you see that? >>>>>>>>>> >>>>>>>>>> Here is the cluster.conf http://pastebin.com/DveLMGXT >>>>>>>>>> >>>>>>>>>> Thanks! >>>>>>>>>> -Paras. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Wed, May 7, 2014 at 1:07 PM, emmanuel segura < >>>>>>>>>> emi2fast at gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> i saw your rgmanager lockspace is there, you see any error in >>>>>>>>>>> your msg? can show your cluster config? >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> 2014-05-07 19:52 GMT+02:00 Paras pradhan >>>>>>>>>> >: >>>>>>>>>>> >>>>>>>>>>> Thats looks good. >>>>>>>>>>>> >>>>>>>>>>>> #cman_tool services >>>>>>>>>>>> >>>>>>>>>>>> type level name id state >>>>>>>>>>>> >>>>>>>>>>>> fence 0 default 00010001 none >>>>>>>>>>>> >>>>>>>>>>>> [1 2 3] >>>>>>>>>>>> >>>>>>>>>>>> dlm 1 clvmd 00020001 none >>>>>>>>>>>> >>>>>>>>>>>> [1 2 3] >>>>>>>>>>>> >>>>>>>>>>>> dlm 1 guest_comp_vms1 00020003 none >>>>>>>>>>>> >>>>>>>>>>>> [1 2 3] >>>>>>>>>>>> >>>>>>>>>>>> dlm 1 guest_comp_vms2 00040003 none >>>>>>>>>>>> >>>>>>>>>>>> [1 2 3] >>>>>>>>>>>> >>>>>>>>>>>> dlm 1 guest_comp_vms3 00060003 none >>>>>>>>>>>> >>>>>>>>>>>> [1 2 3] >>>>>>>>>>>> >>>>>>>>>>>> dlm 1 rgmanager 00030001 none >>>>>>>>>>>> >>>>>>>>>>>> [1 2 3] >>>>>>>>>>>> >>>>>>>>>>>> gfs 2 guest_comp_vms1 00010003 none >>>>>>>>>>>> >>>>>>>>>>>> [1 2 3] >>>>>>>>>>>> >>>>>>>>>>>> gfs 2 guest_comp_vms2 00030003 none >>>>>>>>>>>> >>>>>>>>>>>> [1 2 3] >>>>>>>>>>>> >>>>>>>>>>>> gfs 2 guest_comp_vms3 00050003 none >>>>>>>>>>>> >>>>>>>>>>>> [1 2 3] >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Wed, May 7, 2014 at 12:46 PM, emmanuel segura < >>>>>>>>>>>> emi2fast at gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> cman_tool services? >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> 2014-05-07 19:28 GMT+02:00 hugo aldunce : >>>>>>>>>>>>> >>>>>>>>>>>>> ohh good luck! >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> 2014-05-07 11:14 GMT-04:00 Paras pradhan < >>>>>>>>>>>>>> pradhanparas at gmail.com>: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>> We had some network problem the other day and today I >>>>>>>>>>>>>>> noticed the clusvcadm commands are not working. For example: it does not >>>>>>>>>>>>>>> want to stop the service, migrate a vm etc etc. On one of the nodes clustat >>>>>>>>>>>>>>> does not show any running services. I should restart the rgmanager? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> This is RHEL 5. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks >>>>>>>>>>>>>>> Paras. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> >>>>>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>>>> Hugo Aldunce E >>>>>>>>>>>>>> Tel. 09 82121045 >>>>>>>>>>>>>> mail: haldunce at gmail.com >>>>>>>>>>>>>> >>>>>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster at redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster at redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster at redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster at redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster at redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> >>>> -- >>>> esta es mi vida e me la vivo hasta que dios quiera >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster at redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From lists at alteeve.ca Wed May 7 22:19:19 2014 From: lists at alteeve.ca (Digimer) Date: Wed, 07 May 2014 18:19:19 -0400 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: Message-ID: <536AB167.4070401@alteeve.ca> On 07/05/14 03:05 PM, Paras pradhan wrote: > Well I have a qdisk with vote 3 . Thats why it is 6. > > Here is the log. I see some GFS hung but no issue with GFS mounts at > this time. > > http://pastebin.com/MP4BF86c > > I am seeing this at clumond.log not sure if this is related and what is it. > > Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue > 23340->11670 > > Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue > 23340->11670 > > Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue > 23340->11670 > > > Thanks > Paras Was there a failed fence action prior to this? If so, DLM is probably blocked. Can you post your logs starting from just prior to the network interruption? -- Digimer Papers and Projects: https://alteeve.ca/w/ What if the cure for cancer is trapped in the mind of a person without access to education? From pradhanparas at gmail.com Wed May 7 22:41:17 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 17:41:17 -0500 Subject: [Linux-cluster] clusvcadm In-Reply-To: <536AB167.4070401@alteeve.ca> References: <536AB167.4070401@alteeve.ca> Message-ID: Yes the fencing failed because of the network issues and I had to reboot one of the manually and it came back normal. But now I am seeing this DLM is blocked. I see this in the logs -- Apr 29 10:08:11 vprd2 fenced[7250]: agent "fence_drac5" reports: Unable to connect/login to fencing device Apr 29 10:08:11 vprd2 fenced[7250]: fence "vprd1.domain" failed -- Thanks Paras. On Wed, May 7, 2014 at 5:19 PM, Digimer wrote: > On 07/05/14 03:05 PM, Paras pradhan wrote: > >> Well I have a qdisk with vote 3 . Thats why it is 6. >> >> Here is the log. I see some GFS hung but no issue with GFS mounts at >> this time. >> >> http://pastebin.com/MP4BF86c >> >> I am seeing this at clumond.log not sure if this is related and what is >> it. >> >> Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue >> 23340->11670 >> >> Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue >> 23340->11670 >> >> Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue >> 23340->11670 >> >> >> Thanks >> Paras >> > > Was there a failed fence action prior to this? If so, DLM is probably > blocked. > > Can you post your logs starting from just prior to the network > interruption? > > -- > Digimer > Papers and Projects: https://alteeve.ca/w/ > What if the cure for cancer is trapped in the mind of a person without > access to education? > > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -------------- next part -------------- An HTML attachment was scrubbed... URL: From emi2fast at gmail.com Wed May 7 22:50:35 2014 From: emi2fast at gmail.com (emmanuel segura) Date: Thu, 8 May 2014 00:50:35 +0200 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: <536AB167.4070401@alteeve.ca> Message-ID: Paras, But you told me, that your fencing is working, 2014-05-08 0:41 GMT+02:00 Paras pradhan : > Yes the fencing failed because of the network issues and I had to reboot > one of the manually and it came back normal. But now I am seeing this DLM > is blocked. > > I see this in the logs > -- > > Apr 29 10:08:11 vprd2 fenced[7250]: agent "fence_drac5" reports: Unable to > connect/login to fencing device > > Apr 29 10:08:11 vprd2 fenced[7250]: fence "vprd1.domain" failed > > -- > > Thanks > > Paras. > > > On Wed, May 7, 2014 at 5:19 PM, Digimer wrote: > >> On 07/05/14 03:05 PM, Paras pradhan wrote: >> >>> Well I have a qdisk with vote 3 . Thats why it is 6. >>> >>> Here is the log. I see some GFS hung but no issue with GFS mounts at >>> this time. >>> >>> http://pastebin.com/MP4BF86c >>> >>> I am seeing this at clumond.log not sure if this is related and what is >>> it. >>> >>> Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue >>> 23340->11670 >>> >>> Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue >>> 23340->11670 >>> >>> Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue >>> 23340->11670 >>> >>> >>> Thanks >>> Paras >>> >> >> Was there a failed fence action prior to this? If so, DLM is probably >> blocked. >> >> Can you post your logs starting from just prior to the network >> interruption? >> >> -- >> Digimer >> Papers and Projects: https://alteeve.ca/w/ >> What if the cure for cancer is trapped in the mind of a person without >> access to education? >> >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera -------------- next part -------------- An HTML attachment was scrubbed... URL: From pradhanparas at gmail.com Wed May 7 23:09:21 2014 From: pradhanparas at gmail.com (Paras pradhan) Date: Wed, 7 May 2014 18:09:21 -0500 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: <536AB167.4070401@alteeve.ca> Message-ID: Well yes that was few days back and my impression was after the reboot everything is normal but now this DLM debug is showing the locking. Anything I can do to solve this problem without taking the outage?. services are not migrating and stopping at this time. Thanks Paras. On Wed, May 7, 2014 at 5:50 PM, emmanuel segura wrote: > Paras, > > But you told me, that your fencing is working, > > > 2014-05-08 0:41 GMT+02:00 Paras pradhan : > > Yes the fencing failed because of the network issues and I had to reboot >> one of the manually and it came back normal. But now I am seeing this DLM >> is blocked. >> >> I see this in the logs >> -- >> >> Apr 29 10:08:11 vprd2 fenced[7250]: agent "fence_drac5" reports: Unable >> to connect/login to fencing device >> >> Apr 29 10:08:11 vprd2 fenced[7250]: fence "vprd1.domain" failed >> >> -- >> >> Thanks >> >> Paras. >> >> >> On Wed, May 7, 2014 at 5:19 PM, Digimer wrote: >> >>> On 07/05/14 03:05 PM, Paras pradhan wrote: >>> >>>> Well I have a qdisk with vote 3 . Thats why it is 6. >>>> >>>> Here is the log. I see some GFS hung but no issue with GFS mounts at >>>> this time. >>>> >>>> http://pastebin.com/MP4BF86c >>>> >>>> I am seeing this at clumond.log not sure if this is related and what is >>>> it. >>>> >>>> Mon May 5 21:58:20 2014 clumond: Peer (vprd3.domain): pruning queue >>>> 23340->11670 >>>> >>>> Tue May 6 01:38:57 2014 clumond: Peer (vprd3.domain): pruning queue >>>> 23340->11670 >>>> >>>> Tue May 6 01:39:02 2014 clumond: Peer (vprd1.domain): pruning queue >>>> 23340->11670 >>>> >>>> >>>> Thanks >>>> Paras >>>> >>> >>> Was there a failed fence action prior to this? If so, DLM is probably >>> blocked. >>> >>> Can you post your logs starting from just prior to the network >>> interruption? >>> >>> -- >>> Digimer >>> Papers and Projects: https://alteeve.ca/w/ >>> What if the cure for cancer is trapped in the mind of a person without >>> access to education? >>> >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster at redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> -- >> Linux-cluster mailing list >> Linux-cluster at redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -------------- next part -------------- An HTML attachment was scrubbed... URL: From lists at alteeve.ca Wed May 7 23:13:59 2014 From: lists at alteeve.ca (Digimer) Date: Wed, 07 May 2014 19:13:59 -0400 Subject: [Linux-cluster] clusvcadm In-Reply-To: References: <536AB167.4070401@alteeve.ca> Message-ID: <536ABE37.7060406@alteeve.ca> On 07/05/14 06:41 PM, Paras pradhan wrote: > Yes the fencing failed because of the network issues and I had to reboot > one of the manually and it came back normal. But now I am seeing this > DLM is blocked. Then this is why you're stuck. RGmanager, gfs2 and clvmd all use DLM and they will block when DLM blocks. When a fence is called, fenced informs dlm, and dlm remains blocked until fenced tells dlm that it's safe to recover. If you are *certain* that the one nodes is powered off (or rebooted and cman/rgmanager was not started), you can run 'fence_ack_manual' to tell fenced that the node was fenced. Be very careful with this; improper use can corrupt your cluster. Alternatively, restart all nodes. -- Digimer Papers and Projects: https://alteeve.ca/w/ What if the cure for cancer is trapped in the mind of a person without access to education? From Mark.Vallevand at UNISYS.com Fri May 9 19:04:49 2014 From: Mark.Vallevand at UNISYS.com (Vallevand, Mark K) Date: Fri, 9 May 2014 14:04:49 -0500 Subject: [Linux-cluster] Temporarily moving a resource to another node Message-ID: <99C8B2929B39C24493377AC7A121E21FD9B2B45F07@USEA-EXCH8.na.uis.unisys.com> My resource agent detects an error condition and needs to temporarily move a resource from the current node. So, I have the monitor action return $OCF_ERR_PERM to report a hard error. This is not a permanent move, is it? The resource could return to this node at some time in the future, right? Or, do I need to take explicit actions? Regards. Mark K Vallevand Mark.Vallevand at Unisys.com May you live in interesting times, may you come to the attention of important people and may all your wishes come true. THIS COMMUNICATION MAY CONTAIN CONFIDENTIAL AND/OR OTHERWISE PROPRIETARY MATERIAL and is thus for use only by the intended recipient. If you received this in error, please contact the sender and delete the e-mail and its attachments from all computers. -------------- next part -------------- An HTML attachment was scrubbed... URL: From andrew at beekhof.net Wed May 14 05:21:27 2014 From: andrew at beekhof.net (Andrew Beekhof) Date: Wed, 14 May 2014 15:21:27 +1000 Subject: [Linux-cluster] Temporarily moving a resource to another node In-Reply-To: <99C8B2929B39C24493377AC7A121E21FD9B2B45F07@USEA-EXCH8.na.uis.unisys.com> References: <99C8B2929B39C24493377AC7A121E21FD9B2B45F07@USEA-EXCH8.na.uis.unisys.com> Message-ID: <5FFB5BFF-44C1-49D3-AE88-92761E8D682D@beekhof.net> On 10 May 2014, at 5:04 am, Vallevand, Mark K wrote: > My resource agent detects an error condition and needs to temporarily move a resource from the current node. > So, I have the monitor action return $OCF_ERR_PERM to report a hard error. > This is not a permanent move, is it? Yes > The resource could return to this node at some time in the future, right? Not without a resource cleanup or setting failure-timeout for the resource > Or, do I need to take explicit actions? > > Regards. > Mark K Vallevand Mark.Vallevand at Unisys.com > May you live in interesting times, may you come to the attention of important people and may all your wishes come true. > THIS COMMUNICATION MAY CONTAIN CONFIDENTIAL AND/OR OTHERWISE PROPRIETARY MATERIAL and is thus for use only by the intended recipient. If you received this in error, please contact the sender and delete the e-mail and its attachments from all computers. > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 841 bytes Desc: Message signed with OpenPGP using GPGMail URL: From mgrac at redhat.com Wed May 14 13:00:34 2014 From: mgrac at redhat.com (Marek Grac) Date: Wed, 14 May 2014 15:00:34 +0200 Subject: [Linux-cluster] fence-agents-4.0.9 stable release Message-ID: <537368F2.2060002@redhat.com> Welcome to the fence-agents 4.0.9 release. This release includes new fence agent for Proxmox VE (thanks to Frank Brendel and Ondrej Mular) and several bugfixes: * action 'list' for WTI devices now prints also named groups * fence_ipmilan now works correctly when no password is entered * fix problem in fence_vmware on ESX servers * fence_amt now sends password over environment variables * add support for --delay for all python fence agents * fix problem with new pexpect library * fix problems in fence_ovh * code cleanup * test if XML metadata were not changed unintentionally * test if --delay argument waits before opening connection for python fence agents (others will follow soon) The new source tarball can be downloaded here: https://fedorahosted.org/releases/f/e/fence-agents/fence-agents-4.0.9.tar.xz To report bugs or issues: https://bugzilla.redhat.com/ Would you like to meet the cluster team or members of its community? Join us on IRC (irc.freenode.net #linux-cluster) and share your experience with other sysadministrators or power users. Thanks/congratulations to all people that contributed to achieve this great milestone. m, From devin.bougie at cornell.edu Thu May 15 15:16:52 2014 From: devin.bougie at cornell.edu (Devin A. Bougie) Date: Thu, 15 May 2014 15:16:52 +0000 Subject: [Linux-cluster] VM locking in EL6 clusters Message-ID: <3CD8C126-28A3-4526-8D20-28C280BCA1E8@cornell.edu> Hello, We are running three EL6 clusters using the Red Hat High Availability Add-On, and are encountering issues with locking our clustered VM?s. We followed the virtual machine disk locking documentation (http://libvirt.org/locking.html#sanlockstorage), specifically its recommendation to have sanlock's disk_lease_dir set to a shared GFS2 file system. Ideally we would store the leases on a clustered logical volume block device, but haven?t yet figured out to configure this. Our VM XML definitions are on the same GFS2 file system, and our KVM virtual machines are using clustered logical volumes for their block devices. For example, here are a few configuration excerpts. ??? - From /etc/cluster/cluster.conf: ... Greetings, I am looking to adapt fence_ipmilan to interact with a custom implementation of an IPMI BMC. Doing so requires the use of ipmitool's -t option to bridge IPMI requests to a specified internal (non-networked) hardware address. I do not see this option existing in fence_ipmilan or any of the other fence_agents modules. The ipmitool operation would be '/path/to/ipmitool -t 0x42 chassis power '. No network, IP, Auth, User, Password or other arguments required. I want to check with the developers to see if there is an existing path for this use case before submitting a patch for consideration. Thanks, --Jeff -- ------------------------------ Jeff Johnson Co-Founder Aeon Computing jeff.johnson "at" aeoncomputing dot com www.aeoncomputing.com t: 858-412-3810 x1001 f: 858-412-3845 4170 Morena Boulevard, Suite D - San Diego, CA 92117 High-performance Computing / Lustre Filesystems / Scale-out Storage From yamato at redhat.com Thu May 15 18:49:06 2014 From: yamato at redhat.com (Masatake YAMATO) Date: Fri, 16 May 2014 03:49:06 +0900 Subject: [Linux-cluster] [PATCH] fence-virtd: fix typo in debug message of do_fence_request_tcp Message-ID: <1400179746-17629-1-git-send-email-yamato@redhat.com> fence-virtd: fix typo in debug message of do_fence_request_tcp Signed-off-by: Masatake YAMATO --- server/mcast.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/server/mcast.c b/server/mcast.c index e850ec7..5fbe46a 100644 --- a/server/mcast.c +++ b/server/mcast.c @@ -250,7 +250,7 @@ do_fence_request_tcp(fence_req_t *req, mcast_info *info) fd = connect_tcp(req, info->args.auth, info->key, info->key_len); if (fd < 0) { - dbg_printf(2, "Could call back for fence request: %s\n", + dbg_printf(2, "Could not call back for fence request: %s\n", strerror(errno)); goto out; } -- 1.9.0 From lists at alteeve.ca Fri May 16 00:04:19 2014 From: lists at alteeve.ca (Digimer) Date: Thu, 15 May 2014 20:04:19 -0400 Subject: [Linux-cluster] fence_ipmilan / custom hardware target address (ipmitool -t hexaddr) In-Reply-To: <537509F5.10907@aeoncomputing.com> References: <537509F5.10907@aeoncomputing.com> Message-ID: <53755603.5000509@alteeve.ca> On 15/05/14 02:39 PM, Jeff Johnson wrote: > Greetings, > > I am looking to adapt fence_ipmilan to interact with a custom > implementation of an IPMI BMC. Doing so requires the use of ipmitool's > -t option to bridge IPMI requests to a specified internal > (non-networked) hardware address. > > I do not see this option existing in fence_ipmilan or any of the other > fence_agents modules. > > The ipmitool operation would be '/path/to/ipmitool -t 0x42 chassis power > '. No network, IP, Auth, User, Password or other arguments > required. > > I want to check with the developers to see if there is an existing path > for this use case before submitting a patch for consideration. > > Thanks, > > --Jeff Marek Grac, who I've cc'ed here, would be the best person to give advice on this. As a user, I think a simple patch to add your option would be fine. I do not believe (though stand to be corrected) that address, user or password is currently required with fence_ipmilan. If I am wrong and it is required, then perhaps forking fence_ipmilan to something like fence_ipmihw (or whatever) and then pushing it out as a new agent should be easy and could work. -- Digimer Papers and Projects: https://alteeve.ca/w/ What if the cure for cancer is trapped in the mind of a person without access to education? From christoph at macht-blau.org Wed May 21 16:02:31 2014 From: christoph at macht-blau.org (C. Handel) Date: Wed, 21 May 2014 18:02:31 +0200 Subject: [Linux-cluster] pacemaker location constraint Message-ID: location constraints are somehow not honored by pacemaker 1.1.10 on el6. I have an IP adress which is placed first and then a volumegroup and a filesystem which choose the same node. The IP should be placed on x432, but for some reason it chooses x430. There are additional resources running (also choosing strange nodes). the resource (pcs status): ip_x43c (ocf::heartbeat:IPaddr2): Started x430 the constraint (pcs constraint) Resource: x43c Enabled on: x432 (score:10001) Cluster properties (pcs property) cluster-infrastructure: cman cluster-recheck-interval: 60s default-resource-stickiness: 10 maintenance-mode: false symmetric-cluster: true checking the scoring vim crm_simulate -sL i get native_color: ip_x43c allocation score on x430: 30 native_color: ip_x43c allocation score on x431: 0 native_color: ip_x43c allocation score on x432: -INFINITY the score of 30 on x430 is ok. There is a resourcegroup with two resource with a colocation on the ip. But i can't figure out why x432 get's -INFINITY, there is no further constraint regarding any of the resources in question. I expect them to migrate to x432 after 60 seconds, but nothing happens. I trieds stopping vgfs_service_c, the ip remains. i stopped the ip. Started it again, comes up on x430 again. pacemaker version: pacemaker-libs-1.1.10-14.el6.x86_64 pacemaker-1.1.10-14.el6.x86_64 pacemaker-cli-1.1.10-14.el6.x86_64 pacemaker-cluster-libs-1.1.10-14.el6.x86_64 The stripped config is: node x430 node x431 node x432 primitive fs_service_c ocf:heartbeat:Filesystem \ params device="/dev/mapper/vg_service_c-service_c" directory="/common/service-c" fstype="ext4" \ op start interval="0" timeout="60s" \ op stop interval="0" timeout="60s" \ meta target-role="Started" primitive vg_service_c ocf:heartbeat:LVM \ params volgrpname="vg_service_c" exclusive="true" \ op start interval="0" timeout="120" \ op stop interval="0" timeout="120" \ op monitor interval="10" timeout="120" primitive ip_x43c ocf:heartbeat:IPaddr2 \ params ip="140.181.134.240" \ op monitor interval="30" timeout="20" group vgfs_service_c vg_service_c fs_service_c location location-ip_x43c-x432-10001 ip_x43c 10001: x432 colocation colocation-vgfs_service_c-ip_x43c-INFINITY inf: vgfs_service_c ip_x43c property $id="cib-bootstrap-options" \ dc-version="1.1.10-14.el6-368c726" \ cluster-infrastructure="cman" \ last-lrm-refresh="1400683270" \ stonith-enabled="true" \ stonith-action="poweroff" \ default-resource-stickiness="10" \ cluster-recheck-interval="60s" \ maintenance-mode="false" \ symmetric-cluster="true" Greeings Christoph -------------- next part -------------- An HTML attachment was scrubbed... URL: From andrew at beekhof.net Wed May 21 21:46:34 2014 From: andrew at beekhof.net (Andrew Beekhof) Date: Thu, 22 May 2014 07:46:34 +1000 Subject: [Linux-cluster] pacemaker location constraint In-Reply-To: References: Message-ID: On 22 May 2014, at 2:02 am, C. Handel wrote: > location constraints are somehow not honored by pacemaker 1.1.10 on el6. > I have an IP adress which is placed first and then a volumegroup and a filesystem which choose the same node. The IP should be placed on x432, but for some reason it chooses x430. There are additional resources running (also choosing strange nodes). > > > the resource (pcs status): > > ip_x43c (ocf::heartbeat:IPaddr2): Started x430 > > > > the constraint (pcs constraint) > > Resource: x43c > Enabled on: x432 (score:10001) > > > > Cluster properties (pcs property) > > cluster-infrastructure: cman > cluster-recheck-interval: 60s > default-resource-stickiness: 10 > maintenance-mode: false > symmetric-cluster: true > > > checking the scoring vim crm_simulate -sL i get > > native_color: ip_x43c allocation score on x430: 30 > native_color: ip_x43c allocation score on x431: 0 > native_color: ip_x43c allocation score on x432: -INFINITY > > the score of 30 on x430 is ok. There is a resourcegroup with two resource with a colocation on the ip. But i can't figure out why x432 get's -INFINITY, there is no further constraint colocation or it previously failed on that node > regarding any of the resources in question. I expect them to migrate to x432 after 60 seconds, but nothing happens. > > I trieds stopping vgfs_service_c, the ip remains. i stopped the ip. Started it again, comes up on x430 again. > > > pacemaker version: > > pacemaker-libs-1.1.10-14.el6.x86_64 > pacemaker-1.1.10-14.el6.x86_64 > pacemaker-cli-1.1.10-14.el6.x86_64 > pacemaker-cluster-libs-1.1.10-14.el6.x86_64 > > > > The stripped config is: yeah, don't do that. we need the whole thing (the cibadmin -Ql output in your case since you're using crmsh) > > node x430 > node x431 > node x432 > primitive fs_service_c ocf:heartbeat:Filesystem \ > params device="/dev/mapper/vg_service_c-service_c" directory="/common/service-c" fstype="ext4" \ > op start interval="0" timeout="60s" \ > op stop interval="0" timeout="60s" \ > meta target-role="Started" > primitive vg_service_c ocf:heartbeat:LVM \ > params volgrpname="vg_service_c" exclusive="true" \ > op start interval="0" timeout="120" \ > op stop interval="0" timeout="120" \ > op monitor interval="10" timeout="120" > primitive ip_x43c ocf:heartbeat:IPaddr2 \ > params ip="140.181.134.240" \ > op monitor interval="30" timeout="20" > group vgfs_service_c vg_service_c fs_service_c > location location-ip_x43c-x432-10001 ip_x43c 10001: x432 > colocation colocation-vgfs_service_c-ip_x43c-INFINITY inf: vgfs_service_c ip_x43c > property $id="cib-bootstrap-options" \ > dc-version="1.1.10-14.el6-368c726" \ > cluster-infrastructure="cman" \ > last-lrm-refresh="1400683270" \ > stonith-enabled="true" \ > stonith-action="poweroff" \ > default-resource-stickiness="10" \ > cluster-recheck-interval="60s" \ > maintenance-mode="false" \ > symmetric-cluster="true" > > > > Greeings > Christoph > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 841 bytes Desc: Message signed with OpenPGP using GPGMail URL: From christoph at macht-blau.org Thu May 22 07:35:30 2014 From: christoph at macht-blau.org (C. Handel) Date: Thu, 22 May 2014 09:35:30 +0200 Subject: [Linux-cluster] pacemaker location constraint In-Reply-To: References: Message-ID: > > The stripped config is: > yeah, don't do that. we need the whole thing (the cibadmin -Ql output in > your case since you're using crmsh) i currently mix pcs and crmsh. el6 now includes pcs and no longer crmsh, so i try to learn the new default ;) full output from pcs config below. There are three service groups each with an ip. A) ip_a and nfsserver together with filesystems should run on x430, B) ip_b and service_b with its filesystem and puppet on x431, C) ip_c and service_c with its filesystem and nothing else on x432. Greetings Christoph On Wed, May 21, 2014 at 6:02 PM, C. Handel wrote: > location constraints are somehow not honored by pacemaker 1.1.10 on el6. > I have an IP adress which is placed first and then a volumegroup and a > filesystem which choose the same node. The IP should be placed on x432, but > for some reason it chooses x430. There are additional resources running > (also choosing strange nodes). > > > the resource (pcs status): > > ip_x43c (ocf::heartbeat:IPaddr2): Started x430 > > > > the constraint (pcs constraint) > > Resource: x43c > Enabled on: x432 (score:10001) > > > > Cluster properties (pcs property) > > cluster-infrastructure: cman > cluster-recheck-interval: 60s > default-resource-stickiness: 10 > maintenance-mode: false > symmetric-cluster: true > > > checking the scoring vim crm_simulate -sL i get > > native_color: ip_x43c allocation score on x430: 30 > native_color: ip_x43c allocation score on x431: 0 > native_color: ip_x43c allocation score on x432: -INFINITY > > the score of 30 on x430 is ok. There is a resourcegroup with two resource > with a colocation on the ip. But i can't figure out why x432 get's > -INFINITY, there is no further constraint regarding any of the resources in > question. I expect them to migrate to x432 after 60 seconds, but nothing > happens. > > I trieds stopping vgfs_service_c, the ip remains. i stopped the ip. > Started it again, comes up on x430 again. > > > pacemaker version: > > pacemaker-libs-1.1.10-14.el6.x86_64 > pacemaker-1.1.10-14.el6.x86_64 > pacemaker-cli-1.1.10-14.el6.x86_64 > pacemaker-cluster-libs-1.1.10-14.el6.x86_64 > > > > The stripped config is: > > node x430 > node x431 > node x432 > primitive fs_service_c ocf:heartbeat:Filesystem \ > params device="/dev/mapper/vg_service_c-service_c" > directory="/common/service-c" fstype="ext4" \ > op start interval="0" timeout="60s" \ > op stop interval="0" timeout="60s" \ > meta target-role="Started" > primitive vg_service_c ocf:heartbeat:LVM \ > params volgrpname="vg_service_c" exclusive="true" \ > op start interval="0" timeout="120" \ > op stop interval="0" timeout="120" \ > op monitor interval="10" timeout="120" > primitive ip_x43c ocf:heartbeat:IPaddr2 \ > params ip="140.181.134.240" \ > op monitor interval="30" timeout="20" > group vgfs_service_c vg_service_c fs_service_c > location location-ip_x43c-x432-10001 ip_x43c 10001: x432 > colocation colocation-vgfs_service_c-ip_x43c-INFINITY inf: vgfs_service_c > ip_x43c > property $id="cib-bootstrap-options" \ > dc-version="1.1.10-14.el6-368c726" \ > cluster-infrastructure="cman" \ > last-lrm-refresh="1400683270" \ > stonith-enabled="true" \ > stonith-action="poweroff" \ > default-resource-stickiness="10" \ > cluster-recheck-interval="60s" \ > maintenance-mode="false" \ > symmetric-cluster="true" > > > > Greeings > Christoph > -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- Cluster Name: Corosync Nodes: Pacemaker Nodes: x430m x431m x432m Resources: Resource: nfsserver (class=ocf provider=heartbeat type=nfsserver) Attributes: nfs_shared_infodir=/common/nfsha/shared copy_export=true Operations: start interval=0 timeout=60 (nfsserver-start-0) stop interval=0 timeout=60 (nfsserver-stop-0) Group: vgfs_export Meta Attrs: target-role=Started Resource: vg_export (class=ocf provider=heartbeat type=LVM) Attributes: volgrpname=vg_export exclusive=true Meta Attrs: is-managed=true Operations: start interval=0 timeout=120 (vg_export-start-0) stop interval=0 timeout=120 (vg_export-stop-0) monitor interval=10 timeout=120 (vg_export-monitor-10) Resource: fs_export (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_export-export directory=/common/export fstype=ext4 Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_export-start-0) stop interval=0 timeout=60s (fs_export-stop-0) Resource: fs_tftp (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_export-tftp directory=/common/tftp fstype=ext4 Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_tftp-start-0) stop interval=0 timeout=60s (fs_tftp-stop-0) Resource: fs_fonts (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_export-fonts directory=/common/fonts fstype=ext4 Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_fonts-start-0) stop interval=0 timeout=60s (fs_fonts-stop-0) Resource: fs_nfsha (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_export-nfsha directory=/common/nfsha fstype=ext4 Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_nfsha-start-0) stop interval=0 timeout=60s (fs_nfsha-stop-0) Group: vgfs_fesa Meta Attrs: target-role=Started Resource: vg_fesa (class=ocf provider=heartbeat type=LVM) Attributes: volgrpname=vg_fesa exclusive=true Operations: start interval=0 timeout=120 (vg_fesa-start-0) stop interval=0 timeout=120 (vg_fesa-stop-0) monitor interval=10 timeout=120 (vg_fesa-monitor-10) Resource: fs_fesa (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_fesa-fesa directory=/common/fesa fstype=ext4 options=nodev,nosuid Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_fesa-start-0) stop interval=0 timeout=60s (fs_fesa-stop-0) Resource: fs_fesadata (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_fesa-fesadata directory=/common/fesadata fstype=ext4 options=nodev,nosuid Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_fesa_data-start-0) stop interval=0 timeout=60s (fs_fesa_data-stop-0) Group: vgfs_home Resource: vg_home (class=ocf provider=heartbeat type=LVM) Attributes: volgrpname=vg_home exclusive=true Operations: start interval=0 timeout=120 (vg_home-start-0) stop interval=0 timeout=120 (vg_home-stop-0) monitor interval=10 timeout=120 (vg_home-monitor-10) Resource: fs_home (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_home-home directory=/common/home fstype=ext4 Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_home-start-0) stop interval=0 timeout=60s (fs_home-stop-0) Resource: ip_x43a (class=ocf provider=heartbeat type=IPaddr2) Attributes: ip=192.168.134.236 Operations: monitor interval=30 timeout=20 (ip_x43a-monitor-30) Group: vgfs_log Meta Attrs: target-role=Started Resource: vg_log (class=ocf provider=heartbeat type=LVM) Attributes: volgrpname=vg_log exclusive=true Operations: start interval=0 timeout=120 (vg_log-start-0) stop interval=0 timeout=120 (vg_log-stop-0) monitor interval=10 timeout=120 (vg_log-monitor-10) Resource: fs_log (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_log-log directory=/common/log fstype=ext4 Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_log-start-0) stop interval=0 timeout=60s (fs_log-stop-0) Group: vgfs_software Meta Attrs: target-role=Started Resource: vg_software (class=ocf provider=heartbeat type=LVM) Attributes: volgrpname=vg_software exclusive=true Operations: start interval=0 timeout=120 (vg_software-start-0) stop interval=0 timeout=120 (vg_software-stop-0) monitor interval=10 timeout=120 (vg_software-monitor-10) Resource: fs_software (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_software-software directory=/common/software fstype=ext4 Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_software-start-0) stop interval=0 timeout=60s (fs_software-stop-0) Resource: ip_x43b (class=ocf provider=heartbeat type=IPaddr2) Attributes: ip=192.168.134.238 Operations: monitor interval=30 timeout=20 (ip_x43b-monitor-30) Resource: ip_x43c (class=ocf provider=heartbeat type=IPaddr2) Attributes: ip=192.168.134.240 Operations: monitor interval=30 timeout=20 (ip_x43c-monitor-30) Group: vgfs_service_b Meta Attrs: target-role=Started Resource: vg_service_b (class=ocf provider=heartbeat type=LVM) Attributes: volgrpname=vg_service_b exclusive=true Operations: start interval=0 timeout=120 (vg_service_b-start-0) stop interval=0 timeout=120 (vg_service_b-stop-0) monitor interval=10 timeout=120 (vg_service_b-monitor-10) Resource: fs_service_b (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_service_b-service_b directory=/common/service-b fstype=ext4 Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_service_b-start-0) stop interval=0 timeout=60s (fs_service_b-stop-0) Resource: puppetmaster (class=lsb type=puppetmaster) Operations: start interval=0 timeout=60s (puppetmaster-start-0) stop interval=0 timeout=60s (puppetmaster-stop-0) monitor interval=60s timeout=30s (puppetmaster-monitor-60s) Resource: tftp (class=ocf provider=heartbeat type=tftpd) Attributes: address=192.168.134.236 directory=/common/tftp Meta Attrs: target-role=Started Group: vgfs_service_c Resource: vg_service_c (class=ocf provider=heartbeat type=LVM) Attributes: volgrpname=vg_service_c exclusive=true Operations: start interval=0 timeout=120 (vg_service_c-start-0) stop interval=0 timeout=120 (vg_service_c-stop-0) monitor interval=10 timeout=120 (vg_service_c-monitor-10) Resource: fs_service_c (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/mapper/vg_service_c-service_c directory=/common/service-c fstype=ext4 Meta Attrs: target-role=Started Operations: start interval=0 timeout=60s (fs_service_c-start-0) stop interval=0 timeout=60s (fs_service_c-stop-0) Stonith Devices: Resource: fence_x430 (class=stonith type=fence_ipmilan) Attributes: action=off login=cluster passwd=ecnef ipaddr=x430i lanplus=true privlvl=OPERATOR pcmk_poweroff_action=off pcmk_host_check=static-list pcmk_host_list=x430m Resource: fence_x431 (class=stonith type=fence_ipmilan) Attributes: action=off login=cluster passwd=ecnef ipaddr=x431i lanplus=true privlvl=OPERATOR pcmk_poweroff_action=off pcmk_host_check=static-list pcmk_host_list=x431m Resource: fence_x432 (class=stonith type=fence_ipmilan) Attributes: action=off login=cluster passwd=ecnef ipaddr=x432i lanplus=true privlvl=OPERATOR pcmk_poweroff_action=off pcmk_host_check=static-list pcmk_host_list=x432m Fencing Levels: Location Constraints: Resource: fence_x430 Disabled on: x430m (score:-INFINITY) (id:l_fence_x430) Resource: fence_x431 Disabled on: x431m (score:-INFINITY) (id:l_fence_x431) Resource: fence_x432 Disabled on: x432m (score:-INFINITY) (id:l_fence_x432) Resource: ip_x43a Enabled on: x430m (score:10000) (id:l_x43a) Resource: ip_x43b Enabled on: x431m (score:10000) (id:l_x43b) Resource: ip_x43c Enabled on: x432m (score:10001) (id:location-ip_x43c-x432m-10001) Ordering Constraints: vgfs_service_b then puppetmaster (INFINITY) (id:or-puppetmaster-after-vgfs_service_b) Resource Sets: set vgfs_export vgfs_fesa vgfs_home vgfs_log vgfs_software sequential=false (id:or-nfsserver-after-filesystem-0) set nfsserver (id:or-nfsserver-after-filesystem-1) setoptions score=INFINITY (id:or-nfsserver-after-filesystem) set vgfs_export ip_x43a sequential=false (id:or-tftp-after-vgfs_export-0) set tftp (id:or-tftp-after-vgfs_export-1) setoptions score=INFINITY (id:or-tftp-after-vgfs_export) Colocation Constraints: vgfs_export with ip_x43a (INFINITY) (id:co-ip_x43a-vgfs_export) vgfs_fesa with ip_x43a (INFINITY) (id:co-ip_x43a-vgfs_fesa) vgfs_home with ip_x43a (INFINITY) (id:co-ip_x43a-vgfs_home) vgfs_log with ip_x43a (INFINITY) (id:co-ip_x43a-vgfs_log) vgfs_software with ip_x43a (INFINITY) (id:co-ip_x43a-vgfs_software) vgfs_service_b with ip_x43b (INFINITY) (id:co-ip_x43b-vgfs_service_b) puppetmaster with vgfs_service_b (INFINITY) (id:co-puppetmaster-vgfs_service_b) tftp with vgfs_export (INFINITY) (id:co-tftp-vgfs_export) vgfs_service_c with ip_x43c (INFINITY) (id:colocation-vgfs_service_c-ip_x43c-INFINITY) Resource Sets: set nfsserver (id:co-nfsserver-filesystems-0) set vgfs_export vgfs_fesa vgfs_home vgfs_log vgfs_software sequential=false (id:co-nfsserver-filesystems-1) setoptions score=INFINITY (id:co-nfsserver-filesystems) Cluster Properties: cluster-infrastructure: cman cluster-recheck-interval: 60s dc-version: 1.1.10-14.el6-368c726 default-resource-stickiness: 10 last-lrm-refresh: 1400741976 maintenance-mode: false stonith-action: poweroff stonith-enabled: true symmetric-cluster: true -------------- next part -------------- A non-text attachment was scrubbed... Name: config.xml Type: text/xml Size: 83529 bytes Desc: not available URL: -------------- next part -------------- Current cluster status: Online: [ x430m x431m x432m ] nfsserver (ocf::gsi:nfsserver): Started x431m fence_x430 (stonith:fence_ipmilan): Started x432m fence_x431 (stonith:fence_ipmilan): Started x432m fence_x432 (stonith:fence_ipmilan): Started x430m Resource Group: vgfs_export vg_export (ocf::gsi:LVM): Started x431m fs_export (ocf::heartbeat:Filesystem): Started x431m fs_tftp (ocf::heartbeat:Filesystem): Started x431m fs_fonts (ocf::heartbeat:Filesystem): Started x431m fs_nfsha (ocf::heartbeat:Filesystem): Started x431m Resource Group: vgfs_fesa vg_fesa (ocf::gsi:LVM): Started x431m fs_fesa (ocf::heartbeat:Filesystem): Started x431m fs_fesadata (ocf::heartbeat:Filesystem): Started x431m Resource Group: vgfs_home vg_home (ocf::gsi:LVM): Started x431m fs_home (ocf::heartbeat:Filesystem): Started x431m ip_x43a (ocf::heartbeat:IPaddr2): Started x431m Resource Group: vgfs_log vg_log (ocf::gsi:LVM): Started x431m fs_log (ocf::heartbeat:Filesystem): Started x431m Resource Group: vgfs_software vg_software (ocf::gsi:LVM): Started x431m fs_software (ocf::heartbeat:Filesystem): Started x431m ip_x43b (ocf::heartbeat:IPaddr2): Started x431m ip_x43c (ocf::heartbeat:IPaddr2): Started x430m Resource Group: vgfs_service_b vg_service_b (ocf::gsi:LVM): Started x431m fs_service_b (ocf::heartbeat:Filesystem): Started x431m puppetmaster (lsb:puppetmaster): Started x431m tftp (ocf::gsi:tftpd): Started x431m Resource Group: vgfs_service_c vg_service_c (ocf::gsi:LVM): Started x430m fs_service_c (ocf::heartbeat:Filesystem): Started x430m Allocation scores: group_color: vgfs_export allocation score on x430m: 0 group_color: vgfs_export allocation score on x431m: 0 group_color: vgfs_export allocation score on x432m: 0 group_color: vg_export allocation score on x430m: -INFINITY group_color: vg_export allocation score on x431m: 10 group_color: vg_export allocation score on x432m: 0 group_color: fs_export allocation score on x430m: 0 group_color: fs_export allocation score on x431m: 10 group_color: fs_export allocation score on x432m: 0 group_color: fs_tftp allocation score on x430m: 0 group_color: fs_tftp allocation score on x431m: 10 group_color: fs_tftp allocation score on x432m: 0 group_color: fs_fonts allocation score on x430m: 0 group_color: fs_fonts allocation score on x431m: 10 group_color: fs_fonts allocation score on x432m: 0 group_color: fs_nfsha allocation score on x430m: 0 group_color: fs_nfsha allocation score on x431m: 10 group_color: fs_nfsha allocation score on x432m: 0 native_color: ip_x43a allocation score on x430m: -INFINITY native_color: ip_x43a allocation score on x431m: 210 native_color: ip_x43a allocation score on x432m: 0 native_color: vg_export allocation score on x430m: -INFINITY native_color: vg_export allocation score on x431m: 70 native_color: vg_export allocation score on x432m: -INFINITY native_color: fs_export allocation score on x430m: -INFINITY native_color: fs_export allocation score on x431m: 40 native_color: fs_export allocation score on x432m: -INFINITY native_color: fs_tftp allocation score on x430m: -INFINITY native_color: fs_tftp allocation score on x431m: 30 native_color: fs_tftp allocation score on x432m: -INFINITY native_color: fs_fonts allocation score on x430m: -INFINITY native_color: fs_fonts allocation score on x431m: 20 native_color: fs_fonts allocation score on x432m: -INFINITY native_color: fs_nfsha allocation score on x430m: -INFINITY native_color: fs_nfsha allocation score on x431m: 10 native_color: fs_nfsha allocation score on x432m: -INFINITY group_color: vgfs_fesa allocation score on x430m: 0 group_color: vgfs_fesa allocation score on x431m: 0 group_color: vgfs_fesa allocation score on x432m: 0 group_color: vg_fesa allocation score on x430m: -INFINITY group_color: vg_fesa allocation score on x431m: 10 group_color: vg_fesa allocation score on x432m: 0 group_color: fs_fesa allocation score on x430m: 0 group_color: fs_fesa allocation score on x431m: 10 group_color: fs_fesa allocation score on x432m: 0 group_color: fs_fesadata allocation score on x430m: 0 group_color: fs_fesadata allocation score on x431m: 10 group_color: fs_fesadata allocation score on x432m: 0 native_color: vg_fesa allocation score on x430m: -INFINITY native_color: vg_fesa allocation score on x431m: 40 native_color: vg_fesa allocation score on x432m: -INFINITY native_color: fs_fesa allocation score on x430m: -INFINITY native_color: fs_fesa allocation score on x431m: 20 native_color: fs_fesa allocation score on x432m: -INFINITY native_color: fs_fesadata allocation score on x430m: -INFINITY native_color: fs_fesadata allocation score on x431m: 10 native_color: fs_fesadata allocation score on x432m: -INFINITY group_color: vgfs_home allocation score on x430m: 0 group_color: vgfs_home allocation score on x431m: 0 group_color: vgfs_home allocation score on x432m: 0 group_color: vg_home allocation score on x430m: -INFINITY group_color: vg_home allocation score on x431m: 10 group_color: vg_home allocation score on x432m: 0 group_color: fs_home allocation score on x430m: 0 group_color: fs_home allocation score on x431m: 10 group_color: fs_home allocation score on x432m: 0 native_color: vg_home allocation score on x430m: -INFINITY native_color: vg_home allocation score on x431m: 30 native_color: vg_home allocation score on x432m: -INFINITY native_color: fs_home allocation score on x430m: -INFINITY native_color: fs_home allocation score on x431m: 10 native_color: fs_home allocation score on x432m: -INFINITY group_color: vgfs_log allocation score on x430m: 0 group_color: vgfs_log allocation score on x431m: 0 group_color: vgfs_log allocation score on x432m: 0 group_color: vg_log allocation score on x430m: -INFINITY group_color: vg_log allocation score on x431m: 10 group_color: vg_log allocation score on x432m: 0 group_color: fs_log allocation score on x430m: 0 group_color: fs_log allocation score on x431m: 10 group_color: fs_log allocation score on x432m: 0 native_color: vg_log allocation score on x430m: -INFINITY native_color: vg_log allocation score on x431m: 30 native_color: vg_log allocation score on x432m: -INFINITY native_color: fs_log allocation score on x430m: -INFINITY native_color: fs_log allocation score on x431m: 10 native_color: fs_log allocation score on x432m: -INFINITY group_color: vgfs_software allocation score on x430m: 0 group_color: vgfs_software allocation score on x431m: 0 group_color: vgfs_software allocation score on x432m: 0 group_color: vg_software allocation score on x430m: -INFINITY group_color: vg_software allocation score on x431m: 10 group_color: vg_software allocation score on x432m: 0 group_color: fs_software allocation score on x430m: 0 group_color: fs_software allocation score on x431m: 10 group_color: fs_software allocation score on x432m: 0 native_color: vg_software allocation score on x430m: -INFINITY native_color: vg_software allocation score on x431m: 30 native_color: vg_software allocation score on x432m: -INFINITY native_color: fs_software allocation score on x430m: -INFINITY native_color: fs_software allocation score on x431m: 10 native_color: fs_software allocation score on x432m: -INFINITY native_color: nfsserver allocation score on x430m: -INFINITY native_color: nfsserver allocation score on x431m: 10 native_color: nfsserver allocation score on x432m: -INFINITY native_color: fence_x430 allocation score on x430m: -INFINITY native_color: fence_x430 allocation score on x431m: 0 native_color: fence_x430 allocation score on x432m: 10 native_color: fence_x431 allocation score on x430m: 0 native_color: fence_x431 allocation score on x431m: -INFINITY native_color: fence_x431 allocation score on x432m: 10 native_color: fence_x432 allocation score on x430m: 10 native_color: fence_x432 allocation score on x431m: 0 native_color: fence_x432 allocation score on x432m: -INFINITY native_color: ip_x43b allocation score on x430m: 0 native_color: ip_x43b allocation score on x431m: 10040 native_color: ip_x43b allocation score on x432m: 0 native_color: ip_x43c allocation score on x430m: 30 native_color: ip_x43c allocation score on x431m: 0 native_color: ip_x43c allocation score on x432m: -INFINITY group_color: vgfs_service_b allocation score on x430m: 0 group_color: vgfs_service_b allocation score on x431m: 0 group_color: vgfs_service_b allocation score on x432m: 0 group_color: vg_service_b allocation score on x430m: 0 group_color: vg_service_b allocation score on x431m: 10 group_color: vg_service_b allocation score on x432m: 0 group_color: fs_service_b allocation score on x430m: 0 group_color: fs_service_b allocation score on x431m: 10 group_color: fs_service_b allocation score on x432m: 0 native_color: vg_service_b allocation score on x430m: -INFINITY native_color: vg_service_b allocation score on x431m: 30 native_color: vg_service_b allocation score on x432m: -INFINITY native_color: fs_service_b allocation score on x430m: -INFINITY native_color: fs_service_b allocation score on x431m: 10 native_color: fs_service_b allocation score on x432m: -INFINITY native_color: puppetmaster allocation score on x430m: -INFINITY native_color: puppetmaster allocation score on x431m: 10 native_color: puppetmaster allocation score on x432m: -INFINITY native_color: tftp allocation score on x430m: -INFINITY native_color: tftp allocation score on x431m: 10 native_color: tftp allocation score on x432m: -INFINITY group_color: vgfs_service_c allocation score on x430m: 0 group_color: vgfs_service_c allocation score on x431m: 0 group_color: vgfs_service_c allocation score on x432m: 0 group_color: vg_service_c allocation score on x430m: 10 group_color: vg_service_c allocation score on x431m: 0 group_color: vg_service_c allocation score on x432m: -INFINITY group_color: fs_service_c allocation score on x430m: 10 group_color: fs_service_c allocation score on x431m: 0 group_color: fs_service_c allocation score on x432m: 0 native_color: vg_service_c allocation score on x430m: 20 native_color: vg_service_c allocation score on x431m: -INFINITY native_color: vg_service_c allocation score on x432m: -INFINITY native_color: fs_service_c allocation score on x430m: 10 native_color: fs_service_c allocation score on x431m: -INFINITY native_color: fs_service_c allocation score on x432m: -INFINITY Transition Summary: From andrew at beekhof.net Mon May 26 03:15:40 2014 From: andrew at beekhof.net (Andrew Beekhof) Date: Mon, 26 May 2014 13:15:40 +1000 Subject: [Linux-cluster] pacemaker location constraint In-Reply-To: References: Message-ID: On 22 May 2014, at 5:35 pm, C. Handel wrote: > > > The stripped config is: > > > yeah, don't do that. we need the whole thing (the cibadmin -Ql output in > > your case since you're using crmsh) > > i currently mix pcs and crmsh. el6 now includes pcs and no longer crmsh, so i try to learn the new default ;) > > full output from pcs config below. There are three service groups each with an ip. > > A) ip_a and nfsserver together with filesystems should run on x430, > B) ip_b and service_b with its filesystem and puppet on x431, > C) ip_c and service_c with its filesystem and nothing else on x432. > > Greetings > Christoph > > > On Wed, May 21, 2014 at 6:02 PM, C. Handel wrote: > location constraints are somehow not honored by pacemaker 1.1.10 on el6. > I have an IP adress which is placed first and then a volumegroup and a filesystem which choose the same node. The IP should be placed on x432, but for some reason it chooses x430. There are additional resources running (also choosing strange nodes). The IP prefers not to run on 432 because vg_service_c is collocated with it and vg_service_c cannot run there: vg_service_c: migration-threshold=1000000 fail-count=1000000 last-failure='Wed May 21 18:52:24 2014' (as seen with crm_mon -f) > > > the resource (pcs status): > > ip_x43c (ocf::heartbeat:IPaddr2): Started x430 > > > > the constraint (pcs constraint) > > Resource: x43c > Enabled on: x432 (score:10001) > > > > Cluster properties (pcs property) > > cluster-infrastructure: cman > cluster-recheck-interval: 60s > default-resource-stickiness: 10 > maintenance-mode: false > symmetric-cluster: true > > > checking the scoring vim crm_simulate -sL i get > > native_color: ip_x43c allocation score on x430: 30 > native_color: ip_x43c allocation score on x431: 0 > native_color: ip_x43c allocation score on x432: -INFINITY > > the score of 30 on x430 is ok. There is a resourcegroup with two resource with a colocation on the ip. But i can't figure out why x432 get's -INFINITY, there is no further constraint regarding any of the resources in question. I expect them to migrate to x432 after 60 seconds, but nothing happens. > > I trieds stopping vgfs_service_c, the ip remains. i stopped the ip. Started it again, comes up on x430 again. > > > pacemaker version: > > pacemaker-libs-1.1.10-14.el6.x86_64 > pacemaker-1.1.10-14.el6.x86_64 > pacemaker-cli-1.1.10-14.el6.x86_64 > pacemaker-cluster-libs-1.1.10-14.el6.x86_64 > > > > The stripped config is: > > node x430 > node x431 > node x432 > primitive fs_service_c ocf:heartbeat:Filesystem \ > params device="/dev/mapper/vg_service_c-service_c" directory="/common/service-c" fstype="ext4" \ > op start interval="0" timeout="60s" \ > op stop interval="0" timeout="60s" \ > meta target-role="Started" > primitive vg_service_c ocf:heartbeat:LVM \ > params volgrpname="vg_service_c" exclusive="true" \ > op start interval="0" timeout="120" \ > op stop interval="0" timeout="120" \ > op monitor interval="10" timeout="120" > primitive ip_x43c ocf:heartbeat:IPaddr2 \ > params ip="140.181.134.240" \ > op monitor interval="30" timeout="20" > group vgfs_service_c vg_service_c fs_service_c > location location-ip_x43c-x432-10001 ip_x43c 10001: x432 > colocation colocation-vgfs_service_c-ip_x43c-INFINITY inf: vgfs_service_c ip_x43c > property $id="cib-bootstrap-options" \ > dc-version="1.1.10-14.el6-368c726" \ > cluster-infrastructure="cman" \ > last-lrm-refresh="1400683270" \ > stonith-enabled="true" \ > stonith-action="poweroff" \ > default-resource-stickiness="10" \ > cluster-recheck-interval="60s" \ > maintenance-mode="false" \ > symmetric-cluster="true" > > > > Greeings > Christoph > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 841 bytes Desc: Message signed with OpenPGP using GPGMail URL: From mgrac at redhat.com Mon May 26 08:56:42 2014 From: mgrac at redhat.com (Marek Grac) Date: Mon, 26 May 2014 10:56:42 +0200 Subject: [Linux-cluster] fence_ipmilan / custom hardware target address (ipmitool -t hexaddr) In-Reply-To: <53755603.5000509@alteeve.ca> References: <537509F5.10907@aeoncomputing.com> <53755603.5000509@alteeve.ca> Message-ID: <538301CA.6060707@redhat.com> On 05/16/2014 02:04 AM, Digimer wrote: > On 15/05/14 02:39 PM, Jeff Johnson wrote: >> Greetings, >> >> I am looking to adapt fence_ipmilan to interact with a custom >> implementation of an IPMI BMC. Doing so requires the use of ipmitool's >> -t option to bridge IPMI requests to a specified internal >> (non-networked) hardware address. >> >> I do not see this option existing in fence_ipmilan or any of the other >> fence_agents modules. >> >> The ipmitool operation would be '/path/to/ipmitool -t 0x42 chassis power >> '. No network, IP, Auth, User, Password or other arguments >> required. >> >> I want to check with the developers to see if there is an existing path >> for this use case before submitting a patch for consideration. >> >> Thanks, >> >> --Jeff > > Marek Grac, who I've cc'ed here, would be the best person to give > advice on this. > > As a user, I think a simple patch to add your option would be fine. I > do not believe (though stand to be corrected) that address, user or > password is currently required with fence_ipmilan. > > If I am wrong and it is required, then perhaps forking fence_ipmilan > to something like fence_ipmihw (or whatever) and then pushing it out > as a new agent should be easy and could work. > Currently, the IP address is required because we do not support self-fencing (as node can be in invalid state). How it works? Password and login are not required. m, From christoph at macht-blau.org Mon May 26 13:47:32 2014 From: christoph at macht-blau.org (C. Handel) Date: Mon, 26 May 2014 15:47:32 +0200 Subject: [Linux-cluster] pacemaker location constraint Message-ID: >>>> location constraints are somehow not honored by pacemaker 1.1.10 on el6. >>>> I have an IP adress which is placed first and then a volumegroup and a >>>> filesystem which choose the same node. The IP should be placed on x432, but >>>> for some reason it chooses x430. There are additional resources running >>>> (also choosing strange nodes). >>>> [...] >>>> The stripped config is: >>>> [...] >>> yeah, don't do that. we need the whole thing (the cibadmin -Ql output in your case since you're using crmsh) >> [config] > The IP prefers not to run on 432 because vg_service_c is collocated with it and vg_service_c cannot run there: > vg_service_c: migration-threshold=1000000 fail-count=1000000 last-failure='Wed May 21 18:52:24 2014' > (as seen with crm_mon -f) Thanks a lot. Everything now works as intendend. Greetings Christoph From gianluca.cecchi at gmail.com Thu May 29 08:31:19 2014 From: gianluca.cecchi at gmail.com (Gianluca Cecchi) Date: Thu, 29 May 2014 10:31:19 +0200 Subject: [Linux-cluster] Where logged messages when intra cluster down? Message-ID: Hello, on a two node cluster with CentOS 6.3 I set this in cluster.conf for testing Then I cut the intra cluster lan and only when the configured totem token timeset expires I get inside messages corosync[11945]: [TOTEM ] A processor failed, forming new configuration. corosync[11945]: [QUORUM] Members[1]: 1 corosync[11945]: [TOTEM ] A processor joined or left the membership and a new membership was formed. with every node trying to fence the other one. Where can I find messages during this delay (from cut time to token expired time) that let me know that the nodes are not communicating via the intracluster lan? I presume /var/log/cluster/corosync.log, but this file with debugging on is very verbose and it is not so clear to me how to pinpoint the lines describing that I'm having a problem .... Thanks in advance, Gianluca -------------- next part -------------- An HTML attachment was scrubbed... URL: From amjadcsu at gmail.com Thu May 29 11:23:11 2014 From: amjadcsu at gmail.com (Amjad Syed) Date: Thu, 29 May 2014 14:23:11 +0300 Subject: [Linux-cluster] Heartbeats using CMAN Message-ID: Hello, We are trying to setup a 2 node RHEL 6.5 oracle HA cluster in Active/passive mode. Each node has 2 nics, one connected to outside world and other private VLAN for heartbeat connections. The network admin here is claiming that if the external switch is down or cable is unplugged, how can the heartbeat function ,cause the private VLAN still sees the other node up , but it is actually down and not accessible to network. Apologies if this naive question, but i am not a networking guy and just trying to figure out how to with this. Sincerely, Amjad -------------- next part -------------- An HTML attachment was scrubbed... URL: From emi2fast at gmail.com Thu May 29 11:48:50 2014 From: emi2fast at gmail.com (emmanuel segura) Date: Thu, 29 May 2014 13:48:50 +0200 Subject: [Linux-cluster] Heartbeats using CMAN In-Reply-To: References: Message-ID: in a normal setup, we got two switches, one for public and other for private network, if you have a problem with public network, your cluster nodes can still comunicate with each others, using the private switch 2014-05-29 13:23 GMT+02:00 Amjad Syed : > Hello, > > We are trying to setup a 2 node RHEL 6.5 oracle HA cluster in > Active/passive mode. > > Each node has 2 nics, one connected to outside world and other private > VLAN for heartbeat connections. > > The network admin here is claiming that if the external switch is down or > cable is unplugged, how can the heartbeat function ,cause the private VLAN > still sees the other node up , but it is actually down and not accessible > to network. > > Apologies if this naive question, but i am not a networking guy and just > trying to figure out how to with this. > > Sincerely, > Amjad > > -- > Linux-cluster mailing list > Linux-cluster at redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera -------------- next part -------------- An HTML attachment was scrubbed... URL: From white.heron at yahoo.com Fri May 30 21:55:51 2014 From: white.heron at yahoo.com (YB Tan Sri Dato Sri' Adli a.k.a Dell) Date: Fri, 30 May 2014 14:55:51 -0700 (PDT) Subject: [Linux-cluster] FW: Govt Funded Courses in Kuala Lumpur, PMP, Mobile apps, Agile In-Reply-To: <4ts0xc4ou8lc.lpb638-13h1qex23@api.elasticemail.com> Message-ID: <1401486951.49254.YahooMailIosMobile@web163501.mail.gq1.yahoo.com>

Sent from Yahoo Mail for iPhone
-------------- next part -------------- An HTML attachment was scrubbed... URL: