[Linux-cluster] Linux-cluster Digest, Vol 117, Issue 6
vijay singh
vijay.240385 at gmail.com
Fri Jan 17 17:34:33 UTC 2014
Hi,
According to the clustat status it seems there is some issue with the
multi casting. Plz check the multicast using omping
On Jan 17, 2014 8:04 PM, <linux-cluster-request at redhat.com> wrote:
> Send Linux-cluster mailing list submissions to
> linux-cluster at redhat.com
>
> To subscribe or unsubscribe via the World Wide Web, visit
> https://www.redhat.com/mailman/listinfo/linux-cluster
> or, via email, send a message with subject or body 'help' to
> linux-cluster-request at redhat.com
>
> You can reach the person managing the list at
> linux-cluster-owner at redhat.com
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of Linux-cluster digest..."
>
>
> Today's Topics:
>
> 1. 2 node cluster questions (Benjamin Budts)
> 2. Re: 2 node cluster questions (emmanuel segura)
> 3. Is it gfs2 able to perform on a 24TB partition?
> (Juan Pablo Lorier)
> 4. Re: Is it gfs2 able to perform on a 24TB partition?
> (Steven Whitehouse)
> 5. Re: 2 node cluster questions (Benjamin Budts)
> 6. Re: 2 node cluster questions (emmanuel segura)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Thu, 16 Jan 2014 18:04:18 +0100
> From: "Benjamin Budts" <ben at zentrix.be>
> To: <linux-cluster at redhat.com>
> Subject: [Linux-cluster] 2 node cluster questions
> Message-ID: <002801cf12dc$fe8d1e40$fba75ac0$@zentrix.be>
> Content-Type: text/plain; charset="us-ascii"
>
>
>
> Hey All,
>
>
>
>
>
> About my setup :
>
>
>
> I created a cluster in luci with shared storage, added 2 nodes (reachable),
> added 2 fence devices (idrac).
>
>
>
> Now, I can't seem to add my fence devices to my nodes. If I click on a node
> it times out in the gui.
>
>
>
> So I checked my node logs and found in /var/log/messages ricci hangs on
> /etc/init.d/clvmd status
>
>
>
> When I run the command manually I get the PID and it hangs.
>
>
>
>
>
> I also seem to have a split brain config :
>
>
>
> # clustat on node 1 shows : node 1 Online,Local
>
> Node 2 Offline
>
>
>
> # clustat on node 2 shows : node 1 offline
>
> Node 2 online, Local
>
>
>
>
>
> My next step was testing if my multicast was working correctly (I suspect
> it
> isn't). Would you guys have any recommendations besides the following
> redhat
> multicast test link ? :
>
>
>
> Https://access.redhat.com/site/articles/22304
>
>
>
>
>
>
>
> Some info about my systems :
>
> ----------------------------------
>
>
>
> . Redhat 6.5 on 2 nodes with resilient storage & HA addon licenses
>
> . Mgmt. station redhat 6.5 with Luci
>
>
>
>
>
>
>
> Thx a lot
>
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <
> https://www.redhat.com/archives/linux-cluster/attachments/20140116/1e1c660e/attachment.html
> >
>
> ------------------------------
>
> Message: 2
> Date: Thu, 16 Jan 2014 18:17:11 +0100
> From: emmanuel segura <emi2fast at gmail.com>
> To: linux clustering <linux-cluster at redhat.com>
> Subject: Re: [Linux-cluster] 2 node cluster questions
> Message-ID:
> <
> CAE7pJ3DGsXJx602Tr73UjFFrrYpFi686TJeKQFqybxMv+5D00A at mail.gmail.com>
> Content-Type: text/plain; charset="windows-1252"
>
> If you think your problem is the multicast, try to use unicast
>
> https://access.redhat.com/site/documentation/en-US/Red_Hat_Enterprise_Linux/6/html/Cluster_Administration/s1-unicast-traffic-CA.html
> ,
> if you are using redhat 6.X
>
>
> 2014/1/16 Benjamin Budts <ben at zentrix.be>
>
> >
> >
> > Hey All,
> >
> >
> >
> >
> >
> > About my setup :
> >
> >
> >
> > I created a cluster in luci with shared storage, added 2 nodes
> > (reachable), added 2 fence devices (idrac).
> >
> >
> >
> > Now, I can?t seem to add my fence devices to my nodes. If I click on a
> > node it times out in the gui.
> >
> >
> >
> > So I checked my node logs and found in /var/log/messages ricci hangs on
> > /etc/init.d/clvmd status
> >
> >
> >
> > When I run the command manually I get the PID and it hangs?
> >
> >
> >
> >
> >
> > I also seem to have a split brain config :
> >
> >
> >
> > # clustat on node 1 shows : node 1 Online,Local
> >
> > Node 2 Offline
> >
> >
> >
> > # clustat on node 2 shows : node 1 offline
> >
> > Node 2 online, Local
> >
> >
> >
> >
> >
> > My next step was testing if my multicast was working correctly (I suspect
> > it isn?t). Would you guys have any recommendations besides the following
> > redhat multicast test link ? :
> >
> >
> >
> > *Https://access.redhat.com/site/articles/22304
> > <Https://access.redhat.com/site/articles/22304>*
> >
> >
> >
> >
> >
> >
> >
> > Some info about my systems :
> >
> > ----------------------------------
> >
> >
> >
> > ? Redhat 6.5 on 2 nodes with resilient storage & HA addon
> licenses
> >
> > ? Mgmt. station redhat 6.5 with Luci
> >
> >
> >
> >
> >
> >
> >
> > Thx a lot
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster at redhat.com
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> >
>
>
>
> --
> esta es mi vida e me la vivo hasta que dios quiera
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <
> https://www.redhat.com/archives/linux-cluster/attachments/20140116/b2678815/attachment.html
> >
>
> ------------------------------
>
> Message: 3
> Date: Fri, 17 Jan 2014 09:10:42 -0200
> From: Juan Pablo Lorier <jplorier at gmail.com>
> To: linux-cluster at redhat.com
> Subject: [Linux-cluster] Is it gfs2 able to perform on a 24TB
> partition?
> Message-ID: <52D90FB2.4060906 at gmail.com>
> Content-Type: text/plain; charset=ISO-8859-1
>
> Hi,
>
> I've been using gfs2 on top of a 24 TB lvm volume used as a file server
> for several month now and I see a lot of io related to glock_workqueue
> when there's file transfers. The threads even get to be the top ones in
> read ops in iotop.
> Is that normal? how can I debugg this?
> Regards,
>
>
>
> ------------------------------
>
> Message: 4
> Date: Fri, 17 Jan 2014 11:16:28 +0000
> From: Steven Whitehouse <swhiteho at redhat.com>
> To: Juan Pablo Lorier <jplorier at gmail.com>
> Cc: linux-cluster at redhat.com
> Subject: Re: [Linux-cluster] Is it gfs2 able to perform on a 24TB
> partition?
> Message-ID: <1389957388.2734.20.camel at menhir>
> Content-Type: text/plain; charset="UTF-8"
>
> Hi,
>
> On Fri, 2014-01-17 at 09:10 -0200, Juan Pablo Lorier wrote:
> > Hi,
> >
> > I've been using gfs2 on top of a 24 TB lvm volume used as a file server
> > for several month now and I see a lot of io related to glock_workqueue
> > when there's file transfers. The threads even get to be the top ones in
> > read ops in iotop.
> > Is that normal? how can I debugg this?
> > Regards,
> >
>
> Well the glock_workqueue threads are running the internal state machine
> that controls the caching of data. So depending on the workload in
> question, yes it is expected that you'll see a fair amount of activity
> there.
>
> If the issue is that you are seeing the cache being used inefficiently,
> then it maybe possible to improve that by looking at the i/o pattern
> generated by the application. There are also other things which can make
> a difference, such as setting noatime on the mount.
>
> You don't mention what version of the kernel you are using, but more
> recent kernels have better tools (such as tracepoints) which can assist
> in debugging this kind of issue,
>
> Steve.
>
>
>
>
> ------------------------------
>
> Message: 5
> Date: Fri, 17 Jan 2014 15:01:22 +0100
> From: "Benjamin Budts" <ben at zentrix.be>
> To: <linux-cluster at redhat.com>
> Subject: Re: [Linux-cluster] 2 node cluster questions
> Message-ID: <005e01cf138c$9a981100$cfc83300$@zentrix.be>
> Content-Type: text/plain; charset="us-ascii"
>
>
>
> Found my problem
>
>
>
> When running cman_tool status I saw that each node only saw 1 node.
>
> I have 2 pairs of network on each machine (heartbeat network 1gbit
> multicast
> enabled/application network 10Gbit) with their own host mapping in
> /etc/hosts
>
>
>
> I used the the non-heartbeat hostname to add nodes into cluster via luci
> which caused issues.
>
>
>
> Removed luci & /var/lib/luci on nodes, as well as cat /dev/null
> /etc/cluster/cluster.conf
>
> Reinstalled luci, created cluster, added 2 nodes with correct hostnames,
> problem solved.
>
>
>
> Could anyone point me to a good guide for using cluster lvm ?
>
>
>
> Thx
>
>
>
>
>
>
>
> From: Elvir Kuric [mailto:ekuric at redhat.com]
> Sent: vrijdag 17 januari 2014 10:42
> To: ben at zentrix.be
> Subject: Re: [Linux-cluster] 2 node cluster questions
>
>
>
> On 01/16/2014 06:04 PM, Benjamin Budts wrote:
>
>
>
> Hey All,
>
>
>
>
>
> About my setup :
>
>
>
> I created a cluster in luci with shared storage, added 2 nodes (reachable),
> added 2 fence devices (idrac).
>
>
>
> Now, I can't seem to add my fence devices to my nodes. If I click on a node
> it times out in the gui.
>
>
>
> So I checked my node logs and found in /var/log/messages ricci hangs on
> /etc/init.d/clvmd status
>
>
>
> When I run the command manually I get the PID and it hangs.
>
>
>
>
>
> I also seem to have a split brain config :
>
>
>
> # clustat on node 1 shows : node 1 Online,Local
>
> Node 2 Offline
>
>
>
> # clustat on node 2 shows : node 1 offline
>
> Node 2 online, Local
>
>
>
>
>
> My next step was testing if my multicast was working correctly (I suspect
> it
> isn't). Would you guys have any recommendations besides the following
> redhat
> multicast test link ? :
>
>
>
> Https://access.redhat.com/site/articles/22304
>
>
>
>
>
>
>
> Some info about my systems :
>
> ----------------------------------
>
>
>
> . Redhat 6.5 on 2 nodes with resilient storage & HA addon licenses
>
> . Mgmt. station redhat 6.5 with Luci
>
>
>
>
>
>
>
> Thx a lot
>
>
>
>
>
> Hard to say what issue is based on this without checking fully logs, can
> you
> please open case via Red Hat customer portal https://access.redhat.com ( I
> assume you have valid subscription ) so we can take deeper look into this
> problem.
>
> Thank you in advance,
>
> Kind regards,
>
>
>
>
>
> --
> Elvir Kuric,Sr. TSE / Red Hat / GSS EMEA /
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <
> https://www.redhat.com/archives/linux-cluster/attachments/20140117/6d673d84/attachment.html
> >
>
> ------------------------------
>
> Message: 6
> Date: Fri, 17 Jan 2014 15:27:42 +0100
> From: emmanuel segura <emi2fast at gmail.com>
> To: linux clustering <linux-cluster at redhat.com>
> Subject: Re: [Linux-cluster] 2 node cluster questions
> Message-ID:
> <
> CAE7pJ3Bi7woB42pD_dCS0xBrAs8uMgGCZCmfc9GyBzk3bwNr-A at mail.gmail.com>
> Content-Type: text/plain; charset="windows-1252"
>
>
> https://access.redhat.com/site/documentation/it-IT/Red_Hat_Enterprise_Linux/6/html/Logical_Volume_Manager_Administration/LVM_Cluster_Overview.html
>
>
> 2014/1/17 Benjamin Budts <ben at zentrix.be>
>
> >
> >
> > Found my problem
> >
> >
> >
> > When running cman_tool status I saw that each node only saw 1 node.
> >
> > I have 2 pairs of network on each machine (heartbeat network 1gbit
> > multicast enabled/application network 10Gbit) with their own host mapping
> > in /etc/hosts
> >
> >
> >
> > I used the the non-heartbeat hostname to add nodes into cluster via luci
> > which caused issues.
> >
> >
> >
> > Removed luci & /var/lib/luci on nodes, as well as cat /dev/null
> > /etc/cluster/cluster.conf
> >
> > Reinstalled luci, created cluster, added 2 nodes with correct hostnames,
> > problem solved.
> >
> >
> >
> > Could anyone point me to a good guide for using cluster lvm ?
> >
> >
> >
> > Thx
> >
> >
> >
> >
> >
> >
> >
> > *From:* Elvir Kuric [mailto:ekuric at redhat.com]
> > *Sent:* vrijdag 17 januari 2014 10:42
> > *To:* ben at zentrix.be
> > *Subject:* Re: [Linux-cluster] 2 node cluster questions
> >
> >
> >
> > On 01/16/2014 06:04 PM, Benjamin Budts wrote:
> >
> >
> >
> > Hey All,
> >
> >
> >
> >
> >
> > About my setup :
> >
> >
> >
> > I created a cluster in luci with shared storage, added 2 nodes
> > (reachable), added 2 fence devices (idrac).
> >
> >
> >
> > Now, I can?t seem to add my fence devices to my nodes. If I click on a
> > node it times out in the gui.
> >
> >
> >
> > So I checked my node logs and found in /var/log/messages ricci hangs on
> > /etc/init.d/clvmd status
> >
> >
> >
> > When I run the command manually I get the PID and it hangs?
> >
> >
> >
> >
> >
> > I also seem to have a split brain config :
> >
> >
> >
> > # clustat on node 1 shows : node 1 Online,Local
> >
> > Node 2 Offline
> >
> >
> >
> > # clustat on node 2 shows : node 1 offline
> >
> > Node 2 online, Local
> >
> >
> >
> >
> >
> > My next step was testing if my multicast was working correctly (I suspect
> > it isn?t). Would you guys have any recommendations besides the following
> > redhat multicast test link ? :
> >
> >
> >
> > *Https://access.redhat.com/site/articles/22304
> > <Https://access.redhat.com/site/articles/22304>*
> >
> >
> >
> >
> >
> >
> >
> > Some info about my systems :
> >
> > ----------------------------------
> >
> >
> >
> > ? Redhat 6.5 on 2 nodes with resilient storage & HA addon
> licenses
> >
> > ? Mgmt. station redhat 6.5 with Luci
> >
> >
> >
> >
> >
> >
> >
> > Thx a lot
> >
> >
> >
> > Hard to say what issue is based on this without checking fully logs, can
> > you please open case via Red Hat customer portal
> https://access.redhat.com( I assume you have valid subscription ) so we
> can take deeper look into
> > this problem.
> >
> > Thank you in advance,
> >
> > Kind regards,
> >
> >
> >
> > --
> >
> > Elvir Kuric,Sr. TSE / Red Hat / GSS EMEA /
> >
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster at redhat.com
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> >
>
>
>
> --
> esta es mi vida e me la vivo hasta que dios quiera
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <
> https://www.redhat.com/archives/linux-cluster/attachments/20140117/6fc1d7e7/attachment.html
> >
>
> ------------------------------
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>
> End of Linux-cluster Digest, Vol 117, Issue 6
> *********************************************
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20140117/716edb39/attachment.htm>
More information about the Linux-cluster
mailing list