[Linux-cluster] conga bug or my mistake?
Alex
linux at vfemail.net
Mon Aug 18 13:35:50 UTC 2008
On Monday 18 August 2008 16:05, Grisha G. wrote:
> Show us the logs from your nodes
All right, here comes output from /var/lib/luci/log/event.log when i am
accessing Cluster -> select my cluster (httpcluster) -> Shared Fence Devices
link:
Cluster: httpcluster
Agent type: Global Network Block Device
Name: gnbd_from_shds
Nodes using this device for fencing: 192.168.113.3
and now, hitting 192.168.113.3 link, i'll get the following:
[root at rhclm ~]# tail -f /var/lib/luci/log/event.log
[some old output omitted]
...
2008-08-18T16:25:07 ERROR Zope.SiteErrorLog
https://192.168.113.8:8084/luci/cluster/index_html
Traceback (innermost last):
Module ZPublisher.Publish, line 115, in publish
Module ZPublisher.mapply, line 88, in mapply
Module ZPublisher.Publish, line 41, in call_object
Module Shared.DC.Scripts.Bindings, line 311, in __call__
Module Shared.DC.Scripts.Bindings, line 348, in _bindAndExec
Module Products.PageTemplates.ZopePageTemplate, line 255, in _exec
Module Products.PageTemplates.PageTemplate, line 104, in pt_render
- <ZopePageTemplate at /luci/cluster/index_html>
Module TAL.TALInterpreter, line 238, in __call__
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 728, in do_defineMacro
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 457, in do_optTag_tal
Module TAL.TALInterpreter, line 442, in do_optTag
Module TAL.TALInterpreter, line 437, in no_tag
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 780, in do_defineSlot
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 457, in do_optTag_tal
Module TAL.TALInterpreter, line 442, in do_optTag
Module TAL.TALInterpreter, line 437, in no_tag
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 728, in do_defineMacro
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 749, in do_useMacro
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 457, in do_optTag_tal
Module TAL.TALInterpreter, line 442, in do_optTag
Module TAL.TALInterpreter, line 437, in no_tag
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 715, in do_condition
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 457, in do_optTag_tal
Module TAL.TALInterpreter, line 442, in do_optTag
Module TAL.TALInterpreter, line 437, in no_tag
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 715, in do_condition
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 457, in do_optTag_tal
Module TAL.TALInterpreter, line 442, in do_optTag
Module TAL.TALInterpreter, line 437, in no_tag
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 749, in do_useMacro
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 715, in do_condition
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 457, in do_optTag_tal
Module TAL.TALInterpreter, line 442, in do_optTag
Module TAL.TALInterpreter, line 437, in no_tag
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 691, in do_loop_tal
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 457, in do_optTag_tal
Module TAL.TALInterpreter, line 442, in do_optTag
Module TAL.TALInterpreter, line 437, in no_tag
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 715, in do_condition
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 457, in do_optTag_tal
Module TAL.TALInterpreter, line 442, in do_optTag
Module TAL.TALInterpreter, line 437, in no_tag
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 691, in do_loop_tal
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 457, in do_optTag_tal
Module TAL.TALInterpreter, line 442, in do_optTag
Module TAL.TALInterpreter, line 437, in no_tag
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 749, in do_useMacro
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 715, in do_condition
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 715, in do_condition
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 457, in do_optTag_tal
Module TAL.TALInterpreter, line 442, in do_optTag
Module TAL.TALInterpreter, line 437, in no_tag
Module TAL.TALInterpreter, line 281, in interpret
Module TAL.TALInterpreter, line 735, in do_useMacro
Module Products.PageTemplates.TALES, line 221, in evaluate
- URL: /luci/cluster/fence-macros
- Line 2034, Column 2
- Expression: standard:'here/fence-macros/macros/fence-instance-form-gnbd'
- Names:
{'container': <Folder at /luci/cluster>,
'context': <Folder at /luci/cluster>,
'default': <Products.PageTemplates.TALES.Default instance at
0xb75707ec>, 'here': <Folder at /luci/cluster>,
'loop': <Products.PageTemplates.TALES.SafeMapping object at 0xdcde8ac>,
'modules': <Products.PageTemplates.ZRPythonExpr._SecureModuleImporter
instance at 0xb75123ac>,
'nothing': None,
'options': {'args': ()},
'repeat': <Products.PageTemplates.TALES.SafeMapping object at
0xdcde8ac>, 'request': <HTTPRequest,
URL=https://192.168.113.8:8084/luci/cluster/index_html>,
'root': <Application at >,
'template': <ZopePageTemplate at /luci/cluster/index_html>,
'traverse_subpath': [],
'user': <PropertiedUser 'admin'>}
Module Products.PageTemplates.Expressions, line 185, in __call__
Module Products.PageTemplates.Expressions, line 173, in _eval
Module Products.PageTemplates.Expressions, line 127, in _eval
- __traceback_info__: here
Module Products.PageTemplates.Expressions, line 320, in restrictedTraverse
- __traceback_info__: {'path': ['fence-macros', 'macros',
'fence-instance-form-gnbd'], 'TraversalRequestNameStack': []}
KeyError: 'fence-instance-form-gnbd'
What other logs you need? In syslog (/var/log/messages) does not appear
something related to this event!
Regards,
Alx
>
> On Mon, Aug 18, 2008 at 1:27 PM, Alex <linux at vfemail.net> wrote:
> > Hello all,
> >
> > My current setup si similar with one described here:
> > http://sources.redhat.com/cluster/gnbd/gnbd_usage.txt
> > excepting the fact that i'm having 3 clients and 3 gnbd servers
> > (exporting block devices using gnbd).
> >
> > our gnbd servers have the following IPs: 192.168.113.6 and 192.168.113.7
> > our gnbd clients have the following IPs: 192.168.113.3 and
> > 192.168.113.4and 192.168.113.5
> >
> > On our management machine (other then above gnbd clients and servers) is
> > running:
> > [root at rhclm ~]# rpm -q luci
> > luci-0.12.0-7.el5.centos.3
> > [root at rhclm ~]#
> >
> > On our gnbd clients is running:
> > [root at rs1 ~]# rpm -q ricci
> > ricci-0.12.0-7.el5.centos.3
> > [root at rs1 ~]#
> >
> > Now, i'm trying to do the following operations using conga:
> > Cluster -> Shared Fence Devices -> Add Fence Device
> >
> > added successfully:
> >
> > Fence Type: GNBD
> > Name: gnbd_from_shds
> > Servers: 192.168.113.6 192.168.113.7
> >
> > This will add in our cluster.conf:
> > <fencedevices>
> > <fencedevice agent="fence_gnbd" name="gnbd_from_shds"
> > servers="192.168.113.6 192.168.113.7"/>
> > </fencedevices>
> >
> > Let's try to use it: Cluster -> Nodes hit on 192.168.113.3 and select
> > option
> > "Manage Fencing for this Node" -> "Main Fencing Method" -> "Add a fence
> > device to this level" -> select gnbd_from_shds -> and hit "Update main
> > fence
> > properties"
> >
> > Is not working, all the time i'm getting a javascript window error saying
> > the
> > following:
> >
> > [snip]
> > The following errors were found:
> > An unknown device type was given: "gnbd."
> > ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
> > [end snip]
> >
> > You see, is a dot after "gnbd" which i suppose it causing that error.
> >
> > How can be fixed this error?
> >
> > Now, I edited manually our cluster.conf as following:
> > <clusternode name="192.168.113.3" nodeid="3" votes="1">
> > <fence>
> > <method name="1">
> > <device name="gnbd_from_shds"
> > nodename="192.168.113.3"/>
> > </method>
> > </fence>
> > </clusternode>
> >
> > First Question: In docs, i cannot find any explanation about name="value"
> > in
> > <method> tag. As you see, value is "1": <method name="1">. Is this value
> > valid only inside of <clusternode> section or has global semnification in
> > cluster.conf? Can i name it for example "one" or
> > "first_fence_method_for_this_node"?
> >
> > and run:
> > [root at rs1 ~]# ccs_tool update /etc/cluster/cluster.conf
> > Config file updated from version 28 to 29
> >
> > Update complete.
> > [root at rs1 ~]#
> >
> > Now, i can see using conga in "Shared Fence Devices" section:
> >
> > Shared Fence Devices for Cluster: httpcluster
> > Agent type: Global Network Block Device
> > Name: gnbd_from_shds
> > Nodes using this device for fencing: 192.168.113.3
> >
> > but, if i'm hitting 192.168.113.3 link, i'll get other error:
> >
> > Site error
> >
> > This site encountered an error trying to fulfill your request. The errors
> > were:
> >
> > Error Type
> > KeyError
> > Error Value
> > 'fence-instance-form-gnbd'
> > Request made at
> > 2008/08/18 12:42:45.164 GMT+3
> >
> > Any ideas how to fix it? Is my mistake or is a bug in conga?
> >
> > Second Question: Is correct to add and use for the rest of our client
> > nodes below sintax?
> >
> > For: 192.168.113.4 and 192.168.113.5 client nodes:
> >
> > <clusternode name="192.168.113.4" nodeid="2" votes="1">
> > <fence>
> > <method name="1">
> > <device name="gnbd_from_shds"
> > nodename="192.168.113.4"/>
> > </method>
> > </fence>
> > </clusternode>
> >
> > and
> >
> > <clusternode name="192.168.113.5" nodeid="1" votes="1">
> > <fence>
> > <method name="1">
> > <device name="gnbd_from_shds"
> > nodename="192.168.113.5"/>
> > </method>
> > </fence>
> > </clusternode>
> >
> > For conformity, i am posting below my present cluster.conf file:
> >
> > <?xml version="1.0"?>
> > <cluster alias="httpcluster" config_version="29" name="httpcluster">
> > <fence_daemon clean_start="0" post_fail_delay="0"
> > post_join_delay="3"/>
> > <clusternodes>
> > <clusternode name="192.168.113.5" nodeid="1" votes="1">
> > <fence/>
> > </clusternode>
> > <clusternode name="192.168.113.4" nodeid="2" votes="1">
> > <fence/>
> > </clusternode>
> > <clusternode name="192.168.113.3" nodeid="3" votes="1">
> > <fence>
> > <method name="1">
> > <device name="gnbd_from_shds"
> > nodename="192.168.113.3"/>
> > </method>
> > </fence>
> > </clusternode>
> > <clusternode name="192.168.113.6" nodeid="4" votes="1">
> > <fence/>
> > </clusternode>
> > <clusternode name="192.168.113.7" nodeid="5" votes="1">
> > <fence/>
> > </clusternode>
> > </clusternodes>
> > <cman/>
> > <fencedevices>
> > <fencedevice agent="fence_gnbd" name="gnbd_from_shds"
> > servers="192.168.113.6 192.168.113.7"/>
> > </fencedevices>
> > <rm>
> > <failoverdomains/>
> > <resources/>
> > </rm>
> > </cluster>
> >
> > Regards,
> > Alx
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster at redhat.com
> > https://www.redhat.com/mailman/listinfo/linux-cluster
More information about the Linux-cluster
mailing list