[Linux-cluster] CLVM error 6 nodes

Filipe Miranda filipe.miranda at gmail.com
Sat Oct 14 17:25:09 UTC 2006


You are right Jim,

It's clvm=0, not clvmd=0!!

I wish I could have a document that has all the possible tags, options that
I could use on cluster.conf...

I Found one on the Cluster Projetc, but I think its not complete...

Thanks for the hint!

Regards,
Filipe Miranda

On 10/14/06, Jim Parsons <jparsons at redhat.com> wrote:
>
> Filipe Miranda wrote:
>
> > Alasdair,
> >
> >
> > We did the modifications but still the same error:
> >
> > Since the lock servers does not have access to the LUNs that will be
> > in use by the GFS I used the option:
> >
> ><clusternode name="lock-node3" votes="1" clvmd="0">
> >
> Um, I may be out on a limb here, but I know of no clvmd attribute under
> clusternode in our cluster.conf schema.
>
> -Jim
>
> > Only the rac-nodes have access to the shared LUNs...
> >
> >
> > [root at rac-node1 ~]# lvcreate -L4G -n lv01 vg01
> >
> >   Found duplicate PV o2Xf8uUmskTL5fiVQAgY0nJ1ZJSMA9U3: using /dev/sds1
> > not /dev/emcpowerb1
> >
> >   Found duplicate PV 3yduFLdX3FWPWWMb9lIbtBf3JIPjYnHF: using
> > /dev/emcpowerc1 not /dev/sdr1
> >
> >   Found duplicate PV Kyha3qI6nVE4odg77UXf7igS3FenGJNn: using /dev/sdd1
> > not /dev/emcpowerq1
> >
> >   Found duplicate PV WS1LyhqQ8HaE2fIuSnXNd5sgTRtNzNAJ: using /dev/sdt1
> > not /dev/emcpowera1
> >
> >   Found duplicate PV DuBJ7dZsS3PIO7n5U6hINxPkWorZDzvx: using
> > /dev/emcpowerd1 not /dev/sdq1
> >
> >   Found duplicate PV ZECZzAtbA0e9pFbl9oL0lZg4q7fkS5x4: using /dev/sdk1
> > not /dev/emcpowerj1
> >
> >   Found duplicate PV bnVVmL6WhS2mesnOFUkT4fEfR0cFhybD: using
> > /dev/emcpowerk1 not /dev/sdj1
> >
> >   Found duplicate PV XyXrg2zdxxMS5jo03f9I4QYtGM3ILLGV: using /dev/sdl1
> > not /dev/emcpoweri1
> >
> >   Found duplicate PV SLE5v2eTD7cJlpRUDGG35xfXkRbW86i1: using
> > /dev/emcpowerl1 not /dev/sdi1
> >
> >   Found duplicate PV acGyUd2wX7FnOF94Cbt0ombp10iUWMSf: using /dev/sdm1
> > not /dev/emcpowerh1
> >
> >   Found duplicate PV ll8eNZ0JRh9katV0eui4BcxSc6HBggSI: using
> > /dev/emcpowerm1 not /dev/sdh1
> >
> >   Found duplicate PV ptGubq8R16LxywZ458P7ebmdG3Fq2aJo: using /dev/sdn1
> > not /dev/emcpowerg1
> >
> >   Found duplicate PV PLQ3uON7pYe7nY16gRmAP94WBaEydRwf: using
> > /dev/emcpowern1 not /dev/sdg1
> >
> >   Found duplicate PV PsVYTeKNy6EcqWYbJwQ4KEbPp2Q8HjWv: using /dev/sdo1
> > not /dev/emcpowerf1
> >
> >   Found duplicate PV hvekbzDAltJ3t23QveOMz1axfhj9Mp2j: using
> > /dev/emcpowero1 not /dev/sdf1
> >
> >   Found duplicate PV 5OhUbKbZLW5bTc3tpJeU4YlH0dTttJHF: using /dev/sdp1
> > not /dev/emcpowere1
> >
> >   Found duplicate PV dFtPhq6pkwFdl41NTrAguAEFB3601CTb: using
> > /dev/emcpowerp1 not /dev/sde1
> >
> >   Error locking on node lock-node1: Internal lvm error, check syslog
> >
> >   Error locking on node lock-node2: Internal lvm error, check syslog
> >
> >   Error locking on node lock-node3: Internal lvm error, check syslog
> >
> >   Failed to activate new LV.
> >
> > File:
> > /var/log/messages:
> >
> > Oct 13 21:20:46 lock-node1 lvm[5478]: Volume group for uuid not found:
> > UzvBBmBj7m53APMbye1XXztWjdIavfgX8L5rGTOB3i3KGYPazw1AVaGCmWsXZpqR
> >
> >
> > Here is the cluster.conf
> >
> >[root at rac-node1 ~]# cat /etc/cluster/cluster.conf
> ><?xml version="1.0
> >"?>
> ><cluster alias="cluster" config_version="6" name="cluster">
> >        <fence_daemon post_fail_delay="0" post_join_delay="120"/>
> >        <clusternodes>
> >
> >                <clusternode name="lock-node1" votes="1" clvmd="0">
> >                        <fence>
> >                                <method name="1">
> >
> >                                        <device name="lock-node1-fence"/>
> >                                </method>
> >                        </fence>
> >                </clusternode>
> >
> >                <clusternode name="lock-node2" votes="1" clvmd="0">
> >                        <fence>
> >
> >                                <method name="1">
> >                                        <device name="lock-node2-fence"/>
> >                                </method>
> >                        </fence>
> >
> >                </clusternode>
> >                <clusternode name="lock-node3" votes="1" clvmd="0">
> >                        <fence>
> >                                <method name="1">
> >
> >                                        <device name="lock-node3-fence"/>
> >                                </method>
> >                        </fence>
> >                </clusternode>
> >
> >                <clusternode name="rac-node1" votes="1">
> >                        <fence>
> >                                <method name="1">
> >                                        <device name="rac-node1-fence"/>
> >
> >                                </method>
> >                        </fence>
> >                </clusternode>
> >                <clusternode name="rac-node2" votes="1">
> >
> >                        <fence>
> >                                <method name="1">
> >                                        <device name="rac-node2-fence"/>
> >                                </method>
> >
> >                        </fence>
> >                </clusternode>
> >                <clusternode name="rac-node3" votes="1">
> >                        <fence>
> >                                <method name="1">
> >
> >                                        <device name="rac-node3-fence"/>
> >                                </method>
> >                        </fence>
> >                </clusternode>
> >
> >        </clusternodes>
> >        <gulm>
> >                <lockserver name="lock-node1"/>
> >                <lockserver name="lock-node2"/>
> >                <lockserver name="lock-node3"/>
> >
> >        </gulm>
> >        <fencedevices>
> >                <fencedevice agent="fence_ipmilan" auth="none"
> >ipaddr="20.20.20.4 <http://20.20.20.4>" login="xxxx"
> name="lock-node1-fence" passwd="xxxx"/>
> >
> >                <fencedevice agent="fence_ipmilan" auth="none"
> >ipaddr="20.20.20.5 <http://20.20.20.5>" login="xxxx"
> name="lock-node2-fence" passwd="xxxx"/>
> >
> >                <fencedevice agent="fence_ipmilan" auth="none"
> >ipaddr="20.20.20.6 <http://20.20.20.6>" login="xxxx"
> name="lock-node3-fence" passwd="xxxx"/>
> >
> >                <fencedevice agent="fence_ipmilan" auth="none"
> >ipaddr="20.20.20.1 <http://20.20.20.1>" login="xxxx"
> name="rac-node1-fence" passwd="xxxx"/>
> >
> >                <fencedevice agent="fence_ipmilan" auth="none"
> >ipaddr="20.20.20.2 <http://20.20.20.2>" login="xxxx"
> name="rac-node2-fence" passwd="xxxx"/>
> >
> >                <fencedevice agent="fence_ipmilan" auth="none"
> >ipaddr="20.20.20.3 <http://20.20.20.3>" login="xxxx"
> name="rac-node3-fence" passwd="xxxx"/>
> >
> >        </fencedevices>
> >        <rm>
> >                <failoverdomains/>
> >                <resources/>
> >        </rm>
> ></cluster>
> >
> >
> > /etc/lvm/lvm.conf
> >
> > # By default we accept every block device:
> >
> >     # filter = [ "a/.*/" ]
> >
> >     filter = [ "r|/dev/sda|", "a/.*/" ]
> >
> >
> >
> > Regards,
> > Filipe Miranda
> >
> > On 10/13/06, *Filipe Miranda* <filipe.miranda at gmail.com
> > <mailto:filipe.miranda at gmail.com>> wrote:
> >
> >     Alasdair,
> >
> >     Do I need to reboot the machine to test this configurations
> >     changes? Or is there a way do testing it without rebooting the
> >     machine?
> >
> >     I will try to swap the order of the filters,
> >
> >     Thanks for the hint
> >
> >
> >     On 10/13/06, *Alasdair G Kergon* < agk at redhat.com
> >     <mailto:agk at redhat.com> > wrote:
> >
> >         On Fri, Oct 13, 2006 at 06:50:24PM -0300, Filipe Miranda wrote:
> >> filter = [ "a/.*/", "r|/dev/sda|" ] (forgot to use the close
> >         pipe)
> >
> >         Still won't work - swap the order of the two items; first
> matches
> >         everything so second isn't looked at.
> >         (man lvm.conf)
> >
> >         Alasdair
> >         --
> >         agk at redhat.com <mailto:agk at redhat.com>
> >
> >         --
> >         Linux-cluster mailing list
> >         Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
> >         https://www.redhat.com/mailman/listinfo/linux-cluster
> >         <https://www.redhat.com/mailman/listinfo/linux-cluster>
> >
> >
> >
> >
> >     --
> >     ---
> >     Filipe T Miranda
> >     Red Hat Certified Engineer
> >
> >
> >
> >
> > --
> > ---
> > Filipe T Miranda
> > Red Hat Certified Engineer
> >
> >
> >------------------------------------------------------------------------
> >
> >--
> >Linux-cluster mailing list
> >Linux-cluster at redhat.com
> >https://www.redhat.com/mailman/listinfo/linux-cluster
> >
>
>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>



-- 
---
Filipe T Miranda
Red Hat Certified Engineer
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20061014/48ce0762/attachment.htm>


More information about the Linux-cluster mailing list