[Linux-cluster] CLVM error 6 nodes
Filipe Miranda
filipe.miranda at gmail.com
Sat Oct 14 00:39:25 UTC 2006
Alasdair,
We did the modifications but still the same error:
Since the lock servers does not have access to the LUNs that will be in use
by the GFS I used the option:
<clusternode name="lock-node3" votes="1" clvmd="0">
Only the rac-nodes have access to the shared LUNs...
[root at rac-node1 ~]# lvcreate -L4G -n lv01 vg01
Found duplicate PV o2Xf8uUmskTL5fiVQAgY0nJ1ZJSMA9U3: using /dev/sds1 not
/dev/emcpowerb1
Found duplicate PV 3yduFLdX3FWPWWMb9lIbtBf3JIPjYnHF: using /dev/emcpowerc1
not /dev/sdr1
Found duplicate PV Kyha3qI6nVE4odg77UXf7igS3FenGJNn: using /dev/sdd1 not
/dev/emcpowerq1
Found duplicate PV WS1LyhqQ8HaE2fIuSnXNd5sgTRtNzNAJ: using /dev/sdt1 not
/dev/emcpowera1
Found duplicate PV DuBJ7dZsS3PIO7n5U6hINxPkWorZDzvx: using /dev/emcpowerd1
not /dev/sdq1
Found duplicate PV ZECZzAtbA0e9pFbl9oL0lZg4q7fkS5x4: using /dev/sdk1 not
/dev/emcpowerj1
Found duplicate PV bnVVmL6WhS2mesnOFUkT4fEfR0cFhybD: using /dev/emcpowerk1
not /dev/sdj1
Found duplicate PV XyXrg2zdxxMS5jo03f9I4QYtGM3ILLGV: using /dev/sdl1 not
/dev/emcpoweri1
Found duplicate PV SLE5v2eTD7cJlpRUDGG35xfXkRbW86i1: using /dev/emcpowerl1
not /dev/sdi1
Found duplicate PV acGyUd2wX7FnOF94Cbt0ombp10iUWMSf: using /dev/sdm1 not
/dev/emcpowerh1
Found duplicate PV ll8eNZ0JRh9katV0eui4BcxSc6HBggSI: using /dev/emcpowerm1
not /dev/sdh1
Found duplicate PV ptGubq8R16LxywZ458P7ebmdG3Fq2aJo: using /dev/sdn1 not
/dev/emcpowerg1
Found duplicate PV PLQ3uON7pYe7nY16gRmAP94WBaEydRwf: using /dev/emcpowern1
not /dev/sdg1
Found duplicate PV PsVYTeKNy6EcqWYbJwQ4KEbPp2Q8HjWv: using /dev/sdo1 not
/dev/emcpowerf1
Found duplicate PV hvekbzDAltJ3t23QveOMz1axfhj9Mp2j: using /dev/emcpowero1
not /dev/sdf1
Found duplicate PV 5OhUbKbZLW5bTc3tpJeU4YlH0dTttJHF: using /dev/sdp1 not
/dev/emcpowere1
Found duplicate PV dFtPhq6pkwFdl41NTrAguAEFB3601CTb: using /dev/emcpowerp1
not /dev/sde1
Error locking on node lock-node1: Internal lvm error, check syslog
Error locking on node lock-node2: Internal lvm error, check syslog
Error locking on node lock-node3: Internal lvm error, check syslog
Failed to activate new LV.
File:
/var/log/messages:
Oct 13 21:20:46 lock-node1 lvm[5478]: Volume group for uuid not found:
UzvBBmBj7m53APMbye1XXztWjdIavfgX8L5rGTOB3i3KGYPazw1AVaGCmWsXZpqR
Here is the cluster.conf
[root at rac-node1 ~]# cat /etc/cluster/cluster.conf
<?xml version="1.0"?>
<cluster alias="cluster" config_version="6" name="cluster">
<fence_daemon post_fail_delay="0" post_join_delay="120"/>
<clusternodes>
<clusternode name="lock-node1" votes="1" clvmd="0">
<fence>
<method name="1">
<device name="lock-node1-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="lock-node2" votes="1" clvmd="0">
<fence>
<method name="1">
<device name="lock-node2-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="lock-node3" votes="1" clvmd="0">
<fence>
<method name="1">
<device name="lock-node3-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="rac-node1" votes="1">
<fence>
<method name="1">
<device name="rac-node1-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="rac-node2" votes="1">
<fence>
<method name="1">
<device name="rac-node2-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="rac-node3" votes="1">
<fence>
<method name="1">
<device name="rac-node3-fence"/>
</method>
</fence>
</clusternode>
</clusternodes>
<gulm>
<lockserver name="lock-node1"/>
<lockserver name="lock-node2"/>
<lockserver name="lock-node3"/>
</gulm>
<fencedevices>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.4" login="xxxx" name="lock-node1-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.5" login="xxxx" name="lock-node2-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.6" login="xxxx" name="lock-node3-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.1" login="xxxx" name="rac-node1-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.2" login="xxxx" name="rac-node2-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.3" login="xxxx" name="rac-node3-fence" passwd="xxxx"/>
</fencedevices>
<rm>
<failoverdomains/>
<resources/>
</rm>
</cluster>
/etc/lvm/lvm.conf
# By default we accept every block device:
# filter = [ "a/.*/" ]
filter = [ "r|/dev/sda|", "a/.*/" ]
Regards,
Filipe Miranda
On 10/13/06, Filipe Miranda <filipe.miranda at gmail.com> wrote:
>
> Alasdair,
>
> Do I need to reboot the machine to test this configurations changes? Or is
> there a way do testing it without rebooting the machine?
>
> I will try to swap the order of the filters,
>
> Thanks for the hint
>
> On 10/13/06, Alasdair G Kergon <agk at redhat.com> wrote:
> >
> > On Fri, Oct 13, 2006 at 06:50:24PM -0300, Filipe Miranda wrote:
> > > filter = [ "a/.*/", "r|/dev/sda|" ] (forgot to use the close pipe)
> >
> > Still won't work - swap the order of the two items; first matches
> > everything so second isn't looked at.
> > (man lvm.conf)
> >
> > Alasdair
> > --
> > agk at redhat.com
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster at redhat.com
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> >
>
>
>
> --
> ---
> Filipe T Miranda
> Red Hat Certified Engineer
>
--
---
Filipe T Miranda
Red Hat Certified Engineer
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20061013/8afbe9b1/attachment.htm>
More information about the Linux-cluster
mailing list