[linux-lvm] Having duplicate PV problems, think there's a bug in LVM2 md component detection
go0ogl3
go0ogl3 at gmail.com
Mon Feb 28 11:15:13 UTC 2005
I'm anly a begginer at lvm but...
On Sat, 26 Feb 2005 15:30:08 -0500, Ron Watkins <linux-lvm at malor.com> wrote:
> I'm sorry if this is a FAQ or if I'm being stupid. I saw some mentions to
> this problem on the old mailing list, but it didn't seem to quite cover what
> I'm seeing, and I don't see an archive for this list yet. (and what on
> earth happened to the old list, anyway?)
>
> My problem is this: I'm setting up a software RAID5 across 5 IDE drives.
> I'm running Debian Unstable, using kernel 2.6.8-2-k7. I HAVE set
> md_component_detection to 1 in lvm.conf, and I wiped the drives after
> changing this setting.
>
> I originally set it up as a four-drive RAID, via a 3Ware controller, so my
> original devices were sdb, sdc, sdd, and sde. (the machine also has a
> hardware raid on an ICP Vortex SCSI controller: this is sda.) In this
> mode, it set up and built perfectly. LVM worked exactly as I expected it
> to. I had a test volume running. All the queries and volume management
> worked exactly correctly. All was well.
>
> So then I tried to add one more drive via the motherboard IDE controller, on
> /dev/hda. (note that I stopped the array, wiped the first and last 100 megs
> on the drives, and rebuilt. ). That's when the problems started. The RAID
> itself seems to build and work just fine, although I haven't waited for the
> entire 6 or so hours it will take to completely finish. Build speed is
> good, everything seems normal. But LVM blows up badly in this
> configuration.
>
> When I do a pvcreate on /dev/md0, it succeeds... but if I do a pvdisplay I
> get a bunch of complaints:
>
> jeeves:/etc/lvm# pvdisplay
> Found duplicate PV y8pYTtAg0W703Sc8Wiy79mcWU3gHmCFc: using /dev/sde not
> /dev/hda
> Found duplicate PV y8pYTtAg0W703Sc8Wiy79mcWU3gHmCFc: using /dev/sde not
> /dev/hda
I think you have 2 PV's with the same UUID and that's the problem. You
can even move the drives letters around (hda or sda) as I think it
does not matter for lvm. The only thing it counts it's the "UUID" of
the PV.
You should use pvcreate again on /dev/hda so your last added drive
should have different UUID.
> --- NEW Physical volume ---
> PV Name /dev/hda
> VG Name
> PV Size 931.54 GB
> Allocatable NO
> PE Size (KByte) 0
> Total PE 0
> Free PE 0
> Allocated PE 0
> PV UUID y8pYTt-Ag0W-703S-c8Wi-y79m-cWU3-gHmCFc
>
> It seems to think that /dev/hda is where the PV is, rather than /dev/md0.
>
> (Note, again, I *HAVE* turned the md_component_detection to 1 in lvm.conf!!)
>
> I have erased, using dd, the first and last 100 megs or so on every drive,
> and I get exactly the same results every time... even with all RAID and LVM
> blocks erased, if I use this list of drives:
>
> /dev/hda
> /dev/sdb
> /dev/sdc
> /dev/sdd
> /dev/sde
>
> with the linux MD driver, LVM does not seem to work properly. I think the
> component detection is at least a little buggy. This is what my
> /proc/mdstat looks like:
>
> jeeves:/etc/lvm# cat /proc/mdstat
> Personalities : [raid5]
> md0 : active raid5 sde[5] sdd[3] sdc[2] sdb[1] hda[0]
> 976793600 blocks level 5, 128k chunk, algorithm 2 [5/4] [UUUU_]
> [=>...................] recovery = 6.2% (15208576/244198400)
> finish=283.7min speed=13448K/sec
> unused devices: <none>
>
> I realize that using both IDE and SCSI drives in the same array is
> unusual... but I'm not really using SCSI drives, they just look like that
> because of the 3Ware controller.
>
> Again, this works FINE as long as I just use the (fake) SCSI devices.. it
> doesn't wonk out until I add in /dev/hda.
>
> Any suggestions? Is this a bug?
>
> _______________________________________________
> linux-lvm mailing list
> linux-lvm at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>
More information about the linux-lvm
mailing list