[linux-lvm] Problems with vgimport after software raid initialisation failed.

SystemError system.error at gmx.net
Thu Oct 9 15:26:02 UTC 2003


Hey there,

thanks to the help of Heinz I was able to get my VG online again...
Unfortunately we had a power outage here in Karlsruhe on 02.10.2003 and
so it took a little bit longer than expected.
(Of course the outage came right when it hurt the most)

Now I am able to varryon/varryoff my VG as often I as I like.
(during the chaos after the outage I renamed "datavg" -> "vaultvg")

My "pvscan" looks now like this:
"
[root at athens root]# pvscan 
pvscan -- reading all physical volumes (this may take a while...)
pvscan -- ACTIVE   PV "/dev/md2" of VG "sysvg"   [16 GB / 10 GB free]
pvscan -- ACTIVE   PV "/dev/md3" of VG "vaultvg" [132.25 GB / 0 free]
pvscan -- ACTIVE   PV "/dev/md4" of VG "vaultvg" [57.12 GB / 50.88 GB
free]
pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
pvscan -- WARNING: physical volume "/dev/hdf" belongs to a meta device
pvscan -- total: 5 [205.70 GB] / in use: 5 [205.70 GB] / in no VG: 0 [0]
"

And my here goes my "vgscan":
"
[root at athens root]# vgscan
vgscan -- reading all physical volumes (this may take a while...)
vgscan -- found active volume group "sysvg"
vgscan -- found active volume group "vaultvg"
vgscan -- ERROR "vg_read_with_pv_and_lv(): current PV" can't get data of
volume group "datavg" from physical volume(s)
vgscan -- "/etc/lvmtab" and "/etc/lvmtab.d" successfully created
vgscan -- WARNING: This program does not do a VGDA backup of your volume
groups
"

Here are my remaining questions:
a.) 
Why is "pvscan" complaining about the raid members "hdh" and "hdf" ?
It does not complain about my other raid devices.
Any ideas ?

b.)
And why is "vgscan" complaining about the VG "datavg",
even if this VG does not exist anymore ?

---> Would perhaps a vgexport/vgimport be a solution ?

>>> Great. Sounds like filesystem backup time now ;)
I would like very much to do so, Heinz.
:-)
But unfortunately that is not an option in 
this setup...
:-(
...better don' ask...
;-)

Still grateful for any hints
Lutz Reinegger

Am Don, 2003-10-02 um 15.51 schrieb Heinz J . Mauelshagen:
> On Thu, Oct 02, 2003 at 12:57:18PM +0200, SystemError wrote:
> > Heinz,
> > 
> > I did as instructed and now I am able to access my 
> > precious "datavg" again.
> > Thanks a lot Heinz.
> > You really saved my day. :-)
> 
> Great. Sounds like filesystem backup time now ;)
> 
> > 
> > I also edited /etc/rc.d/rc.sysinit and commented out all vgscans
> > (RedHat does 2 of them), and added a hard coded "vgchange -ay sysvg" in
> > order to bring my other VG with the /usr, /var, etc... 
> > filesystems online during boot time.
> > 
> > Immediately after another reboot I tried to activate "datavg" using
> > "vgchange -ay datavg"...
> > ...which failed.
> > Vgchange complained:
> > "
> > The PVs /dev/md3 and /dev/md4 are not active.
> > Please run vgscan.
> > "
> > (Can't remember the excact error message)
> > 
> > So I gulped,shrugged and decided that it was time for another 
> > cup of coffee...
> > When I returned to my desk I ran the vgchange again in order to fetch
> > the error message for this email...
> > ...et voila:
> > "
> > [root at athens root]# vgchange -ay datavg
> > vgchange -- volume group "datavg" successfully activated
> > "
> > 
> > So I was only able to activate datavg some minutes after my system was
> > back up and running ?
> > I really don't understand this...
> > OK, I am more than just happy to have my data back online, but WHY ?
> > Anyone any ideas on this one ?
> 
> Sounds like a race to me, where the mds aren't active when vgscan runs.
> 
> > 
> > Another question:
> > Is it save to extend lvs and filesystems right now in 
> > this strange situation ?
> 
> No, the activation/metadata problems need to get sorted out first.
> 
> > 
> > Bye
> > Lutz Reinegger
> > 
> > PS: The requested tar archvie is on its way.
> > 
> > PPS: Again: THANK YOU !!!
> > 
> > Am Don, 2003-10-02 um 11.11 schrieb Heinz J . Mauelshagen:
> > > 
> > > Lutz,
> > > 
> > > looks like you hit some strange LVM on top of MD bug :(
> > > 
> > > In order to get your VG active again (which of course is your highest
> > > priority) and before we analyze the vgscan problem, you want to go for
> > > the follwoing workaround (presumably /etc/lvmconf/datavg.conf is the last
> > > correct archive of the metadata):
> > > 
> > > # cp /etc/lvmconf/datavg.conf /etc/lvmtab.d/datavg
> > > # echo -ne "datavg\0" >> /etc/lvmtab
> > > # vgchange -ay datavg
> > > 
> > > Warning: the next vgscan run will remove the above metadata again, so avoid
> > >          running it for now by commenting it out in your boot script.
> > > 
> > > So far about firefighting ;)
> > > 
> > > 
> > > For further analysis, please do the following and send the resulting
> > > bzip2'ed tar archive containing your metadat to me in private mail
> > > <mge at sistina.com>:
> > > 
> > > # for d in md2 md3 md4 hdf hdh
> > > # do
> > > #	dd bs=1k count=4k if=/dev/$d of=$d.vgda
> > > # done
> > > # tar cf Lutz_Reinegger.vgda.tar *.vgda
> > > # rm *.vgda
> > > # bzip2 Lutz_Reinegger.vgda.tar
> > > 
> > > 
> > > Regards,
> > > Heinz    -- The LVM Guy --
> > > 
> > > 
> > > 
> > > On Tue, Sep 30, 2003 at 09:59:19PM +0200, SystemError wrote:
> > > > Hello out there,
> > > > 
> > > > after I migrating my precious volume group "datavg" from unmirrored
> > > > disks to linux software raid devices I ran into serios problems.
> > > > (Although I fear the biggest problem here was my own incompetence...)
> > > > 
> > > > First I moved the data from the old unmirrored disks away, using pvmove.
> > > > No Problems so far.
> > > > 
> > > > At a certain point I had emptied the 2 PVs "/dev/hdh" and "/dev/hdf".
> > > > So I did a vgreduce on them, then created a new raid1 
> > > > "/dev/md4" (containing both "hdf" and "hdh") and added it to my 
> > > > volume group "datavg" using pvcreate(->"/dev/md4") and vgextend.
> > > > No Problems so far.
> > > > 
> > > > Everything looked soooo perfect and so I decided to reboot the system...
> > > > 
> > > > At this point things started to go wrong, during the boot sequence
> > > > "/dev/md4" was not automatically activated and suddenly the PV
> > > > "/dev/hdf" showed up in "datavg", "/dev/md4" was gone.
> > > > 
> > > > Unfortunately I paniced and ran a vgexport on "datavg", fixed the broken
> > > > initialisation of "/dev/md4", and rebooted again. 
> > > > This was a probably a baaad idea. 
> > > > Shame upon me.
> > > > 
> > > > Now my pvscan looks like this:
> > > > "
> > > > [root at athens root]# pvscan 
> > > > pvscan -- reading all physical volumes (this may take a while...)
> > > > pvscan -- ACTIVE   PV "/dev/md2" of VG "sysvg" [16 GB / 10 GB free]
> > > > pvscan -- inactive PV "/dev/md3" is in EXPORTED VG "datavg" [132.25 GB /
> > > > 0 free]
> > > > pvscan -- inactive PV "/dev/md4"  is associated to unknown VG "datavg"
> > > > (run vgscan)
> > > > pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
> > > > pvscan -- inactive PV "/dev/hdf"  is in EXPORTED VG "datavg" [57.12 GB /
> > > > 50.88 GB free]
> > > > pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
> > > > "
> > > > 
> > > > Or with the -u option:
> > > > "
> > > > [root at athens root]# pvscan -u
> > > > pvscan -- reading all physical volumes (this may take a while...)
> > > > pvscan -- ACTIVE   PV "/dev/md2" with UUID
> > > > "g6Au3J-2C4H-Ifjo-iESu-4yp8-aRQv-ozChyW" of VG "sysvg"        [16 GB /
> > > > 10 GB free]
> > > > pvscan -- inactive PV "/dev/md3" with UUID
> > > > "R15mli-TFs2-214J-YTBh-Hatl-erbL-G7WS4b"  is in EXPORTED VG "datavg"
> > > > [132.25 GB / 0 free]
> > > > pvscan -- inactive PV "/dev/md4" with UUID
> > > > "szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
> > > > [57.12 GB / 50.88 GB free]
> > > > pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
> > > > pvscan -- inactive PV "/dev/hdf" with UUID
> > > > "szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
> > > > [57.12 GB / 50.88 GB free]
> > > > pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
> > > > 
> > > > "
> > > > 
> > > > A vgimport using "md3"(no probs with this raid1) and "md4" fails: 
> > > > "
> > > > [root at athens root]# vgimport datavg /dev/md3 /dev/md4
> > > > vgimport -- ERROR "pv_read(): multiple device" reading physical volume
> > > > "/dev/md4"
> > > > "
> > > > 
> > > > Using "md3" and "hdh" also fails:
> > > > "
> > > > [root at athens root]# vgimport datavg /dev/md3 /dev/hdh
> > > > vgimport -- ERROR "pv_read(): multiple device" reading physical volume
> > > > "/dev/hdh"
> > > > "
> > > > 
> > > > It also fails when I try to use "hdf", only the error message is
> > > > different:
> > > > "
> > > > [root at athens root]# vgimport datavg /dev/md3 /dev/hdf
> > > > vgimport -- ERROR: wrong number of physical volumes to import volume
> > > > group "datavg"
> > > > "
> > > > 
> > > > So here I am, with a huge VG an tons of data in it but no way to access
> > > > the VG. Has anybody out there an idea how I can still access
> > > > the data of datavg ?
> > > > 
> > > > By the way: 
> > > > I am using RedHatLinux 9.0 with the lvm-1.0.3-12 binary rpm package
> > > > as provided by RedHat.
> > > > 
> > > > Bye
> > > > In desperation
> > > > Lutz Reinegger
> > > > 
> > > > PS:
> > > > Any comments and suggestions are highly appreciated, even if those
> > > > suggestions include the use of hex editors or sacrificing 
> > > > caffeine to dark and ancient deities.
> > > > ;-)
> > > > 
> > > > 
> > > > 
> > > > _______________________________________________
> > > > linux-lvm mailing list
> > > > linux-lvm at sistina.com
> > > > http://lists.sistina.com/mailman/listinfo/linux-lvm
> > > > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> > > 
> > > *** Software bugs are stupid.
> > >     Nevertheless it needs not so stupid people to solve them ***
> > > 
> > > =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> > > 
> > > Heinz Mauelshagen                                 Sistina Software Inc.
> > > Senior Consultant/Developer                       Am Sonnenhang 11
> > >                                                   56242 Marienrachdorf
> > >                                                   Germany
> > > Mauelshagen at Sistina.com                           +49 2626 141200
> > >                                                        FAX 924446
> > > =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> > > 
> > > _______________________________________________
> > > linux-lvm mailing list
> > > linux-lvm at sistina.com
> > > http://lists.sistina.com/mailman/listinfo/linux-lvm
> > > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> > > 
> > 
> > 
> > 
> > _______________________________________________
> > linux-lvm mailing list
> > linux-lvm at sistina.com
> > http://lists.sistina.com/mailman/listinfo/linux-lvm
> > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> 
> -- 
> 
> Regards,
> Heinz    -- The LVM Guy --
> 
> *** Software bugs are stupid.
>     Nevertheless it needs not so stupid people to solve them ***
> 
> =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> 
> Heinz Mauelshagen                                 Sistina Software Inc.
> Senior Consultant/Developer                       Am Sonnenhang 11
>                                                   56242 Marienrachdorf
>                                                   Germany
> Mauelshagen at Sistina.com                           +49 2626 141200
>                                                        FAX 924446
> =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> 
> _______________________________________________
> linux-lvm mailing list
> linux-lvm at sistina.com
> http://lists.sistina.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> 






More information about the linux-lvm mailing list