[linux-lvm] Lost a full volume group!
Heinz J . Mauelshagen
mauelshagen at sistina.com
Thu Mar 13 17:30:02 UTC 2003
Rodrigo,
which LVM version is this ?
Because we fixed some snapshot bugs in 1.0.7 you should give
that version a try.
In order to get your "vg1" back, you should try to remove the snapshots
with "vgscan -r".
Regards,
Heinz -- The LVM Guy --
On Thu, Mar 13, 2003 at 04:56:37PM +0100, Rodrigo de Salazar wrote:
> Ok, I will paste the output of the relevant commands, if you want more info
> just ask... and thanks for the quick answer, this machine is destined to be
> an important server, and it should be ready ASAP, aka if we can't get it to
> work with LVM soon, we'll have to use a traditional scheme, superior's orders
> /:
> So here it goes:
>
> volatil:~# vgck
> vgck -- VGDA of "vg1" in lvmtab is consistent
> vgck -- VGDA of "vg1" on physical volume is consistent
> vgck -- VGDA of "vg2" in lvmtab is consistent
> vgck -- VGDA of "vg2" on physical volume is consistent
>
> volatil:~# pvscan
> pvscan -- reading all physical volumes (this may take a while...)
> pvscan -- inactive PV "/dev/ide/host0/bus1/target0/lun0/disc" is in no VG
> [8.03 GB]
> pvscan -- ACTIVE PV "/dev/ide/host0/bus1/target0/lun0/part2" of VG "vg2"
> [74.05 GB / 10.05 GB free]
> pvscan -- inactive PV "/dev/ide/host0/bus0/target0/lun0/part4" of VG "vg1"
> [35.74 GB / 2.79 GB free]
> pvscan -- total: 3 [117.83 GB] / in use: 2 [109.80 GB] / in no VG: 1 [8.03 GB]
>
> The 8.03 GB PV makes no sense to me, it is supposedly /dev/hdc, but that disk
> (80 GB) is partitioned, first partition 512 MB swap and the second is the LVM
> partition (the rest of the disk).
>
> volatil:~# vgscan
> vgscan -- reading all physical volumes (this may take a while...)
> vgscan -- found active volume group "vg2"
> vgscan -- found inactive volume group "vg1"
> vgscan -- "/etc/lvmtab" and "/etc/lvmtab.d" successfully created
> vgscan -- WARNING: This program does not do a VGDA backup of your volume
> groups
>
> volatil:~# lvscan
> lvscan -- volume group "vg1" is NOT active; try -D
> lvscan -- ACTIVE Original "/dev/vg2/home" [45 GB]
> lvscan -- ACTIVE "/dev/vg2/usr" [10 GB]
> lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-16" [1008 MB] of /dev/vg2/home
> lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-19" [1008 MB] of /dev/vg2/home
> lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-10" [1008 MB] of /dev/vg2/home
> lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-13" [1008 MB] of /dev/vg2/home
> lvscan -- ACTIVE "/dev/vg2/local" [5 GB]
> lvscan -- 7 logical volumes with 63.94 GB total in 1 volume group
> lvscan -- 7 active logical volumes
>
> volatil:~# lvscan -D
> lvscan -- reading all physical volumes (this may take a while...)
> lvscan -- inactive "/dev/vg1/tmp" [1 GB]
> lvscan -- inactive "/dev/vg1/var" [15 GB]
> lvscan -- inactive Original "/dev/vg1/www" [15 GB]
> lvscan -- inactive Snapshot "/dev/vg1/www-hora-16" [0] of /dev/vg1/www
> lvscan -- inactive Snapshot "/dev/vg1/www-hora-19" [0] of /dev/vg1/www
> lvscan -- inactive Snapshot "/dev/vg1/www-hora-10" [0] of /dev/vg1/www
> lvscan -- inactive Snapshot "/dev/vg1/www-hora-13" [0] of /dev/vg1/www
> lvscan -- inactive Original "/dev/vg2/home" [45 GB]
> lvscan -- inactive "/dev/vg2/usr" [10 GB]
> lvscan -- inactive Snapshot "/dev/vg2/home-hora-16" [0] of /dev/vg2/home
> lvscan -- inactive Snapshot "/dev/vg2/home-hora-19" [0] of /dev/vg2/home
> lvscan -- inactive Snapshot "/dev/vg2/home-hora-10" [0] of /dev/vg2/home
> lvscan -- inactive Snapshot "/dev/vg2/home-hora-13" [0] of /dev/vg2/home
> lvscan -- inactive "/dev/vg2/local" [5 GB]
> lvscan -- 14 logical volumes with 91 GB total in 2 volume groups
> lvscan -- 14 inactive logical volumes
>
> As you can see, there are 4 daily snapshots of /home and /var/www, maybe it
> is too much, but those are the directories where people work and well, files
> get accidentally deleted during the day and so on and so forth (:
> All were available when this happened.
>
> volatil:~# vgchange -a y
> vgchange -- ERROR "Bad address" activating volume group "vg1"
> vgchange -- volume group "vg2" already active
>
> volatil:~# pvdisplay /dev/hda4
> --- Physical volume ---
> PV Name /dev/ide/host0/bus0/target0/lun0/part4
> VG Name vg1
> PV Size 35.75 GB [74970000 secs] / NOT usable 4.19 MB [LVM: 163
> KB]
> PV# 1
> PV Status available
> Allocatable yes
> Cur LV 7
> PE Size (KByte) 4096
> Total PE 9150
> Free PE 714
> Allocated PE 8436
> PV UUID kvIRnL-ZNQm-Ab0g-TSqe-rasN-Hg1C-pHZIkF
>
> volatil:~# vgdisplay -D
> --- Volume group ---
> VG Name vg2
> VG Access read/write
> VG Status NOT available/resizable
> VG # 1
> MAX LV 256
> Cur LV 7
> Open LV 0
> MAX LV Size 255.99 GB
> Max PV 256
> Cur PV 1
> Act PV 1
> VG Size 74.05 GB
> PE Size 4 MB
> Total PE 18956
> Alloc PE / Size 16384 / 64 GB
> Free PE / Size 2572 / 10.05 GB
> VG UUID Nm7brW-Hk8I-hYkt-1LvU-6COW-G5ds-FduaFe
>
> --- Volume group ---
> VG Name vg1
> VG Access read/write
> VG Status NOT available/resizable
> VG # 0
> MAX LV 256
> Cur LV 7
> Open LV 0
> MAX LV Size 255.99 GB
> Max PV 256
> Cur PV 1
> Act PV 1
> VG Size 35.74 GB
> PE Size 4 MB
> Total PE 9150
> Alloc PE / Size 8436 / 32.95 GB
> Free PE / Size 714 / 2.79 GB
> VG UUID GNeKh6-W3TL-AMa6-3E7q-291d-pjm2-av59ou
>
> Another thing, originally /dev/vg1/var was 30 GB, it was resized with:
>
> e2fsadm -L-15G /dev/vg1/var
>
> (only 9 GB were occupied) and then the /dev/vg1/www was created...
> Also, they have ext3 filesystems on them, but with the journal disabled
> (working as ext2). It all worked fine even on reboot, a few days before the
> "accident".
>
> Don't know what else to include... if I missed out something, just ask for
> it, and thanks for the help (:
>
> _______________________________________________
> linux-lvm mailing list
> linux-lvm at sistina.com
> http://lists.sistina.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
*** Software bugs are stupid.
Nevertheless it needs not so stupid people to solve them ***
=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
Heinz Mauelshagen Sistina Software Inc.
Senior Consultant/Developer Am Sonnenhang 11
56242 Marienrachdorf
Germany
Mauelshagen at Sistina.com +49 2626 141200
FAX 924446
=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
More information about the linux-lvm
mailing list