[linux-lvm] Lost a full volume group!
Rodrigo de Salazar
no_id at wanadoo.es
Thu Mar 13 09:57:02 UTC 2003
Ok, I will paste the output of the relevant commands, if you want more info
just ask... and thanks for the quick answer, this machine is destined to be
an important server, and it should be ready ASAP, aka if we can't get it to
work with LVM soon, we'll have to use a traditional scheme, superior's orders
/:
So here it goes:
volatil:~# vgck
vgck -- VGDA of "vg1" in lvmtab is consistent
vgck -- VGDA of "vg1" on physical volume is consistent
vgck -- VGDA of "vg2" in lvmtab is consistent
vgck -- VGDA of "vg2" on physical volume is consistent
volatil:~# pvscan
pvscan -- reading all physical volumes (this may take a while...)
pvscan -- inactive PV "/dev/ide/host0/bus1/target0/lun0/disc" is in no VG
[8.03 GB]
pvscan -- ACTIVE PV "/dev/ide/host0/bus1/target0/lun0/part2" of VG "vg2"
[74.05 GB / 10.05 GB free]
pvscan -- inactive PV "/dev/ide/host0/bus0/target0/lun0/part4" of VG "vg1"
[35.74 GB / 2.79 GB free]
pvscan -- total: 3 [117.83 GB] / in use: 2 [109.80 GB] / in no VG: 1 [8.03 GB]
The 8.03 GB PV makes no sense to me, it is supposedly /dev/hdc, but that disk
(80 GB) is partitioned, first partition 512 MB swap and the second is the LVM
partition (the rest of the disk).
volatil:~# vgscan
vgscan -- reading all physical volumes (this may take a while...)
vgscan -- found active volume group "vg2"
vgscan -- found inactive volume group "vg1"
vgscan -- "/etc/lvmtab" and "/etc/lvmtab.d" successfully created
vgscan -- WARNING: This program does not do a VGDA backup of your volume
groups
volatil:~# lvscan
lvscan -- volume group "vg1" is NOT active; try -D
lvscan -- ACTIVE Original "/dev/vg2/home" [45 GB]
lvscan -- ACTIVE "/dev/vg2/usr" [10 GB]
lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-16" [1008 MB] of /dev/vg2/home
lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-19" [1008 MB] of /dev/vg2/home
lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-10" [1008 MB] of /dev/vg2/home
lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-13" [1008 MB] of /dev/vg2/home
lvscan -- ACTIVE "/dev/vg2/local" [5 GB]
lvscan -- 7 logical volumes with 63.94 GB total in 1 volume group
lvscan -- 7 active logical volumes
volatil:~# lvscan -D
lvscan -- reading all physical volumes (this may take a while...)
lvscan -- inactive "/dev/vg1/tmp" [1 GB]
lvscan -- inactive "/dev/vg1/var" [15 GB]
lvscan -- inactive Original "/dev/vg1/www" [15 GB]
lvscan -- inactive Snapshot "/dev/vg1/www-hora-16" [0] of /dev/vg1/www
lvscan -- inactive Snapshot "/dev/vg1/www-hora-19" [0] of /dev/vg1/www
lvscan -- inactive Snapshot "/dev/vg1/www-hora-10" [0] of /dev/vg1/www
lvscan -- inactive Snapshot "/dev/vg1/www-hora-13" [0] of /dev/vg1/www
lvscan -- inactive Original "/dev/vg2/home" [45 GB]
lvscan -- inactive "/dev/vg2/usr" [10 GB]
lvscan -- inactive Snapshot "/dev/vg2/home-hora-16" [0] of /dev/vg2/home
lvscan -- inactive Snapshot "/dev/vg2/home-hora-19" [0] of /dev/vg2/home
lvscan -- inactive Snapshot "/dev/vg2/home-hora-10" [0] of /dev/vg2/home
lvscan -- inactive Snapshot "/dev/vg2/home-hora-13" [0] of /dev/vg2/home
lvscan -- inactive "/dev/vg2/local" [5 GB]
lvscan -- 14 logical volumes with 91 GB total in 2 volume groups
lvscan -- 14 inactive logical volumes
As you can see, there are 4 daily snapshots of /home and /var/www, maybe it
is too much, but those are the directories where people work and well, files
get accidentally deleted during the day and so on and so forth (:
All were available when this happened.
volatil:~# vgchange -a y
vgchange -- ERROR "Bad address" activating volume group "vg1"
vgchange -- volume group "vg2" already active
volatil:~# pvdisplay /dev/hda4
--- Physical volume ---
PV Name /dev/ide/host0/bus0/target0/lun0/part4
VG Name vg1
PV Size 35.75 GB [74970000 secs] / NOT usable 4.19 MB [LVM: 163
KB]
PV# 1
PV Status available
Allocatable yes
Cur LV 7
PE Size (KByte) 4096
Total PE 9150
Free PE 714
Allocated PE 8436
PV UUID kvIRnL-ZNQm-Ab0g-TSqe-rasN-Hg1C-pHZIkF
volatil:~# vgdisplay -D
--- Volume group ---
VG Name vg2
VG Access read/write
VG Status NOT available/resizable
VG # 1
MAX LV 256
Cur LV 7
Open LV 0
MAX LV Size 255.99 GB
Max PV 256
Cur PV 1
Act PV 1
VG Size 74.05 GB
PE Size 4 MB
Total PE 18956
Alloc PE / Size 16384 / 64 GB
Free PE / Size 2572 / 10.05 GB
VG UUID Nm7brW-Hk8I-hYkt-1LvU-6COW-G5ds-FduaFe
--- Volume group ---
VG Name vg1
VG Access read/write
VG Status NOT available/resizable
VG # 0
MAX LV 256
Cur LV 7
Open LV 0
MAX LV Size 255.99 GB
Max PV 256
Cur PV 1
Act PV 1
VG Size 35.74 GB
PE Size 4 MB
Total PE 9150
Alloc PE / Size 8436 / 32.95 GB
Free PE / Size 714 / 2.79 GB
VG UUID GNeKh6-W3TL-AMa6-3E7q-291d-pjm2-av59ou
Another thing, originally /dev/vg1/var was 30 GB, it was resized with:
e2fsadm -L-15G /dev/vg1/var
(only 9 GB were occupied) and then the /dev/vg1/www was created...
Also, they have ext3 filesystems on them, but with the journal disabled
(working as ext2). It all worked fine even on reboot, a few days before the
"accident".
Don't know what else to include... if I missed out something, just ask for
it, and thanks for the help (:
More information about the linux-lvm
mailing list