[linux-lvm] uuid already in use
Thomas Krichel
krichel at openlib.org
Thu Jan 10 08:46:24 UTC 2008
Thomas Krichel writes
>
> raneb:/etc/lvm/archive# vgdisplay
> --- Volume group ---
> VG Name vg1
> System ID
> Format lvm2
> Metadata Areas 2
> Metadata Sequence No 17
> VG Access read/write
> VG Status resizable
> MAX LV 0
> Cur LV 0
> Open LV 0
> Max PV 0
> Cur PV 2
> Act PV 2
> VG Size 652.06 GB
> PE Size 4.00 MB
> Total PE 166928
> Alloc PE / Size 0 / 0
> Free PE / Size 166928 / 652.06 GB
> VG UUID Hm2mZH-jACj-gxQI-tbZM-H6pm-ovfr-TVgurC
>
> raneb:/etc/lvm/archive# lvdisplay
> raneb:/etc/lvm/archive#
>
> I presume I have to restore the lv somehow. But this
> has got me a step foward.
I could not see how I would restore the lv,
so I created a new one, with the same size and
name as the previous one
raneb:~# lvcreate -n lv1 -L 652.06G vg1
However, checking the volume fails
raneb:~# e2fsck /dev/mapper/vg1-lv1
e2fsck 1.40.2 (12-Jul-2007)
Couldn't find ext2 superblock, trying backup blocks...
e2fsck: Bad magic number in super-block while trying to open /dev/mapper/vg1-lv1
The superblock could not be read or does not describe a correct ext2
filesystem. If the device is valid and it really contains an ext2
filesystem (and not swap or ufs or something else), then the superblock
is corrupt, and you might try running e2fsck with an alternate superblock:
e2fsck -b 8193 <device>
Alternative superblocks, gleaned from
raneb:~# mke2fs -n /dev/mapper/vg1-lv1
mke2fs 1.40.2 (12-Jul-2007)
Filesystem label=
OS type: Linux
Block size=4096 (log=2)
Fragment size=4096 (log=2)
85475328 inodes, 170934272 blocks
8546713 blocks (5.00%) reserved for the super user
First data block=0
Maximum filesystem blocks=0
5217 block groups
32768 blocks per group, 32768 fragments per group
16384 inodes per group
Superblock backups stored on blocks:
32768, 98304, 163840, 229376, 294912, 819200, 884736, 1605632, 2654208,
4096000, 7962624, 11239424, 20480000, 23887872, 71663616, 78675968,
102400000
also fail. I presume all superblocks where on disk /dev/hdb,
and now that disk is gone, it is not possible to recover
data from /dev/hdc and /dev/hdd, the two other disks
in the vg. Thus, a failure on the first disk spills onto
the other disks because that disk hold vital information
for all.
Is that assessment correct?
Note I am desparate to recover data here because I destroyed
the backup through a mistake of mine 10 hours before disk
/dev/hdb crashed. The data represents about 10 years of
work of mine.
Conclusion: next time two backups.
Cheers,
Thomas Krichel http://openlib.org/home/krichel
RePEc:per:1965-06-05:thomas_krichel
skype: thomaskrichel
More information about the linux-lvm
mailing list