[linux-lvm] lvm 1.0.3 kernel panic

Stephanus Fengler fengler at uiuc.edu
Tue Jul 20 01:09:30 UTC 2004


I got a knoppix rescue cd with lvm version 1.0.8 running now
and a sshd server to rescue my data. Problem is that this
version is not compatible with my volume group:

vgscan:

vgscan -- removing "/etc/lvmtab" and "/etc/lvmtab.d"
vgscan -- creating empty "/etc/lvmtab" and "/etc/lvmtab.d"
vgscan -- reading all physical volumes (this may take a while...)
vgscan -- scanning for all active volume group(s) first
vgscan -- reading data of volume group "vg01" from physical
volume(s)
vgscan -- found inactive volume group "vg01"
vgscan -- getting block device numbers for logical volumes

vgscan -- checking block device numbers of logical volumes

vgscan -- inserting "vg01" into lvmtab
vgscan -- backing up volume group "vg01"
vgscan -- checking volume group name "vg01"
vgscan -- checking volume group consistency of "vg01"
vgscan -- checking existence of "/etc/lvmtab.d"
vgscan -- storing volume group data of "vg01" in
"/etc/lvmtab.d/vg01.tmp"
vgscan -- storing physical volume data of "vg01" in
"/etc/lvmtab.d/vg01.tmp"
vgscan -- storing logical volume data of volume group "vg01"
in "/etc/lvmtab.d/vg01.tmp"
vgscan -- ERROR 2 writing volume group backup file
/etc/lvmtab.d/vg01.tmp in vg_cfgbackup.c [line 271]
vgscan -- ERROR: unable to do a backup of volume group "vg01"
vgscan -- ERROR "lvm_tab_vg_remove(): unlink" removing volume
group "vg01" from "/etc/lvmtab"
vgscan -- "/etc/lvmtab" and "/etc/lvmtab.d" successfully created
vgscan -- WARNING: This program does not do a VGDA backup of
your volume group

the error looks the same so I assume that something is
incompatible in my old system. I never changed lvm version as
far as I know of... but it might be possible.

Is there a way to access this volume group now?

Any help is really appreciated.

Stephanus

---- Original message ----
>Date: Mon, 19 Jul 2004 16:46:31 -0500
>From: Stephanus Fengler <fengler at uiuc.edu>  
>Subject: Re: [linux-lvm] lvm 1.0.3 kernel panic  
>To: LVM general discussion and development <linux-lvm at redhat.com>
>
>Here is additionally an output of:
>vgcfgrestore -f /mnt/sysimage/etc/lvmconf/vg01.conf -n vg01 -ll
>
>--- Volume group ---
>VG Name               vg01
>VG Access             read/write
>VG Status             NOT available/resizable
>VG #                  0
>MAX LV                256
>Cur LV                3
>Open LV               0
>MAX LV Size           255.99 GB
>Max PV                256
>Cur PV                4
>Act PV                4
>VG Size               501.92 GB
>PE Size               4 MB
>Total PE              128491
>Alloc PE / Size       128491 / 501.92 GB
>Free  PE / Size       0 / 0
>VG UUID               vsZRhX-6bfh-jqhD-Cn1Z-0h9E-tiE6-isU7hJ
>
>--- Logical volume ---
>LV Name                /dev/vg01/lv_root
>VG Name                vg01
>LV Write Access        read/write
>LV Status              available
>LV #                   1
># open                 0
>LV Size                33.66 GB
>Current LE             8617
>Allocated LE           8617
>Allocation             next free
>Read ahead sectors     10000
>Block device           58:0
>
>--- Logical volume ---
>LV Name                /dev/vg01/lv_data2
>VG Name                vg01
>LV Write Access        read/write
>LV Status              available
>LV #                   2
># open                 0
>LV Size                235.38 GB
>Current LE             60257
>Allocated LE           60257
>Allocation             next free
>Read ahead sectors     10000
>Block device           58:1
>
>--- Logical volume ---
>LV Name                /dev/vg01/lv_data
>VG Name                vg01
>LV Write Access        read/write
>LV Status              available
>LV #                   3
># open                 0
>LV Size                232.88 GB
>Current LE             59617
>Allocated LE           59617
>Allocation             next free
>Read ahead sectors     10000
>Block device           58:2
>
>
>--- Physical volume ---
>PV Name               /dev/hda3
>VG Name               vg01
>PV Size               33.67 GB [70605675 secs] / NOT usable
4.19 MB 
>[LVM: 161 KB]
>PV#                   1
>PV Status             available
>Allocatable           yes (but full)
>Cur LV                1
>PE Size (KByte)       4096
>Total PE              8617
>Free PE               0
>Allocated PE          8617
>PV UUID               ouWY4f-vGwm-Rq3p-eX3M-tDaI-BQ3i-D7gT4f
>
>--- Physical volume ---
>PV Name               /dev/hda1
>VG Name               vg01
>PV Size               2.50 GB [5253192 secs] / NOT usable
4.19 MB [LVM: 
>130 KB]
>PV#                   2
>PV Status             available
>Allocatable           yes (but full)
>Cur LV                1
>PE Size (KByte)       4096
>Total PE              640
>Free PE               0
>Allocated PE          640
>PV UUID               6WZGDT-Sev3-jYXj-VquB-FQj2-Wbeq-cdgNLK
>
>--- Physical volume ---
>PV Name               /dev/hdb
>VG Name               vg01
>PV Size               232.89 GB [488397168 secs] / NOT usable
4.38 MB 
>[LVM: 360 KB]
>PV#                   3
>PV Status             available
>Allocatable           yes (but full)
>Cur LV                1
>PE Size (KByte)       4096
>Total PE              59617
>Free PE               0
>Allocated PE          59617
>PV UUID               nV1chY-OlRj-tLrb-cdSM-D3IN-Mvwb-U2nxfE
>
>--- Physical volume ---
>PV Name               /dev/hdc
>VG Name               vg01
>PV Size               232.89 GB [488397168 secs] / NOT usable
4.38 MB 
>[LVM: 360 KB]
>PV#                   4
>PV Status             NOT available
>Allocatable           yes (but full)
>Cur LV                1
>PE Size (KByte)       4096
>Total PE              59617
>Free PE               0
>Allocated PE          59617
>PV UUID               XMs7mJ-PdbD-voTc-IUY7-sLiu-Gg1P-99nwc5
>
>I have checked the output of some older config files like 
>vg01.conf.[1-6].old and have found one without the additional
hard disk. 
>Hence no data is stored on that disk yet, I wouldn't mind to
overwrite 
>the configuration, but is it save doing it? I definitely
worried about 
>data loss then.
>
>Thanks,
>Stephanus
>
>
>Stephanus Fengler wrote:
>
>> Hi experts,
>>
>> I added a new hard disk to my system and created it
entirely as a new 
>> logical volume. Mounting unmounting everything worked until
reboot. It 
>> stops now with kerel panic and lines like:
>>
>> (null) -- ERROR 2 writing volume group backup file
/etc/lvmtab.d/vg01.tmp
>>       vg_cfgbackup.c [line271]
>>
>> vgscan -- ERROR: unable to do a backup of volume group vg01
>> vgscan -- ERROR: lvm_tab_vg_remove(): unlink" removing
volume group 
>> "vg01"  from  "/etc/lvmtab"
>>
>> ...
>>
>> Activating volume groups
>>    vgchange - no volume groups found
>>
>> I understand the kernel panic if lvm is unable to find the
volume 
>> group vg01 because that's my root system. But I don't get
the first 
>> error.
>> I rebooted with my Redhat Installation disk: linux rescue
>> and can activate the volume group by hand and mount the
file systems. 
>> So it looks to me everything is consistent in the filesystem.
>>
>> So since I am pretty new to lvm, which output do you
additional need 
>> to help me?
>>
>> Thanks in advance,
>> Stephanus
>>
>> lvmdiskscan:
>> lvmdiskscan -- reading all disks / partitions (this may
take a while...)
>> lvmdiskscan -- /dev/hdc   [     232.89 GB] USED LVM whole disk
>> lvmdiskscan -- /dev/hda1  [       2.50 GB] Primary  LVM
partition [0x8E]
>> lvmdiskscan -- /dev/hda2  [     101.97 MB] Primary  LINUX
native 
>> partition [0x83]
>> lvmdiskscan -- /dev/hda3  [      33.67 GB] Primary  LVM
partition [0x8E]
>> lvmdiskscan -- /dev/hda4  [    1019.75 MB] Primary 
Windows98 extended 
>> partition [0x0F]
>> lvmdiskscan -- /dev/hda5  [    1019.72 MB] Extended LINUX swap 
>> partition [0x82]
>> lvmdiskscan -- /dev/hdb   [     232.89 GB] USED LVM whole disk
>> lvmdiskscan -- /dev/loop0 [      59.08 MB] free loop device
>> lvmdiskscan -- 3 disks
>> lvmdiskscan -- 2 whole disks
>> lvmdiskscan -- 1 loop device
>> lvmdiskscan -- 0 multiple devices
>> lvmdiskscan -- 0 network block devices
>> lvmdiskscan -- 5 partitions
>> lvmdiskscan -- 2 LVM physical volume partitions
>>
>> pvscan:
>> pvscan -- reading all physical volumes (this may take a
while...)
>> pvscan -- inactive PV "/dev/hdc"  of VG "vg01" [232.88 GB /
0 free]
>> pvscan -- inactive PV "/dev/hda1" of VG "vg01" [2.50 GB / 0
free]
>> pvscan -- inactive PV "/dev/hda3" of VG "vg01" [33.66 GB /
0 free]
>> pvscan -- inactive PV "/dev/hdb"  of VG "vg01" [232.88 GB /
0 free]
>> pvscan -- total: 4 [501.94 GB] / in use: 4 [501.94 GB] / in
no VG: 0 [0]
>> _______________________________________________
>> linux-lvm mailing list
>> linux-lvm at redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-lvm
>> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>
>
>_______________________________________________
>linux-lvm mailing list
>linux-lvm at redhat.com
>https://www.redhat.com/mailman/listinfo/linux-lvm
>read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/



More information about the linux-lvm mailing list