[linux-lvm] lvm raid5 : drives all present but vg/lvm will not assemble
Andrew Falgout
digitalw00t at gmail.com
Tue Mar 31 04:56:10 UTC 2020
Anyone have any ideas how I can salvage this array?
./drae
On Tue, Mar 24, 2020 at 11:09 AM Andrew Falgout <digitalw00t at gmail.com>
wrote:
> The disks are seen, the volume groups are seen. When I try to activate
> the VG I get this:
>
> vgchange -ay vg1
> device-mapper: reload ioctl on (253:19) failed: Input/output error
> 0 logical volume(s) in volume group "vg1" now active
>
> I executed 'vgchange -ay vg1 -vvvv -dddd' and this is the only time an
> error was thrown.
> 20:53:16.552602 vgchange[10795] device_mapper/libdm-deptree.c:2921 Adding
> target to (253:19): 0 31256068096 raid raid5_ls 3 128 region_size 32768 3
> 253:13 253:14 253:15 253:16 253:17 253:18
> 20:53:16.552609 vgchange[10795] device_mapper/ioctl/libdm-iface.c:1853 dm
> table (253:19) [ opencount flush ] [16384] (*1)
> 20:53:16.552619 vgchange[10795] device_mapper/ioctl/libdm-iface.c:1853 dm
> reload (253:19) [ noopencount flush ] [16384] (*1)
> 20:53:16.572481 vgchange[10795] device_mapper/ioctl/libdm-iface.c:1903
> device-mapper: reload ioctl on (253:19) failed: Input/output error
>
> I've uploaded two very verbose and debug ridden.
> https://pastebin.com/bw5eQBa8
> https://pastebin.com/qV5yft05
>
> Ignore the naming. It's not a gluster. I was planning on making two and
> mirroring them in a gluster.
>
> ./drae
>
>
> On Mon, Mar 23, 2020 at 5:14 AM Bernd Eckenfels <ecki at zusammenkunft.net>
> wrote:
>
>> Do you see any dmesg kernel errors when you try to activate the LVs?
>>
>> Gruss
>> Bernd
>>
>>
>> --
>> http://bernd.eckenfels.net
>> ------------------------------
>> *Von:* linux-lvm-bounces at redhat.com <linux-lvm-bounces at redhat.com> im
>> Auftrag von Andrew Falgout <digitalw00t at gmail.com>
>> *Gesendet:* Saturday, March 21, 2020 4:22:04 AM
>> *An:* linux-lvm at redhat.com <linux-lvm at redhat.com>
>> *Betreff:* [linux-lvm] lvm raid5 : drives all present but vg/lvm will
>> not assemble
>>
>>
>> This started on a Raspberry PI 4 running raspbian. I moved the disks to
>> my Fedora 31 system, that is running the latest updates and kernel. When I
>> had the same issues there I knew it wasn't raspbian.
>>
>> I've reached the end of my rope on this. The disks are there, all three
>> are accounted for, and the LVM data on them can be seen. But it refuses to
>> activate stating I/O errors.
>>
>> [root at hypervisor01 ~]# pvs
>> PV VG Fmt Attr PSize PFree
>> /dev/sda1 local_storage01 lvm2 a-- <931.51g 0
>> /dev/sdb1 local_storage01 lvm2 a-- <931.51g 0
>> /dev/sdc1 local_storage01 lvm2 a-- <931.51g 0
>> /dev/sdd1 local_storage01 lvm2 a-- <931.51g 0
>> /dev/sde1 local_storage01 lvm2 a-- <931.51g 0
>> /dev/sdf1 local_storage01 lvm2 a-- <931.51g <931.51g
>> /dev/sdg1 local_storage01 lvm2 a-- <931.51g <931.51g
>> /dev/sdh1 local_storage01 lvm2 a-- <931.51g <931.51g
>> /dev/sdi3 fedora_hypervisor lvm2 a-- 27.33g <9.44g
>> /dev/sdk1 vg1 lvm2 a-- <7.28t 0
>> /dev/sdl1 vg1 lvm2 a-- <7.28t 0
>> /dev/sdm1 vg1 lvm2 a-- <7.28t 0
>> [root at hypervisor01 ~]# vgs
>> VG #PV #LV #SN Attr VSize VFree
>> fedora_hypervisor 1 2 0 wz--n- 27.33g <9.44g
>> local_storage01 8 1 0 wz--n- <7.28t <2.73t
>> vg1 3 1 0 wz--n- 21.83t 0
>> [root at hypervisor01 ~]# lvs
>> LV VG Attr LSize Pool Origin Data% Meta%
>> Move Log Cpy%Sync Convert
>> root fedora_hypervisor -wi-ao---- 15.00g
>>
>> swap fedora_hypervisor -wi-ao---- 2.89g
>>
>> libvirt local_storage01 rwi-aor--- <2.73t
>> 100.00
>> gluster02 vg1 Rwi---r--- 14.55t
>>
>>
>> The one in question is the vg1/gluster02 lvm group.
>>
>> I try to activate the VG:
>> [root at hypervisor01 ~]# vgchange -ay vg1
>> device-mapper: reload ioctl on (253:19) failed: Input/output error
>> 0 logical volume(s) in volume group "vg1" now active
>>
>> I've got the debugging output from :
>> vgchange -ay vg1 -vvvv -dddd
>> lvchange -ay --partial vg1/gluster02 -vvvv -dddd
>>
>> Just not sure where I should dump the data for people to look at. Is
>> there a way to tell the md system to ignore the metadata since there wasn't
>> an actual disk failure, and rebuild the metadata off what is in the lvm?
>> Or can I even get the LV to mount, so I can pull the data off.
>>
>> Any help is appreciated. If I can save the data great. I'm tossing this
>> to the community to see if anyone else has an idea of what I can do.
>> ./digitalw00t
>> _______________________________________________
>> linux-lvm mailing list
>> linux-lvm at redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-lvm
>> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-lvm/attachments/20200330/5ffcd215/attachment.htm>
More information about the linux-lvm
mailing list