[linux-lvm] After replacing sata card can't read
Glynn Morrison
glymor at hotmail.com
Wed Dec 12 07:06:46 UTC 2007
I had a SATA controller card die on me. After I replaced the card I had to re-create the raid array in order to get mdadm to admit that the two disks previously on the faulty controller where not in fact faulty. This worked fine but after I didn't get anything from pvdisplay etc.
Last lvm backup before this all went south:
raid {
id = "KFk3aC-IYfB-y2VI-R2aZ-xu1M-g0YJ-eYQetH"
seqno = 8
status = ["RESIZEABLE", "READ", "WRITE"]
extent_size = 8192 # 4 Megabytes
max_lv = 0
max_pv = 0
physical_volumes {
pv0 {
id = "uRm1dv-M80B-YRPp-vNM8-WNth-kuke-oNCGHG"
device = "/dev/md2" # Hint only
status = ["ALLOCATABLE"]
pe_start = 384
pe_count = 596171 # 2.27421 Terabytes
}
}
logical_volumes {
all {
id = "jYuHXU-UaNa-cepc-juzv-RebF-zmJL-eJhiZU"
status = ["READ", "WRITE", "VISIBLE"]
segment_count = 1
segment1 {
start_extent = 0
extent_count = 576000 # 2.19727 Terabytes
type = "striped"
stripe_count = 1 # linear
stripes = [
"pv0", 0
]
}
}
}
}
what I did:
mdadm --create --assume-clean /dev/md2 --level=5 --raid-devices=6 /dev/sd[a-f]1
pvcreate -u uRm1dv-M80B-YRPp-vNM8-WNth-kuke-oNCGHG /dev/md2
vgcfgrestore -f lvm-raid raid
Latter I tried: pvcreate --restorefile lvm-raid -u uRm1dv-M80B-YRPp-vNM8-WNth-kuke-oNCGHG /dev/md2 instead to the same affect.
current output:
pvdisplay -v:
--- Physical volume ---
PV Name /dev/md2
VG Name raid
PV Size 2.27 TB / not usable 0
Allocatable yes
PE Size (KByte) 4096
Total PE 596171
Free PE 20171
Allocated PE 576000
PV UUID uRm1dv-M80B-YRPp-vNM8-WNth-kuke-oNCGHG
vgdisplay -v:
--- Volume group ---
VG Name raid
System ID
Format lvm2
Metadata Areas 1
Metadata Sequence No 9
VG Access read/write
VG Status resizable
MAX LV 0
Cur LV 1
Open LV 0
Max PV 0
Cur PV 1
Act PV 1
VG Size 2.27 TB
PE Size 4.00 MB
Total PE 596171
Alloc PE / Size 576000 / 2.20 TB
Free PE / Size 20171 / 78.79 GB
VG UUID KFk3aC-IYfB-y2VI-R2aZ-xu1M-g0YJ-eYQetH
--- Logical volume ---
LV Name /dev/raid/all
VG Name raid
LV UUID jYuHXU-UaNa-cepc-juzv-RebF-zmJL-eJhiZU
LV Write Access read/write
LV Status available
# open 0
LV Size 2.20 TB
Current LE 576000
Segments 1
Allocation inherit
Read ahead sectors 0
Block device 254:0
--- Physical volumes ---
PV Name /dev/md2
PV UUID uRm1dv-M80B-YRPp-vNM8-WNth-kuke-oNCGHG
PV Status allocatable
Total PE / Free PE 596171 / 20171
lvdisplay -v:
--- Logical volume ---
LV Name /dev/raid/all
VG Name raid
LV UUID jYuHXU-UaNa-cepc-juzv-RebF-zmJL-eJhiZU
LV Write Access read/write
LV Status available
# open 0
LV Size 2.20 TB
Current LE 576000
Segments 1
Allocation inherit
Read ahead sectors 0
Block device 254:0
diff /etc/lvm/backup/raid lvm-raid:
...
< seqno = 9
---
> seqno = 8
and the problem:
mount: wrong fs type, bad option, bad superblock on /dev/raid/all
xfs_check: unexpected XFS SB magic number 0x00000000
I'm assuming that all this wont have affected the actual data on the drive, but I'm at a loss of what to do now, as far as I can see everything is now as it was...
I would sincerely appreciate any help.
Rgrds,
Glynn
_________________________________________________________________
Free games, great prizes - get gaming at Gamesbox.
http://www.searchgamesbox.com
More information about the linux-lvm
mailing list