[linux-lvm] Unable to mount LVM partition - table too small

Adam Newham adam at thenewhams.com
Mon Sep 20 20:52:07 UTC 2010


Bump....no one?

On 9/7/2010 10:34 AM, Adam Newham wrote:
>    I didn’t see this getting posted, so re-posting. Sorry if you get this
> twice.
>
> Hi, hopefully somebody on this mailing list will be able to assist. I’ve
> done lots of Googling and tried a few things but with no success.
>
> I recently had multiple hardware failures and had to re-install the OS.
> My server is setup with an OS drive and a data drive. The OS drive is a
> single HDD which had a RHEL5 based distro on it (ClearOS) while the data
> drive consists of a software raid level 5 partition across 4x 1TB drives
> (2.7B available after ext3 format, with 1TB used). On top of this is an
> LVM2 partition with a single PV/LV/LG spanning the whole RAID partition.
>
> The hardware failures that I had were memory&  motherboard with the
> first RMA motherboard powering off sporadically (see note below) .
>
> However after completing the OS re-install, I’m unable to access the LVM
> partition. I’ve originally tried Ubuntu 10.04, which once mdadm/lvm2
> were installed - the distro saw the RAID and LVM container but I’m
> unable to mount the symbolic link (/dev/lvm-raid5/lv0) or the dev mapper
> link (/dev/mapper/lvm-raid5-lvm0). (See logs below) - one thing to note,
> as soon as the distro was installed and the RAID was assembled, a
> re-sync occurred. This wasn’t entirely unexpected as the first RMA’d
> motherboard was defective and would power off during the boot sequence
> and forced a check of the disc during boot which only got a few % into
> the sequence before a kernel panic was observed (/etc/stab was modified
> by booting into rescue mode and disabling this once I realized it was
> happening).
>
> Thinking maybe it was something with the Ubuntu distro, I tried
> installing CentOS 5.5 (and the original ClearOS distro) but both these
> distro’s give the same results. I can auto-create the /etc/mdadm.conf
> file by mdadm –detail –scan or mdadm –examine –scan but they can’t see
> any Physical/Logical volumes. One interesting point to note here is the
> /proc/partitions does not contain /dev/sda1…/dev/sdd1 etc. just the raw
> drives. Fdisk –l however shows all of the partitions information. I
> believe there is an issue with some Redhat based distro’s with how /dev
> is populated – specically it was introduced in FC10/11. I tried FC9 but
> got similar results as the RHEL5 based distro’s.
>
> I’d really like to get this data back, I have some backups (the discs
> contained Video, Music&  Photo’s) in the form of original CD&  DVD’s but
> for the Photo’s due to some other hardware failures, I have a gap from
> March 2008 until around April 2010.
>
> So here are the logs from what I can determine:
>
> Ubuntu 10.04
>
> /proc/partitions
>
> major minor #blocks name
>
> 8 0 976762584 sda
> 8 1 976760001 sda1
> 8 16 976762584 sdb
> 8 17 976760001 sdb1
> 8 32 976762584 sdc
> 8 33 976760001 sdc1
> 8 48 976762584 sdd
> 8 49 976760001 sdd1
> 8 64 58605120 sde
> 8 65 56165376 sde1
> 8 66 1 sde2
> 8 69 2437120 sde5
> 9 0 2930287488 md0
> 259 0 976760001 md0p1
>
> /proc/mdstat
>
> Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5]
> [raid4] [raid10]
> md0 : active raid5 sdc[2] sdb[1] sda[0] sdd[3]
> 2930287488 blocks level 5, 64k chunk, algorithm 2 [4/4] [UUUU]
>
> unused devices:<none>
>
> fdisk –l
>
> Disk /dev/sda: 1000.2 GB, 1000204886016 bytes
> 255 heads, 63 sectors/track, 121601 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
>
> Device Boot Start End Blocks Id System
> /dev/sda1 1 121601 976760001 fd Linux raid autodetect
>
> Disk /dev/sdb: 1000.2 GB, 1000204886016 bytes
> 255 heads, 63 sectors/track, 121601 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
>
> Device Boot Start End Blocks Id System
> /dev/sdb1 1 121601 976760001 fd Linux raid autodetect
>
> Disk /dev/sdc: 1000.2 GB, 1000204886016 bytes
> 255 heads, 63 sectors/track, 121601 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
>
> Device Boot Start End Blocks Id System
> /dev/sdc1 1 121601 976760001 fd Linux raid autodetect
>
> Disk /dev/sdd: 1000.2 GB, 1000204886016 bytes
> 255 heads, 63 sectors/track, 121601 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
>
> Device Boot Start End Blocks Id System
> /dev/sdd1 1 121601 976760001 fd Linux raid autodetect
>
> Disk /dev/sde: 60.0 GB, 60011642880 bytes
> 255 heads, 63 sectors/track, 7296 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x0005cd42
>
> Device Boot Start End Blocks Id System
> /dev/sde1 * 1 6993 56165376 83 Linux
> /dev/sde2 6993 7296 2437121 5 Extended
> /dev/sde5 6993 7296 2437120 82 Linux swap / Solaris
>
> Disk /dev/md0: 3000.6 GB, 3000614387712 bytes
> 255 heads, 63 sectors/track, 364803 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 65536 bytes / 196608 bytes
> Disk identifier: 0x00000000
>
> Device Boot Start End Blocks Id System
> /dev/md0p1 1 121601 976760001 fd Linux raid autodetect
> Partition 1 does not start on physical sector boundary.
>
> pvscan
> PV /dev/md0p1 VG lvm-raid5 lvm2 [2.73 TiB / 0 free]
> Total: 1 [746.53 GiB] / in use: 1 [746.53 GiB] / in no VG: 0 [0 ]
>
> lvscan
> Reading all physical volumes. This may take a while...
> Found volume group "lvm-raid5" using metadata type lvm2
>
> vgscan
> Reading all physical volumes. This may take a while...
> Found volume group "lvm-raid5" using metadata type lvm2
>
> vgdisplay
> --- Volume group ---
> VG Name lvm-raid5
> System ID
> Format lvm2
> Metadata Areas 1
> Metadata Sequence No 2
> VG Access read/write
> VG Status resizable
> MAX LV 0
> Cur LV 1
> Open LV 0
> Max PV 0
> Cur PV 1
> Act PV 1
> VG Size 2.73 TiB
> PE Size 32.00 MiB
> Total PE 89425
> Alloc PE / Size 89425 / 2.73 TiB
> Free PE / Size 0 / 0
> VG UUID wovrCm-knof-Ycdl-LdXt-4t28-mPWq-kngufG
>
> lvmdiskscan
> /dev/ram0 [ 64.00 MiB]
> /dev/md0p1 [ 931.51 GiB] LVM physical volume
> /dev/ram1 [ 64.00 MiB]
> /dev/ram2 [ 64.00 MiB]
> /dev/ram3 [ 64.00 MiB]
> /dev/ram4 [ 64.00 MiB]
> /dev/ram5 [ 64.00 MiB]
> /dev/ram6 [ 64.00 MiB]
> /dev/ram7 [ 64.00 MiB]
> /dev/ram8 [ 64.00 MiB]
> /dev/ram9 [ 64.00 MiB]
> /dev/ram10 [ 64.00 MiB]
> /dev/ram11 [ 64.00 MiB]
> /dev/ram12 [ 64.00 MiB]
> /dev/ram13 [ 64.00 MiB]
> /dev/ram14 [ 64.00 MiB]
> /dev/ram15 [ 64.00 MiB]
> /dev/root [ 53.56 GiB]
> /dev/sde5 [ 2.32 GiB]
> 1 disk
> 17 partitions
> 0 LVM physical volume whole disks
> 1 LVM physical volume
>
> tail /var/log/messages (after mdadm –assemble /dev/md0 and mount
> /dev/lvm-raid5/lvm0 /mnt/lvm-raid5
>
> Sep 3 18:46:13 adam-desktop kernel: [ 479.014444] md: bind<sdb>
> Sep 3 18:46:13 adam-desktop kernel: [ 479.015421] md: bind<sdc>
> Sep 3 18:46:13 adam-desktop kernel: [ 479.015753] md: bind<sdd>
> Sep 3 18:46:13 adam-desktop kernel: [ 479.016272] md: bind<sda>
> Sep 3 18:46:13 adam-desktop kernel: [ 479.022937] raid5: device sda
> operational as raid disk 0
> Sep 3 18:46:13 adam-desktop kernel: [ 479.022944] raid5: device sdd
> operational as raid disk 3
> Sep 3 18:46:13 adam-desktop kernel: [ 479.022950] raid5: device sdc
> operational as raid disk 2
> Sep 3 18:46:13 adam-desktop kernel: [ 479.022955] raid5: device sdb
> operational as raid disk 1
> Sep 3 18:46:13 adam-desktop kernel: [ 479.023690] raid5: allocated
> 4222kB for md0
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024690] 0: w=1 pa=0 pr=4 m=1
> a=2 r=4 op1=0 op2=0
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024697] 3: w=2 pa=0 pr=4 m=1
> a=2 r=4 op1=0 op2=0
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024703] 2: w=3 pa=0 pr=4 m=1
> a=2 r=4 op1=0 op2=0
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024709] 1: w=4 pa=0 pr=4 m=1
> a=2 r=4 op1=0 op2=0
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024715] raid5: raid level 5
> set md0 active with 4 out of 4 devices, algorithm 2
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024719] RAID5 conf printout:
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024722] --- rd:4 wd:4
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024726] disk 0, o:1, dev:sda
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024730] disk 1, o:1, dev:sdb
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024734] disk 2, o:1, dev:sdc
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024737] disk 3, o:1, dev:sdd
> Sep 3 18:46:13 adam-desktop kernel: [ 479.024823] md0: detected capacity
> change from 0 to 3000614387712
> Sep 3 18:46:13 adam-desktop kernel: [ 479.028687] md0: p1
> Sep 3 18:46:13 adam-desktop kernel: [ 479.207359] device-mapper: table:
> 252:0: md0p1 too small for target: start=384, len=5860556800,
> dev_size=1953520002
>
> mdadm –detail /dev/md0
>
> /dev/md0:
> Version : 00.90
> Creation Time : Sat Nov 1 22:14:18 2008
> Raid Level : raid5
> Array Size : 2930287488 (2794.54 GiB 3000.61 GB)
> Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
> Raid Devices : 4
> Total Devices : 4
> Preferred Minor : 0
> Persistence : Superblock is persistent
>
> Update Time : Fri Sep 3 18:39:58 2010
> State : clean
> Active Devices : 4
> Working Devices : 4
> Failed Devices : 0
> Spare Devices : 0
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> UUID : b5e0fcd0:cfadbb04:a5b6f22e:457f47ae
> Events : 0.68
>
> Number Major Minor RaidDevice State
> 0 8 0 0 active sync /dev/sda
> 1 8 16 1 active sync /dev/sdb
> 2 8 32 2 active sync /dev/sdc
> 3 8 48 3 active sync /dev/sdd
>
> mdadm –detail /dev/sda1
> ARRAY /dev/md0 level=raid5 num-devices=4
> UUID=08558923:881d9efd:464c249d:988d2ec6
>
> Note: performing this for /dev/sdb1…/dev/sdd1 produce no output. As the
> UUID for /dev/md0 is above, I remove this line from the mdadm.conf file.
>
> As I don’t have the original /etc/lvm info, here is what I managed to
> recover by doing a dd from the discs/cut pasted into a lvm template.
>
> /etc/lvm/backup/lvm_raid5_00000.vg
> # Generated by LVM2 version 2.02.37-RHEL4 (2008-06-06): Tue Nov 18
> 13:45:06 2008
>
> contents = "Text Format Volume Group"
> version = 1
>
> description = ""
>
> creation_host = "pebblebeach.thenewhams.lan" # Linux
> pebblebeach.thenewhams.lan 2.6.27 #4 SMP Mon Nov 17 11:05:05 PST 2008 i686
> creation_time = 1227044706 # Tue Nov 18 13:45:06 2008
>
> lvm-raid5 {
> id = "wovrCm-knof-Ycdl-LdXt-4t28-mPWq-kngufG"
> seqno = 2
> status = ["RESIZEABLE", "READ", "WRITE"]
> max_lv = 0
> max_pv = 0
>
> physical_volumes {
>
> pv0 {
> id = "aBkcEY-nZho-iWe5-700D-kDSy-pTAK-sJJFYm"
> device = "/dev/md0p1" # Hint only
>
> status = ["ALLOCATABLE"]
> pe_start = 384
> pe_count = 89425
> }
> }
>
> logical_volumes {
>
> lvm0 {
> id = "lzHyck-6X6E-48pC-uW1N-OQmp-Ayjt-vbAvVR"
> status = ["READ", "WRITE", "VISIBLE"]
> segment_count = 1
>
> segment1 {
> start_extent = 0
> extent_count = 89425
>
> type = "striped"
> stripe_count = 1 # linear
>
> stripes = [
> "pv0", 0
> ]
> }
> }
> }
> }
>
>
> Some info from when in Centos/EL5 land…
>
> /proc/partitions (note the missing sub partitions – this I why I belive
> the lv/pv scan’s don’t see any LVM info)
> major minor #blocks name
>
> 3 0 156290904 hda
> 3 1 200781 hda1
> 3 2 4192965 hda2
> 3 3 151894575 hda3
> 8 0 976762584 sda
> 8 16 976762584 sdb
> 8 32 976762584 sdc
> 8 48 976762584 sdd
> 9 0 2930287488 md0
>
> /proc/mdstat
> Personalities : [raid6] [raid5] [raid4]
> md0 : active raid5 sda[0] sdd[3] sdc[2] sdb[1]
> 2930287488 blocks level 5, 64k chunk, algorithm 2 [4/4] [UUUU]
>
> unused devices:<none>
>
> _______________________________________________
> linux-lvm mailing list
> linux-lvm at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
>    




More information about the linux-lvm mailing list