[linux-lvm] Unable to mount LVM partition - table too small
Adam NEWHAM
adam at thenewhams.com
Wed Sep 22 15:39:27 UTC 2010
Thanks for looking into this. Here is the requested info, but I think something might be up with the array. I've captured additional info - I also have a screen capture from the Disk Utility but I will probably have to send that in a private email as it requires an attachment.
Here is my mdadm.conf (I recently uncommented the #DEVICE partitions as I thought Ubuntu might be picking up invalid metadata (see output from examine below). This is the same config used in the RHEL 5 install, again I tried with/without the DEVICE partitions line
#DEVICE partitions
ARRAY /dev/md0 level=raid5 num-devices=4 UUID=b5e0fcd0:cfadbb04:a5b6f22e:457f47ae
Here is what comes from mdmadm --examine --scan
ARRAY /dev/md0 level=raid5 num-devices=4 UUID=b5e0fcd0:cfadbb04:a5b6f22e:457f47ae
ARRAY /dev/md0 level=raid5 num-devices=4 UUID=08558923:881d9efd:464c249d:988d2ec6
This ties in with what the Disk Utility is seeing, hence why I deleted the second line as I think one of the disks has invalid meta data.
Performing examine on each of the RAID members gives:
mdadm: No md superblock detected on /dev/sda.
I've listed the other 3 drives below so that the most relevant info is at the start of this email.
vgchange -a y displays the following at the console:
$ sudo vgchange -a y lvm-raid5
device-mapper: resume ioctl failed: Invalid argument
Unable to resume lvm--raid5-lvm0 (252:0)
1 logical volume(s) in volume group "lvm-raid5" now active
With the following in /var/log/messages
kernel: [ 553.685856] device-mapper: table: 252:0: md0p1 too small for target: start=384, len=5860556800, dev_size=1953520002
But any attempt to mount the LVM results in:
mount: /dev/mapper/lvm--raid5-lvm0 already mounted or //mnt/lvm-raid5 busy
Obviously the mount is failing because something is out of whack.
/dev/sdb:
Magic : a92b4efc
Version : 00.90.03
UUID : b5e0fcd0:cfadbb04:a5b6f22e:457f47ae
Creation Time : Sat Nov 1 22:14:18 2008
Raid Level : raid5
Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
Array Size : 2930287488 (2794.54 GiB 3000.61 GB)
Raid Devices : 4
Total Devices : 4
Preferred Minor : 0
Update Time : Mon Sep 20 19:24:26 2010
State : clean
Active Devices : 4
Working Devices : 4
Failed Devices : 0
Spare Devices : 0
Checksum : e9cec762 - correct
Events : 68
Layout : left-symmetric
Chunk Size : 64K
Number Major Minor RaidDevice State
this 0 8 0 0 active sync /dev/sda
0 0 8 0 0 active sync /dev/sda
1 1 8 16 1 active sync /dev/sdb
2 2 8 32 2 active sync /dev/sdc
3 3 8 48 3 active sync /dev/sdd
/dev/sdc:
Magic : a92b4efc
Version : 00.90.03
UUID : b5e0fcd0:cfadbb04:a5b6f22e:457f47ae
Creation Time : Sat Nov 1 22:14:18 2008
Raid Level : raid5
Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
Array Size : 2930287488 (2794.54 GiB 3000.61 GB)
Raid Devices : 4
Total Devices : 4
Preferred Minor : 0
Update Time : Mon Sep 20 19:24:26 2010
State : clean
Active Devices : 4
Working Devices : 4
Failed Devices : 0
Spare Devices : 0
Checksum : e9cec774 - correct
Events : 68
Layout : left-symmetric
Chunk Size : 64K
Number Major Minor RaidDevice State
this 1 8 16 1 active sync /dev/sdb
0 0 8 0 0 active sync /dev/sda
1 1 8 16 1 active sync /dev/sdb
2 2 8 32 2 active sync /dev/sdc
3 3 8 48 3 active sync /dev/sdd
/dev/sdd:
Magic : a92b4efc
Version : 00.90.03
UUID : b5e0fcd0:cfadbb04:a5b6f22e:457f47ae
Creation Time : Sat Nov 1 22:14:18 2008
Raid Level : raid5
Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
Array Size : 2930287488 (2794.54 GiB 3000.61 GB)
Raid Devices : 4
Total Devices : 4
Preferred Minor : 0
Update Time : Mon Sep 20 19:24:26 2010
State : clean
Active Devices : 4
Working Devices : 4
Failed Devices : 0
Spare Devices : 0
Checksum : e9cec786 - correct
Events : 68
Layout : left-symmetric
Chunk Size : 64K
Number Major Minor RaidDevice State
this 2 8 32 2 active sync /dev/sdc
0 0 8 0 0 active sync /dev/sda
1 1 8 16 1 active sync /dev/sdb
2 2 8 32 2 active sync /dev/sdc
3 3 8 48 3 active sync /dev/sdd
-----Original Message-----
From: bluca at comedia.it [mailto:linux-lvm-bounces at redhat.com] On Behalf Of Luca Berra
Sent: Tuesday, September 21, 2010 11:21 PM
To: linux-lvm at redhat.com
Subject: Re: [linux-lvm] Unable to mount LVM partition - table too small
On Tue, Sep 07, 2010 at 10:34:55AM -0700, Adam Newham wrote:
> vgdisplay
> --- Volume group ---
> VG Name lvm-raid5
> System ID
> Format lvm2
> Metadata Areas 1
> Metadata Sequence No 2
> VG Access read/write
> VG Status resizable
> MAX LV 0
> Cur LV 1
> Open LV 0
> Max PV 0
> Cur PV 1
> Act PV 1
> VG Size 2.73 TiB
> PE Size 32.00 MiB
> Total PE 89425
> Alloc PE / Size 89425 / 2.73 TiB
> Free PE / Size 0 / 0
> VG UUID wovrCm-knof-Ycdl-LdXt-4t28-mPWq-kngufG
does vgchange -a y fail?
is there any error message
> /proc/partitions (note the missing sub partitions – this I why I belive
> the lv/pv scan’s don’t see any LVM info)
> major minor #blocks name
>
> 3 0 156290904 hda
> 3 1 200781 hda1
> 3 2 4192965 hda2
> 3 3 151894575 hda3
> 8 0 976762584 sda
> 8 16 976762584 sdb
> 8 32 976762584 sdc
> 8 48 976762584 sdd
> 9 0 2930287488 md0
the partition info for md component devices is corectly removed from the
kernel, to avoid confusion
the md device itself should be partitionable, can i see your
/etc/mdadm.conf ?
L.
--
Luca Berra -- bluca at comedia.it
Communication Media & Services S.r.l.
/"\
\ / ASCII RIBBON CAMPAIGN
X AGAINST HTML MAIL
/ \
_______________________________________________
linux-lvm mailing list
linux-lvm at redhat.com
https://www.redhat.com/mailman/listinfo/linux-lvm
read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
More information about the linux-lvm
mailing list