[linux-lvm] lvm raid failure

Null EXE nullexe at gmail.com
Thu Dec 13 01:31:16 UTC 2007


>
>
> ------------------------------
>
> Message: 6
> Date: Mon, 10 Dec 2007 21:42:45 -0500
> From: "Null EXE" <nullexe at gmail.com>
> Subject: [linux-lvm] lvm raid failure
> To: linux-lvm at redhat.com
> Message-ID:
>        <1f31f75c0712101842q4d71c03bu75926f1deb026a26 at mail.gmail.com>
> Content-Type: text/plain; charset="iso-8859-1"
>
> Hi everyone,
>
>
> I have been running an Edgy fileserver for about 6 months now. When I came
> home from work last week I found my files all inaccessible. After
> investigation I found issues with my PV. I would like to recover my data
> and
> replace the drives if possible any help is greatly appreciated!
>
> I have one system drive and 6 storage drive, all raid1. The storage layout
> is
> 2x160GB /dev/md0
> 2x250GB /dev/md1
> 2x250GB /dev/md2
> After investigation I found problems with my lvm. pvdisplay showed:
> Couldn't find device with uuid 'xjWU5M-G3WB-tZjB-Tyrk-Q7rN-Yide-1FEVh3'.
>
> I ran the command:
> sudo pvcreate --uuid "xjWU5M-G3WB-tZjB-Tyrk-Q7rN-Yide-1FEVh3"
> --restorefile
> /etc/lvm/archive/fileserver_00006.vg /dev/md2
>
> This seemed to recover /dev/md2. I re-ran pvdisplay and got
> Couldn't find device with uuid '1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo'.
>
> I figured I could run the same command on md1,
> sudo pvcreate --uuid "1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo"
> --restorefile
> /etc/lvm/archive/fileserver_00006.vg /dev/md1
>
> and got the message:
> Couldn't find device with uuid '1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo'.
> Device /dev/md1 not found (or ignored by filtering).
>
> After re-executing the above command with -vvv at the bottom of the output
> I
> get the following message:
> #device/dev-io.c:439         Opened /dev/md1 RO
> #device/dev-io.c:264       /dev/md1: size is 0 sectors
> #filters/filter.c:106         /dev/md1: Skipping: Too small to hold a PV
> #device/dev-io.c:485         Closed /dev/md1
> #pvcreate.c:81   Device /dev/md1 not found (or ignored by filtering).
>
> Here is my pvdisplay. Again, any help is Greatly appreciated
> ***START***
>
> Couldn't find device with uuid '1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo'.
> Couldn't find device with uuid '1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo'.
> --- Physical volume ---
> PV Name /dev/md0
> VG Name fileserver
> PV Size 149.05 GB / not usable 0
> Allocatable yes
> PE Size (KByte) 4096
> Total PE 38156
> Free PE 9996
> Allocated PE 28160
> PV UUID pZV1Ff-Y7fu-S8m1-tVFn-fOMJ-VRls-fLFEov
>
> --- Physical volume ---
> PV Name unknown device
> VG Name fileserver
> PV Size 232.88 GB / not usable 0
> Allocatable yes (but full)
> PE Size (KByte) 4096
> Total PE 59618
> Free PE 0
> Allocated PE 59618
> PV UUID 1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo
>
> --- Physical volume ---
> PV Name /dev/md2
> VG Name fileserver
> PV Size 232.88 GB / not usable 0
> Allocatable yes
> PE Size (KByte) 4096
> Total PE 59618
> Free PE 9156
> Allocated PE 50462
> PV UUID xjWU5M-G3WB-tZjB-Tyrk-Q7rN-Yide-1FEVh3
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL:
> https://www.redhat.com/archives/linux-lvm/attachments/20071210/1c987df9/attachment.html
>
> ------------------------------
>
> Message: 7
> Date: Tue, 11 Dec 2007 09:20:29 +0100
> From: Luca Berra <bluca at comedia.it>
> Subject: Re: [linux-lvm] lvm raid failure
> To: linux-lvm at redhat.com
> Message-ID: <20071211082029.GA1535 at percy.comedia.it>
> Content-Type: text/plain; charset=us-ascii; format=flowed
>
> On Mon, Dec 10, 2007 at 09:42:45PM -0500, Null EXE wrote:
> >get the following message:
> >#device/dev-io.c:439         Opened /dev/md1 RO
> >#device/dev-io.c:264       /dev/md1: size is 0 sectors
> >#filters/filter.c:106         /dev/md1: Skipping: Too small to hold a PV
> >#device/dev-io.c:485         Closed /dev/md1
> >#pvcreate.c:81   Device /dev/md1 not found (or ignored by filtering).
> >
> >Here is my pvdisplay. Again, any help is Greatly appreciated
> you should investigate what's wrong at the md layer, lvm seems to be
> just a victim.
>
> check
> /proc/mdstat
> kernel-logs
> mdadm.conf
> mdadm -Es
>
> L.
>
> --
> Luca Berra -- bluca at comedia.it
>        Communication Media & Services S.r.l.
>  /"\
>  \ /     ASCII RIBBON CAMPAIGN
>  X        AGAINST HTML MAIL
>  / \
>
>
>   ***
/proc/mdstat
Personalities : [raid1]
md2 : active raid1 dm-5[1]
      244195904 blocks [2/1] [_U]

md1 : inactive hdj1[0]
      244195904 blocks super non-persistent

md0 : active raid1 hdc1[0]
      156288256 blocks [2/1] [U_]

unused devices: <none>

/should my dm-5 be displayed or should it be a /dev/hd[a-z] device

***
mdadm -Es
ARRAY /dev/md0 level=raid1 num-devices=2
UUID=eeb0befe:9e50d574:974f5eae:ccb8e527
ARRAY /dev/md2 level=raid1 num-devices=2
UUID=1e5c1ab5:a5a06f84:d722583d:ca0f8cad
ARRAY /dev/md1 level=raid1 num-devices=2
UUID=41fc507e:7293a47c:7f18baeb:89dd5958

***
mdadm.conf
DEVICE partitions

/proc/partitions
   3     0   10022040 hda
   3     1    9574708 hda1
   3     2          1 hda2
   3     5     441756 hda5
  22     0  156290904 hdc
  22     1  156288321 hdc1
  22    64  156290904 hdd
  22    65  156288321 hdd1
  56     0  244198584 hdi
  56     1  244196001 hdi1
  56    64  244198584 hdj
  56    65  244196001 hdj1
  57     0  244198584 hdk
  57     1  244196001 hdk1
  57    64  244198584 hdl
  57    65  244196001 hdl1
   9     0  156288256 md0
   9     2  244195904 md2
 253     0    9574708 dm-0
 253     1     441756 dm-1
 253     2  156288321 dm-2
 253     3  156288321 dm-3
 253     4  244196001 dm-4
 253     5  244196001 dm-5
 253     6  244196001 dm-6
 253     7  244196001 dm-7


***
kernel logs
Nov 28 22:58:46 ark-server kernel: [42949384.540000] raid1: raid set md1
active with 1 out of 2 mirrors
Nov 28 22:58:46 ark-server kernel: [42949384.560000] md: md2 stopped.
Nov 28 22:58:46 ark-server kernel: [42949384.560000] md: bind<hdj1>
Nov 28 22:58:46 ark-server kernel: [42949384.560000] md: hdl1 has same UUID
but different superblock to hdj1
Nov 28 22:58:46 ark-server kernel: [42949384.560000] md: hdl1 has different
UUID to hdj1
Nov 28 22:58:46 ark-server kernel: [42949384.560000] md: export_rdev(hdl1)

***
Looking at all of this.  When I set up the array I remember my devices
ordered with hd[cdefgh] now I'm seeing md1 trying to use hd[jl].  Is this a
problem that it put assigned the drives new letter automatically?
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-lvm/attachments/20071212/c789a574/attachment.htm>


More information about the linux-lvm mailing list