[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]

F9 install onto LVM on RAID1 on RAID5



Hello list,

I am trying to install a new F9 system onto a RAID/LVM setup.

As anaconda doesn't let me create the RAID/LVM configuration I require,
I created these devices using mdadm and lvm in the shell during
installation.  I then created the filesystems and swap space, with
labels, in the LVM volume group and I can mount them, read, write, etc..
So all is well with the underlying setup.

However, returning to the install's "custom layout" partitioning page,
anaconda displays the volume group and the names and sizes of the
members are correct, but in the "TYPE" colume it indicates "foreign" and
the mount point, and other fields are empty.

If I click "LVM" or highlight one of the members and click "Edit" (to
set the mount points and formatting options), anaconda responds with:

--------------
Not enough physical volumes 
(...)
Create a partition or RAID array of type "physical volume (LVM)"
and click "LVM" again.
--------------

Consequently, I cannot edit the member details to set mount points and
formatting options and continue with the installation.

How does anaconda determine the "type" of a RAID array; do md devices
have types (as partitions do)?  How can I satisfy it there are indeed
physical volumes for an LVM (and that I have already configured them)?

Alternately, how can I definitely tell anaconda to simply skip all
partitioning and let me tell it which /dev entries to use for whichever
partitions?


Details
=======

There are four disks in the system, I will add three more.  The
intention is the have a mirror three disks, for six total, plus one
spare on one half.  So:

md0 is a raid1 (mirror) of four small partitions of each disk.
  mdadm --create /dev/md0 --level=raid1 --raid-devices=4 /dev/sd[abcd]1

md1 is a raid5 of the remaining portion of three disks plus a spare.
  mdadm --create /dev/md1 --level=raid5 --raid-devices=3 --spare-devices=1 --assume-clean /dev/sd[abcd]2

md4 is a raid1 (mirror), degraded because disks are currently missing,
of md0 and "missing".
  mdadm --create /dev/md4 --level=raid1 --raid-devices=2 /dev/md0 missing

md5 is a raid1 (mirror), degraded because disks are currently missing,
of md1 and "missing".
  mdadm --create /dev/md5 --level=raid1 --raid-devices=2 /dev/md1 missing

I configured my lvm volumes on md5, and as I mentioned, anaconda does
see the members.

  lvm> pvcreate /dev/md5
  lvm> vgcreate -s 32m vg0 /dev/md5
  lvm> lvcreate -L 1024m -n root vg0 ; ...
  lvm> vgchange -a y vg0


Aside, anaconda displays md0 and md1 in the list of RAID volumes (both
as type "foreign"), but *not* md4 and md5 -- even though they are just
normal mirror RAIDs.  Is this because they are degraded?

I suspect anaconda lists the lvm members because it notices which vg's
are active.  It doesn't believe md5 contains a physical volume suitable
for LVM use.  (In fact, I don't think anaconda believes there any
physical volumes for lvm on the system at all and, as above, it doesn't
show md5 at all.)

If I cannot get anaconda to cooperate, I'll install onto a raid5 array
on temporary disks, then move the entire system into the proper
nested-RAID5/RAID1/LVM setup.

Thanks for suggestions.


Regards,
Richard


[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]