[Linux-cluster] CLVM not activating LVs
Christine Caulfield
ccaulfie at redhat.com
Wed Aug 26 14:09:11 UTC 2009
On 26/08/09 14:19, Jakov Sosic wrote:
> Hi! CLVM not activating logical volumes.
>
>
>
> I have a major issues with CLVM. It is not activating volumes in my
> VG's. I have 2 iSCSI volumes, and one SAS volume with 3 VG's. On
> node01, all logical volumes in 1 iSCSI and on SAS are activated. Other
> iSCSI - zero. On node02 same situation. On node03 only lv's from SAS
> are activated. lvm.conf is same on the machines....
>
> This is very strange, because when I boot the machines, all the
> services are stopped, so logical volumes shouldn't be activated.
>
> Here is the situation:
>
> [root at node01 lvm]# vgs
> VG #PV #LV #SN Attr VSize VFree
> VolGroupC0 1 7 0 wz--nc 3.41T 1.48T
> VolGroupC1 1 0 0 wz--nc 3.41T 3.41T
> VolGroupSAS 1 2 0 wz--nc 20.63G 4.63G
>
> [root at node01 lvm]# lvs
> LV VG Attr LSize
> nered1 VolGroupC0 -wi-a- 200.00G
> nered2 VolGroupC0 -wi-a- 200.00G
> nered3 VolGroupC0 -wi-a- 1.46T
> nered4 VolGroupC0 -wi-a- 20.00G
> nered5 VolGroupC0 -wi-a- 20.00G
> nered6 VolGroupC0 -wi-a- 20.00G
> nered7 VolGroupC0 -wi-a- 20.00G
> sasnered0 VolGroupSAS -wi-a- 8.00G
> sasnered1 VolGroupSAS -wi-a- 8.00G
>
> [root at node03 cache]# vgs
> VG #PV #LV #SN Attr VSize VFree
> VolGroupC0 1 0 0 wz--nc 3.41T 3.41T
> VolGroupC1 1 0 0 wz--nc 3.41T 3.41T
> VolGroupSAS 1 2 0 wz--nc 20.63G 4.63G
>
> [root at node03 lvm]# lvs
> LV VG Attr LSize Origin
> sasnered0 VolGroupSAS -wi-a- 8.00G
> sasnered1 VolGroupSAS -wi-a- 8.00G
>
>
> here is my lvm.conf:
>
> [root at node01 lvm]# lvm dumpconfig
> devices {
> dir="/dev"
> scan="/dev"
> preferred_names=[]
> filter=["a|^/dev/mapper/controller0$|",
> "a|^/dev/mapper/controller1$|", "a|^/dev/mapper/sas-xen$|", "r|.*|"]
> cache_dir="/etc/lvm/cache"
> cache_file_prefix=""
> write_cache_state=0
> sysfs_scan=1
> md_component_detection=1
> md_chunk_alignment=1
> ignore_suspended_devices=0
> }
> dmeventd {
> mirror_library="libdevmapper-event-lvm2mirror.so"
> snapshot_library="libdevmapper-event-lvm2snapshot.so"
> }
> activation {
> missing_stripe_filler="error"
> reserved_stack=256
> reserved_memory=8192
> process_priority=-18
> mirror_region_size=512
> readahead="auto"
> mirror_log_fault_policy="allocate"
> mirror_device_fault_policy="remove"
> }
> global {
> library_dir="/usr/lib64"
> umask=63
> test=0
> units="h"
> activation=1
> proc="/proc"
> locking_type=3
> fallback_to_clustered_locking=1
> fallback_to_local_locking=1
> locking_dir="/var/lock/lvm"
> }
> shell {
> history_size=100
> }
> backup {
> backup=1
> backup_dir="/etc/lvm/backup"
> archive=1
> archive_dir="/etc/lvm/archive"
> retain_min=10
> retain_days=30
> }
> log {
> verbose=0
> syslog=1
> overwrite=0
> level=0
> indent=1
> command_names=0
> prefix=" "
> }
>
>
> Note that logical volumes form C1 were present on node01 and node02,
> but after the node03 joined cluster they dissapeared. I'm running
> CentOS 5.3.
>
> This is really dissapointing. Enterprise Linux? Linux maybe but not
> Enterprise... After much trouble with linux dm-multipath issues with my
> storage - which are unresolved and are waiting for RHEL 5.4, now clvmd.
>
> Note that locking (DLM), cman, rgmanager, qdisk and all the other
> cluster services are working without problems. I just don't get it why
> is CLVM behaving this way?
>
> I'm thinking about switching to non-clustered LVM - but are there
> issues with possible corruption of metadata? I won't create any new
> volumes nor snapshots or anything similar. Setup is done and it should
> work like this for the extended period of time.... But are there issues
> with activation or something else changing metadata?
>
You need to mark the shared VGs clustered using the command
# vgchange -cy <VGname>
If you created them while clvmd was active then this is the default. If
not then you will have to add it yourself as above.
Chrissie
More information about the Linux-cluster
mailing list