[Linux-cluster] CLVM error 6 nodes

Jim Parsons jparsons at redhat.com
Sat Oct 14 16:16:48 UTC 2006


Filipe Miranda wrote:

> Alasdair,
>
>
> We did the modifications but still the same error:
>
> Since the lock servers does not have access to the LUNs that will be 
> in use by the GFS I used the option:
>
><clusternode name="lock-node3" votes="1" clvmd="0">
>
Um, I may be out on a limb here, but I know of no clvmd attribute under 
clusternode in our cluster.conf schema.

-Jim

> Only the rac-nodes have access to the shared LUNs...
>
>
> [root at rac-node1 ~]# lvcreate -L4G -n lv01 vg01
>
>   Found duplicate PV o2Xf8uUmskTL5fiVQAgY0nJ1ZJSMA9U3: using /dev/sds1 
> not /dev/emcpowerb1
>
>   Found duplicate PV 3yduFLdX3FWPWWMb9lIbtBf3JIPjYnHF: using 
> /dev/emcpowerc1 not /dev/sdr1
>
>   Found duplicate PV Kyha3qI6nVE4odg77UXf7igS3FenGJNn: using /dev/sdd1 
> not /dev/emcpowerq1
>
>   Found duplicate PV WS1LyhqQ8HaE2fIuSnXNd5sgTRtNzNAJ: using /dev/sdt1 
> not /dev/emcpowera1
>
>   Found duplicate PV DuBJ7dZsS3PIO7n5U6hINxPkWorZDzvx: using 
> /dev/emcpowerd1 not /dev/sdq1
>
>   Found duplicate PV ZECZzAtbA0e9pFbl9oL0lZg4q7fkS5x4: using /dev/sdk1 
> not /dev/emcpowerj1
>
>   Found duplicate PV bnVVmL6WhS2mesnOFUkT4fEfR0cFhybD: using 
> /dev/emcpowerk1 not /dev/sdj1
>
>   Found duplicate PV XyXrg2zdxxMS5jo03f9I4QYtGM3ILLGV: using /dev/sdl1 
> not /dev/emcpoweri1
>
>   Found duplicate PV SLE5v2eTD7cJlpRUDGG35xfXkRbW86i1: using 
> /dev/emcpowerl1 not /dev/sdi1
>
>   Found duplicate PV acGyUd2wX7FnOF94Cbt0ombp10iUWMSf: using /dev/sdm1 
> not /dev/emcpowerh1
>
>   Found duplicate PV ll8eNZ0JRh9katV0eui4BcxSc6HBggSI: using 
> /dev/emcpowerm1 not /dev/sdh1
>
>   Found duplicate PV ptGubq8R16LxywZ458P7ebmdG3Fq2aJo: using /dev/sdn1 
> not /dev/emcpowerg1
>
>   Found duplicate PV PLQ3uON7pYe7nY16gRmAP94WBaEydRwf: using 
> /dev/emcpowern1 not /dev/sdg1
>
>   Found duplicate PV PsVYTeKNy6EcqWYbJwQ4KEbPp2Q8HjWv: using /dev/sdo1 
> not /dev/emcpowerf1
>
>   Found duplicate PV hvekbzDAltJ3t23QveOMz1axfhj9Mp2j: using 
> /dev/emcpowero1 not /dev/sdf1
>
>   Found duplicate PV 5OhUbKbZLW5bTc3tpJeU4YlH0dTttJHF: using /dev/sdp1 
> not /dev/emcpowere1
>
>   Found duplicate PV dFtPhq6pkwFdl41NTrAguAEFB3601CTb: using 
> /dev/emcpowerp1 not /dev/sde1
>
>   Error locking on node lock-node1: Internal lvm error, check syslog
>
>   Error locking on node lock-node2: Internal lvm error, check syslog
>
>   Error locking on node lock-node3: Internal lvm error, check syslog
>
>   Failed to activate new LV.
>
> File:
> /var/log/messages:
>
> Oct 13 21:20:46 lock-node1 lvm[5478]: Volume group for uuid not found: 
> UzvBBmBj7m53APMbye1XXztWjdIavfgX8L5rGTOB3i3KGYPazw1AVaGCmWsXZpqR
>
>
> Here is the cluster.conf
>
>[root at rac-node1 ~]# cat /etc/cluster/cluster.conf
><?xml version="1.0
>"?>
><cluster alias="cluster" config_version="6" name="cluster">
>        <fence_daemon post_fail_delay="0" post_join_delay="120"/>
>        <clusternodes>
>
>                <clusternode name="lock-node1" votes="1" clvmd="0">
>                        <fence>
>                                <method name="1">
>
>                                        <device name="lock-node1-fence"/>
>                                </method>
>                        </fence>
>                </clusternode>
>
>                <clusternode name="lock-node2" votes="1" clvmd="0">
>                        <fence>
>
>                                <method name="1">
>                                        <device name="lock-node2-fence"/>
>                                </method>
>                        </fence>
>
>                </clusternode>
>                <clusternode name="lock-node3" votes="1" clvmd="0">
>                        <fence>
>                                <method name="1">
>
>                                        <device name="lock-node3-fence"/>
>                                </method>
>                        </fence>
>                </clusternode>
>
>                <clusternode name="rac-node1" votes="1">
>                        <fence>
>                                <method name="1">
>                                        <device name="rac-node1-fence"/>
>
>                                </method>
>                        </fence>
>                </clusternode>
>                <clusternode name="rac-node2" votes="1">
>
>                        <fence>
>                                <method name="1">
>                                        <device name="rac-node2-fence"/>
>                                </method>
>
>                        </fence>
>                </clusternode>
>                <clusternode name="rac-node3" votes="1">
>                        <fence>
>                                <method name="1">
>
>                                        <device name="rac-node3-fence"/>
>                                </method>
>                        </fence>
>                </clusternode>
>
>        </clusternodes>
>        <gulm>
>                <lockserver name="lock-node1"/>
>                <lockserver name="lock-node2"/>
>                <lockserver name="lock-node3"/>
>
>        </gulm>
>        <fencedevices>
>                <fencedevice agent="fence_ipmilan" auth="none"
>ipaddr="20.20.20.4 <http://20.20.20.4>" login="xxxx" name="lock-node1-fence" passwd="xxxx"/>
>
>                <fencedevice agent="fence_ipmilan" auth="none"
>ipaddr="20.20.20.5 <http://20.20.20.5>" login="xxxx" name="lock-node2-fence" passwd="xxxx"/>
>
>                <fencedevice agent="fence_ipmilan" auth="none"
>ipaddr="20.20.20.6 <http://20.20.20.6>" login="xxxx" name="lock-node3-fence" passwd="xxxx"/>
>
>                <fencedevice agent="fence_ipmilan" auth="none"
>ipaddr="20.20.20.1 <http://20.20.20.1>" login="xxxx" name="rac-node1-fence" passwd="xxxx"/>
>
>                <fencedevice agent="fence_ipmilan" auth="none"
>ipaddr="20.20.20.2 <http://20.20.20.2>" login="xxxx" name="rac-node2-fence" passwd="xxxx"/>
>
>                <fencedevice agent="fence_ipmilan" auth="none"
>ipaddr="20.20.20.3 <http://20.20.20.3>" login="xxxx" name="rac-node3-fence" passwd="xxxx"/>
>
>        </fencedevices>
>        <rm>
>                <failoverdomains/>
>                <resources/>
>        </rm>
></cluster>
>
>
> /etc/lvm/lvm.conf
>
> # By default we accept every block device:
>
>     # filter = [ "a/.*/" ]
>
>     filter = [ "r|/dev/sda|", "a/.*/" ]
>
>  
>
> Regards,
> Filipe Miranda
>
> On 10/13/06, *Filipe Miranda* <filipe.miranda at gmail.com 
> <mailto:filipe.miranda at gmail.com>> wrote:
>
>     Alasdair,
>
>     Do I need to reboot the machine to test this configurations
>     changes? Or is there a way do testing it without rebooting the
>     machine?
>
>     I will try to swap the order of the filters,
>
>     Thanks for the hint
>
>
>     On 10/13/06, *Alasdair G Kergon* < agk at redhat.com
>     <mailto:agk at redhat.com> > wrote:
>
>         On Fri, Oct 13, 2006 at 06:50:24PM -0300, Filipe Miranda wrote:
>> filter = [ "a/.*/", "r|/dev/sda|" ] (forgot to use the close
>         pipe)
>
>         Still won't work - swap the order of the two items; first matches
>         everything so second isn't looked at.
>         (man lvm.conf)
>
>         Alasdair
>         --
>         agk at redhat.com <mailto:agk at redhat.com>
>
>         --
>         Linux-cluster mailing list
>         Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
>         https://www.redhat.com/mailman/listinfo/linux-cluster
>         <https://www.redhat.com/mailman/listinfo/linux-cluster>
>
>
>
>
>     -- 
>     ---
>     Filipe T Miranda
>     Red Hat Certified Engineer
>
>
>
>
> -- 
> ---
> Filipe T Miranda
> Red Hat Certified Engineer
>
>
>------------------------------------------------------------------------
>
>--
>Linux-cluster mailing list
>Linux-cluster at redhat.com
>https://www.redhat.com/mailman/listinfo/linux-cluster
>






More information about the Linux-cluster mailing list