[Linux-cluster] 2 node cluster gets fenced despite qdisk

Flavio Junior billpp at gmail.com
Tue Aug 26 19:03:10 UTC 2008


Hi ;)

I'm a newbie in this subject but, maybe it could help you:

http://sources.redhat.com/cluster/doc/cluster_schema_rhel5.html

http://sources.redhat.com/cluster/wiki/FAQ/CMAN#cman_quorum - I'm
reading exactly about it right now :)

Ignores me if you had already read this links.

--

Flávio do Carmo Júnior aka waKKu

On Tue, Aug 26, 2008 at 1:26 PM, Gerhard Spiegl <gspiegl at gmx.at> wrote:
> Hi all!
>
> I'm trying to get a 2 node cluster w/ quorum device running, most of its running
> fine.
> The cluster has a public net interface (bond0) and a private one (bond1).
> When the clusterinterconnect gets lost (ifconfig down the underlying eth devs),
> the two nodes immediatly fence each other and the cluster goes down. Is this some
> sort of expected behavior?
>
> I assumed the master node (qdiskd) {w,c,sh}ould stay alive and provide services,
> as the cluster still has one communication channel (the quorum disk).
>
> Below is the cluster.conf, OS used is RHEL5.2 with the latest patches.
>
> Or is the only use of the quorum disk not to get a split brain condition?
>
> If one could point me to a good resource for RHCS cluster configuration (eg
> comprehensive explaination of cluster.conf options) I would much apreciate this.
>
> kind regards
> Gerhard
>
>
>
> [root at ols011p yum.repos.d]# cat /etc/cluster/cluster.conf
> <?xml version="1.0"?>
> <cluster alias="fusi01" config_version="12" name="fusi01">
>        <fence_daemon post_fail_delay="0" post_join_delay="10"/>
>        <cman expected_votes="3" two_node="0"/>
>        <clusternodes>
>                <clusternode name="ols011p" nodeid="1" votes="1">
>                        <fence>
>                                <method name="1">
>                                        <device lanplus="1" name="OLS011-m"/>
>                                </method>
>                        </fence>
>                </clusternode>
>                <clusternode name="ols012p" nodeid="2" votes="1">
>                        <fence>
>                                <method name="1">
>                                        <device lanplus="1" name="OLS012-m"/>
>                                </method>
>                        </fence>
>                </clusternode>
>        </clusternodes>
>        <quorumd device="/dev/mapper/HDS-00F9p1" votes="1">
>
>            <heuristic interval="2" tko="3" program="ping -c1 -t3 172.27.111.254" score="1"/>
>
>        </quorumd>
>
>        <fencedevices>
>                <fencedevice agent="fence_ipmilan" option="off" auth="password" ipaddr="ols011-m" login="root" name="OLS011-m" passwd="changeme"/>
>                <fencedevice agent="fence_ipmilan" option="off" auth="password" ipaddr="ols012-m" login="root" name="OLS012-m" passwd="changeme"/>
>        </fencedevices>
>
>        <rm>
>                <failoverdomains>
>                        <failoverdomain name="fusi01_hvm_dom" nofailback="0" ordered="1" restricted="1">
>                                <failoverdomainnode name="ols011p" priority="2"/>
>                                <failoverdomainnode name="ols012p" priority="1"/>
>                        </failoverdomain>
>                        <failoverdomain name="fusi01_pvm_dom" nofailback="0" ordered="1" restricted="1">
>                                <failoverdomainnode name="ols011p" priority="1"/>
>                                <failoverdomainnode name="ols012p" priority="2"/>
>                        </failoverdomain>
>                </failoverdomains>
>                <resources/>
>                <vm autostart="1" domain="fusi01_pvm_dom" exclusive="0" migrate="live" name="fusi01pvm" path="/global/xenconfig" recovery="restart"/>
>                <vm autostart="1" domain="fusi01_hvm_dom" exclusive="0" migrate="live" name="fusi01hvm" path="/global/xenconfig" recovery="restart"/>
>        </rm>
> </cluster>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>




More information about the Linux-cluster mailing list