[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]

Re: [Linux-cluster] Rebooting the Master Node in an RHCS Cluster



Here is our cluster.conf, edited for security reasons:

----------------------------------------
<?xml version="1.0"?>
<cluster name="WFF_RHCS" config_version="34">
        <fence_domain clean_start="0" post_fail_delay="20" post_join_delay="3"/>
        <clusternodes>
                <clusternode name="node1" nodeid="1" votes="1">
                        <fence>
                                <method name="1">
                                        <device name="node1-ilo-fence"/>
                                </method>
                        </fence>
                </clusternode>
                <clusternode name="node2" nodeid="2" votes="1">
                        <fence>
                                <method name="1">
                                        <device name="node2-ilo-fence"/>
                                </method>
                        </fence>
                </clusternode>
                <clusternode name="node3" nodeid="3" votes="1">
                        <fence>
                                <method name="1">
                                        <device name="node3-ilo-fence"/>
                                </method>
                        </fence>
                </clusternode>
        </clusternodes>
        <fencedevices>
                <fencedevice agent="fence_ilo" hostname="xxx.xxx.xxx.xxx" login="login" name="node1-ilo-fence" passwd="passwd"/>
                <fencedevice agent="fence_ilo" hostname="xxx.xxx.xxx.xxx" login="login" name="node2-ilo-fence" passwd="passwd"/>
                <fencedevice agent="fence_ilo" hostname="xxx.xxx.xxx.xxx" login="login" name="node3-ilo-fence" passwd="passwd"/>
        </fencedevices>
        <rm log_facility="local5" log_level="7">
                <failoverdomains>
                        <failoverdomain name="rhcstestdomain" nofailback="0" ordered="0" restricted="0">
                                <failoverdomainnode name="node1" priority="1"/>
                                <failoverdomainnode name="node2" priority="1"/>
                                <failoverdomainnode name="node3" priority="1"/>
                        </failoverdomain>
                </failoverdomains>
                <resources>
                <script file="ora_cluster.sh" name="oracle_init_script"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <lvm lv_name="lv_a001_001_oradata" name="ora-001-a001-lv" vg_name="/dev/vg_oracle_001"/>
                <lvm lv_name="lv_b001_001_oradata" name="ora-001-b001-lv" vg_name="/dev/vg_oracle_001"/>
                <lvm lv_name="lv_u001_001_oradata" name="ora-001-u001-lv" vg_name="/dev/vg_oracle_001"/>
                <lvm lv_name="lv_u001_001_admin" name="ora-001-u001-admin-lv" vg_name="/dev/vg_oracle_001"/>
                <fs device="/dev/vg_oracle_001/lv_a001_001_oradata" force_fsck="0" force_unmount="1" fsid="50280" fstype="ext3" mountpoint="/a001/oradata/001" name="ora-001-a001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_001/lv_b001_001_oradata" force_fsck="0" force_unmount="1" fsid="57077" fstype="ext3" mountpoint="/b001/oradata/001" name="ora-001-b001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_001/lv_u001_001_oradata" force_fsck="0" force_unmount="1" fsid="31925" fstype="ext3" mountpoint="/u001/oradata/001" name="ora-u001-001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_001/lv_u001_001_admin" force_fsck="0" force_unmount="1" fsid="14919" fstype="ext3" mountpoint="/u001/app/oracle/admin/001" name="ora-u001-001-admin-fs" self_fence="0"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <lvm lv_name="lv_a001_002_oradata" name="ora-002-a001-lv" vg_name="/dev/vg_oracle_002"/>
                <lvm lv_name="lv_b001_002_oradata" name="ora-002-b001-lv" vg_name="/dev/vg_oracle_002"/>
                <lvm lv_name="lv_u001_002_oradata" name="ora-002-u001-lv" vg_name="/dev/vg_oracle_002"/>
                <lvm lv_name="lv_u001_002_admin" name="ora-002-u001-admin-lv" vg_name="/dev/vg_oracle_002"/>
                <fs device="/dev/vg_oracle_002/lv_a001_002_oradata" force_fsck="0" force_unmount="1" fsid="50280" fstype="ext3" mountpoint="/a001/oradata/002" name="ora-002-a001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_002/lv_b001_002_oradata" force_fsck="0" force_unmount="1" fsid="57077" fstype="ext3" mountpoint="/b001/oradata/002" name="ora-002-b001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_002/lv_u001_002_oradata" force_fsck="0" force_unmount="1" fsid="31925" fstype="ext3" mountpoint="/u001/oradata/002" name="ora-002-u001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_002/lv_u001_002_admin" force_fsck="0" force_unmount="1" fsid="14919" fstype="ext3" mountpoint="/u001/app/oracle/admin/002" name="ora-002-u001-admin-fs" self_fence="0"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <lvm lv_name="lv_a001_003_oradata" name="ora-003-a001-lv" vg_name="/dev/vg_oracle_003"/>
                <lvm lv_name="lv_b001_003_oradata" name="ora-003-b001-lv" vg_name="/dev/vg_oracle_003"/>
                <lvm lv_name="lv_u001_003_oradata" name="ora-003-u001-lv" vg_name="/dev/vg_oracle_003"/>
                <lvm lv_name="lv_u001_003_admin" name="ora-003-u001-admin-lv" vg_name="/dev/vg_oracle_003"/>
                <fs device="/dev/vg_oracle_003/lv_a001_003_oradata" force_fsck="0" force_unmount="1" fsid="50280" fstype="ext3" mountpoint="/a001/oradata/003" name="ora-003-a001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_003/lv_b001_003_oradata" force_fsck="0" force_unmount="1" fsid="57077" fstype="ext3" mountpoint="/b001/oradata/003" name="ora-003-b001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_003/lv_u001_003_oradata" force_fsck="0" force_unmount="1" fsid="31925" fstype="ext3" mountpoint="/u001/oradata/003" name="ora-003-u001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_003/lv_u001_003_admin" force_fsck="0" force_unmount="1" fsid="14919" fstype="ext3" mountpoint="/u001/app/oracle/admin/003" name="ora-003-u001-admin-fs" self_fence="0"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <lvm lv_name="lv_a001_004_oradata" name="ora-004-a001-lv" vg_name="/dev/vg_oracle_004"/>
                <lvm lv_name="lv_b001_004_oradata" name="ora-004-b001-lv" vg_name="/dev/vg_oracle_004"/>
                <lvm lv_name="lv_u001_004_oradata" name="ora-004-u001-lv" vg_name="/dev/vg_oracle_004"/>
                <lvm lv_name="lv_u001_004_admin" name="ora-004-u001-admin-lv" vg_name="/dev/vg_oracle_004"/>
                <fs device="/dev/vg_oracle_004/lv_a001_004_oradata" force_fsck="0" force_unmount="1" fsid="50280" fstype="ext3" mountpoint="/a001/oradata/004" name="ora-004-a001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_004/lv_b001_004_oradata" force_fsck="0" force_unmount="1" fsid="57077" fstype="ext3" mountpoint="/b001/oradata/004" name="ora-004-b001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_004/lv_u001_004_oradata" force_fsck="0" force_unmount="1" fsid="31925" fstype="ext3" mountpoint="/u001/oradata/004" name="ora-004-u001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_004/lv_u001_004_admin" force_fsck="0" force_unmount="1" fsid="14919" fstype="ext3" mountpoint="/u001/app/oracle/admin/004" name="ora-004-u001-admin-fs" self_fence="0"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <lvm lv_name="lv_a001_005_oradata" name="ora-005-a001-lv" vg_name="/dev/vg_oracle_005"/>
                <lvm lv_name="lv_b001_005_oradata" name="ora-005-b001-lv" vg_name="/dev/vg_oracle_005"/>
                <lvm lv_name="lv_u001_005_oradata" name="ora-005-u001-lv" vg_name="/dev/vg_oracle_005"/>
                <lvm lv_name="lv_u001_005_admin" name="ora-005-u001-admin-lv" vg_name="/dev/vg_oracle_005"/>
                <fs device="/dev/vg_oracle_005/lv_a001_005_oradata" force_fsck="0" force_unmount="1" fsid="50280" fstype="ext3" mountpoint="/a001/oradata/005" name="ora-005-a001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_005/lv_b001_005_oradata" force_fsck="0" force_unmount="1" fsid="57077" fstype="ext3" mountpoint="/b001/oradata/005" name="ora-005-b001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_005/lv_u001_005_oradata" force_fsck="0" force_unmount="1" fsid="31925" fstype="ext3" mountpoint="/u001/oradata/005" name="ora-005-u001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_005/lv_u001_005_admin" force_fsck="0" force_unmount="1" fsid="14919" fstype="ext3" mountpoint="/u001/app/oracle/admin/005" name="ora-005-u001-admin-fs" self_fence="0"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <lvm lv_name="lv_a001_006_oradata" name="ora-006-a001-lv" vg_name="/dev/vg_oracle_006"/>
                <lvm lv_name="lv_b001_006_oradata" name="ora-006-b001-lv" vg_name="/dev/vg_oracle_006"/>
                <lvm lv_name="lv_u001_006_oradata" name="ora-006-u001-lv" vg_name="/dev/vg_oracle_006"/>
                <lvm lv_name="lv_u001_006_admin" name="ora-006-u001-admin-lv" vg_name="/dev/vg_oracle_006"/>
                <fs device="/dev/vg_oracle_006/lv_a001_006_oradata" force_fsck="0" force_unmount="1" fsid="50280" fstype="ext3" mountpoint="/a001/oradata/006" name="ora-006-a001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_006/lv_b001_006_oradata" force_fsck="0" force_unmount="1" fsid="57077" fstype="ext3" mountpoint="/b001/oradata/006" name="ora-006-b001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_006/lv_u001_006_oradata" force_fsck="0" force_unmount="1" fsid="31925" fstype="ext3" mountpoint="/u001/oradata/006" name="ora-006-u001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_006/lv_u001_006_admin" force_fsck="0" force_unmount="1" fsid="14919" fstype="ext3" mountpoint="/u001/app/oracle/admin/006" name="ora-006-u001-admin-fs" self_fence="0"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <lvm lv_name="lv_a001_007_oradata" name="ora-007-a001-lv" vg_name="/dev/vg_oracle_007"/>
                <lvm lv_name="lv_b001_007_oradata" name="ora-007-b001-lv" vg_name="/dev/vg_oracle_007"/>
                <lvm lv_name="lv_u001_007_oradata" name="ora-007-u001-lv" vg_name="/dev/vg_oracle_007"/>
                <lvm lv_name="lv_u001_007_admin" name="ora-007-u001-admin-lv" vg_name="/dev/vg_oracle_007"/>
                <fs device="/dev/vg_oracle_007/lv_a001_007_oradata" force_fsck="0" force_unmount="1" fsid="50280" fstype="ext3" mountpoint="/a001/oradata/007" name="ora-007-a001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_007/lv_b001_007_oradata" force_fsck="0" force_unmount="1" fsid="57077" fstype="ext3" mountpoint="/b001/oradata/007" name="ora-007-b001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_007/lv_u001_007_oradata" force_fsck="0" force_unmount="1" fsid="31925" fstype="ext3" mountpoint="/u001/oradata/007" name="ora-007-u001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_007/lv_u001_007_admin" force_fsck="0" force_unmount="1" fsid="14919" fstype="ext3" mountpoint="/u001/app/oracle/admin/007" name="ora-007-u001-admin-fs" self_fence="0"/>
                <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/>
                <lvm lv_name="lv_a001_008_oradata" name="ora-008-a001-lv" vg_name="/dev/vg_oracle_008"/>
                <lvm lv_name="lv_b001_008_oradata" name="ora-008-b001-lv" vg_name="/dev/vg_oracle_008"/>
                <lvm lv_name="lv_u001_008_oradata" name="ora-008-u001-lv" vg_name="/dev/vg_oracle_008"/>
                <lvm lv_name="lv_u001_008_admin" name="ora-008-u001-admin-lv" vg_name="/dev/vg_oracle_008"/>
                <fs device="/dev/vg_oracle_008/lv_a001_008_oradata" force_fsck="0" force_unmount="1" fsid="50280" fstype="ext3" mountpoint="/a001/oradata/008" name="ora-008-a001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_008/lv_b001_008_oradata" force_fsck="0" force_unmount="1" fsid="57077" fstype="ext3" mountpoint="/b001/oradata/008" name="ora-008-b001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_008/lv_u001_008_oradata" force_fsck="0" force_unmount="1" fsid="31925" fstype="ext3" mountpoint="/u001/oradata/008" name="ora-008-u001-fs" self_fence="0"/>
                <fs device="/dev/vg_oracle_008/lv_u001_008_admin" force_fsck="0" force_unmount="1" fsid="14919" fstype="ext3" mountpoint="/u001/app/oracle/admin/008" name="ora-008-u001-admin-fs" self_fence="0"/>
                </resources>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_009" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_010" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_011" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xx.xxx"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_601" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                        <script ref="oracle_init_script"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_001" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                        <lvm ref="ora-001-a001-lv">
                                <fs ref="ora-001-a001-fs"/>
                        </lvm>
                        <lvm ref="ora-001-b001-lv">
                                <fs ref="ora-001-b001-fs"/>
                        </lvm>
                        <lvm ref="ora-001-u001-lv">
                                <fs ref="ora-001-u001-fs"/>
                        </lvm>
                        <lvm ref="ora-001-u001-admin-lv">
                                <fs ref="ora-001-u001-admin-fs"/>
                        </lvm>
                        <script ref="oracle_init_script"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_002" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                        <lvm ref="ora-002-a001-lv">
                                <fs ref="ora-002-a001-fs"/>
                        </lvm>
                        <lvm ref="ora-002-b001-lv">
                                <fs ref="ora-002-b001-fs"/>
                        </lvm>
                        <lvm ref="ora-002-u001-lv">
                                <fs ref="ora-002-u001-fs"/>
                        </lvm>
                        <lvm ref="ora-002-u001-admin-lv">
                                <fs ref="ora-002-u001-admin-fs"/>
                        </lvm>
                        <script ref="oracle_init_script"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_003" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                        <lvm ref="ora-003-a001-lv">
                                <fs ref="ora-003-a001-fs"/>
                        </lvm>
                        <lvm ref="ora-003-b001-lv">
                                <fs ref="ora-003-b001-fs"/>
                        </lvm>
                        <lvm ref="ora-003-u001-lv">
                                <fs ref="ora-003-u001-fs"/>
                        </lvm>
                        <lvm ref="ora-003-u001-admin-lv">
                                <fs ref="ora-003-u001-admin-fs"/>
                        </lvm>
                        <script ref="oracle_init_script"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_004" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                        <lvm ref="ora-004-a001-lv">
                                <fs ref="ora-004-a001-fs"/>
                        </lvm>
                        <lvm ref="ora-004-b001-lv">
                                <fs ref="ora-004-b001-fs"/>
                        </lvm>
                        <lvm ref="ora-004-u001-lv">
                                <fs ref="ora-004-u001-fs"/>
                        </lvm>
                        <lvm ref="ora-004-u001-admin-lv">
                                <fs ref="ora-004-u001-admin-fs"/>
                        </lvm>
                        <script ref="oracle_init_script"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_005" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                        <lvm ref="ora-005-a001-lv">
                                <fs ref="ora-005-a001-fs"/>
                        </lvm>
                        <lvm ref="ora-005-b001-lv">
                                <fs ref="ora-005-b001-fs"/>
                        </lvm>
                        <lvm ref="ora-005-u001-lv">
                                <fs ref="ora-005-u001-fs"/>
                        </lvm>
                        <lvm ref="ora-005-u001-admin-lv">
                                <fs ref="ora-005-u001-admin-fs"/>
                        </lvm>
                        <script ref="oracle_init_script"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_006" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                        <lvm ref="ora-006-a001-lv">
                                <fs ref="ora-006-a001-fs"/>
                        </lvm>
                        <lvm ref="ora-006-b001-lv">
                                <fs ref="ora-006-b001-fs"/>
                        </lvm>
                        <lvm ref="ora-006-u001-lv">
                                <fs ref="ora-006-u001-fs"/>
                        </lvm>
                        <lvm ref="ora-006-u001-admin-lv">
                                <fs ref="ora-006-u001-admin-fs"/>
                        </lvm>
                        <script ref="oracle_init_script"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_007" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                        <lvm ref="ora-007-a001-lv">
                                <fs ref="ora-007-a001-fs"/>
                        </lvm>
                        <lvm ref="ora-007-b001-lv">
                                <fs ref="ora-007-b001-fs"/>
                        </lvm>
                        <lvm ref="ora-007-u001-lv">
                                <fs ref="ora-007-u001-fs"/>
                        </lvm>
                        <lvm ref="ora-007-u001-admin-lv">
                                <fs ref="ora-007-u001-admin-fs"/>
                        </lvm>
                        <script ref="oracle_init_script"/>
                </service>
                <service autostart="1" domain="rhcstestdomain" exclusive="0" max_restarts="0" name="oracle_008" recovery="restart" restart_expire_time="0">
                        <ip ref="xxx.xxx.xxx.xxx"/>
                        <lvm ref="ora-008-a001-lv">
                                <fs ref="ora-008-a001-fs"/>
                        </lvm>
                        <lvm ref="ora-008-b001-lv">
                                <fs ref="ora-008-b001-fs"/>
                        </lvm>
                        <lvm ref="ora-008-u001-lv">
                                <fs ref="ora-008-u001-fs"/>
                        </lvm>
                        <lvm ref="ora-008-u001-admin-lv">
                                <fs ref="ora-008-u001-admin-fs"/>
                        </lvm>
                        <script ref="oracle_init_script"/>
                </service>
        </rm>
        <totem consensus="4800" join="60" token="100000" token_retransmits_before_loss_const="20" secauth="off"/>
        <quorumd interval="2" min_score="1" tko="15" votes="1" device="/dev/mapper/qdisk" status_file="/var/log/qdisk.status" log_level="7" log_facility="local5">
        </quorumd>
        <logging syslog_facility="local5" syslog_priority="debug" to_syslog="on" debug="on"/>
---------------------------------------------------------------------------------------------


Greg Charles
Mid Range Systems
gcharles ups com

-----Original Message-----
From: linux-cluster-bounces redhat com [mailto:linux-cluster-bounces redhat com] On Behalf Of Juan Ramon Martin Blanco
Sent: Tuesday, October 26, 2010 9:27 AM
To: linux clustering
Subject: Re: [Linux-cluster] Rebooting the Master Node in an RHCS Cluster

On Tue, Oct 26, 2010 at 2:52 PM,  <gcharles ups com> wrote:
> Hello,
>
> Was wondering if anyone else has ever run into this.  We have a
> three-node RHCS cluster:
>
> Three Proliant DL380-G6s, 48G memory
> Dual network, power, QLogic HBAs for redundancy EMC SAN RHEL 5.5
> kernel 2.6.18-194.el5
>
> All three in an RHCS cluster, 12 Oracle database services.  The
> cluster itself runs fine under normal conditions, and all failovers
> function as expected.  There is only one failover domain configured,
> and all three nodes are members of that domain.  Four of the Oracle
> database services contain
> GFS2 file systems; the rest are ext3.
>
> The problem is when we attempt a controlled shutdown of the current
> master node.  We have tested in the following situations:
>
> 1.  Node 1 is the current master and not running any services.  Node 2
> is also not running any services.  Node 3 is running all 12 services.
> We hard-fail node 1 (by logging into the ILO and clicking on "Reset"
> in power
> management) and node 2 immediately takes over the master role and the
> services stay where they are and continue to function.  I believe this
> is the expected behavior.
>
> 2.  Node 1 is the current master and not running any services.  Three
> services are on node 2, and node 3 is running the rest.  Again, we
> hard-fail node 1 as described above and node 2 assumes the master role
> and the services stay where they are and continue to function.
>
> 3.  Repeating the same steps as above; node 1 is the master and not
> running any services, three services on node 2 and the rest on node
> three.  This time we perform a controlled shutdown of node 1 to
> "properly" remove it from the cluster (let's say we're doing a rolling
> patch of the OS on the nodes) with the following steps on the master node:
>  - Unmount any GFS file systems.
>  - service rgmanager stop; service gfs2 stop; service gfs stop
> (clustat shows node1 Online but no rgmanager, as expected)
>  - fence_tool leave    (this removes node 1 from the fence group in
> the hopes that the other nodes don't try to fence it as it is
> rebooting)
>  - service clvmd stop
>  - cman_tool leave remove
>  - service qdiskd stop
>  - shutdown
> Everything appears normal until we execute the 'cman_tool leave
> remove'.  At that point the cluster log on node 2 and node 3 shows
> "Lost contact with quorum device" (we expect that) but also shows "Emergency stop of services"
> for all 12 services.  While access to the quorum device is restored
> almost immediately (node 2 takes over the master role), rgmanager is
> temporarily unavailable on nodes 2 and 3 while the cluster basically
> reconfigures itself, restarting all 12 services.  Eventually all 12
> services properly restart (not necessarily on the original node they
> were on) and when node 1 finishes rebooting, it properly rejoins
> itself to the cluster.  Node 2 retains itself as Master.
>
> If I do the same tests as above and reboot a node that is NOT the
> master, the services remain where they are and the cluster does not
> reconfigure itself or restart any services.
>
> My questions are, Why does the cluster reconfigure itself and restart
> ALL services regardless of what node they are on when I do a
> controlled shutdown of the current Master node?  Do I have to
> hard-reset the Master node in an RHCS cluster so the remaining
> services don't get restarted?  Why does the cluster completely reconfigure itself when the Master node is 'properly'
> removed.
>
Hi, could you please show us your cluster.conf?

Regards,
Juanra
> Thanks for your help, and any suggestions would be appreciated.
>
> Greg Charles
> Mid Range Systems
>
> gcharles ups com
>
>
> --
> Linux-cluster mailing list
> Linux-cluster redhat com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>

--
Linux-cluster mailing list
Linux-cluster redhat com
https://www.redhat.com/mailman/listinfo/linux-cluster


[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]