[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]

Re: [Linux-cluster] fencing problem in 2 node cluster using apc fence device



Hi Lon, thanks for the reply. My cluster.conf is as follows.
coral1 and coral2 are my 2 nodes.


<?xml version="1.0"?>
<cluster config_version="264" name="MF_Cluster">
	<fence_daemon clean_start="1" post_fail_delay="0" post_join_delay="3"/>
	<clusternodes>
		<clusternode name="coral1" votes="1">
			<fence>
				<method name="1">
					<device name="pdu1" option="off" port="13" switch="1"/>
					<device name="pdu2" option="off" port="13" switch="2"/>
					<device name="pdu1" option="on" port="13" switch="1"/>
					<device name="pdu2" option="on" port="13" switch="2"/>
				</method>
			</fence>
		</clusternode>
		<clusternode name="coral2" votes="1">
			<fence>
				<method name="1">
					<device name="pdu1" option="off" port="20" switch="1"/>
					<device name="pdu2" option="off" port="20" switch="2"/>
					<device name="pdu1" option="on" port="20" switch="1"/>
					<device name="pdu2" option="on" port="20" switch="2"/>
				</method>
			</fence>
		</clusternode>
	</clusternodes>
	<fencedevices>
		<fencedevice agent="fence_apc" ipaddr="10.10.50.100" login="apc" name="pdu1" passwd="apc"/>
		<fencedevice agent="fence_apc" ipaddr="10.10.50.101" login="apc" name="pdu2" passwd="apc"/>
	</fencedevices>
	<rm>
		<failoverdomains>
			<failoverdomain name="MF_Failover" ordered="0" restricted="1">
				<failoverdomainnode name="coral2" priority="1"/>
				<failoverdomainnode name="coral1" priority="1"/>
			</failoverdomain>
		</failoverdomains>
		<resources>
			<fs device="/dev/sda1" force_fsck="0" force_unmount="1" fstype="ext3" mountpoint="/MF/MF_v1.1/shared" name="testmount" options="" self_fence="0"/>
			<script file="/etc/rc.d/init.d/msgfwd" name="Message Forwarder"/>
			<ip address="10.10.50.22" monitor_link="1"/>
			<script file="/etc/rc.d/init.d/namesvc" name="Name Service"/>
		</resources>
		<service autostart="1" domain="MF_Failover" name="msgfwd" recovery="relocate">
			<fs ref="testmount"/>
			<script ref="Message Forwarder"/>
			<ip ref="10.10.50.22"/>
		</service>
	</rm>
	<cman expected_votes="1" two_node="1"/>
</cluster>




--- Lon Hohberger <lhh redhat com> wrote:

> On Wed, 2006-08-16 at 22:37 +0800, Gary Lua wrote:
> > Hi,
> > 
> > I'm currently configuring fencing devices for my 2 nodes on a
> RHEL4 
> > cluster. The problem is quite long, so please bear with me.
> > 
> > I have 2 nodes (let's call them stone1 and stone2) and 2 APC
> fencing 
> > devices (pdu1 and pdu2, both apc 7952 devices). Both stone1 and
> stone2 
> > has dual power supplies. Stone1's power supplies are connected to
> outlet 
> > 13 of pdu1 and pdu2. Stone2's power supplies are connected to
> outlet 20 
> > of both the pdus. My question is: during the fencing configuration
> for 
> > each node, i need to specify which fence device to add to the
> fence 
> > level of each node. Is it correct to specify for stone1 as follows
> : 
> > pdu1 -> port=13, switch=1, pdu2-> port=13, switch=2? The same
> applies to 
> > stone 2 : pdu1-> port=20, switch=1, pdu2-> port=20, switch=2?
> > 
> > After configuring as mentioned above, with both nodes on the
> cluster 
> > running and my application running on stone1, i pull out the
> ethernet 
> > cables for stone1 to simulate that the server is down. By right,
> my 
> > application should fail over to stone2 and fencing should occur to
> 
> > stone1 (ie, stone1 should be rebooted/shutdown). However, what
> happened 
> > is that my application is started on stone2, and stone1 is not
> fenced. 
> > In fact, when i reconnect by cables, my application is still
> running on 
> > stone1! Seems that there are 2 instances of my application
> running, each 
> > on stone1 and stone2.
> 
> Post the cluster configuration.
> 
> -- Lon
> 
> 
> --
> Linux-cluster mailing list
> Linux-cluster redhat com
> https://www.redhat.com/mailman/listinfo/linux-cluster
> 


[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]