[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]

Re: [Linux-cluster] error clusvcadm



Hello Delphine

your problem as you know is here
===============================================================
<fs device="LABEL=postfix" mountpoint="/var/spool/
postfix" force_unmount="1" fstype="ext3" name="mgmtha5" options=""/>
===============================================================

I don't know if you are using lvm or partition, but you should look for the device corresponding to that LABEL, if you are using lvm use vgs and lvs to see if your volume are actived

Thanks



2013/5/13 Delphine Ramalingom <delphine ramalingom univ-reunion fr>
Hi,

I used it :

rg_test test /etc/cluster/cluster.conf start service HA_MGMT
Running in test mode.
Starting HA_MGMT...

<err>    startFilesystem: Could not match LABEL=postfix with a real device
Failed to start HA_MGMT

But it gives me the same message.

Regards
Delphine

Le 13/05/13 11:47, emmanuel segura a écrit :
Hello

If you would like see why your service doens't start, you should use "rg_test test /etc/cluster/cluster.conf start service HA_MGMT"




2013/5/13 Delphine Ramalingom <delphine ramalingom univ-reunion fr>
Hi,

This is the cluster.conf :

[root titan0 11:29:14 ~]# cat /etc/cluster/cluster.conf
<?xml version="1.0" ?>
<cluster config_version="7" name="HA_MGMT">
        <fence_daemon clean_start="1" post_fail_delay="0" post_join_delay="60"/>
        <clusternodes>
                <clusternode name="titan0"  nodeid="1" votes="1">
                        <fence>
                                <method name="1">
                                        <device name="titan0fence" option="reboot"/>
                                </method>
                        </fence>
                </clusternode>
                <clusternode name="titan1" nodeid="2" votes="1">
                        <fence>
                                <method name="1">
                                        <device name="titan1fence" option="reboot"/>
                                </method>
                        </fence>
                </clusternode>
        </clusternodes>
        <cman  cluster_id="0" expected_votes="1" two_node="1"/>
        <fencedevices>
                <fencedevice agent="fence_ipmilan" ipaddr="172.17.0.101" login="administrator" name="titan0fence" passwd="administrator"/>
                <fencedevice agent="fence_ipmilan" ipaddr="172.17.0.102" login="administrator" name="titan1fence" passwd="administrator"/>
        </fencedevices>
        <rm>
                <failoverdomains>
                        <failoverdomain name="titan0_heuristic" ordered="0" restricted="1">
                                <failoverdomainnode name="titan0" priority="1"/>
                        </failoverdomain>
                        <failoverdomain name="titan1_heuristic" ordered="0" restricted="1">
                                <failoverdomainnode name="titan1" priority="1"/>
                        </failoverdomain>
                        <failoverdomain name="MgmtNodes" ordered="0" restricted="0">
                                <failoverdomainnode name="titan0" priority="1"/>
                                <failoverdomainnode name="titan1" priority="2"/>
                        </failoverdomain>
            <failoverdomain name="NFSHA" ordered="0" restricted="0">
                <failoverdomainnode name="titan0" priority="2"/>
                <failoverdomainnode name="titan1" priority="1"/>
            </failoverdomain>
                </failoverdomains>
            <service domain="titan0_heuristic" name="ha_titan0_check" autostart="1" checkinterval="10">
                    <script file="/usr/sbin/ha_titan0_check" name="ha_titan0_check"/>
            </service>
            <service domain="titan1_heuristic" name="ha_titan1_check" autostart="1" checkinterval="10">
                    <script file="/usr/sbin/ha_titan1_check" name="ha_titan1_check"/>
            </service>
                <service domain="MgmtNodes" name="HA_MGMT" autostart="0" recovery="relocate">
            <!-- ip addresses lines mgmt -->
                                <ip address="172.17.0.99/16" monitor_link="1"/>
                                <ip address="10.90.0.99/24" monitor_link="1"/>
            <!-- devices lines mgmt -->
                       <fs device="LABEL=postfix" mountpoint="/var/spool/postfix" force_unmount="1" fstype="ext3" name="mgmtha5" options=""/>
                       <fs device="LABEL=bigimage" mountpoint="/var/lib/systemimager" force_unmount="1" fstype="ext3" name="mgmtha4" options=""/>
                       <clusterfs device="LABEL=HA_MGMT:conman" mountpoint="/var/log/conman" force_unmount="0" fstype="gfs2" name="mgmtha3" options=""/>
                       <clusterfs device="LABEL=HA_MGMT:ganglia" mountpoint="/var/lib/ganglia/rrds" force_unmount="0" fstype="gfs2" name="mgmtha2" options=""/>
                       <clusterfs device="LABEL=HA_MGMT:syslog" mountpoint="/var/log/HOSTS" force_unmount="0" fstype="gfs2" name="mgmtha1" options=""/>
                       <clusterfs device="LABEL=HA_MGMT:cdb" mountpoint="/var/lib/pgsql/data" force_unmount="0" fstype="gfs2" name="mgmtha0" options=""/>
                        <script file="/usr/sbin/haservices" name="haservices"/>
                </service>
        <service domain="NFSHA" name="HA_NFS" autostart="0" checkinterval="60">
            <!-- ip addresses lines nfs -->
                                <ip address="10.31.0.99/16" monitor_link="1"/>
                                <ip address="10.90.0.88/24" monitor_link="1"/>
                                <ip address="172.17.0.88/16" monitor_link="1"/>
            <!-- devices lines nfs -->
                       <fs device="LABEL=PROGS" mountpoint="/programs" force_unmount="1" fstype="ext3" name="nfsha4" options=""/>
                       <fs device="LABEL=WRKTMP" mountpoint="/worktmp" force_unmount="1" fstype="ext3" name="nfsha3" options=""/>
                       <fs device="LABEL=LABOS" mountpoint="/labos" force_unmount="1" fstype="xfs" name="nfsha2" options="ikeep"/>
                       <fs device="LABEL=OPTINTEL" mountpoint="/opt/intel" force_unmount="1" fstype="ext3" name="nfsha1" options=""/>
                       <fs device="LABEL=HOMENFS" mountpoint="/home_nfs" force_unmount="1" fstype="ext3" name="nfsha0" options=""/>
            <script file="/etc/init.d/nfs" name="nfs_service"/>
        </service>
        </rm>
    <totem token="21000" />
</cluster>
<!-- !!!!! DON'T REMOVE OR CHANGE ANYTHING IN PARAMETERS SECTION BELOW
node_name=titan0
node_ipmi_ipaddr=172.17.0.101
node_hwmanager_login=administrator
node_hwmanager_passwd=administrator
ipaddr1_for_heuristics=172.17.0.200
node_ha_name=titan1
node_ha_ipmi_ipaddr=172.17.0.102
node_ha_hwmanager_login=administrator
node_ha_hwmanager_passwd=administrator
ipaddr2_for_heuristics=172.17.0.200
mngt_virt_ipaddr_for_heuristics=not used on this type of node
END OF SECTION !!!!! -->


The var/log/messages is too long and have some messages repeated :
May 13 11:30:33 s_sys titan0 snmpd[4584]: Connection from UDP: [10.40.20.30]:39198
May 13 11:30:33 s_sys titan0 snmpd[4584]: Connection from UDP: [10.40.20.30]:39198
May 13 11:30:33 s_sys titan0 snmpd[4584]: Connection from UDP: [10.40.20.30]:39198
May 13 11:30:33 s_sys titan0 snmpd[4584]: Connection from UDP: [10.40.20.30]:39198
May 13 11:30:33 s_sys titan0 snmpd[4584]: Connection from UDP: [10.40.20.30]:39198
May 13 11:30:33 s_sys titan0 snmpd[4584]: Connection from UDP: [10.40.20.30]:39198
May 13 11:30:33 s_sys titan0 snmpd[4584]: Connection from UDP: [10.40.20.30]:39198
May 13 11:30:34 s_sys titan0 snmpd[4584]: Connection from UDP: [10.40.20.30]:53030
May 13 11:30:34 s_sys titan0 snmpd[4584]: Received SNMP packet(s) from UDP: [10.40.20.30]:53030
May 13 11:30:34 s_sys titan0 snmpd[4584]: Connection from UDP: [10.40.20.30]:41083
May 13 11:30:34 s_sys titan0 snmpd[4584]: Received SNMP packet(s) from UDP: [10.40.20.30]:41083

Regards
Delphine



Le 13/05/13 10:37, Rajveer Singh a écrit :
Hi Delphine,
It seems there is some filesystem crash. Please share your /var/log/messages and /etc/cluster/cluster.conf file to help you futher.

Regards,
Rajveer Singh


On Mon, May 13, 2013 at 11:58 AM, Delphine Ramalingom <delphine ramalingom univ-reunion fr> wrote:
Hello,

I have a problem and I need some help.

Our cluster linux have been stopped for maintenance in the room server butr, an error was occured during the stopping procedure :
Local machine disabling service:HA_MGMT...Failure

The cluster was electrically stopped. But since the restart, I don't succed to restart services with command clussvcadm.
I have this message :

clusvcadm -e HA_MGMT
Local machine trying to enable service:HA_MGMT...Aborted; service failed
and
<err>    startFilesystem: Could not match LABEL=postfix with a real device

Do you have a solution for me ?

Thanks a lot in advance.

Regards
Delphine

--
Linux-cluster mailing list
Linux-cluster redhat com
https://www.redhat.com/mailman/listinfo/linux-cluster





--
Linux-cluster mailing list
Linux-cluster redhat com
https://www.redhat.com/mailman/listinfo/linux-cluster



--
esta es mi vida e me la vivo hasta que dios quiera




--
Linux-cluster mailing list
Linux-cluster redhat com
https://www.redhat.com/mailman/listinfo/linux-cluster



--
esta es mi vida e me la vivo hasta que dios quiera

[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]