[Linux-cluster] error clusvcadm

Delphine Ramalingom delphine.ramalingom at univ-reunion.fr
Mon May 13 08:35:41 UTC 2013


Hi,

I used it :
rg_test test /etc/cluster/cluster.conf start service HA_MGMT
Running in test mode.
Starting HA_MGMT...
<err>    startFilesystem: Could not match LABEL=postfix with a real device
Failed to start HA_MGMT

But it gives me the same message.

Regards
Delphine

Le 13/05/13 11:47, emmanuel segura a écrit :
> Hello
>
> If you would like see why your service doens't start, you should use 
> "rg_test test /etc/cluster/cluster.conf start service HA_MGMT"
>
>
>
>
> 2013/5/13 Delphine Ramalingom <delphine.ramalingom at univ-reunion.fr 
> <mailto:delphine.ramalingom at univ-reunion.fr>>
>
>     Hi,
>
>     This is the cluster.conf :
>
>     [root at titan0 11:29:14 ~]# cat /etc/cluster/cluster.conf
>     <?xml version="1.0" ?>
>     <cluster config_version="7" name="HA_MGMT">
>             <fence_daemon clean_start="1" post_fail_delay="0"
>     post_join_delay="60"/>
>             <clusternodes>
>                     <clusternode name="titan0" nodeid="1" votes="1">
>                             <fence>
>                                     <method name="1">
>                                             <device name="titan0fence"
>     option="reboot"/>
>                                     </method>
>                             </fence>
>                     </clusternode>
>                     <clusternode name="titan1" nodeid="2" votes="1">
>                             <fence>
>                                     <method name="1">
>                                             <device name="titan1fence"
>     option="reboot"/>
>                                     </method>
>                             </fence>
>                     </clusternode>
>             </clusternodes>
>             <cman  cluster_id="0" expected_votes="1" two_node="1"/>
>             <fencedevices>
>                     <fencedevice agent="fence_ipmilan"
>     ipaddr="172.17.0.101" login="administrator" name="titan0fence"
>     passwd="administrator"/>
>                     <fencedevice agent="fence_ipmilan"
>     ipaddr="172.17.0.102" login="administrator" name="titan1fence"
>     passwd="administrator"/>
>             </fencedevices>
>             <rm>
>                     <failoverdomains>
>                             <failoverdomain name="titan0_heuristic"
>     ordered="0" restricted="1">
>                                     <failoverdomainnode name="titan0"
>     priority="1"/>
>                             </failoverdomain>
>                             <failoverdomain name="titan1_heuristic"
>     ordered="0" restricted="1">
>                                     <failoverdomainnode name="titan1"
>     priority="1"/>
>                             </failoverdomain>
>                             <failoverdomain name="MgmtNodes"
>     ordered="0" restricted="0">
>                                     <failoverdomainnode name="titan0"
>     priority="1"/>
>                                     <failoverdomainnode name="titan1"
>     priority="2"/>
>                             </failoverdomain>
>                 <failoverdomain name="NFSHA" ordered="0" restricted="0">
>                     <failoverdomainnode name="titan0" priority="2"/>
>                     <failoverdomainnode name="titan1" priority="1"/>
>                 </failoverdomain>
>                     </failoverdomains>
>                 <service domain="titan0_heuristic"
>     name="ha_titan0_check" autostart="1" checkinterval="10">
>                         <script file="/usr/sbin/ha_titan0_check"
>     name="ha_titan0_check"/>
>                 </service>
>                 <service domain="titan1_heuristic"
>     name="ha_titan1_check" autostart="1" checkinterval="10">
>                         <script file="/usr/sbin/ha_titan1_check"
>     name="ha_titan1_check"/>
>                 </service>
>                     <service domain="MgmtNodes" name="HA_MGMT"
>     autostart="0" recovery="relocate">
>                 <!-- ip addresses lines mgmt -->
>                                     <ip address="172.17.0.99/16
>     <http://172.17.0.99/16>" monitor_link="1"/>
>                                     <ip address="10.90.0.99/24
>     <http://10.90.0.99/24>" monitor_link="1"/>
>                 <!-- devices lines mgmt -->
>                            <fs device="LABEL=postfix"
>     mountpoint="/var/spool/postfix" force_unmount="1" fstype="ext3"
>     name="mgmtha5" options=""/>
>                            <fs device="LABEL=bigimage"
>     mountpoint="/var/lib/systemimager" force_unmount="1" fstype="ext3"
>     name="mgmtha4" options=""/>
>                            <clusterfs device="LABEL=HA_MGMT:conman"
>     mountpoint="/var/log/conman" force_unmount="0" fstype="gfs2"
>     name="mgmtha3" options=""/>
>                            <clusterfs device="LABEL=HA_MGMT:ganglia"
>     mountpoint="/var/lib/ganglia/rrds" force_unmount="0" fstype="gfs2"
>     name="mgmtha2" options=""/>
>                            <clusterfs device="LABEL=HA_MGMT:syslog"
>     mountpoint="/var/log/HOSTS" force_unmount="0" fstype="gfs2"
>     name="mgmtha1" options=""/>
>                            <clusterfs device="LABEL=HA_MGMT:cdb"
>     mountpoint="/var/lib/pgsql/data" force_unmount="0" fstype="gfs2"
>     name="mgmtha0" options=""/>
>                             <script file="/usr/sbin/haservices"
>     name="haservices"/>
>                     </service>
>             <service domain="NFSHA" name="HA_NFS" autostart="0"
>     checkinterval="60">
>                 <!-- ip addresses lines nfs -->
>                                     <ip address="10.31.0.99/16
>     <http://10.31.0.99/16>" monitor_link="1"/>
>                                     <ip address="10.90.0.88/24
>     <http://10.90.0.88/24>" monitor_link="1"/>
>                                     <ip address="172.17.0.88/16
>     <http://172.17.0.88/16>" monitor_link="1"/>
>                 <!-- devices lines nfs -->
>                            <fs device="LABEL=PROGS"
>     mountpoint="/programs" force_unmount="1" fstype="ext3"
>     name="nfsha4" options=""/>
>                            <fs device="LABEL=WRKTMP"
>     mountpoint="/worktmp" force_unmount="1" fstype="ext3"
>     name="nfsha3" options=""/>
>                            <fs device="LABEL=LABOS"
>     mountpoint="/labos" force_unmount="1" fstype="xfs" name="nfsha2"
>     options="ikeep"/>
>                            <fs device="LABEL=OPTINTEL"
>     mountpoint="/opt/intel" force_unmount="1" fstype="ext3"
>     name="nfsha1" options=""/>
>                            <fs device="LABEL=HOMENFS"
>     mountpoint="/home_nfs" force_unmount="1" fstype="ext3"
>     name="nfsha0" options=""/>
>                 <script file="/etc/init.d/nfs" name="nfs_service"/>
>             </service>
>             </rm>
>         <totem token="21000" />
>     </cluster>
>     <!-- !!!!! DON'T REMOVE OR CHANGE ANYTHING IN PARAMETERS SECTION
>     BELOW
>     node_name=titan0
>     node_ipmi_ipaddr=172.17.0.101
>     node_hwmanager_login=administrator
>     node_hwmanager_passwd=administrator
>     ipaddr1_for_heuristics=172.17.0.200
>     node_ha_name=titan1
>     node_ha_ipmi_ipaddr=172.17.0.102
>     node_ha_hwmanager_login=administrator
>     node_ha_hwmanager_passwd=administrator
>     ipaddr2_for_heuristics=172.17.0.200
>     mngt_virt_ipaddr_for_heuristics=not used on this type of node
>     END OF SECTION !!!!! -->
>
>
>     The var/log/messages is too long and have some messages repeated :
>     May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>     [10.40.20.30]:39198
>     May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>     [10.40.20.30]:39198
>     May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>     [10.40.20.30]:39198
>     May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>     [10.40.20.30]:39198
>     May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>     [10.40.20.30]:39198
>     May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>     [10.40.20.30]:39198
>     May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
>     [10.40.20.30]:39198
>     May 13 11:30:34 s_sys at titan0 snmpd[4584]: Connection from UDP:
>     [10.40.20.30]:53030
>     May 13 11:30:34 s_sys at titan0 snmpd[4584]: Received SNMP packet(s)
>     from UDP: [10.40.20.30]:53030
>     May 13 11:30:34 s_sys at titan0 snmpd[4584]: Connection from UDP:
>     [10.40.20.30]:41083
>     May 13 11:30:34 s_sys at titan0 snmpd[4584]: Received SNMP packet(s)
>     from UDP: [10.40.20.30]:41083
>
>     Regards
>     Delphine
>
>
>
>     Le 13/05/13 10:37, Rajveer Singh a écrit :
>>     Hi Delphine,
>>     It seems there is some filesystem crash. Please share your
>>     /var/log/messages and /etc/cluster/cluster.conf file to help you
>>     futher.
>>
>>     Regards,
>>     Rajveer Singh
>>
>>
>>     On Mon, May 13, 2013 at 11:58 AM, Delphine Ramalingom
>>     <delphine.ramalingom at univ-reunion.fr
>>     <mailto:delphine.ramalingom at univ-reunion.fr>> wrote:
>>
>>         Hello,
>>
>>         I have a problem and I need some help.
>>
>>         Our cluster linux have been stopped for maintenance in the
>>         room server butr, an error was occured during the stopping
>>         procedure :
>>         Local machine disabling service:HA_MGMT...Failure
>>
>>         The cluster was electrically stopped. But since the restart,
>>         I don't succed to restart services with command clussvcadm.
>>         I have this message :
>>
>>         clusvcadm -e HA_MGMT
>>         Local machine trying to enable service:HA_MGMT...Aborted;
>>         service failed
>>         and
>>         <err>    startFilesystem: Could not match LABEL=postfix with
>>         a real device
>>
>>         Do you have a solution for me ?
>>
>>         Thanks a lot in advance.
>>
>>         Regards
>>         Delphine
>>
>>         -- 
>>         Linux-cluster mailing list
>>         Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
>>         https://www.redhat.com/mailman/listinfo/linux-cluster
>>
>>
>>
>>
>
>
>     --
>     Linux-cluster mailing list
>     Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
>     https://www.redhat.com/mailman/listinfo/linux-cluster
>
>
>
>
> -- 
> esta es mi vida e me la vivo hasta que dios quiera
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20130513/80b342a9/attachment.htm>


More information about the Linux-cluster mailing list