[Linux-cluster] error clusvcadm
Delphine Ramalingom
delphine.ramalingom at univ-reunion.fr
Mon May 13 08:35:41 UTC 2013
Hi,
I used it :
rg_test test /etc/cluster/cluster.conf start service HA_MGMT
Running in test mode.
Starting HA_MGMT...
<err> startFilesystem: Could not match LABEL=postfix with a real device
Failed to start HA_MGMT
But it gives me the same message.
Regards
Delphine
Le 13/05/13 11:47, emmanuel segura a écrit :
> Hello
>
> If you would like see why your service doens't start, you should use
> "rg_test test /etc/cluster/cluster.conf start service HA_MGMT"
>
>
>
>
> 2013/5/13 Delphine Ramalingom <delphine.ramalingom at univ-reunion.fr
> <mailto:delphine.ramalingom at univ-reunion.fr>>
>
> Hi,
>
> This is the cluster.conf :
>
> [root at titan0 11:29:14 ~]# cat /etc/cluster/cluster.conf
> <?xml version="1.0" ?>
> <cluster config_version="7" name="HA_MGMT">
> <fence_daemon clean_start="1" post_fail_delay="0"
> post_join_delay="60"/>
> <clusternodes>
> <clusternode name="titan0" nodeid="1" votes="1">
> <fence>
> <method name="1">
> <device name="titan0fence"
> option="reboot"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="titan1" nodeid="2" votes="1">
> <fence>
> <method name="1">
> <device name="titan1fence"
> option="reboot"/>
> </method>
> </fence>
> </clusternode>
> </clusternodes>
> <cman cluster_id="0" expected_votes="1" two_node="1"/>
> <fencedevices>
> <fencedevice agent="fence_ipmilan"
> ipaddr="172.17.0.101" login="administrator" name="titan0fence"
> passwd="administrator"/>
> <fencedevice agent="fence_ipmilan"
> ipaddr="172.17.0.102" login="administrator" name="titan1fence"
> passwd="administrator"/>
> </fencedevices>
> <rm>
> <failoverdomains>
> <failoverdomain name="titan0_heuristic"
> ordered="0" restricted="1">
> <failoverdomainnode name="titan0"
> priority="1"/>
> </failoverdomain>
> <failoverdomain name="titan1_heuristic"
> ordered="0" restricted="1">
> <failoverdomainnode name="titan1"
> priority="1"/>
> </failoverdomain>
> <failoverdomain name="MgmtNodes"
> ordered="0" restricted="0">
> <failoverdomainnode name="titan0"
> priority="1"/>
> <failoverdomainnode name="titan1"
> priority="2"/>
> </failoverdomain>
> <failoverdomain name="NFSHA" ordered="0" restricted="0">
> <failoverdomainnode name="titan0" priority="2"/>
> <failoverdomainnode name="titan1" priority="1"/>
> </failoverdomain>
> </failoverdomains>
> <service domain="titan0_heuristic"
> name="ha_titan0_check" autostart="1" checkinterval="10">
> <script file="/usr/sbin/ha_titan0_check"
> name="ha_titan0_check"/>
> </service>
> <service domain="titan1_heuristic"
> name="ha_titan1_check" autostart="1" checkinterval="10">
> <script file="/usr/sbin/ha_titan1_check"
> name="ha_titan1_check"/>
> </service>
> <service domain="MgmtNodes" name="HA_MGMT"
> autostart="0" recovery="relocate">
> <!-- ip addresses lines mgmt -->
> <ip address="172.17.0.99/16
> <http://172.17.0.99/16>" monitor_link="1"/>
> <ip address="10.90.0.99/24
> <http://10.90.0.99/24>" monitor_link="1"/>
> <!-- devices lines mgmt -->
> <fs device="LABEL=postfix"
> mountpoint="/var/spool/postfix" force_unmount="1" fstype="ext3"
> name="mgmtha5" options=""/>
> <fs device="LABEL=bigimage"
> mountpoint="/var/lib/systemimager" force_unmount="1" fstype="ext3"
> name="mgmtha4" options=""/>
> <clusterfs device="LABEL=HA_MGMT:conman"
> mountpoint="/var/log/conman" force_unmount="0" fstype="gfs2"
> name="mgmtha3" options=""/>
> <clusterfs device="LABEL=HA_MGMT:ganglia"
> mountpoint="/var/lib/ganglia/rrds" force_unmount="0" fstype="gfs2"
> name="mgmtha2" options=""/>
> <clusterfs device="LABEL=HA_MGMT:syslog"
> mountpoint="/var/log/HOSTS" force_unmount="0" fstype="gfs2"
> name="mgmtha1" options=""/>
> <clusterfs device="LABEL=HA_MGMT:cdb"
> mountpoint="/var/lib/pgsql/data" force_unmount="0" fstype="gfs2"
> name="mgmtha0" options=""/>
> <script file="/usr/sbin/haservices"
> name="haservices"/>
> </service>
> <service domain="NFSHA" name="HA_NFS" autostart="0"
> checkinterval="60">
> <!-- ip addresses lines nfs -->
> <ip address="10.31.0.99/16
> <http://10.31.0.99/16>" monitor_link="1"/>
> <ip address="10.90.0.88/24
> <http://10.90.0.88/24>" monitor_link="1"/>
> <ip address="172.17.0.88/16
> <http://172.17.0.88/16>" monitor_link="1"/>
> <!-- devices lines nfs -->
> <fs device="LABEL=PROGS"
> mountpoint="/programs" force_unmount="1" fstype="ext3"
> name="nfsha4" options=""/>
> <fs device="LABEL=WRKTMP"
> mountpoint="/worktmp" force_unmount="1" fstype="ext3"
> name="nfsha3" options=""/>
> <fs device="LABEL=LABOS"
> mountpoint="/labos" force_unmount="1" fstype="xfs" name="nfsha2"
> options="ikeep"/>
> <fs device="LABEL=OPTINTEL"
> mountpoint="/opt/intel" force_unmount="1" fstype="ext3"
> name="nfsha1" options=""/>
> <fs device="LABEL=HOMENFS"
> mountpoint="/home_nfs" force_unmount="1" fstype="ext3"
> name="nfsha0" options=""/>
> <script file="/etc/init.d/nfs" name="nfs_service"/>
> </service>
> </rm>
> <totem token="21000" />
> </cluster>
> <!-- !!!!! DON'T REMOVE OR CHANGE ANYTHING IN PARAMETERS SECTION
> BELOW
> node_name=titan0
> node_ipmi_ipaddr=172.17.0.101
> node_hwmanager_login=administrator
> node_hwmanager_passwd=administrator
> ipaddr1_for_heuristics=172.17.0.200
> node_ha_name=titan1
> node_ha_ipmi_ipaddr=172.17.0.102
> node_ha_hwmanager_login=administrator
> node_ha_hwmanager_passwd=administrator
> ipaddr2_for_heuristics=172.17.0.200
> mngt_virt_ipaddr_for_heuristics=not used on this type of node
> END OF SECTION !!!!! -->
>
>
> The var/log/messages is too long and have some messages repeated :
> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
> [10.40.20.30]:39198
> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
> [10.40.20.30]:39198
> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
> [10.40.20.30]:39198
> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
> [10.40.20.30]:39198
> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
> [10.40.20.30]:39198
> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
> [10.40.20.30]:39198
> May 13 11:30:33 s_sys at titan0 snmpd[4584]: Connection from UDP:
> [10.40.20.30]:39198
> May 13 11:30:34 s_sys at titan0 snmpd[4584]: Connection from UDP:
> [10.40.20.30]:53030
> May 13 11:30:34 s_sys at titan0 snmpd[4584]: Received SNMP packet(s)
> from UDP: [10.40.20.30]:53030
> May 13 11:30:34 s_sys at titan0 snmpd[4584]: Connection from UDP:
> [10.40.20.30]:41083
> May 13 11:30:34 s_sys at titan0 snmpd[4584]: Received SNMP packet(s)
> from UDP: [10.40.20.30]:41083
>
> Regards
> Delphine
>
>
>
> Le 13/05/13 10:37, Rajveer Singh a écrit :
>> Hi Delphine,
>> It seems there is some filesystem crash. Please share your
>> /var/log/messages and /etc/cluster/cluster.conf file to help you
>> futher.
>>
>> Regards,
>> Rajveer Singh
>>
>>
>> On Mon, May 13, 2013 at 11:58 AM, Delphine Ramalingom
>> <delphine.ramalingom at univ-reunion.fr
>> <mailto:delphine.ramalingom at univ-reunion.fr>> wrote:
>>
>> Hello,
>>
>> I have a problem and I need some help.
>>
>> Our cluster linux have been stopped for maintenance in the
>> room server butr, an error was occured during the stopping
>> procedure :
>> Local machine disabling service:HA_MGMT...Failure
>>
>> The cluster was electrically stopped. But since the restart,
>> I don't succed to restart services with command clussvcadm.
>> I have this message :
>>
>> clusvcadm -e HA_MGMT
>> Local machine trying to enable service:HA_MGMT...Aborted;
>> service failed
>> and
>> <err> startFilesystem: Could not match LABEL=postfix with
>> a real device
>>
>> Do you have a solution for me ?
>>
>> Thanks a lot in advance.
>>
>> Regards
>> Delphine
>>
>> --
>> Linux-cluster mailing list
>> Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>
>>
>>
>>
>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
> https://www.redhat.com/mailman/listinfo/linux-cluster
>
>
>
>
> --
> esta es mi vida e me la vivo hasta que dios quiera
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20130513/80b342a9/attachment.htm>
More information about the Linux-cluster
mailing list