Red Hat Pacemaker cluster resource fail to startup

Latest response

Hi There,

We have encountered this error when starting the pcs resource. ClusterSrcIP (ocf::heartbeat:IPsrcaddr): FAILED node01 (blocked)

Please help if similar problem are reported in buzilla?

[root@ cluster]# pcs status
Cluster name:
Stack: corosync
Current DC: hostname (version 1.1.16-12.el7_4.2-94ff4df) - partition with quorum
Last updated: Fri Oct 13 14:59:03 2017
Last change: Thu Oct 12 21:09:01 2017 by hacluster via crmd on node2

2 nodes configured
11 resources configured

Online: [ node01 node02 ]

Full list of resources:

Master/Slave Set: ClusterDataClone [ClusterData]
Masters: [ node01 ]
Slaves: [ node02 ]
ClusterFS (ocf::heartbeat:Filesystem): Started node01
ClusterIP (ocf::heartbeat:IPaddr2): Started node01
ClusterSrcIP (ocf::heartbeat:IPsrcaddr): FAILED node01 (blocked)
WebServer (ocf::heartbeat:apache): Stopped
ClusterStatus (ocf::pacemaker:ClusterMon): Stopped
SyslogUdp1 (ocf::arcsight:SmartConnector): Stopped
Windowsfg1 (ocf::arcsight:SmartConnector): Stopped
VmwareWS1 (ocf::arcsight:SmartConnector): Stopped
arc-syslogdaemon-1 (ocf::arcsight:SmartConnector): Stopped

Failed Actions:
* ClusterSrcIP_stop_0 on node01 'unknown error' (1): call=68, status=complete, exitreason='command 'ip route replace dev eno16780032' failed',
last-rc-change='Thu Oct 12 21:42:00 2017', queued=0ms, exec=37ms
* WebServer_start_0 on node01 'unknown error' (1): call=54, status=complete, exitreason='Failed to access httpd status page.',
last-rc-change='Thu Oct 12 21:41:57 2017', queued=0ms, exec=3404ms
* Windowsfg1_monitor_10000 on node01 'not running' (7): call=59, status=complete, exitreason='none',
last-rc-change='Thu Oct 12 21:41:58 2017', queued=0ms, exec=12ms

Daemon Status:
corosync: active/enabled
pacemaker: active/enabled
pcsd: inactive/disabled
[root@node01 cluster]#



Hi All,

i have clean up vmfence resource, but still shown resource done state. Please help if you have any suggestion/ recommendation. Please be informed cluster running fine, no issue at DB level. thank you in advance

pcs resource cleanup vmfence Waiting for 2 replies from the CRMd.. OK Cleaning up vmfence on a0410pcrmdbs19, removing fail-count-vmfence Cleaning up vmfence on a0410pcrmdbs20, removing fail-count-vmfence

Online: [ a0410pcrmdbs19 a0410pcrmdbs20 ]

Full list of resources:

vmfence (stonith:fence_vmware_soap): Stopped Resource Group: mysqldb_group mysqldb_vip (ocf::heartbeat:IPaddr2): Started a0410pcrmdbs19 halvm (ocf::heartbeat:LVM): Started a0410pcrmdbs19 ext4fs_logs (ocf::heartbeat:Filesystem): Started a0410pcrmdbs19 ext4fs_data01 (ocf::heartbeat:Filesystem): Started a0410pcrmdbs19 smsapp_mysql (ocf::heartbeat:mysql): Started a0410pcrmdbs19

Failed Actions: * vmfence_start_0 on a0410pcrmdbs20 'unknown error' (1): call=175, status=Error, exitreason='none', last-rc-change='Thu Apr 12 09:54:39 2018', queued=0ms, exec=13957ms * vmfence_start_0 on a0410pcrmdbs19 'unknown error' (1): call=751, status=Timed Out, exitreason='none', last-rc-change='Thu Apr 12 09:54:18 2018', queued=0ms, exec=20361ms

PCSD Status: a0410pcrmdbs19: Online a0410pcrmdbs20: Online a0410pcrmdbs19 ( Online a0410pcrmdbs20 ( Online

Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled