pcs resource cleanup doesn't clean up all fail counts when run from a Pacemaker remote node
Issue
-
I ran
pcs resource cleanup
from a Pacemaker remote node. It cleared its own fail counts, and it cleared the failed resource actions for all nodes, but it didn't clear the fail counts for any other nodes.[root@node1 pacemaker]# pcs status --full ... Migration Summary: * Node: node1 (1): * dummy: migration-threshold=1000000 fail-count=2 last-failure='Wed Dec 16 22:57:38 2020' * Node: node2 (2): * dummy: migration-threshold=1000000 fail-count=2 last-failure='Wed Dec 16 22:57:38 2020' * Node: node3-rem: * dummy: migration-threshold=1000000 fail-count=1 last-failure='Wed Dec 16 22:57:38 2020' Failed Resource Actions: * dummy_asyncmon_0 on node1 'error' (1): call=16, status='complete', exitreason='Simulated failure', last-rc-change='2020-12-16 22:57:38 -08:00', queued=0ms, exec=0ms * dummy_asyncmon_0 on node2 'error' (1): call=18, status='complete', exitreason='Simulated failure', last-rc-change='2020-12-16 22:57:38 -08:00', queued=0ms, exec=0ms * dummy_asyncmon_0 on node3-rem 'error' (1): call=70, status='complete', exitreason='Simulated failure', last-rc-change='2020-12-16 22:57:38 -08:00', queued=0ms, exec=0ms [root@node3-rem pacemaker]# pcs resource cleanup Cleaned up all resources on all nodes Waiting for 3 replies from the controller ... got reply ... got reply ... got reply (done) [root@node1 pacemaker]# pcs status --full ... Migration Summary: * Node: node1 (1): * dummy: migration-threshold=1000000 fail-count=2 last-failure='Wed Dec 16 22:57:38 2020' * Node: node2 (2): * dummy: migration-threshold=1000000 fail-count=2 last-failure='Wed Dec 16 22:57:38 2020'
-
When run from a Pacemaker remote node,
pcs resource cleanup
behaves likepcs resource cleanup --node <local_node_name>
.
Environment
- Red Hat Enterprise Linux 8 (with the High Availability Add-on)
Subscriber exclusive content
A Red Hat subscription provides unlimited access to our knowledgebase, tools, and much more.