How can I avoid fencing loops with 2 node clusters and Red Hat High Availability clusters?

Solution Verified - Updated -

Issue

  • Why should I have my fencing devices on the same network as my cluster communication with Red Hat High Availability clusters?
  • How can I prevent fence loops with my cluster?
  • If one node in a 2-node cluster is up and running, and the other node boots up while there is a network issue, that booting node fences the active node
  • In a 2-node cluster configured with fence_scsi, if a node gets fenced due to a network issue and I then reboot it, the "active" node reports SCSI reservation conflicts and path failures as the rebooted node starts its services.
  • Server hangs during boot time with error message "Joining fence domain".

Environment

  • Red Hat Enterprise Linux (RHEL) 5 and later with the High Availability Add-On
  • Red Hat High Availability Cluster with 2 nodes (this issue does not apply to 3 or more node clusters).
  • cman-based clusters: No quorum disk is configured. Cluster must have only 2 votes with two_node="1" set.
  • corosync+votequorum-based clusters: wait_for_all is not set (it is set by default if two_node is enabled).
  • Cluster fencing devices are IP-based and are accessed over the network.
    • Fencing devices are reached over a different network to the network that the cluster communicates over. This effectively means both nodes can reach the fencing devices when the cluster interconnect is unavailable.

Subscriber exclusive content

A Red Hat subscription provides unlimited access to our knowledgebase, tools, and much more.

Current Customers and Partners

Log in for full access

Log In

New to Red Hat?

Learn more about Red Hat subscriptions

Using a Red Hat product through a public cloud?

How to access this content