- Previously, the
statusaction in the netfs interface failed to write any output to the
/var/log/cluster/rgmanager.logfile. Consequently, it was not possible to verify if the status check of an NFS mount was successful. The bug has been fixed, and results of the status check are now properly stored in the log file.
- For HA-LVM to work properly, the
/boot/initrd.imgfile, which is used during the boot process, must be synchronized with the
/etc/lvm/lvm.conffile. Previously, the HA-LVM startup failed when
lvm.confwas changed without updating
initrd.img. With this update, this behavior has been modified. A warning message is now displayed, but the startup is no longer terminated in the described case.
- Prior to this update, occasional service failures occurred when starting the
clvmdvariant of the HA-LVM service on multiple nodes in a cluster at the same time. The start of an HA-LVM resource coincided with another node initializing that same HA-LVM resource. With this update, a patch has been introduced to synchronize the initialization of both resources. As a result, services no longer fail due to the simultaneous initialization.
- When the oracledb.sh script was called with the
statusargument, it restarted the database after checking its status without any notification to the rgmanager application. This bug has been fixed, and the unwanted restart no longer occurs.
- Previously, the /usr/sbin/tomcat-6.sh script parsed configuration files and set shell variables before starting the Apache Tomcat 6 servlet container. Consequently, the default configuration was ignored. This bug has been fixed and the aforementioned problem no longer occurs.
- Previously, an output of HA-LVM commands that contained more than one word, was not correctly parsed. Consequently, starting an HA-LVM service with the
rg_testcommand occasionally failed with the following message:
too many argumentsWith this update, the underlying source code has been modified to add quotation marks around variables that expand to more than one word. As a result, the aforementioned startup errors no longer occur.
- If the contents of the
/proc/mountsfile changed during a status check operation of the file system resource agent, the status check could incorrectly detect a missing mount and mark the service as failed. This bug has been fixed and rgmanager's file system resource agent no longer reports false failures in the described scenario.
- Previously, rgmanager did not recognize CIFS (Common Internet File System) mounts in case their corresponding entries in the device field of the
/proc/mountsfile contained trailing slashes. With this update, a patch has been introduced to remove trailing slashes from device names when reading the contents of
/proc/mounts. As a result, CIFS mounts are now recognized properly.
- Prior to this update, when running a file system depending on an LVM resource in a service, and that LVM resource failed to start, the subsequent attempt to unregister the file system resource failed. This bug has been fixed, and a file system resource can now be successfully unregistered after a failed mount operation.
- Previously, when using the HA-LVM resource agent in the Pacemaker cluster environment, several errors and failed actions occurred. With this update, several scripts have been added to prevent these errors. These scripts repair the treatment of whitespace within HA-LVM and the processing of non-zero codes in rgmanager. In addition, the member_util utility has been updated to use Corosync and Pacemaker when rgmanager is not present on the system.
- Previously, when a node lost access to the storage device, HA-LVM was unable to deactivate the volume group for the services running in that node. The underlying source code has been modified to allow services to migrate to other machines that still have access to storage devices, thus preventing this bug.
- Previously, SAP instances started by the SAPInstance cluster resource agent inherited limits on system resources for the root user. Higher limits were needed on the maximum number of open files (
ulimit -n), the maximum stack size (
ulimit -s), and the maximum size of data segments (
ulimit -d). With this update, the SAPInstance agent has been modified to accept limits specified in the
/usr/sap/servicesfile. As a result, system resources limits can now be specified manually.
- With this update, the /usr/share/cluster/script.sh resource, used mainly by the rgmanager application, has been enhanced to provide more informative reports on causes of internal errors.
- With this update, the
nfsrestartoption has been added to both the fs and clusterfs resource agents. This option provides a way to forcefully restart NFS servers and allow a clean unmount of an exported file system.
- The pacemaker SAPInstance and SAPDatabase resource agents have been updated with the latest upstream patches.
- A new
prefer_interfaceparameter has been added to the rgmanager ip.sh resource agent. This parameter is used for adding an IP address to a particular network interface when a cluster node has multiple active interfaces with IP addresses on the same subnetwork.
- This update adds support for the Pacemaker resource agents under the Heartbeat OCF provider.
- Usage of lvm.sh with tags resulted in the stripping of cluster tags when the node rejoined the cluster. This was because the lvm.sh agent was unable to accurately detect the tag represented by a cluster node. Thus, the active logical volume on a cluster node failed when another node re-joined the cluster. This update properly detects whether tags represent a cluster node, the node-name, or, if fqdn is returned by the corosync-quorumtool -l output. When nodes re-join the cluster, tags are not stripped of LVM volume groups, and the volume group no longer fails on other nodes.
- Previous versions of the Oracle Resource Agent were only tested against Oracle 10. With this update, support for the Oracle Database 11g has been added to the oracledb, orainstance, and oralistener resource agents.