Chapter 3. New features

This section lists all major updates, enhancements, and new features introduced in this release of Red Hat Ceph Storage.

3.1. The Ceph Ansible utility

Users can now purge the dashboard and monitoring stack only

Previously, users could not purge only the Ceph Manager Dashboard and Monitoring stack components such as Alertmanager, Prometheus, Grafana, and node-exporter separately.

With the `purge-dashboard.yml ` playbook, users can remove only the dashboard and the monitoring stack components.

Purging the storage cluster with osd_auto_discovery: true now purges the cluster and removes the Ceph OSDs

Previously, purging the storage cluster deployed with osd_auto_discovery: true would not purge the Ceph OSDs. With this release, the purge playbook works as expected and removes the Ceph OSDs when the storage cluster is deployed with osd_auto_discovery: true scenario.

The Alertmanager configuration is customizable

With this release, you can customize the Alertmanager configuration using the alertmanager_conf_overrides parameter in the /group_vars/all.yml file.

The Red Hat Ceph Storage Dashboard deployment is supported on a dedicated network

Previously, ceph-ansible stated the address that should be used for deploying the dashboard was on the same subnet as the public_network.

With this release, you can override the default dedicated subnet for the dashboard by setting the dashboard_network parameter in the /group_vars/all.yml file with the CIDR subnet address.

Setting the global NFS options in the configuration file is supported

Previously, ceph-ansible would not allow overriding any parameter in the configuration file.

With this release, you can override any parameter in the NFS_CORE_PARAM block section in the ganesha.conf file by setting the variable ganesha_core_param_overrides in group_vars/all.yml and update client-related configuration.

ceph-ansible checks for the Ceph Monitor quorum before starting the upgrade

Previously, when the storage cluster was in a HEALTH ERR or HEALTH WARN state due to one of the Ceph monitors being down, the rolling_upgrade.yml playbook would run. However, the upgrade would fail and the quorum was lost resulting in I/O down or a cluster failure.

With this release, an additional condition occurs where ceph-ansible checks the Ceph Monitor quorum before starting the upgrade.

The systemd target units for containerized deployments are now supported

Previously, there was no way to stop all Ceph daemons on a node in a containerized deployment.

With this release, systemd target units for containerized deployments are supported and you can stop all the Ceph daemons on a host or specific Ceph daemons similar to bare-metal deployments.

ceph-ansible now checks the relevant release version during an upgrade before executing the playbook

With this release, during a storage cluster upgrade, ceph-ansible first checks for the relevant release version and the playbook fails with an error message if a wrong Ceph version is provided.

3.2. Ceph Management Dashboard

A new Grafana Dashboard to display graphs for Ceph Object Gateway multi-site setup

With this release, a new Grafana dashboard is now available and displays graphs for Ceph Object Gateway multisite sync performance including two-way replication throughput, polling latency, and unsuccessful replications.

See the Monitoring Ceph object gateway daemons on the dashboard section in the Red Hat Ceph Storage Dashboard Guide for more information.

3.3. Ceph File System

Use max_concurrent_clones option to configure the number of clone threads

Previously, the number of concurrent clones was not configurable and the default was 4.

With this release, the maximum number of concurrent clones is configurable using the manager configuration option:


ceph config set mgr mgr/volumes/max_concurrent_clones VALUE

Increasing the maximum number of concurrent clones could improve the performance of the storage cluster.

3.4. Ceph Object Gateway

The role name and the role session information is displayed in the ops log for S3 operations

With this release, you get information such as the role name and the role session in the ops log for all the S3 operations that use temporary credentials returned by AssumeRole* operations for debugging and auditing purposes.

3.5. Multi-site Ceph Object Gateway

Data sync logging experienced delays in processing

Previously, data sync logging could be subject to delays in processing large backlogs of log entries.

With this release, data sync includes caching for bucket sync status. The addition of the cache speeds the processing of duplicate datalog entries when a backlog exists.