Red Hat Training

A Red Hat training course is available for Red Hat Gluster Storage

Monitoring Guide

Red Hat Gluster Storage 3.4

Monitoring Gluster Cluster

Rakesh Ghatvisave

Red Hat Customer Content Services

Abstract

This guide provides essential information on how to import a Gluster cluster using Red Hat Gluster Storage Web Administration and how to monitor your Gluster cluster health, performance, and status. The monitoring and metrics visualization is provided by the Grafana monitoring platform which is integrated into the Web Administration interface.

Chapter 1. Overview

Red Hat Gluster Storage Web Administration provides visual monitoring and metrics infrastructure for Red Hat Gluster Storage 3.4 and is the primary method to monitor your Red Hat Gluster Storage environment. The Red Hat Gluster Storage Web Administration is based on the Tendrl upstream project and utilizes Ansible automation for installation. The key goal of Red Hat Gluster Storage Web Administration is to provide deep metrics and visualization of Red Hat Storage Gluster clusters and the associated physical storage elements such as storage nodes, volumes, and bricks.

Key Features

  1. Monitoring dashboards for Clusters, Hosts, Volumes, and Bricks
  2. Top-level list views of Clusters, Hosts, and Volumes
  3. SNMPv3 Configuration and alerting
  4. User Management
  5. Importing Gluster cluster

Chapter 2. Import Cluster

You can import and start managing a cluster once Web Administration is installed and ready. For installation instructions, see the Installing Web Administration chapter of the Red Hat Gluster Storage Web Administration Quick Start Guide.

2.1. Importing Cluster

The following procedure outlines the steps to import a Gluster cluster.

Procedure. Importing Cluster

  1. Log in to the Web Administration interface. Enter the username and password, and click Log In.

    Note

    The default username is admin and the default password is adminuser.

    Figure 2.1. Login Page

    Login Page
  2. In the default landing interface, a list of all the clusters ready to be imported is displayed. Locate the cluster to be imported and click Import.

    Figure 2.2. Import Cluster

    Import Cluster
  3. Enter a user-friendly Cluster name. By default, the Enable for all volumes option is selected.

    Figure 2.3. Cluster name

    Import Cluster
    Note

    If the cluster name is not provided, the system will assign a randomly generated uuid as a cluster name. However, it is advisable to enter a user-friendly cluster name to easily locate the cluster from the clusters list.

  4. Click Import to continue.
Import Cluster
  1. The cluster import request is submitted. To view the task progress, click View Task Progress.

    Figure 2.4. Task Detail

    Import Cluster
  2. Navigate to the All Clusters interface view. The Cluster is successfully imported and ready for use.

Figure 2.5. Cluster Ready

Cluster Ready

2.1.1. Troubleshooting Import Cluster

Scenario: The Import cluster UI button is disabled after a failed cluster import operation.

In this scenario, when cluster import fails, the Import button is disabled.

Resolution

Resolve the issue by investigating why the import cluster operation failed. To see details of the failed operation, navigate to the All Clusters interface of the Web Administration environment. In the clusters list, locate the cluster that you attempted to import and click on View Details next to the Import Failed status label.

Examine the reason of the failed cluster import operation and resolve the issue. After resolving the issue, unmanage the cluster and then reimport the cluster. For unmanaging cluster instructions, navigate to the Unmanaging Cluster section of this Guide.

2.1.2. Volume Profiling

Volume profiling enables additional telemetry information to be collected on a per volume basis for a given cluster, which helps in troubleshooting, capacity planning, and performance tuning.

Volume profiling can be enabled or disabled on a per cluster basis and per volume basis when a cluster is actively managed and monitored using the Web Administration interface.

Note

Enabling volume profiling results in richer set of metrics being collected which may cause performance degradation to occur as system resources, for example, CPU and memory, may get used for volume profiling data collection.

Volume Profiling at Cluster Level

To enable or disable volume profiling at cluster level:

  1. Log in to the Web Administration interface
  2. From the Clusters list, locate the cluster to disable Volume Profiling.

    Note

    Clusters list is the default landing interface after login and the Interface switcher is on All Clusters.

  3. At the right-hand side, next to the Dashboard button, click the vertical ellipsis.
  4. An inline menu is opened. Click Disable Profiling or Enable Profiling depending on the current state. In the example screen below, Volume Profiling option is enabled. Click Disable Profiling to disable.

    Figure 2.6. Disable Volume Profiling

    Disable Volume Profiling
  5. The disable profiling task is submitted and processed. After processing, Volume Profiling is successfully disabled.

Figure 2.7. Disable Volume Profiling

Disable Volume Profiling

Volume Profiling at Volume Level

To enable or disable volume profiling at Volume level:

  1. Log in to the Web Administration interface and select the specific cluster from the Interface switcher drop-down.
  2. After selecting the specific cluster, the left vertical navigation pane is exposed.
  3. From the navigation pane, click Volumes. The Volumes view is displayed listing all the Volumes part of the cluster.

    Figure 2.8. Volumes View

    Disable Volume Profiling
  4. Locate the volume and click Disable Profiling or Enable Profiling depending on the current state. In the example screen below, Volume Profiling is enabled. To disable volume profiling, click Disable Profiling.

    Figure 2.9. Disable Volume Profiling

    Disable Volume Profiling
  5. The disable profiling task is submitted and processed. After processing, Volume Profiling is successfully disabled.

Figure 2.10. Disable Volume Profiling

Disable Volume Profiling

Volume Profiling Metrics

When volume profiling is disabled, the following metrics will not be displayed in the Grafana Dashboard. Based on the metrics required to view, enable or disable volume profiling accordingly.

For detailed information on Volume Profiling, see the Monitoring Red Hat Gluster Storage Gluster Workload chapter of the Red Hat Gluster Storage Administration Guide.

Table 2.1. Volume Profiling Metrics

Grafana Dashboard Level Dashboard SectionPanel and Metrics

Cluster Dashboard

At-a-glance

IOPS

Host Dashboard

At-a-glance

Brick IOPS

Volume Dashboard

Performance

IOPS

Volume Dashboard

Profiling Information

File Operations For Locks

Volume Dashboard

Profiling Information

Top File Operations

Volume Dashboard

Profiling Information

File Operations for Read/Write

Volume Dashboard

Profiling Information

File Operations for Inode Operations

Volume Dashboard

Profiling Information

File Operations for Entry Operations

Brick Dashboard

At-a-glance

IOPS

Chapter 3. Cluster Expansion

To expand an existing Gluster cluster already imported and managed by the Web Administration environment, perform the following sequence of actions:

Create New Nodes

Before initiating cluster expansion in the Web Administration interface, create new Gluster storage nodes to be imported in Web Administration. For detailed instructions, see the Expanding Volumes section in the Red Hat Gluster Storage Administration Guide.

New Alert

After successfully expanding the cluster with new nodes, an alert is generated in the Notifications panel of the Web Administration interface as displayed in the following screen:

expansion alert

In this screen, new peers represents new storage nodes that were created in the previous cluster expansion process.

New Event

Additionally, a new Event is also generated about the availability of new storage nodes for expansion as displayed in the following screen:

expansion event

Run Tendrl-ansible installation for new nodes

After the storage nodes are identified by Web Administration, run tendrl-ansible installation by adding the new hosts in the inventory file. For detailed installation instructions, see the Installing Web Administration chapter in the Red Hat Gluster Storage Web Administration Quick Start Guide.

Important

If you have implemented TLS-based client-server authentication, install and deploy TLS encryption certificates on the new nodes before running tendrl-ansible installation to avoid cluster expansion failure. For detailed information on configuring TLS encryption, see the TLS Encryption Configuration chapter of the Red Hat Gluster Storage Web Administration Quick Start Guide.

Expand Cluster in Web Administration Environment

After tendrl-ansible is executed, the expand cluster option becomes available in the Web Administration interface. Follow these instructions to expand the cluster:

Procedure. Expanding Cluster

  1. Log in to the Web Administration interface and click All Clusters interface from the interface selector drop-down.
  2. In the Clusters view, locate the cluster to be expanded. It will be labelled as Expansion required.

    expansion1
  3. At the far right of the cluster row, click the vertical ellipsis and then click Expand.

    expansion2
  4. The available hosts to be added are listed. Click Expand.

    expand hypothethica
  5. The cluster expansion task is submitted.

    expansion4
  6. The cluster is successfully expanded and ready for use.
expansion5

3.1. Troubleshooting Cluster Expansion

Scenario: Cluster Expansion task failed

In this scenario, the cluster explansion task fails.

Resolution

If cluster expansion fails, check if tendrl-ansible was executed successfully and ensure the node agents are correctly configured. If cluster expansion failed due to errors, resolve the errors on affected nodes and re-initiate the Expand Cluster action.

To verify tendrl-ansible execution steps, see the Web Administration Installation section of the Red Hat Gluster Storage Web Administration Quick Start Guide.

For detailed instructions on expanding cluster, navigate to the Cluster Expansion section of this Guide.

Chapter 4. Unmanaging Cluster

The following are the reasons to unmanage a cluster:

  1. Import cluster fails
  2. Removal of a cluster that is no longer under management and used for monitoring
  3. A way to remove orphaned cluster entries

Unmanage a cluster from Web Administration

To unmanage a cluster:

  1. Log in to the Web Administration interface and select the All Clusters view from drop-down.
  2. Locate the cluster from the list of imported clusters. At the far right of the cluster row, click the vertical ellipsis and then click Unmanage.

    unmanage cluster2
  3. A confirmation box appears. Click Unmanage to proceed.

    unmanage cluster3
  4. The unmanage cluster task is submitted. To view the task progress, click View Task Progress.

    unmanage cluster final
  5. The status shows Completed after the unmanage task is successfully completed.
unmanage cluster 5
Note

After a cluster is unmanaged, it will not reappear immediately for reimport in the Web Administration environment. When the cluster is unmanaged, the tendrl-node-agent service undergoes synchronization with the nodes in the cluster. Based on the number of nodes in the cluster, expect a delay ranging from 60 seconds to a few minutes for the unmanaged cluster to reappear in the Web Administration environment.

4.1. Graphite and Carbon Data Archival

Graphite and Carbon Data Archival

When the cluster unmanage process is initiated, a concurrent process of cluster metrics data archival also begins. Once the cluster unmanage operation is successfully concluded, the Graphite and Carbon metrics data is archived to the following path on the Web Administration server:

/var/lib/carbon/whisper/tendrl/archive/clusters

The size of the archived data will depend on the size of the cluster that is unmanaged. You can either store the archived data or delete it to free some disk space.

4.2. Graphite Database Metrics Retention

The Graphite web service is a repository of telemetry data collected using collectd.

Changing metrics retention period

By default, Graphite retains the cluster metrics for a period of 180 days. To change the default metrics retention period for a given cluster, follow these steps:

  1. Unmanage the cluster by following the Unmanage Cluster procedure outlined in the Red Hat Gluster Storage Web Administration Monitoring Guide.
  2. After the cluster is unmanaged, stop the carbon-cache service on the Web Administration server:

    # systemctl stop carbon-cache
  3. Access the storage-schemas.conf file at:

    /etc/tendrl/monitoring-integration/storage-schemas.conf
  4. Change the retention period in the retentions parameter under [tendrl] section and save the changes.

    Example:

    If you want to change the retention value to 90 days, set the value in the retentions parameter:

    [tendrl]
    pattern = ^tendrl\.
    retentions = 60s:90d

    Where:

    retentions = Each datapoint represents 60 seconds, and we want to keep enough datapoints so that they add up to 90 days of data.

  5. Start the carbon-cache service on the Web Administration server:

    # systemctl start carbon-cache
  6. Import the cluster again. For instructions, see the Import Cluster chapter in the Red Hat Gluster Storage Web Administration Monitoring Guide.

Chapter 5. Web Administration Log Levels

In Web Administration, the log messages of different components and services are logged by tendrl-node-agent. It receives all the log messages from other components via a socket and logs the messages using syslog.

tendrl-node-agent logs messages for the following Web Administration components:

Server-side Web Administration components:

  • tendrl-monitoring-integration
  • tendrl-notifier

Storage node-side Web Administration components:

  • tendrl-gluster-integration

These are the available log levels for Web Administration:

  1. DEBUG
  2. INFO
  3. WARNING
  4. ERROR (Default)
  5. CRITICAL
Note

The log levels are case-sensitive. Input the log levels in uppercase format to avoid any error messages.

To change the default log level (ERROR), follow these steps:

  1. Open the logging configuration yaml file in an editor:

    /etc/tendrl/node-agent/node-agent_logging.yaml
  2. Change the log level by modifying all instances of level variables under handlers and root:

    handlers:
        console:
            class: logging.StreamHandler
            level: ERROR 1
            stream: ext://sys.stdout
    
        info_file_handler:
            class: logging.handlers.SysLogHandler
            facility: local5
            address: /dev/log
            level: ERROR 2
    
    root:
        level: ERROR 3
        handlers: [console, info_file_handler]
    1
    Enter new log level
    2
    Enter new log level
    3
    Enter new log level
  3. Restart the tendrl-node-agent service by:

    # service tendrl-node-agent restart

Chapter 6. Monitoring and Metrics

Gluster Web Administration provides deep metrics and visualization of Gluster clusters, the physical server nodes and the storage elements (disks) through the Grafana open-source monitoring platform.

Chapter 7. Monitoring Dashboard and Concepts

The Monitoring Dashboard provides high level visual information on health, performance and utilization of cluster wide resources.

7.1. Dashboard Selector

The Dashboard Selector is the primary navigation tool to move between different dashboards.

Figure 7.1. Dashboard Selector

Dashboard Selector

7.2. Dashboard Panels

The Dashboard is composed of individual visualization blocks displaying different metrics and statistics termed as Panels. The panels exhibit different data based on the current status of the cluster component. Panels can be dragged and dropped and rearranged on the Dashboard.

The following are the types of panels available to visualize monitoring data:

  1. Single Status Panel: The status panel displays the aggregated value of a series in e a single number data. For example, the Health, volume, snapshots are Singlestat panels.

    Figure 7.2. Single Status Panel Example

    Singes Status Panel Example
  2. Multiple Status Panel: The Multiple status panel displays multiple values of a data source. The multiple status panel displays:

    • The severity of the component
    • If the component is disabled
    • Extra data in the panel about the component

      Figure 7.3. Multiple Status Panel Example

      Multiple Status Panel Example
      Note

      If a multitple status panel displays Invalid Number as a status for a component, it indicates there is no data available to display from the time series database.

  3. Graph Panel: The Graph panel allows to visualize unrestrained amounts of metrics. The Connection Trend and the Throughput Trend are examples of Graph panel.

    Figure 7.4. Graph Panel Example

    Graph Panel Example
  4. Table Panel: The Graph panel allows to visualize unrestrained amounts of metrics. The Connection Trend and the Throughput Trend are examples of Graph panel.

Figure 7.5. Table Panel Example

Graph Panel Example

7.3. Dashboard Rows

A row is a logical divider in a given Dashboard. The panels of the dashboard are arranged and organized in rows to give a streamlined look and visual.

7.4. Dashboard Color Codes

The Dashboard panels text displays the following color codes to represent health status information:

  • Green: Healthy
  • Orange: Degraded
  • Red: Unhealthy, Down, or Unavailable

Chapter 8. Monitoring Dashboard Features

8.2. Dashboard Time Range

The Grafana interface provides time range management of the the data being visualized. You can change the time range for a graph to view the data at different points in time

At the top right, you can access the master Dashboard time picker. It shows the currently selected time range and the refresh interval.

time range menu

Clicking the master Dashboard time picker, toggles a menu for time range controls.

time control

Time range

The time range filter allows to mix both explicit and relative time ranges. The explicit time range format is YYYY-MM-DD HH:MM:SS.

Quick Range

Quick ranges are preset values to choose a relative time.

Refreshing every

When enabled, auto-refresh will reload the dashboard at the specified time range.

8.3. Dashboard Sharing

The Dashboard Selector is the primary navigation tool to move between different dashboards.

Figure 8.2. Dashboard Selector

Dashboard Selector

Chapter 9. Monitoring Dashboard Navigation

To access the Monitoring Dashboard, follow these steps:

  1. Log in to the Web Administration interface at http://web-admin-server.example.com.
  2. In the default Cluster view, locate the cluster and click Launch Dashboard.
  3. The Cluster dashboard showing the aggregated metrics view is opened in a new window.

9.1. Cluster View Dashboard

The Cluster view dashboard allows the Gluster Administrator to:

  • View at-a-glance information about the Gluster cluster that includes health and status information, key performance indicators such as IOPS, throughput, etc, and alerts that can highlight attention to potential issues in the cluster, host, volume, and brick.
  • Compare a metric such as IOPS, CPU, Memory, Network Load across hosts within the cluster.
  • Compare utilization across bricks within a volume, for example, IOPS, capacity, etc.

Figure 9.1. Cluster View Dashboard

Cluster View Dashboard

9.2. Hosts View Dashboard

The Host view Dashboard allows the Gluster Administrator to:

  • View at-a-glance information about the Gluster host that includes health and status information, key performance indicators such as IOPS, throughput, etc and alerts that highlights attention to potential issues in the host, volume, brick, and disk.
  • Compare one or more metrics such as IOPS, CPU, Memory, Network Load across bricks within the host.
  • Compare utilization such as IOPS, capacity, etc across bricks within a host.
host view dash

9.3. Volume View Dashboard

The Volume view Dashboard allows the Gluster Administrator to:

  • View at-a-glance information about the Gluster Volume that includes health and status information, key performance indicators that highlights attention to potential issues in the volume, brick, and disk.
volume view dash

9.4. Brick View Dashboard

The Brick view dashboard allows the Gluster Administrator to:

  • View at-a-glance information about the Gluster brick that includes health and status information, key performance indicators such as IOPS, throughput, latency, etc and alerts that can highlight attention to potential issues in the brick and underlying disks.
  • Look at performance by brick to address diagnosing of RAID 6 disk failure/rebuild/degradation poor performance on one brick.
brick view dash

Chapter 10. Monitoring Cluster Metrics

10.1. Cluster Level Dashboard

This is the default dashboard of the Monitoring interface that shows the overview of the selected cluster.

10.1.1. Monitoring and Viewing Cluster Health

To monitor the Cluster health status and the metrics associated with it, view the panels in the Cluster Dashboard. For detailed panel descriptions and health indicators, see Table 7.1. Cluster Health Panel Descriptions.

10.1.1.1. Health and Snapshots

The Health panel displays the overall health of the selected cluster and the Snapshots panel shows the active number of snapshots.

health snapshot

10.1.1.2. Hosts, Volumes and Bricks

The Hosts, Volumes, and Bricks panels displays status information. The following is an example screen displaying the respective status information.

hosts volumes bricks
  • Hosts: In total, there are 3 online Hosts
  • Volumes: In total, there are 9 Volumes
  • Bricks: In total, there are 44 Bricks

10.1.1.3. Geo-Replication Session

The Geo-Replication Session panel displays geo-replication session information from a given cluster, including the total number of geo-replication session and a count of geo-replication sessions by status.

geo replication session

10.1.1.4. Health Panel Descriptions

The following table lists the Panels and the descriptions.

Table 10.1. Cluster Health Panel Descriptions

PanelDescriptionHealth Indicator

Health

The Health panel displays the overall health of the selected cluster, which is either Healthy or Unhealthy

Green: Healthy

Red: Unhealthy

Orange: Degraded

Snapshots

The Snapshots panel displays the count of the active snapshots

 

Hosts

The Hosts panel displays host status information including the total number of hosts and a count of hosts by status

 

Volume

The Volumes panel displays volume status information for the selected cluster, including the total number of volumes and a count of volumes by status

 

Bricks

The Bricks panel displays brick status information for the selected cluster, including the total number of bricks in the cluster, and a count of bricks by status

 

Geo-Replication Session

The Geo-Replication Session panel displays geo-replication session information from a given cluster, including the total number of geo-replication session and a count of geo-replication sessions by status

 

10.1.2. Monitoring and Viewing Cluster Performance

Cluster performance metrics can be monitored by the data displayed in the following panels.

Connection Trend

The Connection Trend panel displays the total number of client connections to bricks in the volumes for the selected cluster over a period of time. Typical statistics may look like this:

cluster connection trend

IOPS

The IOPS panel displays IOPS for the selected cluster over a period of time. IOPS is based on the aggregated brick level read and write operations collected using gluster volume profile info.

cluster iops

Capacity Utilization and Capacity Available

The Capacity Utilization panel displays the capacity utilized across all volumes for the selected cluster.

The Capacity Available panel displays the available capacity across all volumes for the selected cluster.

capacity uti available

Weekly Growth Rate

The Weekly Growth Rate panel displays the forecasted weekly growth rate for capacity utilization computed based on daily capacity utilization.

weekly growth rate

Weeks Remaining

The Weeks Remaining panel displays the estimated time remaining in weeks till volumes reach full capacity based on the forecasted Weekly Growth Rate.

weeks remaining

Throughput Trend

The Throughput Trend panel displays the network throughput for the selected cluster over a period of time.

throughput trend

10.1.3. Top Consumers

The Top Consumers panels displays the highest capacity utilization by the cluster resources.

To view the top consumers of the cluster:

  1. In the Cluster level dashboard, at the bottom, click Top Consumers to expand the menu.

    cluster status bottom

Top 5 Utilization By Bricks

The Top 5 Utilization By Bricks panel displays the bricks with the highest capacity utilization.

top bricks

Top 5 Utilization by Volume

The Top 5 Utilization By Volumes panel displays the volumes with the highest capacity utilization.

top volume

CPU Utilization by Host

The CPU Utilization by Host panel displays the CPU utilization of  each node in the cluster.

top cpu host

Memory Utilization By Host

The Memory Utilization by Hosts panel displays memory utilization of  each node in the cluster.

top memory host

Ping Latency Trend

The Ping Latency Trend panel displays the ping latency for each host in a given cluster.

ping latency trend

10.1.4. Monitoring and Viewing Cluster Status

To view the status of the overall cluster:

  1. In the Cluster level dashboard, at the bottom, click Status to expand the menu.

    cluster status bottom
  2. The Volume, Host, and Brick status are displayed in the panels.
expanded cluster status

Volume Status

The Volume Status panel displays the status code of each volume for the selected cluster.

voume cluster status

The volume status is displayed in numerals and colors. The following are the corresponding status of the numerals.

  • 0 = Up
  • 3 = Up (Degraded)
  • 4 = Up (Partial)
  • 5 = Unknown
  • 8 = Down

Host Status

The Host Status panel displays the status code of each host for the selected cluster.

host cluster status

The Host status is displayed in numeric codes:

  • 0 = Up
  • 8 = Down

Brick Status

The Brick Status panel displays the status code of each brick for the selected cluster.

brick cluster status

The Brick status is displayed in numeric codes:

  • 1 = Started
  • 10 = Stopped

10.2. Host Level Dashboard

10.2.1. Monitoring and Viewing Health and Status

To monitor the Cluster Hosts status and the metrics associated with it, navigate to the Hosts Level Dashboard and view the panels.

Health

The Health panel displays the overall health for a given host.

host health

Bricks and Bricks Status

The Bricks panel displays brick status information for a given host, including the total number of bricks in the host, and a count of bricks by status.

host bricks

The Brick Status panel displays the status code of each brick for a given host.

hosts bricks status
  • 1 = Started
  • 10 = Stopped

10.2.2. Monitoring and Viewing Performance

10.2.2.1. Memory and CPU Utilization

Memory Available

The Memory Available panel displays the sum of memory free and memory cached.

memeory available

Memory Utilization

The Memory Utilization panel displays memory utilization percentage for a given host that includes buffers and caches used by the kernel over a period of time.

memeory utilization
  • Buffered: Amount of memory used for buffering, mostly for I/O operations
  • Cached: Memory used for caching disk data for reads, memory-mapped files or tmpfs data
  • Slab Rec: Amount of reclaimable memory used for slab kernel allocations
  • Slab Unrecl: Amount of unreclaimable memory used for slab kernel allocations
  • Used: Amount of memory used, calculated as Total - Free (Unused Memory) - Buffered - Cache
  • Total: Total memory used

Swap Free

The Swap Free panel displays the available swap space in percent for a given host.

swap free

Swap Utilization

The Swap Utilization panel displays the used swap space in percent for a given host.

swap utilization

CPU Utilization

The CPU utilization panel displays the CPU utilization for a given host over a period of time.

cpu uti host

IOPS

The IOPS panel displays IOPS for a given host over a period of time.  IOPS is based on the aggregated brick level read and write operations.

host iops

10.2.2.2. Capacity and Disk Load

Total Brick Capacity Utilization Trend

The Total Brick Capacity Utilization Trend panel displays the capacity utilization for all bricks on a given for a period of time.

brick cap trend

Total Brick Capacity Utilization

The Total Brick Capacity Utilization panel displays the current percent capacity utilization for a given host.

brick cap uti

Total Brick Capacity Available

The Total Brick Capacity Available panel displays the current available capacity for a given host.

brick cap avail

Weekly Growth Rate

The Weekly Growth Rate panel displays the forecasted weekly growth rate for capacity utilization computed based on daily capacity utilization.

weekly growth

Weeks Remaining

The Weeks Remaining panel displays the estimated time remaining in weeks till host capacity reaches full capacity based on the forecasted Weekly Growth Rate.

week remaining

Brick Utilization

The Brick Utilization panel displays the utilization of each brick for a given host.

host brick uti

Brick Capacity

The Brick Capacity panel displays the total capacity of each brick for a given host.

host bricks cap

Brick Capacity Used

The Brick Capacity Used panel displays the used capacity of each brick for a given host.

host brick used

Disk Load

The Disk Load panel shows the host’s aggregated read and writes from/to disks over a period of time.

disk load

Disk Operation

The Disk Operations panel shows the host’s aggregated read and writes disk operations over a period of time.

disk ops

Disk IO

The Disk IO panel shows the host’s aggregated I/O time over a period of time.

disk io

10.2.2.3. Network

Throughput

The Throughput panel displays the network throughput for a given host over a period of time.

network trhoughout

Dropped Packets Per Second

The Dropped Packets Per Second panel displays dropped network packets for the host over a period of time.  Typically, dropped packets indicates network congestion, for example, the queue on the switch port your host is connected to is full and packets are dropped because it cannot transmit data fast enough.

dropped packet

Errors Per Second

The Errors Per Second panel displays network errors for a given host over a period of time. Typically, the errors indicate issues that occurred while transmitting packets due to carrier errors (duplex mismatch, faulty cable), fifo errors, heartbeat errors, and window errors, CRC errors too short frames, and/or too long frames. In short, errors typically result from faulty hardware, and/or speed mismatch.

errors per second

10.2.3. Host Dashboard Metric Units

The following table shows the metrics and their corresponding measurement units.

Table 10.2. Host Dashboard Metric Units

MetricsUnits

Memory Available

Megabyte/Gigabyte/Terabyte

Memory Utilization

Percentage %

Swap free

Percentage %

Swap Utilization

Percentage %

CPU Utilization

Percentage %

Total Brick Capacity Utilization

Percentage %

Total Brick Capacity

MB/GB/TB

Weekly Growth Rate

MB/GB/TB

Disk Load

kbps

Disk IO

millisecond ms

Network Throughput

kbps

10.3. Volume Level Dashboard

The Volume view dashboard allows the Gluster Administrator to:

  • View at-a-glance information about the Gluster volume that includes health and status information, key performance indicators such as IOPS, throughput, etc, and alerts that can highlight attention to potential issues in the volume, brick, and disk.
  • Compare 1 or more metrics such as IOPS, CPU, Memory, Network Load across bricks within the volume.
  • Compare utilization such as IOPS, capacity, etc, across bricks within a volume.
  • View performance metrics by brick (within a volume) to address diagnosing of failure, rebuild, degradation, and poor performance on one brick.

When all the Gluster storage nodes are shut down or offline, Time to live (TTL) will delete the volume details from etcd as per the TTL value measured in seconds. The TTL value for volumes is set based on the number of volumes and bricks in the system. The formula to calculate the TTL value to delete volume details is:

Time to Live (seconds) = synchronization interval (60 seconds) + number of volumes * 20 + number of bricks * 10 + 160.

In Web Administration environment

  • Cluster will show status as unhealthy and all hosts will be marked as down
  • No display of Volumes and Bricks
  • The Events view will reflect the relevant status

In Grafana Dashboard

  • In Cluster level Dashboard, the Host, Volumes, and Bricks panels reflects the relevant updated counts with status.
  • In Cluster, Volume, and Brick level dashboards, some panels will be marked as N/A, indicating no data is available.

10.3.1. Monitoring and Viewing Health

Health

The Health panel displays the overall health for a given volume.

volume health

Snapshots

The Snapshots panel displays the count of active snapshots for the selected cluster.

volume snapshot

Brick Status

The Brick Status panel displays the status code of each brick for a given volume.

volume brick status
  • 1 = Started
  • 10 = Stopped

Bricks

The Bricks panel displays brick status information for a given volume, including the total number of bricks in the volume, and a count of bricks by status.

volume bricks

Subvolumes

The Subvolumes panel displays subvolume status information for a given volume.

subloume

Geo-Replication Sessions

The Geo-Replication Session panel displays geo-replication session information from a given volumes, including the total number of geo-replication session and a count of geo-replication sessions by status.

geo rep status

Rebalance

The Rebalance panel displays rebalance progress information for a given volume, which is applicable when rebalancing is underway.

reblance

Rebalance Status:

The Rebalance Status panel displays the status of rebalancing for a given volume, which is applicable when rebalancing is underway.

rebalance status

10.3.2. Monitoring and Viewing Performance

Capacity Utilization

The Capacity Utilization panel displays the used capacity for a given volume.

vol capacity uti

Capacity Available

The Capacity Available panel displays the available capacity for a given volume.

vol cap avail

Weekly Growth Rate

The Weekly Growth Rate panel displays the forecasted weekly growth rate for capacity utilization computed based on daily capacity utilization.

vol weekly growth rate

Weeks Remaining

The Weeks Remaining panel displays the estimated time remaining in weeks till volume reaches full capacity based on the forecasted Weekly Growth Rate.

weeks remaining

Capacity Utilization Trend

The Capacity Utilization Trend panel displays the volume capacity utilization over a period of time.

caputi

Inode Utilization

The Inode Utilization panel displays inodes used for bricks in the volume over a period of time.

inode uti

Inode Available

The Inode Available panel displays inodes free for bricks in the volume.

inode avail

Throughput

The Throughput panel displays volume throughput based on brick-level read and write operations fetched using gluster volume profile.

through

LVM Thin Pool Metadata %

The LVM Thin Pool Metadata % panel displays the utilization of LVM thin pool metadata for a given volume. Monitoring the utilization of LVM thin pool metadata and data usage is important to ensure they do not run out of space. If the data space is exhausted, I/O operations are either queued or failing based on the configuration. If metadata space is exhausted, you will observe error I/O’s until the LVM pool is taken offline and repair is performed to fix potential inconsistencies. Moreover, due to the metadata transaction being aborted and the pool doing caching there might be uncommitted (to disk) I/O operations that were acknowledged to the upper storage layers (file system) so those layers will need to have checks/repairs performed as well.

lvm metadata

LVM Thin Pool Data Usage %

The LVM Thin Pool Data Usage % panel displays the LVM thin pool data usage for a given volume. Monitoring the utilization of LVM thin pool metadata and data usage is important to ensure they do not run out of space. If the data space is exhausted , I/O operations are either queued or failing based on the configuration. If metadata space is exhausted, you will observe error I/O’s until the LVM pool is taken offline and repair is performed to fix potential inconsistencies. Moreover, due to the metadata transaction being aborted and the pool doing caching there might be uncommitted (to disk) I/O operations that were acknowledged to the upper storage layers (file system) so those layers will need to have checks/repairs performed as well.

lvm usage

10.3.3. Monitoring File Operations

Top File Operations

The Top File Operations panel displays the top 5 FOP (file operations) with the highest % latency, wherein the % latency is the fraction of the FOP response time that is consumed by the FOP.

top fop

File Operations for Locks Trend

The File Operations for Locks Trend panel displays the average latency, maximum latency, call rate for each FOP for Locks over a period of time.

fop trend

File Operations for Read/Write

The File Operations for Read/Write panel displays the average latency, maximum latency, call rate for each FOP for Read/Write Operations over a period of time.

fop wr

File Operations for Inode Operations

The File Operations for Inode Operations panel displays the average latency, maximum latency, call rate for each FOP for Inode Operations over a period of time.

dop inode

File Operations for Entry Operations

The File Operations for Entry Operations panel displays the average latency, maximum latency, call rate for each FOP for Entry Operations over a period of time.

fop entry

10.3.4. Volume Dashboard Metric Units

The following table shows the metrics and their corresponding measurement units.

Table 10.3. Volume Dashboard Metric Units

MetricsUnits

Capacity Utilization

Percentage %

Capacity Available

Megabyte/Gigabyte/Terabyte

Weekly Growth Rate

Megabyte/Gigabyte/Terabyte

Capacity Utilization Trend

Percentage %

Inode Utilization

Percentage %

Lvm Thin Pool Metadata

Percentage %

Lvm Thin Pool Data Usage

Percentage %

File Operations for Locks Trend

MB/GB/TB

File Operations for Read/Write

K

File Operations for Inode Operation Trend

K

File Operations for Entry Operations

K

10.4. Brick Level Dashboard

10.4.1. Monitoring and Viewing Brick Status

The Status panel displays the status for a given brick.

brick status

10.4.2. Monitoring and Viewing Brick Performance

Capacity Utilization

The Capacity Utilization panel displays the percentage of capacity utilization for a given brick.

br cap uti

Capacity Available

The Capacity Available panel displays the available capacity for a given volume.

brick cap uti

Capacity Utilization Trend

The Capacity Utilization Trend panel displays the brick capacity utilization over a period of time.

brick capa uti

Weekly Growth Rate

The Weekly Growth Rate panel displays the forecasted weekly growth rate for capacity utilization computed based on daily capacity utilization.

br week rate

Weeks Remaining

The Weeks Remaining panel displays the estimated time remaining in weeks till brick reaches full capacity based on the forecasted Weekly Growth Rate.

reaminig weeks

Healing

The Healing panel displays healing information for a given volume based on healinfo.

volume healing
Note

The Healing panel will not show any data for volumes without replica.

IOPS

The IOPS panel displays IOPS for a brick over a period of time. IOPS is based on brick level read and write operations.

brick iops

LVM Thin Pool Metadata %

The LVM Thin Pool Metadata % panel displays the utilization of LVM thin pool metadata for a given brick. Monitoring the utilization of LVM thin pool metadata and data usage is important to ensure they don’t run out of space. If the data space is exhausted , I/O operations are either queued or failing based on the configuration. If metadata space is exhausted, you will observe error I/O’s until the LVM pool is taken offline and repair is performed to fix potential inconsistencies. Moreover, due to the metadata transaction being aborted and the pool doing caching there might be uncommitted (to disk) I/O operations that were acknowledged to the upper storage layers (file system) so those layers will need to have checks/repairs performed as well.

lvm meta

LVM Thin Pool Data Usage %

The LVM Thin Pool Data Usage % panel displays the LVM thin pool data usage for a given brick. Monitoring the utilization of LVM thin pool metadata and data usage is important to ensure they don’t run out of space. If the data space is exhausted , I/O operations are either queued or failing based on the configuration. If metadata space is exhausted, you will observe error I/O’s until the LVM pool is taken offline and repair is performed to fix potential inconsistencies. Moreover, due to the metadata transaction being aborted and the pool doing caching there might be uncommitted (to disk) I/O operations that were acknowledged to the upper storage layers (file system) so those layers will need to have repairs performed as well.

lvm pool

Throughput

The Throughput panel displays brick-level read and write operations fetched using “gluster volume profile.”

brick throuout

Latency

The Latency panel displays latency for a brick  over a period of time.  Latency is based on the average amount of time a brick spends doing a read or write operation.

br latency

10.4.3. Brick Dashboard Metric Units

The following table shows the metrics and their corresponding measurement units.

Table 10.4. Brick Dashboard Metric Units

MetricsUnits

Capacity Utilization

Percentage %

Capacity Available

Megabyte/Gigabyte/Terabyte

Weekly Growth Rate

Megabyte/Gigabyte/Terabyte

Capacity Utilization Trend

Percentage %

Inode Utilization

Percentage %

Lvm Thin Pool Metadata

Percentage %

Lvm Thin Pool Data Usage

Percentage %

Disk Throughput

Percentage %

Chapter 11. Users and Roles Administration

11.1. User Roles

There are three user roles available for Web Administration.

  1. Admin: The Admin role gives complete rights to the user to manage all Web Administration operations.
  2. Normal User: The Normal User role authorizes the user to perform operations such as importing cluster and enabling or disabling volume profiling but restricts managing users and other administrative operations.
  3. Read-only User: Read-only: The Read-only User role authorizes the user to only view and monitor cluster-wide metrics and readable data. The user can launch Grafana dashboards from the Web Administration interface but is restricted to perform any storage operations. This role is suited for users performing monitoring tasks.

11.2. Configuring Roles

To add and configure a new user, follow these steps:

  1. Log In the Web Administration interface and in navigation pane, click Admin > Users.
  2. The users list is displayed. To add a new user, click Add at the right-hand side.

    add user
  3. Enter the user information in the given fields. To enable or disable email notifications, toggle the ON-OFF button.

    add user1
  4. Select a Role from the available three roles and click Save.

    add user3
  5. The new user is successfully created.
add user4

11.2.1. Editing Users

To edit an existing user:

  1. Navigate to the user view by clicking Admin > Users from the interface navigation.
  2. Locate the user to be edited and click Edit at the right-hand side.

    add user4
  3. Edit the required information and click Save.
edit user

11.2.2. Disabling Notifications and Deleting User

Enabling and Disabling Notifications

To enable notifications:

  1. Navigate to the user view by clicking Admin > Users from the interface navigation.

    add user4
  2. Click the vertical elipsis next to the Edit button and click Disable Email Notification from the callout menu.

    disable notif
  3. Email notification is successfuly disabled for the user.
notification disabled

Deleting User

To delete an existing user:

  1. Navigate to the user view by clicking Admin > Users from the interface navigation.
  2. Locate the user to be deleted and click the vertical elipsis next to the Edit button. A callout menu opens, click Delete User.

    disable notif
  3. A confirmation box appears. Click Delete.

    delete use

Chapter 12. Alerts and Notifications

Alerts are current problems and critical conditions that occur in the system and notified to the user. The Grafana monitoring platform generates alerts based on severity levels.

You can configure alerts via SMTP and SNMP protocols. SMTP configuration will send email alerts to users that have email notifications enabled. SNMPv3 configuration will send SNMP trap alerts to the Alerts notifications drawer of the Web Administration environment.

12.1. Types of Alerts

The alerts triggered by the dashboard are classified in the following categories:

  • Status alerts : Alerts arising when a cluster resource undergoes a change of state. For example, Healthy to Unhealthy.
  • Utilization alerts: Alerts arising after a cluster resource exceed the set threshold and after it reverts to the normal state. For example, when the Host CPU utilization is breached, an alert is triggered notifying the user about the event.

12.2. List of Alerts

The list of Web Administration alerts are given in the tables below.

Status Alerts

Table 12.1. Status Alerts

AlertSystem Resource(s)

volume status

Volume and Cluster

volume state

Volume and Cluster

brick status

Volume, Host, and Cluster

peer status

Cluster

rebalance status

Volume and Cluster

Geo-Replication status

Cluster

quorum of volume lost

Volume and Cluster

quorum of volume regained

Volume and Cluster

svc connected

Cluster

svc disconnected

Cluster

minimum number of bricks not up in EC subvolume

Volume and Cluster

minimum number of bricks up in EC subvolume

Volume and Cluster

afr quorum met for subvolume

Volume and Cluster

afr quorum fail for subvolume

Volume and Cluster

afr subvolume up

Volume and Cluster

afr subvolume down

Volume and Cluster

Utilization Alerts

Table 12.2. Utilization Alerts

AlertSystem Resource

cpu utilization

Host

memory utilization

host

swap utilization

host

volume utilization

Volume and Cluster

brick utilization

Volume and Cluster

12.3. Alerts Notifications Drawer

Alerts drawer is a notification indicator embedded in the Web Administration interface to display the system wide alerts.

Accessing Alerts Drawer

  1. To access the Alerts drawer, log in the Web Administration interface. In the default landing interface, locate and click on the interactive bell icon on the header bar at the top right-hand side.

    headerbar
  2. The drawer is opened displaying the number of alerts generated.
alert drw

To filter alerts, click on the status icon at the right.

alerts filter

12.4. SMTP Notifications Configuration

Tendrl-ansible installs and configures tendrl-notifier. After the tendrl-notifier file is configured, configure SMTP email notifications:

  1. Open the /etc/tendrl/notifier/email.conf.yaml file.
  2. Update the parameters:

    email_id = <The sender email id>
    
    email_smtp_server = <The smtp server>
    
    email_smtp_port = <The smtp port>
  3. If the SMTP server supports only authenticated email, follow the template in the /etc/tendrl/notifier/email_auth.conf.yaml file and accordingly enable the following:

        auth = <ssl/tls>
    
        email_pass = <password corresponding to email_id for authenticating to smtp server>
  4. Restart the tendrl-notifier service:

    systemctl restart tendrl-notifier

12.5. SNMPv3 Notification Configuration

Configure SNMP

Tendrl-ansible installs and configures tendrl-notifier. After the tendrl-notifier file is configured, configure SNMPv3 trap notifications:

  1. Open the tendrl-notifier configuration file:

    # cat /etc/tendrl/notifier/snmp.conf.yaml
  2. Update the parameters in the file for v3 trap alerts:

    For v3_endpoint:
    # For more hosts you can add more entry with endpoint2, endpoint3, etc
    endpoint1:
    
    # Name or IP address of the remote SNMP host.
            host_ip: <Receiving machine ip>
    # Name of the user on the host that connects to the agent.
            username: <Username of receiver>
    # Enables the agent to receive packets from the host.
            auth_key: <md5 password>
    # The private user password
            priv_key: <des password>
    
    
    # For v2_endpoint:
    
    # For more hosts you can add more entry with endpoint2, endpoint3, etc
    
    endpoint1:
    # Name or IP address of the remote SNMP host.
         host_ip: <Receiving machine ip>
         community: <community name>
  3. Restart the tendrl-notifier service:

    systemctl restart tendrl-notifier

Legal Notice

Copyright © 2019 Red Hat, Inc.
The text of and illustrations in this document are licensed by Red Hat under a Creative Commons Attribution–Share Alike 3.0 Unported license ("CC-BY-SA"). An explanation of CC-BY-SA is available at http://creativecommons.org/licenses/by-sa/3.0/. In accordance with CC-BY-SA, if you distribute this document or an adaptation of it, you must provide the URL for the original version.
Red Hat, as the licensor of this document, waives the right to enforce, and agrees not to assert, Section 4d of CC-BY-SA to the fullest extent permitted by applicable law.
Red Hat, Red Hat Enterprise Linux, the Shadowman logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries.
Linux® is the registered trademark of Linus Torvalds in the United States and other countries.
Java® is a registered trademark of Oracle and/or its affiliates.
XFS® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries.
MySQL® is a registered trademark of MySQL AB in the United States, the European Union and other countries.
Node.js® is an official trademark of Joyent. Red Hat Software Collections is not formally related to or endorsed by the official Joyent Node.js open source or commercial project.
The OpenStack® Word Mark and OpenStack logo are either registered trademarks/service marks or trademarks/service marks of the OpenStack Foundation, in the United States and other countries and are used with the OpenStack Foundation's permission. We are not affiliated with, endorsed or sponsored by the OpenStack Foundation, or the OpenStack community.
All other trademarks are the property of their respective owners.