Verifying Intel® VROC RAID Status on RHEL 9 with shell script

Posted on

Hi everyone,

I’m currently working on a Red Hat Enterprise Linux 9 environment and managing an Intel® VROC RAID setup. I want to verify and test the following RAID statuses/events with BASH Shell.

  • Fail: A member drive of Intel® VROC RAID is faulty.
  • FailSpare: The spare drive being rebuilt has failed.
  • DeviceDisappeared: A RAID volume disappeared or was removed.
  • DegradedArray: A RAID array is running in degraded mode.
  • RebuildStarted: Rebuilding or recovery of a degraded RAID has started.
  • RebuildNN: Notification of rebuild progress (e.g., 20%, 40%).
  • RebuildFinished: RAID rebuild completed or aborted.
  • SparesMissing: One or more spare drives are missing.

Here’s my current setup:
- RAID Configuration:

[root@localhost ~]# mdadm --detail-platform
mdadm: imsm capabilities not found for controller: /sys/devices/pci0000:00/0000:00:17.0 (type SATA)
mdadm: imsm capabilities not found for controller: /sys/devices/pci0000:00/0000:00:19.0 (type SATA)
       Platform : Intel(R) Virtual RAID on CPU
        Version : 8.5.0.1096
    RAID Levels : raid0 raid1 raid10
    Chunk Sizes : 4k 8k 16k 32k 64k 128k
    2TB volumes : supported
      2TB disks : supported
      Max Disks : 96
    Max Volumes : 2 per array, 24 per controller
 3rd party NVMe : supported
 I/O Controller : /sys/devices/pci0000:9a/0000:9a:00.5 (VMD)
 NVMe under VMD : /dev/nvme0n1 (S64FNN0X507593)
                  Encryption(Ability|Status): SED|Unencrypted
 NVMe under VMD : /dev/nvme2n1 (S64FNN0X507591)
                  Encryption(Ability|Status): SED|Unencrypted
 NVMe under VMD : /dev/nvme1n1 (S64FNN0X507583)
                  Encryption(Ability|Status): SED|Unencrypted
 I/O Controller : /sys/devices/pci0000:c6/0000:c6:00.5 (VMD)
  $ cat /proc/mdstat  
  Personalities : [raid1]  
  md125 : inactive nvme2n1[0](S)  
        1105 blocks super external:imsm  

  md126 : active raid1 nvme0n1[1] nvme1n1[0]  
        890806272 blocks super external:/md127/0 [2/2] [UU]  

  md127 : inactive nvme1n1[1](S) nvme0n1[0](S)  
        10402 blocks super external:imsm  

  unused devices: <none>  

I attempted to check status using the following command:

[root@localhost ~]# sudo mdadm --detail /dev/md126
/dev/md126:
         Container : /dev/md/imsm, member 0
        Raid Level : raid1
        Array Size : 890806272 (849.54 GiB 912.19 GB)
     Used Dev Size : 890806272 (849.54 GiB 912.19 GB)
      Raid Devices : 2
     Total Devices : 2

             State : active
    Active Devices : 2
   Working Devices : 2
    Failed Devices : 0

Consistency Policy : resync


              UUID : 991e034f:faaee5eb:fc8c7c50:8482898a
    Number   Major   Minor   RaidDevice State
       1     259        0        0      active sync   /dev/nvme0n1
       0     259        2        1      active sync   /dev/nvme1n1

My goals are:
1. Understand how to manually verify these all mentioned RAID statuses and events on RHEL with BASH Shell.

I would appreciate detailed guidance, including commands, configurations, or best practices.

Thank you in advance for your support!
Best regards,
Samadhan

Responses