Multipath shows "failed faulty running" status for all luns on Hitachi DF600F storage

Solution Verified - Updated -

Environment

  • Red Hat Enterprise Linux 6
  • Hitachi DF600F storage array
  • multipath

Issue

  • Allocated Hitachi HUS150 storage for the new RHEL 6 servers and the new luns were visible after the system reboot. However there are some weird issues with the multipath configuration: it shows the status as failed faulty for all the SAN luns. multipath -v3 reports checker msg is "readsector0 checker reports path is down"

  • The luns have the correct number of paths on the storage side.

mpathg (360060e80101af5a0058baf2a00000096) dm-12 HITACHI,DF600F
size=50G features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=0 status=enabled
  |- 1:0:0:1  sdc  8:32   failed faulty running
  `- 2:0:0:1  sdr  65:16  failed faulty running
...
lpfc 0000:06:00.1: 1:(0):2757 Protocol failure detected during processing of FCP I/O op: tgt x0 lun xb cmnd x28 rspInfo3 x2
lpfc 0000:06:00.1: 1:(0):2757 Protocol failure detected during processing of FCP I/O op: tgt x0 lun xc cmnd x28 rspInfo3 x2
lpfc 0000:06:00.1: 1:(0):2757 Protocol failure detected during processing of FCP I/O op: tgt x0 lun xd cmnd x28 rspInfo3 x2
lpfc 0000:06:00.1: 1:(0):2757 Protocol failure detected during processing of FCP I/O op: tgt x0 lun xe cmnd x28 rspInfo3 x2

Resolution

  • Add a space after "HITACHI" on the "vendor" line in multipath.conf. For example:
vendor                  "HITACHI "

Root Cause

The "vendor" value for the Hitachi storage is required to be 8 characters long.

This solution is part of Red Hat’s fast-track publication program, providing a huge library of solutions that Red Hat engineers have created while supporting our customers. To give you the knowledge you need the instant it becomes available, these articles may be presented in a raw and unedited form.

1 Comments

In my case, only one path was in 'failed faulty running' state.

multipath -ll

mpath999 (3600a0b80000b5c9c0000044d3b667c19) dm-0 STORAGE,VV
size=200G features='1 queue_if_no_path' hwhandler='0' wp=rw
-+- policy='round-robin 0' prio=1 status=active
|- 4:0:0:0 sdb 8:16 failed faulty running
- 1:0:0:0 sdc 8:32 active ready running

I have unbind and bind host4 as per https://access.redhat.com/solutions/287303 and it was resolved.