Problem with Multipath FC and Ethernus DX200S4

nextit

Member
May 11, 2020
2
0
21
37
Hi,

we have a cluster with 3 nodes that the storage is based on Ethernus DX200S4 connected via FC.

The SAN have 2 controller that are connected to each nodes via 2 link.

The multipath view the LUN on the storage, but when we connect alla the cable one path flap on each node concurrently.

multipath -ll

3600000e00d2c0000002c1b2000000000 dm-2 FUJITSU,ETERNUS_DXL
size=26T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
|-+- policy='service-time 0' prio=50 status=active
| `- 1:0:0:0 sdb 8:16 active ready running
`-+- policy='service-time 0' prio=10 status=enabled
|- 1:0:1:0 sdc 8:32 active ready running
`- 2:0:1:0 sdd 8:48 active ready running

/var/log/messages

May 11 17:50:16 pve03-fi kernel: [1745898.894241] sd 2:0:0:0: Attached scsi generic sg4 type 0
May 11 17:50:16 pve03-fi kernel: [1745898.905579] sd 2:0:0:0: alua: transition timeout set to 60 seconds
May 11 17:50:16 pve03-fi kernel: [1745898.905585] sd 2:0:0:0: alua: port group 8080 state A preferred supports toluSNA
May 11 17:50:16 pve03-fi kernel: [1745898.995235] sd 2:0:0:0: Power-on or device reset occurred
May 11 17:50:22 pve03-fi kernel: [1745904.941800] sd 2:0:0:0: [sde] 55029268480 512-byte logical blocks: (28.2 TB/25.6 TiB)
May 11 17:50:22 pve03-fi kernel: [1745904.941813] sd 2:0:0:0: [sde] 4096-byte physical blocks
May 11 17:50:22 pve03-fi kernel: [1745905.045139] sd 2:0:0:0: Power-on or device reset occurred
May 11 17:50:22 pve03-fi kernel: [1745905.045219] sd 2:0:0:0: [sde] Write Protect is off
May 11 17:50:22 pve03-fi kernel: [1745905.045343] sd 2:0:0:0: [sde] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA
May 11 17:50:22 pve03-fi kernel: [1745905.047626] sd 2:0:0:0: [sde] Attached SCSI disk
May 11 17:50:23 pve03-fi kernel: [1745905.561638] sd 1:0:0:0: alua: port group 8081 state A preferred supports toluSNA
May 11 17:50:23 pve03-fi kernel: [1745905.561654] sd 2:0:0:0: alua: port group 8080 state A preferred supports toluSNA
May 11 17:50:23 pve03-fi kernel: [1745905.561723] sd 1:0:0:0: alua: port group 8081 state A preferred supports toluSNA
May 11 17:50:23 pve03-fi kernel: [1745905.561738] sd 2:0:0:0: alua: port group 8080 state A preferred supports toluSNA
May 11 17:50:33 pve03-fi kernel: [1745915.229870] sd 2:0:0:0: Power-on or device reset occurred
May 11 17:50:49 pve03-fi kernel: [1745932.144922] sd 2:0:0:0: Power-on or device reset occurred
May 11 17:51:12 pve03-fi kernel: [1745954.511377] sd 2:0:0:0: Power-on or device reset occurred
May 11 17:51:22 pve03-fi kernel: [1745964.732790] sd 2:0:0:0: Power-on or device reset occurred
May 11 17:51:30 pve03-fi kernel: [1745972.435144] sd 2:0:0:0: Power-on or device reset occurred
May 11 17:51:36 pve03-fi kernel: [1745978.644466] sd 2:0:0:0: Power-on or device reset occurred
May 11 17:51:49 pve03-fi kernel: [1745991.444330] sd 2:0:0:0: Power-on or device reset occurred
May 11 17:51:56 pve03-fi kernel: [1745999.042402] device-mapper: multipath: Failing path 8:64.
May 11 17:52:42 pve03-fi kernel: [1746045.097180] sd 2:0:0:0: [sde] Synchronizing SCSI cache
May 11 17:52:42 pve03-fi kernel: [1746045.097208] sd 2:0:0:0: [sde] Synchronize Cache(10) failed: Result: hostbyte=DID_NO_CONNECT driverbyte=DRIVER_OK
May 11 17:52:42 pve03-fi kernel: [1746045.178017] scsi 2:0:0:0: alua: Detached
May 11 17:52:43 pve03-fi kernel: [1746045.209730] sd 1:0:0:0: alua: port group 8081 state A preferred supports toluSNA
May 11 17:52:43 pve03-fi kernel: [1746045.209806] sd 1:0:0:0: alua: port group 8081 state A preferred supports toluSNA

Does some one have the same problem or a solution with this product?

Thanks in advance.

Best regards
 
Why haven't you configured ALUA? The flapping would stop and you would have access to both paths in a round robin fashion.
Hi,
we haven't configured nothing because the default configs of multipath.

Do you have any example that you can post us to see the correct configuration?
 
Do you have any example that you can post us to see the correct configuration?

It is configured on the SAN side, you need to change to host response in "Volume" -> "Performance (HOST I/O)":

1589292191837.png

I also recommend using Wide Stripe Volumes.

In general, it's best to ask the company who sold the SAN to you.
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!