PVE version 5.2-2
Kernel Version
- Linux 4.15.17-1-pve #1 SMP PVE 4.15.17-9 (Wed, 9 May 2018 13:31:43 +0200)
CPU
- 24 x Intel(R) Xeon(R) CPU E5-2620 v3 @ 2.40GHz (2 Sockets)
Storage:
- 8 x SSD SATA3 INTEL S3520 800GB 6Gb/s 3D MLC in a ZFS pool (RAID 5 HW, ...ye we didn't know ZFS in the beginning...)
- NO ZIL (same issue with 32 GB ZIL)
- NO L2ARC
- RAID bus controller: LSI Logic / Symbios Logic MegaRAID SAS-3 3108 [Invader] (rev 02)
RAM:
- 256GB
There is no need to comment what complains I get from my users (about 20 Windows VMs) when this happens:
I understand that I can expect some punctual latencies or general speed down if intensive IO processes are working (inside a VM, or migrating VMs between nodes, for instance)... but this behaviour totally stucks all VMs in the node for big whiles... Maybe something regarding unproper IO queue management... ?
Any advise (besides recreating a pure SW ZFS RAID...) ?
Thx in advance.
Kernel Version
- Linux 4.15.17-1-pve #1 SMP PVE 4.15.17-9 (Wed, 9 May 2018 13:31:43 +0200)
CPU
- 24 x Intel(R) Xeon(R) CPU E5-2620 v3 @ 2.40GHz (2 Sockets)
Storage:
- 8 x SSD SATA3 INTEL S3520 800GB 6Gb/s 3D MLC in a ZFS pool (RAID 5 HW, ...ye we didn't know ZFS in the beginning...)
- NO ZIL (same issue with 32 GB ZIL)
- NO L2ARC
- RAID bus controller: LSI Logic / Symbios Logic MegaRAID SAS-3 3108 [Invader] (rev 02)
RAM:
- 256GB
There is no need to comment what complains I get from my users (about 20 Windows VMs) when this happens:
I understand that I can expect some punctual latencies or general speed down if intensive IO processes are working (inside a VM, or migrating VMs between nodes, for instance)... but this behaviour totally stucks all VMs in the node for big whiles... Maybe something regarding unproper IO queue management... ?
Any advise (besides recreating a pure SW ZFS RAID...) ?
Thx in advance.