Hi,
We have a 6-node cluster, each connected with 2x1Gbit trunk to a 4x1Gbit trunk ZFS NFS server loaded with SSD's.
The disk IO on the ZFS server is really high, it has 12 SSD's in RAIDz2 (RAID6).
We have another ZFS SSD server, also connected with 4x1Gbit.
Now we want to migrate some disks to this new storage server with live migration. When we start the migration, all VM's on the node that initiates the move get really slow. Login on RDP takes about 5 minutes on any VM, while this normally takes several seconds. Webservers crash or respond really slow.
So when we move 1 disk of 1 VM, all VM's on that node become unusable.
How can we limit the disk move, so it will only use like 80% of the 1000Mbit link? The SSD's are almost idle when the disk moves, it's the network link.
The costs to replace all 1Gbit to 10Gbit, including a redundant managed switch, is really high. So that's not a option.
We have a 6-node cluster, each connected with 2x1Gbit trunk to a 4x1Gbit trunk ZFS NFS server loaded with SSD's.
The disk IO on the ZFS server is really high, it has 12 SSD's in RAIDz2 (RAID6).
We have another ZFS SSD server, also connected with 4x1Gbit.
Now we want to migrate some disks to this new storage server with live migration. When we start the migration, all VM's on the node that initiates the move get really slow. Login on RDP takes about 5 minutes on any VM, while this normally takes several seconds. Webservers crash or respond really slow.
So when we move 1 disk of 1 VM, all VM's on that node become unusable.
How can we limit the disk move, so it will only use like 80% of the 1000Mbit link? The SSD's are almost idle when the disk moves, it's the network link.
The costs to replace all 1Gbit to 10Gbit, including a redundant managed switch, is really high. So that's not a option.