Hi there,
I have a 10 node cluster running v6.3-3. Each node has access to an SSD Ceph pool (pool_ssd). Each node has a separate ceph cluster network, proxmox cluster network, and proxmox data network. I have recently setup a new node with local SSDs in raid 10. The local-lvm storage shows and is sufficient (4.67 TiB of 6.83 TiB). The new node has 2 x Xeon(R) CPU E5-2630 v3 with 188GB RAM. Disk tests with fio in a VM running on this node yielded speeds in excess of 3000 MBps.
Currently I am trying to move a VM disk (1TB) from the pool_ssd to the local-lvm, but after about 10 seconds all VM's on that node become unresponsive. There is nothing apparent in the /var/log/syslog. The same issue occurs when backups of the VMs run. The local-lvm disk is definitely fast enough at 3000 MBps. IO delay ends up being quite high at 14+.
Can someone perhaps help me shed some light on the matter?
Cheers,
Curt
I have a 10 node cluster running v6.3-3. Each node has access to an SSD Ceph pool (pool_ssd). Each node has a separate ceph cluster network, proxmox cluster network, and proxmox data network. I have recently setup a new node with local SSDs in raid 10. The local-lvm storage shows and is sufficient (4.67 TiB of 6.83 TiB). The new node has 2 x Xeon(R) CPU E5-2630 v3 with 188GB RAM. Disk tests with fio in a VM running on this node yielded speeds in excess of 3000 MBps.
Currently I am trying to move a VM disk (1TB) from the pool_ssd to the local-lvm, but after about 10 seconds all VM's on that node become unresponsive. There is nothing apparent in the /var/log/syslog. The same issue occurs when backups of the VMs run. The local-lvm disk is definitely fast enough at 3000 MBps. IO delay ends up being quite high at 14+.
Can someone perhaps help me shed some light on the matter?
Cheers,
Curt
Last edited: