I have a three node Proxmox cluster in my lab currently, the hardware specifications are completely identical, running Proxmox 4.1 with all of the latest updates installed (with reboot performed post-upgrade).
The hardware specs:
DL380 G7
LSI HBA
2x 300G 10K SAS in ZFS Raid 1 (setup as "local" storage)
6x 300G 10K SAS for Ceph OSD (setup for distributed VM storage)
96 GB RAM
Firmware is up-to-date on all 3 nodes.
The Ceph cluster is healthy, each node is serving OSDs and MON functionality. I have two VMs running on the nodes, live migration and HA appears to work successfully. The combined RAM allocation for the VMs is only 12 GB (less than 4% of the cluster capacity).
When I attempt to backup either of the VMs to the "local" storage device (Mode: Snapshot, Compression: LZO (fast)), the backup job either stalls at around 40% or completes successfully. However, the node performing the backup then reboots itself.
Has anyone experienced this behavior before? I thought at first it was bad hardware, but I was able to reproduce the issue 100% of the time on all 3 nodes.
The hardware specs:
DL380 G7
LSI HBA
2x 300G 10K SAS in ZFS Raid 1 (setup as "local" storage)
6x 300G 10K SAS for Ceph OSD (setup for distributed VM storage)
96 GB RAM
Firmware is up-to-date on all 3 nodes.
The Ceph cluster is healthy, each node is serving OSDs and MON functionality. I have two VMs running on the nodes, live migration and HA appears to work successfully. The combined RAM allocation for the VMs is only 12 GB (less than 4% of the cluster capacity).
When I attempt to backup either of the VMs to the "local" storage device (Mode: Snapshot, Compression: LZO (fast)), the backup job either stalls at around 40% or completes successfully. However, the node performing the backup then reboots itself.
Has anyone experienced this behavior before? I thought at first it was bad hardware, but I was able to reproduce the issue 100% of the time on all 3 nodes.
Last edited: