Slave node failure on cluster, how to start VMs?

rjacobsen

New Member
Jan 7, 2011
6
0
1
Hello everyone,

We have a 4 node proxmox cluster, currently at version 1.6. VM disks are located on a NAS device and accessed via LVM and iSCSI. Recently a node failed with a hardware problem, at which point I could not start any VMs from the failed node on another node, nor could I migrate them or even see them in the WebUI. In order to start the VMs I had to copy files from /etc/qemu-system/ on the failed node manually to a different node.

Luckily the disk hadn't failed catastrophically and I was able to restore the files, as there were no backups of the slave node.

Is there another way to do this in the WebUI that I missed? Are these files now stored/synchronized between nodes perhaps in 1.7? If not, I suppose backups should be in order.

Here's some more information about my system:

Proxmox 1.6
Code:
prox1:~#uname -a
Linux prox1 2.6.24-11-pve #1 SMP PREEMPT Fri May 14 09:28:08 CEST 2010 x86_64 GNU/Linux

prox1:~# pveversion -v
pve-manager: 1.6-2 (pve-manager/1.6/5087)
running kernel: 2.6.24-11-pve
proxmox-ve-2.6.24: 1.5-23
pve-kernel-2.6.24-11-pve: 2.6.24-23
pve-kernel-2.6.24-8-pve: 2.6.24-16
qemu-server: 1.1-18
pve-firmware: 1.0-7
libpve-storage-perl: 1.0-13
vncterm: 0.9-2
vzctl: 3.0.24-1pve4
vzdump: 1.2-7
vzprocps: 2.0.11-1dso2
vzquota: 3.0.11-1
pve-qemu-kvm: 0.12.5-1
Thanks,
Richard