Proxmox Node Failed Ceph but had to do manual recover of VM

jimvman

Active Member
Oct 16, 2019
17
0
41
53
Hi,
I have a 3-node Ceph Cluster and a bunch of VMs scattered between hosts. Today one of the hosts boot drive failed, and all of its VMs were down.
1. I had to manually restore those VMs from NFS backup and start them up on one of the 2 working hosts (that was time consuming)
2. I could not figure out how to just start them up right away even though the configurations in /etc/pve/nodes/proxmoxservername/qemu_server/* files were there. I tried to move those to a working proxmoxservername from command line but it said file exists
3. I was hoping that there was automatic fail-over if one of the hosts fails. Is there something I missed that allows this to happen?

I'm going to have to rebuild these hosts anyway so I want to make sure it's done properly.

Thanks for all your help.