"Start at boot" not working on one node...

christophe

Active Member
Mar 31, 2011
169
2
38
Hi all,

While rebooting a 5 nodes cluster (after upgrade to 2.3), on one node (the first one), none of the KVM VMs started at boot.
On the other nodes, all went OK.

Every VM on each node as option "Start at boot" set to "Yes", "Start / shutdown order" set to "order=any".

Manual startu went OK.

I can find in daemon.log entries from pvedaemon when a manual start occurs :

pvedaemon[9110]: start VM 118: UPID:px1:00002396:0000EA09:51387C91:qmstart:1
18:coucou@ldap:

But no entry about automatic start at boot (on the other nodes).

root@px1:/var/log# pveversion -v
pve-manager: 2.3-13 (pve-manager/2.3/7946f1f1)
running kernel: 2.6.32-18-pve
proxmox-ve-2.6.32: 2.3-88
pve-kernel-2.6.32-16-pve: 2.6.32-82
pve-kernel-2.6.32-18-pve: 2.6.32-88
lvm2: 2.02.95-1pve2
clvm: 2.02.95-1pve2
corosync-pve: 1.4.4-4
openais-pve: 1.1.4-2
libqb: 0.10.1-2
redhat-cluster-pve: 3.1.93-2
resource-agents-pve: 3.9.2-3
fence-agents-pve: 3.1.9-1
pve-cluster: 1.0-36
qemu-server: 2.3-17
pve-firmware: 1.0-21
libpve-common-perl: 1.0-48
libpve-access-control: 1.0-26
libpve-storage-perl: 2.3-6
vncterm: 1.0-3
vzctl: 4.0-1pve2
vzprocps: 2.0.11-2
vzquota: 3.1-1
pve-qemu-kvm: 1.4-6
ksm-control-daemon: 1.1-1
root@px1:/var/log#

Any idea why?

Thanks,

Christophe.
 

tom

Proxmox Staff Member
Staff member
Aug 29, 2006
15,523
908
163
check syslog, any hint?
 

christophe

Active Member
Mar 31, 2011
169
2
38
Not really, neither about a fail on node 1, nor about explicit "start at boot" on other nodes...

Christophe.
 

dietmar

Proxmox Staff Member
Staff member
Apr 28, 2005
17,113
513
133
Austria
www.proxmox.com
You probably lost quorum? When you reboot a 5 node cluster, you should be carefull that at least 3 nodes stay online - else you loose quorum and VMs cannot be started.
 

christophe

Active Member
Mar 31, 2011
169
2
38
Yes, quorum was lost. This is precisely why reboot was planified.

See this thread : http://forum.proxmox.com/threads/13044-Multiple-Clusters-destroyed-at-the-same-time

I still don't know why an update - dist-upgrade one one cluster (from v2.2 to v2.3) "killed" (cman) the other, but this is a fact.

Quorum was lost, but all other nodes did start automatically VMs.

Not a big problem : I had a listing of VMs to check and start.

And now, each cluster runs fine, testing qcow2 and snapshots!

Christophe.
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get your own in 60 seconds.

Buy now!