Proxmox v3.3 may have old bug with PVEstatd

Raymond Burns

Member
Apr 2, 2013
333
3
18
Houston, Texas, United States
I noticed a red node this morning.
This is after some things went haywire with NFS shares failing in the middle of backups. The solution posted here works.
http://forum.proxmox.com/threads/10013-Cluster-Nodes-become-red-without-any-reason-once-a-week

The server was running properly, and all of the VMs were active and running, but the GUI showed the node as down, with all VM's as off.
This is from the node in the down state:
Code:
root@zwtPROX3:~# cat /etc/pve/.members{
"nodename": "zwtPROX3",
"version": 7,
"cluster": { "name": "ZWT", "version": 9, "nodes": 3, "quorate": 1 },
"nodelist": {
  "zwtprox2": { "id": 1, "online": 1, "ip": "192.168.222.13"},
  "zwtPROX3": { "id": 2, "online": 1, "ip": "192.168.222.16"},
  "zwtPROX4": { "id": 5, "online": 1, "ip": "192.168.222.17"}
  }
}

Also:
Code:
root@zwtPROX3:~# pveversion -v
proxmox-ve-2.6.32: 3.3-138 (running kernel: 2.6.32-33-pve)
pve-manager: 3.3-2 (running version: 3.3-2/995e687e)
pve-kernel-2.6.32-33-pve: 2.6.32-138
pve-kernel-2.6.32-30-pve: 2.6.32-130
pve-kernel-2.6.32-29-pve: 2.6.32-126
pve-kernel-2.6.32-26-pve: 2.6.32-114
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.7-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.10-1
pve-cluster: 3.0-15
qemu-server: 3.1-35
pve-firmware: 1.1-3
libpve-common-perl: 3.0-19
libpve-access-control: 3.0-15
libpve-storage-perl: 3.0-23
pve-libspice-server1: 0.12.4-3
vncterm: 1.1-8
vzctl: 4.0-1pve6
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 2.1-9
ksm-control-daemon: 1.1-1
glusterfs-client: 3.5.2-1