I noticed a red node this morning.
This is after some things went haywire with NFS shares failing in the middle of backups. The solution posted here works.
http://forum.proxmox.com/threads/10013-Cluster-Nodes-become-red-without-any-reason-once-a-week
The server was running properly, and all of the VMs were active and running, but the GUI showed the node as down, with all VM's as off.
This is from the node in the down state:
Also:
This is after some things went haywire with NFS shares failing in the middle of backups. The solution posted here works.
http://forum.proxmox.com/threads/10013-Cluster-Nodes-become-red-without-any-reason-once-a-week
The server was running properly, and all of the VMs were active and running, but the GUI showed the node as down, with all VM's as off.
This is from the node in the down state:
Code:
root@zwtPROX3:~# cat /etc/pve/.members{
"nodename": "zwtPROX3",
"version": 7,
"cluster": { "name": "ZWT", "version": 9, "nodes": 3, "quorate": 1 },
"nodelist": {
"zwtprox2": { "id": 1, "online": 1, "ip": "192.168.222.13"},
"zwtPROX3": { "id": 2, "online": 1, "ip": "192.168.222.16"},
"zwtPROX4": { "id": 5, "online": 1, "ip": "192.168.222.17"}
}
}
Also:
Code:
root@zwtPROX3:~# pveversion -v
proxmox-ve-2.6.32: 3.3-138 (running kernel: 2.6.32-33-pve)
pve-manager: 3.3-2 (running version: 3.3-2/995e687e)
pve-kernel-2.6.32-33-pve: 2.6.32-138
pve-kernel-2.6.32-30-pve: 2.6.32-130
pve-kernel-2.6.32-29-pve: 2.6.32-126
pve-kernel-2.6.32-26-pve: 2.6.32-114
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.7-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.10-1
pve-cluster: 3.0-15
qemu-server: 3.1-35
pve-firmware: 1.1-3
libpve-common-perl: 3.0-19
libpve-access-control: 3.0-15
libpve-storage-perl: 3.0-23
pve-libspice-server1: 0.12.4-3
vncterm: 1.1-8
vzctl: 4.0-1pve6
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 2.1-9
ksm-control-daemon: 1.1-1
glusterfs-client: 3.5.2-1