Hi,
I have set up a two nodes cluster following the wiki guidance, but I keep facing the same problem.
I'm able to see both nodes within the web GUI and spawn VMs while everything after my initial configuration.
But after a while (a few hours), one or both of the nodes seems to leave the cluster. Browsing the GUI from node1, node2 is marked as 'offline' (red) but node2's stats are correctly displayed, expect the graphs which are empty. I can even connect to node2's VMs console from node1's GUI.
Connecting on node2's GUI expose the exact same problem but reversed.
My
Please note that I added the
Here are the versions of pve-related packages installed on both nodes:
and a different Membership information on node2:
As we can see, both nodes seem to not belong to the same cluster anymore.
While setting up the cluster (I re-installed it a few times) everything was working fine and
(More information below, but waiting for moderation approval)
I have set up a two nodes cluster following the wiki guidance, but I keep facing the same problem.
I'm able to see both nodes within the web GUI and spawn VMs while everything after my initial configuration.
But after a while (a few hours), one or both of the nodes seems to leave the cluster. Browsing the GUI from node1, node2 is marked as 'offline' (red) but node2's stats are correctly displayed, expect the graphs which are empty. I can even connect to node2's VMs console from node1's GUI.
Connecting on node2's GUI expose the exact same problem but reversed.
My
corosync
config is the following:
Code:
logging {
debug: off
to_syslog: yes
}
nodelist {
node {
name: rd-srv-front-pmx01
nodeid: 1
quorum_votes: 1
ring0_addr: 10.26.52.31
}
node {
name: rd-srv-front-pmx02
nodeid: 2
quorum_votes: 1
ring0_addr: 10.26.52.32
}
}
quorum {
provider: corosync_votequorum
two_node: 1
wait_for_all: 0
}
totem {
cluster_name: rd-front
config_version: 2
interface {
linknumber: 0
}
ip_version: ipv4-6
secauth: on
version: 2
}
Please note that I added the
two_node
and wait_for_all
flags.Here are the versions of pve-related packages installed on both nodes:
Code:
proxmox-ve: 6.0-2 (running kernel: 5.0.15-1-pve)
pve-manager: 6.0-4 (running version: 6.0-4/2a719255)
pve-kernel-5.0: 6.0-5
pve-kernel-helper: 6.0-5
pve-kernel-5.0.15-1-pve: 5.0.15-1
ceph-fuse: 12.2.11+dfsg1-2.1+b1
corosync: 3.0.2-pve2
criu: 3.11-3
glusterfs-client: 5.5-3
ksm-control-daemon: 1.3-1
libjs-extjs: 6.0.1-10
libknet1: 1.10-pve1
libpve-access-control: 6.0-2
libpve-apiclient-perl: 3.0-2
libpve-common-perl: 6.0-2
libpve-guest-common-perl: 3.0-1
libpve-http-server-perl: 3.0-2
libpve-storage-perl: 6.0-5
libqb0: 1.0.5-1
lvm2: 2.03.02-pve3
lxc-pve: 3.1.0-61
lxcfs: 3.0.3-pve60
novnc-pve: 1.0.0-60
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.0-5
pve-cluster: 6.0-4
pve-container: 3.0-3
pve-docs: 6.0-4
pve-edk2-firmware: 2.20190614-1
pve-firewall: 4.0-5
pve-firmware: 3.0-2
pve-ha-manager: 3.0-2
pve-i18n: 2.0-2
pve-qemu-kvm: 4.0.0-3
pve-xtermjs: 3.13.2-1
qemu-server: 6.0-5
smartmontools: 7.0-pve2
spiceterm: 3.1-1
vncterm: 1.6-1
zfsutils-linux: 0.8.1-pve1
pvecm status
returns the following output on node1:
Code:
Quorum information
------------------
Date: Tue Oct 22 11:23:55 2019
Quorum provider: corosync_votequorum
Nodes: 1
Node ID: 0x00000001
Ring ID: 1/654004
Quorate: Yes
Votequorum information
----------------------
Expected votes: 2
Highest expected: 2
Total votes: 1
Quorum: 1
Flags: 2Node Quorate
Membership information
----------------------
Nodeid Votes Name
0x00000001 1 10.26.52.31 (local)
and a different Membership information on node2:
Code:
Quorum information
------------------
Date: Tue Oct 22 11:25:42 2019
Quorum provider: corosync_votequorum
Nodes: 1
Node ID: 0x00000002
Ring ID: 2/828872
Quorate: Yes
Votequorum information
----------------------
Expected votes: 2
Highest expected: 2
Total votes: 1
Quorum: 1
Flags: 2Node Quorate
Membership information
----------------------
Nodeid Votes Name
0x00000002 1 10.26.52.32 (local)
As we can see, both nodes seem to not belong to the same cluster anymore.
While setting up the cluster (I re-installed it a few times) everything was working fine and
pvecm status
was showing the same status on both nodes (until the problem occurs).(More information below, but waiting for moderation approval)
Last edited: