Error with Cluster System

SossenSystems

Member
Dec 13, 2017
5
0
6
26
hello,

i have two servers on one running Proxmox Virtual Environment 5.1-43 on the other Virtual Environment 5.1-46.
I wanted to connect the second server to server two so that I can see and manage the other node even on the first host.

root@vHost-2:~# pvecm add 164.132.***.*
The authenticity of host '164.132.205.* (164.132.***.*)' can't be established.
ECDSA key fingerprint is SHA256:OxLAFdQSI5OBiRw7kUbAnQ83Sj8GZhjyy8****P6o.
Are you sure you want to continue connecting (yes/no)? yes
root@164.132.20*.*'s password:
copy corosync auth key
stopping pve-cluster service
backup old database
Job for corosync.service failed because the control process exited with error code.
See "systemctl status corosync.service" and "journalctl -xe" for details.
waiting for quorum...
^C

That's vHost 1 the default cluster:
Bildschirmfoto vom 2018-03-19 22-17-51.png

Please help me :(
 
Code:
root@vHost-2:~# systemctl status corosync.service
● corosync.service - Corosync Cluster Engine
   Loaded: loaded (/lib/systemd/system/corosync.service; enabled; vendor preset:
   Active: failed (Result: exit-code) since Mon 2018-03-19 20:59:20 UTC; 23min a
     Docs: man:corosync
           man:corosync.conf
           man:corosync_overview
  Process: 1942 ExecStart=/usr/sbin/corosync -f $COROSYNC_OPTIONS (code=exited,
 Main PID: 1942 (code=exited, status=20)
      CPU: 48ms

Mär 19 20:59:20 vHost-2 corosync[1942]: info    [WD    ] no resources configured
Mär 19 20:59:20 vHost-2 corosync[1942]: notice  [SERV  ] Service engine loaded:
Mär 19 20:59:20 vHost-2 corosync[1942]: notice  [QUORUM] Using quorum provider c
Mär 19 20:59:20 vHost-2 corosync[1942]: crit    [QUORUM] Quorum provider: corosy
Mär 19 20:59:20 vHost-2 corosync[1942]: error   [SERV  ] Service engine 'corosyn
Mär 19 20:59:20 vHost-2 corosync[1942]: error   [MAIN  ] Corosync Cluster Engine
Mär 19 20:59:20 vHost-2 systemd[1]: corosync.service: Main process exited, code=
Mär 19 20:59:20 vHost-2 systemd[1]: Failed to start Corosync Cluster Engine.
Mär 19 20:59:20 vHost-2 systemd[1]: corosync.service: Unit entered failed state.
Mär 19 20:59:20 vHost-2 systemd[1]: corosync.service: Failed with result 'exit-c
 
I have a similar problem. 6 nodes, 4 connected normally and 2 not connecting. all completely virgin installs- only modification is kernel 4.15-3.

The two nodes that refuse to connect are Epyc based; the other 4 intel. logs show
Code:
Mar 20 14:04:05 nvme6 pmxcfs[1709961]: [quorum] crit: quorum_initialize failed: 2                                                     x
xMar 20 14:04:05 nvme6 pmxcfs[1709961]: [confdb] crit: cmap_initialize failed: 2                                                       x
xMar 20 14:04:05 nvme6 pmxcfs[1709961]: [dcdb] crit: cpg_initialize failed: 2                                                          x
xMar 20 14:04:05 nvme6 pmxcfs[1709961]: [status] crit: cpg_initialize failed: 2                                                        x

repeatedly following pvecm add. nodes are visible normally and ping normally, and are attached to a common switch.

pveversion -v, epyc system
Code:
proxmox-ve: 5.1-42 (running kernel: 4.15.3-1-pve)
pve-manager: 5.1-46 (running version: 5.1-46/ae8241d4)
pve-kernel-4.13: 5.1-42
pve-kernel-4.15: 5.1-1
pve-kernel-4.15.3-1-pve: 4.15.3-1
pve-kernel-4.13.13-6-pve: 4.13.13-42
pve-kernel-4.13.13-2-pve: 4.13.13-33
corosync: 2.4.2-pve3
criu: 2.11.1-1~bpo90
glusterfs-client: 3.8.8-1
ksm-control-daemon: 1.2-2
libjs-extjs: 6.0.1-2
libpve-access-control: 5.0-8
libpve-common-perl: 5.0-28
libpve-guest-common-perl: 2.0-14
libpve-http-server-perl: 2.0-8
libpve-storage-perl: 5.0-17
libqb0: 1.0.1-1
lvm2: 2.02.168-pve6
lxc-pve: 2.1.1-3
lxcfs: 2.0.8-2
novnc-pve: 0.6-4
openvswitch-switch: 2.7.0-2
proxmox-widget-toolkit: 1.0-11
pve-cluster: 5.0-20
pve-container: 2.0-19
pve-docs: 5.1-16
pve-firewall: 3.0-5
pve-firmware: 2.0-4
pve-ha-manager: 2.0-5
pve-i18n: 1.0-4
pve-libspice-server1: 0.12.8-3
pve-qemu-kvm: 2.9.1-9
pve-xtermjs: 1.0-2
qemu-server: 5.0-22
smartmontools: 6.5+svn4324-1
spiceterm: 3.0-5
vncterm: 1.5-3
zfsutils-linux: 0.7.6-pve1~bpo9

pveversion -v, intel system:
Code:
proxmox-ve: 5.1-42 (running kernel: 4.15.3-1-pve)
pve-manager: 5.1-46 (running version: 5.1-46/ae8241d4)
pve-kernel-4.13: 5.1-42
pve-kernel-4.15: 5.1-1
pve-kernel-4.15.3-1-pve: 4.15.3-1
pve-kernel-4.13.13-6-pve: 4.13.13-42
pve-kernel-4.13.13-2-pve: 4.13.13-33
corosync: 2.4.2-pve3
criu: 2.11.1-1~bpo90
glusterfs-client: 3.8.8-1
ksm-control-daemon: 1.2-2
libjs-extjs: 6.0.1-2
libpve-access-control: 5.0-8
libpve-common-perl: 5.0-28
libpve-guest-common-perl: 2.0-14
libpve-http-server-perl: 2.0-8
libpve-storage-perl: 5.0-17
libqb0: 1.0.1-1
lvm2: 2.02.168-pve6
lxc-pve: 2.1.1-3
lxcfs: 2.0.8-2
novnc-pve: 0.6-4
openvswitch-switch: 2.7.0-2
proxmox-widget-toolkit: 1.0-11
pve-cluster: 5.0-20
pve-container: 2.0-19
pve-docs: 5.1-16
pve-firewall: 3.0-5
pve-firmware: 2.0-4
pve-ha-manager: 2.0-5
pve-i18n: 1.0-4
pve-libspice-server1: 0.12.8-3
pve-qemu-kvm: 2.9.1-9
pve-xtermjs: 1.0-2
qemu-server: 5.0-22
smartmontools: 6.5+svn4324-1
spiceterm: 3.0-5
vncterm: 1.5-3
zfsutils-linux: 0.7.6-pve1~bpo9