Error with Cluster System

SossenSystems

Member
Dec 13, 2017
5
0
6
25
hello,

i have two servers on one running Proxmox Virtual Environment 5.1-43 on the other Virtual Environment 5.1-46.
I wanted to connect the second server to server two so that I can see and manage the other node even on the first host.

root@vHost-2:~# pvecm add 164.132.***.*
The authenticity of host '164.132.205.* (164.132.***.*)' can't be established.
ECDSA key fingerprint is SHA256:OxLAFdQSI5OBiRw7kUbAnQ83Sj8GZhjyy8****P6o.
Are you sure you want to continue connecting (yes/no)? yes
root@164.132.20*.*'s password:
copy corosync auth key
stopping pve-cluster service
backup old database
Job for corosync.service failed because the control process exited with error code.
See "systemctl status corosync.service" and "journalctl -xe" for details.
waiting for quorum...
^C

That's vHost 1 the default cluster:
Bildschirmfoto vom 2018-03-19 22-17-51.png

Please help me :(
 
Code:
root@vHost-2:~# systemctl status corosync.service
● corosync.service - Corosync Cluster Engine
   Loaded: loaded (/lib/systemd/system/corosync.service; enabled; vendor preset:
   Active: failed (Result: exit-code) since Mon 2018-03-19 20:59:20 UTC; 23min a
     Docs: man:corosync
           man:corosync.conf
           man:corosync_overview
  Process: 1942 ExecStart=/usr/sbin/corosync -f $COROSYNC_OPTIONS (code=exited,
 Main PID: 1942 (code=exited, status=20)
      CPU: 48ms

Mär 19 20:59:20 vHost-2 corosync[1942]: info    [WD    ] no resources configured
Mär 19 20:59:20 vHost-2 corosync[1942]: notice  [SERV  ] Service engine loaded:
Mär 19 20:59:20 vHost-2 corosync[1942]: notice  [QUORUM] Using quorum provider c
Mär 19 20:59:20 vHost-2 corosync[1942]: crit    [QUORUM] Quorum provider: corosy
Mär 19 20:59:20 vHost-2 corosync[1942]: error   [SERV  ] Service engine 'corosyn
Mär 19 20:59:20 vHost-2 corosync[1942]: error   [MAIN  ] Corosync Cluster Engine
Mär 19 20:59:20 vHost-2 systemd[1]: corosync.service: Main process exited, code=
Mär 19 20:59:20 vHost-2 systemd[1]: Failed to start Corosync Cluster Engine.
Mär 19 20:59:20 vHost-2 systemd[1]: corosync.service: Unit entered failed state.
Mär 19 20:59:20 vHost-2 systemd[1]: corosync.service: Failed with result 'exit-c
 
I have a similar problem. 6 nodes, 4 connected normally and 2 not connecting. all completely virgin installs- only modification is kernel 4.15-3.

The two nodes that refuse to connect are Epyc based; the other 4 intel. logs show
Code:
Mar 20 14:04:05 nvme6 pmxcfs[1709961]: [quorum] crit: quorum_initialize failed: 2                                                     x
xMar 20 14:04:05 nvme6 pmxcfs[1709961]: [confdb] crit: cmap_initialize failed: 2                                                       x
xMar 20 14:04:05 nvme6 pmxcfs[1709961]: [dcdb] crit: cpg_initialize failed: 2                                                          x
xMar 20 14:04:05 nvme6 pmxcfs[1709961]: [status] crit: cpg_initialize failed: 2                                                        x

repeatedly following pvecm add. nodes are visible normally and ping normally, and are attached to a common switch.

pveversion -v, epyc system
Code:
proxmox-ve: 5.1-42 (running kernel: 4.15.3-1-pve)
pve-manager: 5.1-46 (running version: 5.1-46/ae8241d4)
pve-kernel-4.13: 5.1-42
pve-kernel-4.15: 5.1-1
pve-kernel-4.15.3-1-pve: 4.15.3-1
pve-kernel-4.13.13-6-pve: 4.13.13-42
pve-kernel-4.13.13-2-pve: 4.13.13-33
corosync: 2.4.2-pve3
criu: 2.11.1-1~bpo90
glusterfs-client: 3.8.8-1
ksm-control-daemon: 1.2-2
libjs-extjs: 6.0.1-2
libpve-access-control: 5.0-8
libpve-common-perl: 5.0-28
libpve-guest-common-perl: 2.0-14
libpve-http-server-perl: 2.0-8
libpve-storage-perl: 5.0-17
libqb0: 1.0.1-1
lvm2: 2.02.168-pve6
lxc-pve: 2.1.1-3
lxcfs: 2.0.8-2
novnc-pve: 0.6-4
openvswitch-switch: 2.7.0-2
proxmox-widget-toolkit: 1.0-11
pve-cluster: 5.0-20
pve-container: 2.0-19
pve-docs: 5.1-16
pve-firewall: 3.0-5
pve-firmware: 2.0-4
pve-ha-manager: 2.0-5
pve-i18n: 1.0-4
pve-libspice-server1: 0.12.8-3
pve-qemu-kvm: 2.9.1-9
pve-xtermjs: 1.0-2
qemu-server: 5.0-22
smartmontools: 6.5+svn4324-1
spiceterm: 3.0-5
vncterm: 1.5-3
zfsutils-linux: 0.7.6-pve1~bpo9

pveversion -v, intel system:
Code:
proxmox-ve: 5.1-42 (running kernel: 4.15.3-1-pve)
pve-manager: 5.1-46 (running version: 5.1-46/ae8241d4)
pve-kernel-4.13: 5.1-42
pve-kernel-4.15: 5.1-1
pve-kernel-4.15.3-1-pve: 4.15.3-1
pve-kernel-4.13.13-6-pve: 4.13.13-42
pve-kernel-4.13.13-2-pve: 4.13.13-33
corosync: 2.4.2-pve3
criu: 2.11.1-1~bpo90
glusterfs-client: 3.8.8-1
ksm-control-daemon: 1.2-2
libjs-extjs: 6.0.1-2
libpve-access-control: 5.0-8
libpve-common-perl: 5.0-28
libpve-guest-common-perl: 2.0-14
libpve-http-server-perl: 2.0-8
libpve-storage-perl: 5.0-17
libqb0: 1.0.1-1
lvm2: 2.02.168-pve6
lxc-pve: 2.1.1-3
lxcfs: 2.0.8-2
novnc-pve: 0.6-4
openvswitch-switch: 2.7.0-2
proxmox-widget-toolkit: 1.0-11
pve-cluster: 5.0-20
pve-container: 2.0-19
pve-docs: 5.1-16
pve-firewall: 3.0-5
pve-firmware: 2.0-4
pve-ha-manager: 2.0-5
pve-i18n: 1.0-4
pve-libspice-server1: 0.12.8-3
pve-qemu-kvm: 2.9.1-9
pve-xtermjs: 1.0-2
qemu-server: 5.0-22
smartmontools: 6.5+svn4324-1
spiceterm: 3.0-5
vncterm: 1.5-3
zfsutils-linux: 0.7.6-pve1~bpo9
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!