Error with adding a new node in the cluster.

Alexandros85

New Member
Feb 3, 2024
2
0
1
Hello, I'm facing the problem with adding the new node to existing PVE cluster. After adding the node through the web-interface, it shows as offline:1767746884393.png
And on the new node all other nodes are shown offline:
1767746942642.png

And the error I'm getting is:
1767746991544.png

I've tried to update certificates on the new node, but I'm getting an error:

Code:
dc1pve6# pvecm updatecerts -F
waiting for pmxcfs mount to appear and get quorate...
waiting for pmxcfs mount to appear and get quorate...
waiting for pmxcfs mount to appear and get quorate...
waiting for pmxcfs mount to appear and get quorate...
waiting for pmxcfs mount to appear and get quorate...
waiting for pmxcfs mount to appear and get quorate...
got timeout when trying to ensure cluster certificates and base file hierarchy is set up - no quorum (yet) or hung pmxcfs?

Pvecm status on existing node:

Code:
dc1pve6# pvecm status
Cluster information
-------------------
Name:             dc1
Config Version:   9
Transport:        knet
Secure auth:      on

Quorum information
------------------
Date:             Tue Jan  6 16:52:11 2026
Quorum provider:  corosync_votequorum
Nodes:            4
Node ID:          0x00000004
Ring ID:          1.48d
Quorate:          Yes

Votequorum information
----------------------
Expected votes:   5
Highest expected: 5
Total votes:      4
Quorum:           3
Flags:            Quorate

Membership information
----------------------
    Nodeid      Votes Name
0x00000001          1 10.0.1.2
0x00000002          1 10.0.1.1
0x00000003          1 10.0.1.3
0x00000004          1 10.0.1.5 (local)

pvecm status on new node:
Code:
dc1pve6# pvecm status
Cluster information
-------------------
Name:             dc1
Config Version:   9
Transport:        knet
Secure auth:      on

Quorum information
------------------
Date:             Tue Jan  6 16:53:29 2026
Quorum provider:  corosync_votequorum
Nodes:            1
Node ID:          0x00000005
Ring ID:          5.14
Quorate:          No

Votequorum information
----------------------
Expected votes:   5
Highest expected: 5
Total votes:      1
Quorum:           3 Activity blocked
Flags:

Membership information
----------------------
    Nodeid      Votes Name
0x00000005          1 10.0.1.6 (local)

Any suggestion is going to be highly appreciated, thank you!
 
Your network is, for a lack of a better word, broken.

are you using bonds for your corosync interface(s)? do the individual interfaces have a path to ALL members of bonds on the OTHER nodes?
Yes, we are using bonds for our corosync interfaces. We have 5 nodes with the similar network settings: 1 Gb 10.0.0.255 network for cluster switching, and 10.0.2.255 10 Gb network for ceph. 4 nodes are working fine.
 
Yes, we are using bonds for our corosync interfaces.
yeah that was pretty much a given with your problem description :)

4 nodes are working fine.
They probably arent; you're just not aware of the problem because their active bond interfaces all connect to the same switch.

1st order of business- get your switches to talk to each other. 2nd- dont use bonds for corosync traffic; use 2 rings with a single interface each.
 
  • Like
Reactions: Johannes S