Delete node - Node IP still in corosync.conf in totem section?

Feb 14, 2024
7
0
1
Hi,
I have a little Proxmox 7.4 cluster I set up some time ago. Now my fist node died and I deleted it from the cluster. I plan to freshly install the node and add it to the cluster, but with the same name / IP address.

For that reason I checked every file in the /etc directory if there are any remains of the old hostname / ip address.

Interestingly, I still find the old IP address in corosync.conf (both in /etc/pve and /etc/corosync). The node is gone, but it is still in the totem section:

totem {
cluster_name: my_cluster
config_version: 11
interface {
bindnetaddr: 192.168.0.21
ringnumber: 0
}
ip_version: ipv4
secauth: on
version: 2
}

"192.168.0.21" is the IP address of the deleted node and is therefore unreachable.

Does anyone know why this (inexisten) IP address ist still in corosync.conf? Should I alter this IP address to some other Node IP?

Is there anything else to think about when I add the new node with the old hostname / IP Address? Maybe there are some other remains in /var/lib/pve-cluster/ I should be aware of?

Best Regards,
Hermann
 
What's the full /etc/pve/corosync.conf ?
Normally first remove any (lxc+vm) ha configs, then remove old node there, restart corosync (rewrites /etc/corosync/corosync.conf), then reconfig ha settings. Later reinstall your old pve node and rejoin it.
 
Thanks for the quick reply! Unfortunately my issue was that the node had a single SSD as root disk and this SSD died. So there was no way to remove the node while still running. And unfortunately, I have only the /etc directory of the dead node backed up and not /var/lib, so there is no way to reconstruct it.

I don't use HA at all as I am unsure if it would work properly (split brain scenarios etc.).

My full corosync.conf looks like this:

---------- snip ------------
logging {
debug: off
to_syslog: yes
}

nodelist {
node {
name: node2
nodeid: 2
quorum_votes: 1
ring0_addr: 192.168.0.22
}
node {
name: node3
nodeid: 3
quorum_votes: 1
ring0_addr: 192.168.0.23
}
node {
name: node4
nodeid: 4
quorum_votes: 1
ring0_addr: 192.168.0.24
}
node {
name: node5
nodeid: 5
quorum_votes: 1
ring0_addr: 192.168.0.25
}
node {
name: node6
nodeid: 6
quorum_votes: 1
ring0_addr: 192.168.0.26
}
node {
name: node7
nodeid: 7
quorum_votes: 1
ring0_addr: 192.168.0.27
}
node {
name: node8
nodeid: 8
quorum_votes: 1
ring0_addr: 192.168.0.28
}
}

quorum {
provider: corosync_votequorum
}

totem {
cluster_name: pmx1
config_version: 11
interface {
bindnetaddr: 192.168.0.21
ringnumber: 0
}
ip_version: ipv4
secauth: on
version: 2
}

---------- snip ------------
 
Looks nearly like ours but the end is different:
...
totem {
cluster_name: pvelab
config_version: 6
interface {
linknumber: 0
}
ip_version: ipv4-6
link_mode: passive
secauth: on
version: 2
}

So I would remove your line "bindnetaddr" and maybe exchange "ringnumber: 0" to "linknumber: 0", we have link_mode passive also.
One very IMPORTANT thing I forgot - you MUST do your changes, then switch your line "config_version: <+1>" - then save ( not save earlier) !!!
 
Last edited:
Thank you for your reply, I altered my config as suggested (but the /etc/pve/corosync.conf) and it seems the cluster keeps running flawlessly.

So, problem is solved.
 
  • Like
Reactions: waltar

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!