Hello,
I have a problem creating a proxmox cluster on hetzner bare metal machines. I tried everything, but the process is getting blocked for some reason. I will provide the details of my configuration bellow and the errors that I get.
Server 1 network config :
# network interface settings; autogenerated
# Please do NOT modify this file directly, unless you know what
# you're doing.
source /etc/network/interfaces.d/*
auto lo
iface lo inet loopback
iface lo inet6 loopback
iface eno1 inet manual
auto vmbr0
iface vmbr0 inet static
address xxx.xxx.xxx.xxx
netmask xxx.xxx.xxx.xxx
gateway xxx.xxx.xxx.xxx
bridge-ports eno1
bridge-stp off
bridge-fd 0
up sysctl -p
up sysctl -w net.ipv4.ip_forward=1
up sysctl -w net.ipv4.conf.eno1.send_redirects=0
up sysctl -w net.ipv6.conf.all.forwarding=1
post-up echo 2048 > /sys/class/net/vmbr0/bridge/hash_max
post-up echo 1 > /sys/class/net/vmbr0/bridge/multicast_snooping
post-up echo 0 > /proc/sys/net/ipv6/conf/vmbr0/accept_ra
#vlan between nodes
auto vmbr4001
iface vmbr4001 inet static
bridge_ports eno1.4001
bridge_stp off
bridge_fd 0
address 10.0.100.10
netmask 24
#COROSYNC1
iface eth0 inet manual
iface eth1 inet manual
in /etc/hosts for server 1 I have - 10.0.100.10 pve1.mydomain.mydomain pve1
Server 2 config
source /etc/network/interfaces.d/*
auto lo
iface lo inet loopback
iface lo inet6 loopback
iface enp4s0 inet manual
auto vmbr0
iface vmbr0 inet static
address xxx
netmask xxx
gateway xxx
bridge-ports enp4s0
bridge-stp off
bridge-fd 0
up sysctl -p
up sysctl -w net.ipv4.ip_forward=1
up sysctl -w net.ipv4.conf.enp4s0.send_redirects=0
up sysctl -w net.ipv6.conf.all.forwarding=1
post-up echo 2048 > /sys/class/net/vmbr0/bridge/hash_max
post-up echo 1 > /sys/class/net/vmbr0/bridge/multicast_snooping
post-up echo 0 > /proc/sys/net/ipv6/conf/vmbr0/accept_ra
#vlan between nodes
auto vmbr4001
iface vmbr4001 inet static
bridge_ports enp4s0.4001
bridge_stp off
bridge_fd 0
address 10.0.100.11
netmask 24
#COROSYNC1
iface eth0 inet manual
The clock is in sync and show the same values on both nodes.
I creat the cluster on server 1.
in /etc/hosts for server 2 I have - 10.0.100.11 pve2.mydomain.mydomain pve2
I am creating the cluster from command line :
root@pve2 ~ # pvecm create prod
perl: warning: Setting locale failed.
perl: warning: Please check that your locale settings:
LANGUAGE = (unset),
LC_ALL = (unset),
LC_CTYPE = "UTF-8",
LANG = "en_US.UTF-8"
are supported and installed on your system.
perl: warning: Falling back to a fallback locale ("en_US.UTF-8").
Corosync Cluster Engine Authentication key generator.
Gathering 2048 bits for key from /dev/urandom.
Writing corosync key to /etc/corosync/authkey.
Writing corosync config to /etc/pve/corosync.conf
Restart corosync and cluster filesystem
root@pve2 ~ # pvecm status
perl: warning: Setting locale failed.
perl: warning: Please check that your locale settings:
LANGUAGE = (unset),
LC_ALL = (unset),
LC_CTYPE = "UTF-8",
LANG = "en_US.UTF-8"
are supported and installed on your system.
perl: warning: Falling back to a fallback locale ("en_US.UTF-8").
Cluster information
-------------------
Name: prod
Config Version: 1
Transport: knet
Secure auth: on
Quorum information
------------------
Date: Mon Mar 18 21:05:47 2024
Quorum provider: corosync_votequorum
Nodes: 1
Node ID: 0x00000001
Ring ID: 1.5
Quorate: Yes
Votequorum information
----------------------
Expected votes: 1
Highest expected: 1
Total votes: 1
Quorum: 1
Flags: Quorate
Membership information
----------------------
Nodeid Votes Name
0x00000001 1 10.0.100.11 (local)
Now i try to connect a node. First i try via cli :
pvecm add 10.0.100.11 - this command fails with :
500 Can't connect to 10.0.100.11:8006 (hostname verification failed)
end task UPIDve5:00000BF0:00011642:65F89F26:clusterjoin::root@pam: 500 Can't connect to 10.0.100.11:8006 (hostname verification failed)
Then i try from the GUI using join cluster. First i get :
Establishing API connection with host '10.0.100.11'
Login succeeded.
check cluster join API version
Request addition of this node
Join request OK, finishing setup locally
stopping pve-cluster service
Then on the node i get : permission denied - invalid PVE ticket (401) and now i cannot access the proxmox GUI, i cannot login. I can access the server via ssh but not using the ssh key, using the password. So, something overwrite my ssh key on the node.
An on the master, i get the error : '/etc/pve/nodes/pve5/pve-ssl.pem' does not exist! (500) because the files are not in /etc/pve...
What can i do to fix this issue? Any ideas, tried everything from the internet.
regards.
I have a problem creating a proxmox cluster on hetzner bare metal machines. I tried everything, but the process is getting blocked for some reason. I will provide the details of my configuration bellow and the errors that I get.
Server 1 network config :
# network interface settings; autogenerated
# Please do NOT modify this file directly, unless you know what
# you're doing.
source /etc/network/interfaces.d/*
auto lo
iface lo inet loopback
iface lo inet6 loopback
iface eno1 inet manual
auto vmbr0
iface vmbr0 inet static
address xxx.xxx.xxx.xxx
netmask xxx.xxx.xxx.xxx
gateway xxx.xxx.xxx.xxx
bridge-ports eno1
bridge-stp off
bridge-fd 0
up sysctl -p
up sysctl -w net.ipv4.ip_forward=1
up sysctl -w net.ipv4.conf.eno1.send_redirects=0
up sysctl -w net.ipv6.conf.all.forwarding=1
post-up echo 2048 > /sys/class/net/vmbr0/bridge/hash_max
post-up echo 1 > /sys/class/net/vmbr0/bridge/multicast_snooping
post-up echo 0 > /proc/sys/net/ipv6/conf/vmbr0/accept_ra
#vlan between nodes
auto vmbr4001
iface vmbr4001 inet static
bridge_ports eno1.4001
bridge_stp off
bridge_fd 0
address 10.0.100.10
netmask 24
#COROSYNC1
iface eth0 inet manual
iface eth1 inet manual
in /etc/hosts for server 1 I have - 10.0.100.10 pve1.mydomain.mydomain pve1
Server 2 config
source /etc/network/interfaces.d/*
auto lo
iface lo inet loopback
iface lo inet6 loopback
iface enp4s0 inet manual
auto vmbr0
iface vmbr0 inet static
address xxx
netmask xxx
gateway xxx
bridge-ports enp4s0
bridge-stp off
bridge-fd 0
up sysctl -p
up sysctl -w net.ipv4.ip_forward=1
up sysctl -w net.ipv4.conf.enp4s0.send_redirects=0
up sysctl -w net.ipv6.conf.all.forwarding=1
post-up echo 2048 > /sys/class/net/vmbr0/bridge/hash_max
post-up echo 1 > /sys/class/net/vmbr0/bridge/multicast_snooping
post-up echo 0 > /proc/sys/net/ipv6/conf/vmbr0/accept_ra
#vlan between nodes
auto vmbr4001
iface vmbr4001 inet static
bridge_ports enp4s0.4001
bridge_stp off
bridge_fd 0
address 10.0.100.11
netmask 24
#COROSYNC1
iface eth0 inet manual
The clock is in sync and show the same values on both nodes.
I creat the cluster on server 1.
in /etc/hosts for server 2 I have - 10.0.100.11 pve2.mydomain.mydomain pve2
I am creating the cluster from command line :
root@pve2 ~ # pvecm create prod
perl: warning: Setting locale failed.
perl: warning: Please check that your locale settings:
LANGUAGE = (unset),
LC_ALL = (unset),
LC_CTYPE = "UTF-8",
LANG = "en_US.UTF-8"
are supported and installed on your system.
perl: warning: Falling back to a fallback locale ("en_US.UTF-8").
Corosync Cluster Engine Authentication key generator.
Gathering 2048 bits for key from /dev/urandom.
Writing corosync key to /etc/corosync/authkey.
Writing corosync config to /etc/pve/corosync.conf
Restart corosync and cluster filesystem
root@pve2 ~ # pvecm status
perl: warning: Setting locale failed.
perl: warning: Please check that your locale settings:
LANGUAGE = (unset),
LC_ALL = (unset),
LC_CTYPE = "UTF-8",
LANG = "en_US.UTF-8"
are supported and installed on your system.
perl: warning: Falling back to a fallback locale ("en_US.UTF-8").
Cluster information
-------------------
Name: prod
Config Version: 1
Transport: knet
Secure auth: on
Quorum information
------------------
Date: Mon Mar 18 21:05:47 2024
Quorum provider: corosync_votequorum
Nodes: 1
Node ID: 0x00000001
Ring ID: 1.5
Quorate: Yes
Votequorum information
----------------------
Expected votes: 1
Highest expected: 1
Total votes: 1
Quorum: 1
Flags: Quorate
Membership information
----------------------
Nodeid Votes Name
0x00000001 1 10.0.100.11 (local)
Now i try to connect a node. First i try via cli :
pvecm add 10.0.100.11 - this command fails with :
500 Can't connect to 10.0.100.11:8006 (hostname verification failed)
end task UPIDve5:00000BF0:00011642:65F89F26:clusterjoin::root@pam: 500 Can't connect to 10.0.100.11:8006 (hostname verification failed)
Then i try from the GUI using join cluster. First i get :
Establishing API connection with host '10.0.100.11'
Login succeeded.
check cluster join API version
Request addition of this node
Join request OK, finishing setup locally
stopping pve-cluster service
Then on the node i get : permission denied - invalid PVE ticket (401) and now i cannot access the proxmox GUI, i cannot login. I can access the server via ssh but not using the ssh key, using the password. So, something overwrite my ssh key on the node.
An on the master, i get the error : '/etc/pve/nodes/pve5/pve-ssl.pem' does not exist! (500) because the files are not in /etc/pve...
What can i do to fix this issue? Any ideas, tried everything from the internet.
regards.
Last edited: