Cluster Master can't access node

netbone

Member
Feb 5, 2009
93
0
6
I have added a node.

Both (cluster and node are working well).

Node is shown in clusters webpanel, but shown as offline:

Error: unable to get IP for node 'kemp2' - node offline? (500)

How to solve that?

Thank you.
 
Hi,
sorry not read the error messages right.
Do you have the right entry for kemp2 (and i guess for kemp1) in both /etc/hosts files?

Udo

Yes Udo. On both engines there is a etc hosts entry for both servers.
 
pvecm status

Code:
Version: 6.2.0
Config Version: 1
Cluster Name: kemp010
Cluster Id: 2970
Cluster Member: Yes
Cluster Generation: 4
Membership state: Cluster-Member
Nodes: 1
Expected votes: 1
Total votes: 1
Node votes: 1
Quorum: 1
Active subsystems: 5
Flags:
Ports Bound: 0
Node name: ks010
Node ID: 1
Multicast addresses: 39.92.11.165
Node addresses: 146.67.17.38

Code:
Version: 6.2.0
Config Version: 2
Cluster Name: kemp070
Cluster Id: 2982
Cluster Member: Yes
Cluster Generation: 4
Membership state: Cluster-Member
Nodes: 1
Expected votes: 1
Total votes: 1
Node votes: 1
Quorum: 1
Active subsystems: 5
Flags:
Ports Bound: 0
Node name: ks070
Node ID: 1
Multicast addresses: 39.92.11.177
Node addresses: 146.67.17.10
 
yep. this is what i did - and as result i have two clusters, but i want a cluster and a node.

Additional - running (!) - containers disappeared from webgui on kemp010

Additional in ssh all additional ip addresses disappeared when I take a look with ipconfig, but containers with not listed IPs are reachable and running.

vzlist (on kemp10)
Unable to open /etc/pve/openvz/100.conf: No such file or directory
Unable to open /etc/pve/openvz/101.conf: No such file or directory
Unable to open /etc/pve/openvz/102.conf: No such file or directory
CTID NPROC STATUS IP_ADDR HOSTNAME
100 46 running 10.19.100.11 -
101 53 running 10.19.100.17 -
102 93 running 10.19.100.50 -

root@kemp010 /etc/pve/openvz $ ls -all
total 0
drwxr-x--- 2 root www-data 0 May 20 18:32 .
drwxr-x--- 2 root www-data 0 May 20 18:32 ..
root@kemp010 /etc/pve/openvz $

root@kemp010 /etc/pve/openvz $ pvecm nodes
Node Sts Inc Joined Name
1 M 4 2012-05-19 23:22:17 kemp010
root@kemp010 /etc/pve/openvz $


root@kemp070:~# pvecm nodes
Node Sts Inc Joined Name
1 M 4 2012-05-19 23:17:42 kemp070
2 X 0 kemp010
root@kemp070:~#



root@kemp010 /etc/pve/openvz $ pvecm add kemp070 -force
The authenticity of host 'kemp070 (a.b.c.d)' can't be established.
RSA key fingerprint is xx:xx:xx:xx:xx:xx.
Are you sure you want to continue connecting (yes/no)? yes
node kemp010 already defined
copy corosync auth key
stopping pve-cluster service
Stopping pve cluster filesystem: pve-cluster.
backup old database
Starting pve cluster filesystem : pve-clusterfuse: failed to access mountpoint /etc/pve: Transport endpoint is not connected
[main] crit: fuse_mount error: Transport endpoint is not connected
[main] notice: exit proxmox configuration filesystem (-1)
(warning).
cannot stat initial working directory for /etc/pve/nodes/kemp010/openvz: Transport endpoint is not connected at /usr/bin/pvecm line 476
starting pve-cluster failed



root@kemp010 /etc/pve/openvz $ pvecm updatecerts
pve configuration filesystem not mounted



Completly confused now.

Simply want to have one cluster and one node.
 
Last edited:
hell - now vzlist is empty.

but all containers are still here:

drwxr-xr-x 20 root root 4096 May 6 00:05 100
drwxr-xr-x 20 root root 4096 May 6 09:21 101
drwxr-xr-x 21 root root 4096 May 10 16:15 102

how to start or reacitvate the containers now?
 
What output do you get when starting the cluster filesystem (pve-cluster) manually in debug mode:

# pmxcfs -f -d

Does it crash?