Problem creating a 2 nodes cluster with proxmox VE 2.3

G

Géraud Tillit

Guest
Hello,

I'm trying to create a 2 nodes cluster on proxmox VE 2.3, using unicast (my hosting company does not support multicast) on a dedicated network interface

Here are the steps I've done :

1. fresh installation of proxmox VE then :

Code:
root@node1:~# aptitude update && aptitude full-upgrade -y

2. configuration of /etc/hosts :

Code:
root@node1:~# cat /etc/hosts
127.0.0.1       localhost

88.190.xx.xx   sd-xxxxx.dedibox.fr node1ext
10.90.44.xx    node1 pvelocalhost
10.90.44.xx   node2
root@node1:~# hostname
node1



3. Creation of the cluster

Code:
root@node1:~# pvecm create dataexperience
Restarting pve cluster filesystem: pve-cluster[dcdb] notice: wrote new cluster config '/etc/cluster/cluster.conf'
.
Starting cluster:
   Checking if cluster has been disabled at boot... [  OK  ]
   Checking Network Manager... [  OK  ]
   Global setup... [  OK  ]
   Loading kernel modules... [  OK  ]
   Mounting configfs... [  OK  ]
   Starting cman... [  OK  ]
   Waiting for quorum... [  OK  ]
   Starting fenced... [  OK  ]
   Starting dlm_controld... [  OK  ]
   Tuning DLM kernel config... [  OK  ]
   Unfencing self... [  OK  ]

4. Modification of the cluster.conf

Code:
root@node1:~# cp /etc/pve/cluster.conf /etc/pve/cluster.conf.new
root@node1:~# vi /etc/pve/cluster.conf.new


Code:
<?xml version="1.0"?>
<cluster name="dataexperience" config_version="2">

  <cman keyfile="/var/lib/pve-cluster/corosync.authkey" transport="udpu" two_node="1" expected_votes="1">
  </cman>

  <clusternodes>
  <clusternode name="node1" votes="1" nodeid="1"/>
  </clusternodes>

</cluster>

Code:
root@node1:~# ccs_config_validate -v -f /etc/pve/cluster.conf.new
Creating temporary file: /tmp/tmp.BopYEiuGdz
Config interface set to:
Configuration stored in temporary file
Updating relaxng schema
Validating..
Configuration validates
Validation completed

Then I activated the cluster configuration through GUI

In order to ensure the new cluster configuration is taken into account, I rebooted the master (a bit overkilling) and following reboot :

Code:
root@node1:~# pvecm status
cman_tool: Cannot open connection to cman, is it running ?

Here is what I've seen in the log :

Code:
May  9 11:39:51 node1 pmxcfs[1457]: [dcdb] crit: cpg_initialize failed: 6
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:57 node1 pmxcfs[1457]: [quorum] crit: quorum_initialize failed: 6
May  9 11:39:57 node1 pmxcfs[1457]: [confdb] crit: confdb_initialize failed: 6
May  9 11:39:57 node1 pmxcfs[1457]: [dcdb] crit: cpg_initialize failed: 6
May  9 11:39:57 node1 pmxcfs[1457]: [dcdb] crit: cpg_initialize failed: 6

That's my 3rd try with exactly the same result
 
Here is the output of pveversion :

Code:
root@node1:~# pveversion -v
pve-manager: 2.3-13 (pve-manager/2.3/7946f1f1)
running kernel: 2.6.32-19-pve
proxmox-ve-2.6.32: 2.3-95
pve-kernel-2.6.32-16-pve: 2.6.32-82
pve-kernel-2.6.32-19-pve: 2.6.32-95
lvm2: 2.02.95-1pve2
clvm: 2.02.95-1pve2
corosync-pve: 1.4.4-4
openais-pve: 1.1.4-2
libqb: 0.10.1-2
redhat-cluster-pve: 3.1.93-2
resource-agents-pve: 3.9.2-3
fence-agents-pve: 3.1.9-1
pve-cluster: 1.0-36
qemu-server: 2.3-20
pve-firmware: 1.0-21
libpve-common-perl: 1.0-49
libpve-access-control: 1.0-26
libpve-storage-perl: 2.3-7
vncterm: 1.0-4
vzctl: 4.0-1pve2
vzprocps: 2.0.11-2
vzquota: 3.1-1
pve-qemu-kvm: 1.4-10
ksm-control-daemon: 1.1-1
 
Code:
root@node1:~# cat /etc/hosts
127.0.0.1       localhost

88.190.xxx.xxx   xxx.dedibox.fr node1ext
10.90.44.xxx    node1 pvelocalhost
10.90.44.xxx    node2

Which is coherent with my hostname :

Code:
root@node1:~# hostname
node1

And with my interface configuration :

Code:
root@node1:~# ifconfig eth1
eth1      Link encap:Ethernet  HWaddr 28:92:4a:33:xx:xx
          inet adr:10.90.44.xxx  Bcast:10.90.44.xxx  Masque:255.255.255.xxx
          UP BROADCAST RUNNING MULTICAST  MTU:9000  Metric:1
          RX packets:34 errors:0 dropped:0 overruns:0 frame:0
          TX packets:5 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 lg file transmission:1000
          RX bytes:2620 (2.5 KiB)  TX bytes:810 (810.0 B)
          Interruption:17 Mémoire:fbce0000-fbd00000
 
I would suggest a sanity check at this time to isolate the problem whether this is Unicast vs Multicast issue.
I know you said your company does not allow multicast, possible to get a different switch and plug these servers into this different switch and do multicast?
 
I would suggest a sanity check at this time to isolate the problem whether this is Unicast vs Multicast issue.
I know you said your company does not allow multicast, possible to get a different switch and plug these servers into this different switch and do multicast?

Unfortunatly no, multicast is not supported on none of the 2 interfaces. That's why I'm trying unicast
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!