Problem creating a 2 nodes cluster with proxmox VE 2.3

  • Thread starter Thread starter Géraud Tillit
  • Start date Start date
G

Géraud Tillit

Guest
Hello,

I'm trying to create a 2 nodes cluster on proxmox VE 2.3, using unicast (my hosting company does not support multicast) on a dedicated network interface

Here are the steps I've done :

1. fresh installation of proxmox VE then :

Code:
root@node1:~# aptitude update && aptitude full-upgrade -y

2. configuration of /etc/hosts :

Code:
root@node1:~# cat /etc/hosts
127.0.0.1       localhost

88.190.xx.xx   sd-xxxxx.dedibox.fr node1ext
10.90.44.xx    node1 pvelocalhost
10.90.44.xx   node2
root@node1:~# hostname
node1



3. Creation of the cluster

Code:
root@node1:~# pvecm create dataexperience
Restarting pve cluster filesystem: pve-cluster[dcdb] notice: wrote new cluster config '/etc/cluster/cluster.conf'
.
Starting cluster:
   Checking if cluster has been disabled at boot... [  OK  ]
   Checking Network Manager... [  OK  ]
   Global setup... [  OK  ]
   Loading kernel modules... [  OK  ]
   Mounting configfs... [  OK  ]
   Starting cman... [  OK  ]
   Waiting for quorum... [  OK  ]
   Starting fenced... [  OK  ]
   Starting dlm_controld... [  OK  ]
   Tuning DLM kernel config... [  OK  ]
   Unfencing self... [  OK  ]

4. Modification of the cluster.conf

Code:
root@node1:~# cp /etc/pve/cluster.conf /etc/pve/cluster.conf.new
root@node1:~# vi /etc/pve/cluster.conf.new


Code:
<?xml version="1.0"?>
<cluster name="dataexperience" config_version="2">

  <cman keyfile="/var/lib/pve-cluster/corosync.authkey" transport="udpu" two_node="1" expected_votes="1">
  </cman>

  <clusternodes>
  <clusternode name="node1" votes="1" nodeid="1"/>
  </clusternodes>

</cluster>

Code:
root@node1:~# ccs_config_validate -v -f /etc/pve/cluster.conf.new
Creating temporary file: /tmp/tmp.BopYEiuGdz
Config interface set to:
Configuration stored in temporary file
Updating relaxng schema
Validating..
Configuration validates
Validation completed

Then I activated the cluster configuration through GUI

In order to ensure the new cluster configuration is taken into account, I rebooted the master (a bit overkilling) and following reboot :

Code:
root@node1:~# pvecm status
cman_tool: Cannot open connection to cman, is it running ?

Here is what I've seen in the log :

Code:
May  9 11:39:51 node1 pmxcfs[1457]: [dcdb] crit: cpg_initialize failed: 6
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:55 node1 pmxcfs[1457]: [status] crit: cpg_send_message failed: 9
May  9 11:39:57 node1 pmxcfs[1457]: [quorum] crit: quorum_initialize failed: 6
May  9 11:39:57 node1 pmxcfs[1457]: [confdb] crit: confdb_initialize failed: 6
May  9 11:39:57 node1 pmxcfs[1457]: [dcdb] crit: cpg_initialize failed: 6
May  9 11:39:57 node1 pmxcfs[1457]: [dcdb] crit: cpg_initialize failed: 6

That's my 3rd try with exactly the same result
 
Here is the output of pveversion :

Code:
root@node1:~# pveversion -v
pve-manager: 2.3-13 (pve-manager/2.3/7946f1f1)
running kernel: 2.6.32-19-pve
proxmox-ve-2.6.32: 2.3-95
pve-kernel-2.6.32-16-pve: 2.6.32-82
pve-kernel-2.6.32-19-pve: 2.6.32-95
lvm2: 2.02.95-1pve2
clvm: 2.02.95-1pve2
corosync-pve: 1.4.4-4
openais-pve: 1.1.4-2
libqb: 0.10.1-2
redhat-cluster-pve: 3.1.93-2
resource-agents-pve: 3.9.2-3
fence-agents-pve: 3.1.9-1
pve-cluster: 1.0-36
qemu-server: 2.3-20
pve-firmware: 1.0-21
libpve-common-perl: 1.0-49
libpve-access-control: 1.0-26
libpve-storage-perl: 2.3-7
vncterm: 1.0-4
vzctl: 4.0-1pve2
vzprocps: 2.0.11-2
vzquota: 3.1-1
pve-qemu-kvm: 1.4-10
ksm-control-daemon: 1.1-1
 
Code:
root@node1:~# cat /etc/hosts
127.0.0.1       localhost

88.190.xxx.xxx   xxx.dedibox.fr node1ext
10.90.44.xxx    node1 pvelocalhost
10.90.44.xxx    node2

Which is coherent with my hostname :

Code:
root@node1:~# hostname
node1

And with my interface configuration :

Code:
root@node1:~# ifconfig eth1
eth1      Link encap:Ethernet  HWaddr 28:92:4a:33:xx:xx
          inet adr:10.90.44.xxx  Bcast:10.90.44.xxx  Masque:255.255.255.xxx
          UP BROADCAST RUNNING MULTICAST  MTU:9000  Metric:1
          RX packets:34 errors:0 dropped:0 overruns:0 frame:0
          TX packets:5 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 lg file transmission:1000
          RX bytes:2620 (2.5 KiB)  TX bytes:810 (810.0 B)
          Interruption:17 Mémoire:fbce0000-fbd00000
 
I would suggest a sanity check at this time to isolate the problem whether this is Unicast vs Multicast issue.
I know you said your company does not allow multicast, possible to get a different switch and plug these servers into this different switch and do multicast?
 
I would suggest a sanity check at this time to isolate the problem whether this is Unicast vs Multicast issue.
I know you said your company does not allow multicast, possible to get a different switch and plug these servers into this different switch and do multicast?

Unfortunatly no, multicast is not supported on none of the 2 interfaces. That's why I'm trying unicast