Problems after updating cluster

grl

Renowned Member
Oct 20, 2011
24
1
68
[solved] Problems after updating cluster

Hi! I've set up a proxmox 2.0 cluster with drbd following http://pve.proxmox.com/wiki/DRBD using the first PVE 2.0 beta ISO.

After a while I updated to beta3 using apt-get update; apt-get dist-upgrade.
After that I had to set locking_type = 3 in /etc/lvm/lvm.conf on both nodes to get it up and running again. (Otherwise the drbd-storage would not get active).

Today I did another apt-get update; apt-get dist-upgrade and now drbd doesn't work again. Issuing /etc/init.d/clvm restart i get "Starting Cluster LVM Daemon: clvmclvmd could not connect to cluster manager" and in syslog:
Code:
Dec 26 19:04:29 vm02 kernel: dlm: Using TCP for communications 
Dec 26 19:04:29 vm02 dlm_controld[4797]: dlm_join_lockspace no fence domain 
Dec 26 19:04:29 vm02 dlm_controld[4797]: process_uevent online@ error -1 errno 11 
Dec 26 19:04:29 vm02 clvmd: Unable to create DLM lockspace for CLVM: Operation not permitted 
Dec 26 19:04:29 vm02 kernel: dlm: clvmd: group join failed -1 -1 
Dec 26 19:04:29 vm02 kernel: dlm: Using TCP for communications 
Dec 26 19:04:29 vm02 dlm_controld[4797]: dlm_join_lockspace no fence domain 
Dec 26 19:04:29 vm02 dlm_controld[4797]: process_uevent online@ error -1 errno 11 
Dec 26 19:04:29 vm02 kernel: dlm: clvmd: group join failed -1 -1 
Dec 26 19:04:29 vm02 clvmd: Unable to create DLM lockspace for CLVM: Operation not permitted

A pveversion -v reports:
Code:
pve-manager: 2.0-18 (pve-manager/2.0/16283a5a) 
running kernel: 2.6.32-6-pve 
proxmox-ve-2.6.32: 2.0-55 
pve-kernel-2.6.32-6-pve: 2.6.32-55 
lvm2: 2.02.88-2pve1 
clvm: 2.02.88-2pve1 
corosync-pve: 1.4.1-1 
openais-pve: 1.1.4-1 
libqb: 0.6.0-1 
redhat-cluster-pve: 3.1.8-3 
pve-cluster: 1.0-17 
qemu-server: 2.0-13 
pve-firmware: 1.0-14 
libpve-common-perl: 1.0-11 
libpve-access-control: 1.0-5 
libpve-storage-perl: 2.0-9 
vncterm: 1.0-2 
vzctl: 3.0.29-3pve8 
vzprocps: 2.0.11-2 
vzquota: 3.0.12-3 
pve-qemu-kvm: 1.0-1 
ksm-control-daemon: 1.1-1

AFAIK I need fencing only for HA features, in which I'm not interested so far.
I just want to get back what I already had: A cluster with drbd.

Although it's only used for testing I would not be very happy to loose all the testing machines installed on the cluster -

so, is there a way out?

regards Lukas
 
Last edited:
Re: [solved] Problems after updating cluster

Hi,

a "fence_tool join" on both nodes brought clvm up again.

I don't understand why this is necessary without any fencing device or fencing configuration, but it helped for the actual problem.

regards
Lukas
 
Re: [solved] Problems after updating cluster

We neither use clvm nor fenced by default in beta3.