Problem permission /etc/pve/nodes

Jorge Fernando

New Member
Sep 21, 2017
6
0
1
38
La Habana, Cuba
www.facebook.com
In my proxmox 3.4 installation I can not write to the /etc/pve directory.


Code:
root@proxmox26:/etc/pve# pveversion
pve-manager/3.4-1/3f2d890e (running kernel: 2.6.32-37-pve)

Code:
root@proxmox26:/etc/pve# pvecm status
cman_tool: Cannot open connection to cman, is it running ?

Please Help!!!
 
Last edited:
  • Like
Reactions: Jorge Fernando
Thanks for your reply, t.lamprecht.

Answering your question from:
https://forum.proxmox.com/threads/problem-permission-etc-pve-nodes.26999/#post-185203

Code:
root@proxmox26:/etc/pve# service cman restart
Stopping cluster:
   Stopping dlm_controld... [  OK  ]
   Stopping fenced... [  OK  ]
   Stopping cman... [  OK  ]
   Unloading kernel modules... [  OK  ]
   Unmounting configfs... [  OK  ]
Starting cluster:
   Checking if cluster has been disabled at boot... [  OK  ]
   Checking Network Manager... [  OK  ]
   Global setup... [  OK  ]
   Loading kernel modules... [  OK  ]
   Mounting configfs... [  OK  ]
   Starting cman... /usr/sbin/ccs_config_validate: line 186: 969541 Segmentation fault      (core dumped) ccs_config_dump > $tempfile

Unable to get the configuration
corosync [MAIN  ] Corosync Cluster Engine ('1.4.7'): started and ready to provide service.
corosync [MAIN  ] Corosync built-in features: nss
corosync [MAIN  ] Successfully read config from /etc/cluster/cluster.conf
corosync died with signal: 11 Check cluster logs for details
[FAILED]

Code:
root@proxmox26:/etc/pve# service pve-cluster restart
Restarting pve cluster filesystem: pve-cluster.

Another additional information is that I have a cluster with two nodes and from the web interface I only see the main in green. The other is shown in red, proxmox26 is the main node.

This happened a few weeks ago and I think it is related to loss of information due to electrical failures.
 
Last edited:
Starting cman... /usr/sbin/ccs_config_validate: line 186: 969541 Segmentation fault (core dumped) ccs_config_dump > $tempfile

Hmm, can you post your cluster configuration please? "/etc/pve/cluster.conf"
 
  • Like
Reactions: Jorge Fernando
/etc/pve/cluster.conf

Code:
<?xml version="1.0"?>
<cluster name="clusterHP" config_version="19">

  <cman keyfile="/var/lib/pve-cluster/corosync.authkey">
  </cman>

  <clusternodes>

  <clusternode name="proxmox27" votes="1" nodeid="2"/></clusternodes>

</cluster>

After an incorrect shutdown, due to a blackout the file /etc/pve/cluster.conf was unconfigured and I always modified it and the problem of the red node was solved, but now as I can not write to the /etc/pve directory I can do it.

I know that eventually I will reinstall that cluster, but I want to solve this problem if it has viable solution.

The main problem is I can not change /etc/pve/storage.cfg, and I need to add a new Storage.
 
Last edited:
<clusternodes> <clusternode name="proxmox27" votes="1" nodeid="2"/></clusternodes> </cluster>

OK, there is just one cluster node in there, which is a bit strange as you said you've got an two node cluster.

Try the cman config validator "ccs_config_validate" manually? I'm not really into the PVE 3.4 cluster stack, so I do not see directly the problem here.

Else dismantle the cluster for now and rebuild it:
Code:
service pve-cluster stop
pmxcfs -l      # <- starts the cluster file system in local mode where it should be readable
rm /etc/pve/cluster.conf # (there should be also such a file under /etc/cman or /etc/corosync or/var/lib/pve-cluster/ (sorry I pull this info from top of my head))
rm /var/lib/pve-cluster/corosync.authkey
killall pmxcfs
service pve-cluster start

Do this first on one node only, then when /etc/pve is OK after that and contains still all configs/... on the other node too.
You may then rebuild the cluster with `pvecm create clusterHP` on one and `pvecm add <node> --force` on the other.
Good luck I'm off for for a few days, so I hope someone else can help you if ther's still need.
Oh, and I really suggest going for 5.1 now, you get a bit more saner cluster stack and almost three years free of (big) update worries :)
 
Last edited:
pvecm create clusterHP Error in main node (proxmox26)


Code:
root@proxmox26:/etc/pve# pvecm create clusterHP
Restarting pve cluster filesystem: pve-cluster[dcdb] crit: local cluster.conf is newer
.
Starting cluster:
   Checking if cluster has been disabled at boot... [  OK  ]
   Checking Network Manager... [  OK  ]
   Global setup... [  OK  ]
   Loading kernel modules... [  OK  ]
   Mounting configfs... [  OK  ]
   Starting cman... /usr/sbin/ccs_config_validate: line 186: 263606 Segmentation fault      ccs_config_dump > $tempfile

Unable to get the configuration
corosync [MAIN  ] Corosync Cluster Engine ('1.4.7'): started and ready to provide service.
corosync [MAIN  ] Corosync built-in features: nss
corosync [MAIN  ] Successfully read config from /etc/cluster/cluster.conf
corosync died with signal: 11 Check cluster logs for details
[FAILED]
root@proxmox26:/etc/pve#


2nd node proxmox27 , 10.10.88.10 is proxmox26

Code:
root@proxmox27:/mnt/pve/discoexterno# pvecm add --force 10.10.88.10
400 unable to parse boolean option
pvecm add <hostname> [OPTIONS]


without --force

Code:
root@proxmox27:/mnt/pve/discoexterno# pvecm add 10.10.88.10
this host already contains virtual machines - please remove the first

:(:eek: Help me please!!
 
Last edited:

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!