Proxmox 7.1 UI Password Fails

Seed

Active Member
Oct 18, 2019
98
59
38
124
I setup a second node and temporarily added it to a cluster. I then removed it via the docs and am trying to login to it as a standalone. I can ssh as root fine but I can't login as root under PAM via the console UI. I have tried the standard things like resetting the passwd and such but it won't take. Any ideas?
 
Can you please post the output of pvecm status? As on a hunch I'd think that the cluster isn't yet separated cleanly and thus the node has no quorum.
 
Can you please post the output of pvecm status? As on a hunch I'd think that the cluster isn't yet separated cleanly and thus the node has no quorum.
I ran this on the node that I can't login to via the UI:

Cluster information
-------------------
Name: proxmox
Config Version: 2
Transport: knet
Secure auth: on

Quorum information
------------------
Date: Tue Dec 7 20:52:13 2021
Quorum provider: corosync_votequorum
Nodes: 1
Node ID: 0x00000002
Ring ID: 2.1d
Quorate: No

Votequorum information
----------------------
Expected votes: 2
Highest expected: 2
Total votes: 1
Quorum: 2 Activity blocked
Flags:

Membership information
----------------------
Nodeid Votes Name
0x00000002 1 192.168.7.241 (local)
 
Highest expected: 2
Total votes: 1
Quorum: 2 Activity blocked
My hunch was true, your node has still a cluster setup and missing the other one so it not quorate which explains your login issue.

What's the status on the other node that was previously in that cluster?
 
My hunch was true, your node has still a cluster setup and missing the other one so it not quorate which explains your login issue.

What's the status on the other node that was previously in that cluster?

Cluster information
-------------------
Name: proxmox
Config Version: 3
Transport: knet
Secure auth: on

Quorum information
------------------
Date: Tue Dec 7 22:32:26 2021
Quorum provider: corosync_votequorum
Nodes: 1
Node ID: 0x00000001
Ring ID: 1.d
Quorate: Yes

Votequorum information
----------------------
Expected votes: 1
Highest expected: 1
Total votes: 1
Quorum: 1
Flags: Quorate

Membership information
----------------------
Nodeid Votes Name
0x00000001 1 192.168.7.240 (local)
 
Ok, so that one is quorate and separations seems to have worked there, but they both have the same cluster name, that's not so ideal.

Now, I'd remove the remaining cluster config on the semi-broken node

Bash:
systemctl stop pve-cluster corosync
# start pmxcfs temporary in local mode to make it ignore the cluster config
pmxcfs -l
# remove cluster configs
rm -f /etc/pve/corosync.conf /etc/corosync/corosync.conf /etc/corosync/authkey
# stop local mode pmxcfs again
killall pmxcfs
# start it normally and cleanly via its service
systemctl start pve-cluster

With that it should work again, and the node is not in a broken cluster anymore, albeit there may still be some left-over configurations from the other node from the time they were in a cluster.
 
  • Like
Reactions: zeuxprox
Ok, so that one is quorate and separations seems to have worked there, but they both have the same cluster name, that's not so ideal.

Now, I'd remove the remaining cluster config on the semi-broken node

Bash:
systemctl stop pve-cluster corosync
# start pmxcfs temporary in local mode to make it ignore the cluster config
pmxcfs -l
# remove cluster configs
rm -f /etc/pve/corosync.conf /etc/corosync/corosync.conf /etc/corosync/authkey
# stop local mode pmxcfs again
killall pmxcfs
# start it normally and cleanly via its service
systemctl start pve-cluster

With that it should work again, and the node is not in a broken cluster anymore, albeit there may still be some left-over configurations from the other node from the time they were in a cluster.
You're the man. Thank you so much for this. Should be tagged or something.

There is one strange artifact however. When logging into the broken node that I ran the above on, it still has some metadata of the other node:

Screen Shot 2021-12-08 at 9.44.34 PM.png
 
Last edited:
There is one strange artifact however. When logging into the broken node that I ran the above on, it still has some metadata of the other node:
Yeah, that would be that part I described:
albeit there may still be some left-over configurations from the other node from the time they were in a cluster.
Basically the configuration files from VMs that where on the "pve" node are still there so the Proxmox VE web-interface needs to map them somewhere, it's just a pseudo entry.

You can remove it by deleting the configs from the other node (be sure you got the correct one), e.g.:
rm -rf /etc/pve/nodes/<NODENAME> replace <NODENAME> with the actual node, in your specific case that'd be pve

After that reload the web-interface completely.
 
Yeah, that would be that part I described:

Basically the configuration files from VMs that where on the "pve" node are still there so the Proxmox VE web-interface needs to map them somewhere, it's just a pseudo entry.

You can remove it by deleting the configs from the other node (be sure you got the correct one), e.g.:
rm -rf /etc/pve/nodes/<NODENAME> replace <NODENAME> with the actual node, in your specific case that'd be pve

After that reload the web-interface completely.
Great. Works a treat. Thank you so much!
 
Ok, so that one is quorate and separations seems to have worked there, but they both have the same cluster name, that's not so ideal.

Now, I'd remove the remaining cluster config on the semi-broken node

Bash:
systemctl stop pve-cluster corosync
# start pmxcfs temporary in local mode to make it ignore the cluster config
pmxcfs -l
# remove cluster configs
rm -f /etc/pve/corosync.conf /etc/corosync/corosync.conf /etc/corosync/authkey
# stop local mode pmxcfs again
killall pmxcfs
# start it normally and cleanly via its service
systemctl start pve-cluster

With that it should work again, and the node is not in a broken cluster anymore, albeit there may still be some left-over configurations from the other node from the time they were in a cluster.
THANK YOU! That little recipe is now saved in my notes!
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!