Denied login to proxmox

bolar

New Member
Apr 23, 2021
2
0
1
55
Hi

i have a small cluster, only two servers with two/one votes each. I arranged it like that for easier administration.
Recently cluster rejects login to GUI. From other posts I've found out that pvescheduler is failing.

On server 1 status for pvescheduler is:
pvescheduler[1239003]: replication: cfs-lock 'file-replication_cfg' error: got lock request timeout
pvescheduler[1239004]: jobs: cfs-lock 'file-jobs_cfg' error: got lock request timeout

Server 2

pvescheduler[306092]: ERROR: can't lock file '/var/lock/pvesr.lck' - got timeout
pve2 pvescheduler[306092]: got shutdown request, signal running jobs to stop
pvescheduler[306092]: server stopped

Proxmox version is 7.1

After restart of both pvescheduler it's possible to login to server 1, not two.

Does anybody have an idea about this? Maybe delete cluster as a quick fix?
 
hi,

Does anybody have an idea about this? Maybe delete cluster as a quick fix?
i wouldn't do that

From other posts I've found out that pvescheduler is failing.

On server 1 status for pvescheduler is:
pvescheduler[1239003]: replication: cfs-lock 'file-replication_cfg' error: got lock request timeout
pvescheduler[1239004]: jobs: cfs-lock 'file-jobs_cfg' error: got lock request timeout

Server 2

pvescheduler[306092]: ERROR: can't lock file '/var/lock/pvesr.lck' - got timeout
pve2 pvescheduler[306092]: got shutdown request, signal running jobs to stop
pvescheduler[306092]: server stopped

Proxmox version is 7.1

After restart of both pvescheduler it's possible to login to server 1, not two.
it's possible that there's another underlying problem as to why the scheduler is failing. i'd suggest you to check the journals on your machines with journalctl -b0 since there might be other services that are also failing for one or the other reason.

i'd also check systemctl | grep fail and get the statuses of those services with journalctl -u <servicename>

hope this helps
 
Thanx for helping!

I have tried searching in log, can't find anything pointing to this issue. But I believe that server 2 started all this.

Anyway, I removed the cluster, seems to work fine now.