Error with dist-upgrade - pvestatd.service failed

odimarco

New Member
Nov 24, 2023
3
0
1
Hello,

a few days ago the hard disk of our proxmox server was full due to a snapshot that was too large.
Since we could no longer log in to the web frontend, we restarted the server.

We then deleted unnecessary log files on the cmd to free up at least 2 GB.
However, after restarting the server, the PVE services no longer ran.
The first thing we did was apt-get update && apt-get upgrade
This command was successful.
After restarting the server, the PVE services were still not running.
We then ran the command apt-get update && apt-get dist-upgrade.
However, this runs into an error:

Code:
Hit:1 http://asi-fs-d.contabo.net/debian bullseye InRelease
Hit:2 http://asi-fs-d.contabo.net/debian bullseye-updates InRelease
Hit:3 http://security.debian.org/debian-security bullseye-security InRelease
Hit:4 http://download.proxmox.com/debian/pue bullseye InRelease Beading package lists... Done Reading package lists... Done Building dependency tree... Done Reading state information... Done Calculating upgrade... Done The following package was automatically installed and is no longer required: pve-kernel-5.15.104-1-pve Use 'apt autoremoue' to remove it.
0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded. 1 not fully installed or removed.
After this operation, 0 B of additional disk space will be used. Do you want to continue? [Yin] Y
Setting up pve-manager (7.4-17) ... Job for pvestatd.service failed because the control process exited with error code.
See "systemctl status pvestatd.service" and "journalctl -xe" for details.
dpkg: error processing package pve-manager (--configure): installed pve-manager package post-installation script subprocess returned error exit status 1
Errors were encountered while processing: pve-manager E: Sub-process /usr/bin/dpkg returned an error code (1)

systemctl status pvestatd.service shows:

Code:
ipcc_send_rec failed: Connection refused
Unable to load access control list: Connection refused
pvestatd.service: Control process exited, code=exited, status=111/n/a

journalctl -xe shows some erros like:

Code:
/etc/pve/local/pve-ssl.key failed to laod local private key

and

Code:
cron[1409]: (*system*vzdump) Can't open symlink (/etc/cron.d/vzdump)

apt autoremove shows:

Code:
...
Warning: os-prober will not be executed to detect other bootable partitions.
Systems on them will not be added to the GRUB boot configuration.
Check GRUB_DISABLE_OS_PROBER documention entry.
done
Setting up pve-manager (7.4-17)
Job for pvestatd.service failed because the control process exited with error code.
dbkg: error processing package pve-manager (--configure):
installed pve-manager package post-installation script subprocess returned error exit status 1
Errors were encountered while processing:
pve-manager
E: Sub-process /usr/bin/dbkg returnded an error code (1)

We are really having problems with it and are getting nowhere. However, it must definitely have something to do with the hard disk overflowing, as everything always worked perfectly before.

I hope you can help me.

Best Regards
 
Last edited:
Update:
I have checked the /etc/pve directory.
This is empty!

journalctl -u pve-cluster shows:

Code:
systemd[1]: Failed to start The Proxmox VE cluster filesystem.
systemd[1]: pue-cluster.seruice: Scheduled restart job, restart counter is at 2.
systemd[1]: Stopped The Proxmox VE cluster filesystem.
systemd[1]: Starting The Proxmox VE cluster filesystem...
pmxcfs[1680]: fuse: mountpoint is not empty
pmxcfs[1680]: fuse: if you are sure this is safe, use the 'nonempty' mount option
pmxcfs[1680]: [main] crit: fuse_mount error: File exists
pmxcfs[1680]: [main] notice: exit proxmox configuration filesystem (-1)
pmxcfs[1680]: [main] crit: fuse_mount error: File exists
pmxcfs[1680]: [main] notice: exit proxmox configuration filesystem (-1)
systemd[1]: pue-cluster.seruice: Control process exited, code=exited, status=255/EXCEPR
systemd[1]: pue-cluster.seruice: Failed with result 'exit-code'.
systemd[1]: Failed to start The Proxmox VE cluster filesystem.
systemdEll: pue-cluster.seruice: Scheduled restart job, restart counter is at 3.
 
Last edited:
I did it myself.
the /etc/pve directory was not really empty. there was still no .swp file in it, which of course only became visible with ls -la. I deleted it and restarted the pve-cluster. the PVE directory was recreated and now the web interface appears again and everything works.