recently i have deployed allot of small Proxmox VE Clusters which now have a need for monitoring. I have used Influx and Grafana to monitor Containers and VM's in the past and was very surprised to see that the External Metric Server doesn't supply any Metrics about the Cluster status...
This has occurred from the later PVE 6.1 updates (I think) and definitely all throughout PVE 6.2 (including 6.2-12) since about March 2020. Prior to this the system was rock-solid for about 1.5 years.
Normal operation is ~10-12Gig usage (of 32G total). See attached picture for the cycle...
I've just added a 4th node to my existing proxmox cluster, and I'm running into an intermittent issue with the status of the node and VMs on that node turning to the gray question mark, storage stats not loading. When I run `systemctl status pvestatd` the service shows as running. When I...
Hello, I need some help please.
Today after deleting a volume from a storage array, and later the lvm from proxmox, the service pvestatd isn't working. If I enter to the web gui, I see everything greyed with question marks.
When I restart the service pvestatd it works for a minute, but later...
Yesterday, 3 storage units destined to NFS backup lost connection and their status are gone. With the intention of recovering those states again, they were remounted with the only command that allowed it: umount -f -l ... what happened next, is what I show in the attached picture...
Recently we upgrade 5.4 to 6.1 and replace ifupdown to ifupdown2.
After changes - pvestatd and pve-firewall doesn`t start after reboot, logs showing only - timeout.
If 'systemctl start pvestatd' invoked by hands - service starts.
Can someone help us - find the right direction to debug this ...
I have been having an issue with the web interface of proxmox showing my node and all vm/containers running on it as having an "unknown" status with grey ? marks.
This seems to happen a few hours after every reboot of the server. Restarting pvedaemon, pveproxy, pvestatd does not seem to...
Hi. I'm running into an issue where when I try to stop a container. In my syslog I get this error:
Aug 07 18:27:40 vmworld pvestatd: unable to get PID for CT 122 (not running?)
Aug 07 18:27:41 vmworld pvestatd: unable to get PID for CT 122 (not running?)
Aug 07 18:27:50 vmworld...
you might have the problem that you want to spin down some disks, but that doesn't work as planned.
I just spend 2 hours debugging the problem, and so decided to write this short tutorial.
You have multiple drives in your proxmox machines, and want some (like all HDDs) to spin down...
Mar 19 01:10:30 freddy pvestatd: lxc status update error: can't open '/sys/fs/cgroup/blkio/lxc/117/ns/blkio.throttle.io_service_bytes' - No such file or directory
Mar 19 01:10:34 freddy pvestatd: lxc console cleanup error: can't open...
Kind of an ongoing issue that I have been dealing with. It would seem that my pvestatd.service randomly crashes at or very near midnight on one of my nodes. Sometimes this causes only the daemon to crash, other times it will cause the node to become unresponsive until hard-reboot...
I've got an Archival HDD installed in my system, additionally to a RAID Z-1 out of 4x 500 GB SSDs.
The only issue is that the Archival HDD keeps getting waken up by pvestatd, when stopping the service it usally stays in sleep mode.
Any ideas how to fix this issue?
It appears that pvestatd doesn't track all the traffic on all the interfaces in the host summary page, only the "primary" interface.
Any pointers how to tell it to track all/other interfaces too?
Updated and rebooted pve. Now getting logs full of these. Sparse information on the googles. Most seems more oriented to someone in the rarefied zfs guru realms. Me? I'm not even sure where to start.
May 13 14:07:59 pve pvestatd: could not...
Hello, I'm adding pvestatd's stats into influxdb+grafana. status.cfg works with influxdb config with no problem, but I have several partial behaviors depending on (I guess) kernel version. With last version (4.4.49-1-pve) I have no diskio stats, and with previous 4.4.XX versions I have no stats...
I'm receiving this error message on /var/log/syslog of all nodes of my cluster:
This event repeats in brief or long intervals (1-30 minutes) and might happen every couple of days.
I believe it's impossible that the network is to be blamed. At most, this could be a problem on the NAS...
I got Proxmox running on my homeserver. Because I don't need the VM with my storage hdds (passtroughed to the VM) much often, I would like to spindown these hdds when they're not used.
I tried hdparm and hd-idle, but both are not able to spindown the hdds. Even if I try to spindown them...