Problem with WEB GUI

silver_surfer9

Active Member
Jul 29, 2019
11
0
41
33
Hello,
I run a 3-node Proxmox VE cluster. Everything works (VMs run, storage is fine, quorum OK), but since last week the web GUI cluster tree visualization is broken.

1757484362954.png
pvecm status
Cluster information
-------------------
Name: proxmox-cluster
Config Version: 5
Transport: knet
Secure auth: on

Quorum information
------------------
Date: Wed Sep 10 09:01:36 2025
Quorum provider: corosync_votequorum
Nodes: 3
Node ID: 0x00000001
Ring ID: 1.53f
Quorate: Yes

Votequorum information
----------------------
Expected votes: 3
Highest expected: 3
Total votes: 3
Quorum: 2
Flags: Quorate

Membership information
----------------------
Nodeid Votes Name
0x00000001 1 192.168.100.248 (local)
0x00000002 1 192.168.100.249
0x00000003 1 192.168.100.247

proxmox-ve: 8.4.0 (running kernel: 6.8.12-14-pve)
pve-manager: 8.4.13 (running version: 8.4.13/5b08ebc2823dd9cb)

Any advice on what to check?
 
Check and restart -if it is not "active (running)"- pveproxy:
Code:
~# systemctl  status  pveproxy.service
~# systemctl  restart pveproxy.service
~# systemctl  status  pveproxy.service
 
It was in an error state. Now it's ok, but the web is still not good...
If I reduce the number of workers?
In /etc/default/pveproxy:
PVE_PROXY_WORKERS=2



root@pve1:~# systemctl status pveproxy.service
● pveproxy.service - PVE API Proxy Server
Loaded: loaded (/lib/systemd/system/pveproxy.service; enabled; preset: enabled)
Active: active (running) since Wed 2025-09-10 08:53:47 EEST; 2h 22min ago
Process: 5856 ExecStartPre=/usr/bin/pvecm updatecerts --silent (code=exited, status=0/SUCCESS)
Process: 5858 ExecStart=/usr/bin/pveproxy start (code=exited, status=0/SUCCESS)
Main PID: 5859 (pveproxy)
Tasks: 5 (limit: 76433)
Memory: 204.3M
CPU: 56.001s
CGroup: /system.slice/pveproxy.service
├─ 5859 pveproxy
├─12954 "pveproxy worker (shutdown)"
├─31103 "pveproxy worker"
├─31110 "pveproxy worker"
└─31166 "pveproxy worker"

Sep 10 11:08:02 pve1 pveproxy[5859]: worker 31110 started
Sep 10 11:08:02 pve1 pveproxy[31109]: got inotify poll request in wrong process - disabling inotify
Sep 10 11:08:02 pve1 pveproxy[31102]: got inotify poll request in wrong process - disabling inotify
Sep 10 11:08:05 pve1 pveproxy[31102]: worker exit
Sep 10 11:08:06 pve1 pveproxy[31109]: worker exit
Sep 10 11:08:20 pve1 pveproxy[5859]: worker 21970 finished
Sep 10 11:08:20 pve1 pveproxy[5859]: starting 1 worker(s)
Sep 10 11:08:20 pve1 pveproxy[5859]: worker 31166 started
Sep 10 11:08:22 pve1 pveproxy[31165]: got inotify poll request in wrong process - disabling inotify
Sep 10 11:08:24 pve1 pveproxy[31165]: worker exit
root@pve1:~# systemctl restart pveproxy.service
root@pve1:~# systemctl status pveproxy.service
● pveproxy.service - PVE API Proxy Server
Loaded: loaded (/lib/systemd/system/pveproxy.service; enabled; preset: enabled)
Active: active (running) since Wed 2025-09-10 11:18:36 EEST; 21s ago
Process: 32962 ExecStartPre=/usr/bin/pvecm updatecerts --silent (code=exited, status=0/SUCCESS)
Process: 32967 ExecStart=/usr/bin/pveproxy start (code=exited, status=0/SUCCESS)
Main PID: 32968 (pveproxy)
Tasks: 4 (limit: 76433)
Memory: 186.6M
CPU: 1.323s
CGroup: /system.slice/pveproxy.service
├─32968 pveproxy
├─32969 "pveproxy worker"
├─32970 "pveproxy worker"
└─32971 "pveproxy worker"

Sep 10 11:18:35 pve1 systemd[1]: Starting pveproxy.service - PVE API Proxy Server...
Sep 10 11:18:36 pve1 pveproxy[32968]: starting server
Sep 10 11:18:36 pve1 pveproxy[32968]: starting 3 worker(s)
Sep 10 11:18:36 pve1 pveproxy[32968]: worker 32969 started
Sep 10 11:18:36 pve1 pveproxy[32968]: worker 32970 started
Sep 10 11:18:36 pve1 pveproxy[32968]: worker 32971 started
Sep 10 11:18:36 pve1 systemd[1]: Started pveproxy.service - PVE API Proxy Server.
root@pve1:~# systemctl restart pveproxy
root@pve1:~# systemctl status pveproxy.service
● pveproxy.service - PVE API Proxy Server
Loaded: loaded (/lib/systemd/system/pveproxy.service; enabled; preset: enabled)
Active: active (running) since Wed 2025-09-10 11:19:41 EEST; 17s ago
Process: 33179 ExecStartPre=/usr/bin/pvecm updatecerts --silent (code=exited, status=0/SUCCESS)
Process: 33188 ExecStart=/usr/bin/pveproxy start (code=exited, status=0/SUCCESS)
Main PID: 33190 (pveproxy)
Tasks: 4 (limit: 76433)
Memory: 187.6M
CPU: 1.344s
CGroup: /system.slice/pveproxy.service
├─33190 pveproxy
├─33191 "pveproxy worker"
├─33192 "pveproxy worker"
└─33193 "pveproxy worker"

Sep 10 11:19:40 pve1 systemd[1]: Starting pveproxy.service - PVE API Proxy Server...
Sep 10 11:19:41 pve1 pveproxy[33190]: starting server
Sep 10 11:19:41 pve1 pveproxy[33190]: starting 3 worker(s)
Sep 10 11:19:41 pve1 pveproxy[33190]: worker 33191 started
Sep 10 11:19:41 pve1 pveproxy[33190]: worker 33192 started
Sep 10 11:19:41 pve1 pveproxy[33190]: worker 33193 started
Sep 10 11:19:41 pve1 systemd[1]: Started pveproxy.service - PVE API Proxy Server.
root@pve1:~#
 
The "?" in the GUI can indicate two things:
a) "pvestatd" daemon has trouble collecting information. This is usually related to storage misconfiguration (network or otherwise)
b) trouble connecting to that particular node

You can perform the following tests:
a) ssh from each node to all other nodes. Does it work?
b) run "pvesm status" on each node. Does it work? Does one take longer than others?
c) check the logs on each node: journalctl -n 100 ; journalctl -f


Blockbridge : Ultra low latency all-NVME shared storage for Proxmox - https://www.blockbridge.com/proxmox
 
  • Like
Reactions: UdoB
a) No problem with ssh connection.
b) On the problem node "pvesm status" freezes, and nothing shows. On the other nodes, it is ok.
c) This is on the problem node:

1757571393246.png


Do I need to reconnect the storage?
 
Last edited:
On the problem node "pvesm status" freezes, and nothing shows.
This is likely the source of your issue.
Do I need to reconnect the storage?
It is impossible to answer your question since we don't know anything about your storage.
Start with examining the /etc/pve/storage.cfg. Continue with making a backup of it, follow on with disabling/deleting all storage pools but one. Probe the storage. Keep testing with one pool at a time. Identify the problematic one/s.
Start troubleshooting why the probe of that storage/s fail.


Blockbridge : Ultra low latency all-NVME shared storage for Proxmox - https://www.blockbridge.com/proxmox
 
Ok, the problem is with one of the storage. We use this storage for a backup VM. When I remove it and reboot the node, everything works properly. Today I reconnected the storage, and again it is with a "?" mark. After 5 minutes, all problems appear again. On the other two nodes, it is fine. The storage is connected via NFS, and it has worked normally for 5+ years. What can cause this?
 
What can cause this?
You have a network problem. When investigating a network issue, there are three major components to consider: the Server (NFS), the Intermediate devices (NIC, cable, switch), and the Client (PVE).

PVE performs frequent health checks on NFS storage by running various RPC commands (e.g. rpcinfo, showmount, etc.). You would troubleshoot this in the same way as any other NFS client/server issue: check the logs, run the commands manually, verify MTUs, and capture a network trace if needed.


Blockbridge : Ultra low latency all-NVME shared storage for Proxmox - https://www.blockbridge.com/proxmox
 
  • Like
Reactions: Johannes S