Ok, I tried to get another try in this case.
I have another node that is for testing only and is not part of the productive cluster.
This node had also the same CIFS shares.
Same initial situation there:
Network drive dropped.
UI got unstable with the grey icons and no live data.
The old mount...
Now I removed the node 3 from the cluster in the hope that the web ui is working again.
Nope.
I can access node 3 via SSH and I see the web UI login screen but when I try to login, it fails.
Thanks for your answer!
Tried all three steps but none of them helped.
Still no direct WebUI login on node 3 and node 4
And no live VM data is shown for those two nodes.
And another problematic thing is that I need to restore a VM today and the backup is on the storage that can't be mounted...
to sum it up:
on 2 of 4 nodes the following problems appear:
1.) direct login on the web ui fails with an error: "Login failed. Please try again"
2.) cluster remote access works and it all green. Selecting one VM only configuration data shows up but the live data (CPU load, remote console, ...)...
Still, another problem:
Allthough I am able to access the terminal via the PVE UI of other nodes, I am not able to login directly to node 3 and node 4.
but both nodes are shown green now in the ui of node 1 and node 2.
But also the system tab can't be shown (communication failure (0) )
Maybe...
Found something else:
PVE has left the mount directorys under /mnt/pve/...
On good node I am able to access the (now empty) directroy by:
cd /mnt/pve/SwFileServer-cifs
On the other nodes the same command leads to a stuck terminal.
I can ls in /mnt/pve but when going in a subfolder the...
Now I had to remove the lock files on all 3 nodes but afterwards the pvestatd service could be started and now all nodes are green again.
However, without any CIFS or NFS share mounted.
Seems that mounting has been broken on 3 of 4 nodes.
or maybe a problem with the cluster syncing? but the...
Tested the same with the old NFS share.
Exactly the same results.
first working node:
other three nodes:
After waiting several minutes the whole nodes are also gone again.
3 of 4 nodes have grey question marks again.
1 is just completely green.
I guess I can do the same procedure again...
Hmmm...
re enabled the CIFS share from the Datacenter storage UI but it failed.
p-virt-sw-1 works
The other 3 nodes have the grey question mark and are inaccessible.
So the CIFS connection is still broken on 3 of 4 machines.
Took a long time but now I stopped and started the service on both machines and now everything seems to be fine again.
Without a reboot of any node.
Thanks a log Moayad!
Now I try to start the CIFS share again.
Let's see if this also works again.
Thanks.
I did so and now I fixed p-virt-sw-2.
So 2 of 4 nodes are completely up and running again.
However the other two just have the grey question mark and the UI is not updated any more.
But I still have terminal access.
But on these two machines pvestatd ist already running.
No idea how...
I guess this line could be of interest?
pvestatd[2893887]: start failed - can't acquire lock '/var/run/pvestatd.pid.lock' - Resource temporarily unavailable
manual mount:
not sure about that. I always used the PVE UI. I need to find out the equal command.
I only can access syslog for 2 of 4...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.