All OSD graphs are created from smartmon values including the temp. graph i posted earlier. I am pulling varieties of data such as read/write/sector errors etc. into graphs. Is it what you are talking about? So far it helped us identify 6 aging OSDs which probably would have gone unnoticed...
I am still using Zabbix and loving it! Over last few weeks i got to know more about Zabbix and added numerous networks to it. From monitoring to graphing to notification, all working just great. Recently i added Hard Drive temperature monitor for ALL drives with max temperature threshold. Now i...
I understand for Home environment you dont necessarily want to invest great deal of money on. Thats probably the reason you are going with 8 port switch. But this wont give you proper learning edge as you will be limited to few things. Such as Ceph will need a 2nd switch for proper setup for...
I have noticed that in the Proxmox corosync.conf secauth option is on by default. If i am not mistaking secauth has been deprecated. It is suggested to use crypto_cipher and crypto_hash instead.
Will this change in near future updates or should we manually use it without breaking the cluster?
Since you are looking at a 8 port switch i am assuming your cluster is extremely small with no plans of expansion in near future. I am fan of Netgear. My personal experience, i find Netgear interface much more polished and reliability rate much higher. For 8 port with features you are looking...
Nothing in dpkg.log.
Strangely like Ovidiu mine also started working this morning on its own. But as soon as pvedaemon started all VMs on the node had severe lag issue. So i had to force reboot the node. Now the node working just fine.
No doubt the issue is directly related to LXC. I will not...
The node only have 1 Ceph and 1 NFS shared storage. Both working as far as i can tell. VMs stored on Ceph are still running on the node. I can access the NFS share in /mnt/pve and see all the files. So storages are definitely working.
I interrupted the upgrade and ran dpkg --configure -a but...
My package upgrade seems to be stuck at
"Preparing to unpack .../lxc-pve_1.1.5-5_amd64.deb ..."
Should i Ctrl+Z and see if i can restart the upgrade process without breaking the node any further ? All VMs on the node still working and SSH access still available.
Ah i see. Glad to see it is a legit issue.
I agree with NFS becoming stale during the pvedaemon crash. It also crashes Ceph OSDs on the same node. Hopefully a patch will be released soon. Would like to give LXC real chance.
Unfortunately even after few days mine still not working on its own. Pvedaemon remains stubborn and wont start. All VMs on it still working though. I cant even migrate them to another node since cluster thinks the node is down.
Open for more suggestions. Really trying to figure this out without...
The command also showed me many lines of Close-Wait message on localhost.
Looks like Ovidiu has exactly the same problem. Mine too started after stopping a newly created LXC. Looks like i should not see Close-Wait messages as shown in the screenshot below:
Trying my best to find a solution...
Pvedaemon wont run from command line. It fails with instruction to see journalctl -xn for details. Following is the screenshot of the journalctl details:
The issue seems to be narrowing down to LXC container in my case. I was able to reproduce the error on a different node with no VMs. I tried to create a new LXC container on that node with storage on a Ceph KRBD enabled pool. The node got stuck on creating LXC killing services. The node can...
Hello,
While trying to stop a LXC container it crashed pvedaemon and pvestatd services. The node has several KVM VMs which seems to be running just fine but WebGUI showed the node is offline. I restarted pveproxy and pvestatd but when trying to restart pvedaemon it gives error
start failed -...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.