Thanks Udo and dcsapak. The issue seems to be related to bad RAM. I swapped the RAM and deleted the lxc containers that were misbehaving and the issue hasnt occurred since last Friday. Will close this thread.
So I noticed a strange behavior. When the crash does occur for the KVM's some LXC containers are running wild with CPU using the most CPU time. I cannot SSH into those LXC containers to see what's using up all processes. The only way I can get it to stop is to kill it from the web gui :8006...
This also affects windows machines. There is only one Ubuntu KVM and others are windows. I will be replacing some of the RAM sticks on that box by Monday and will see if that resolves the issue. With the latest update I applied today, the issue still occurred and when it does, the syslog shows...
Thanks for the response udo. I already applied the update. I have the VMs currently running and will update this thread. Those usually occurs with a VM crashing after about 8hrs so I should know more by this time tomorrow.
Thanks
I have finished applying the upgrade. Below is latest pveversion. I will monitor and see if that resolved the issue.
proxmox-ve: 4.1-39 (running kernel: 4.2.8-1-pve)
pve-manager: 4.1-22 (running version: 4.1-22/aca130cf)
pve-kernel-4.2.6-1-pve: 4.2.6-36
pve-kernel-4.2.8-1-pve: 4.2.8-39...
Thats correct that I upgraded using the below command
apt-get update
apt-get dist-upgrade
The patches I applied previously were from a post I saw regarding backups failing due to containers waiting for timeout. After I applied that patch, ti seemed to resolve the issue of the backups failing...
Hi dcsapak,
Thanks for the response. Yes I applied some update to the proxmox server yesterday hoping that would resolve the issue. Please see below for updated pveversion. Also, please see below for answers to your questions.
how much ram do you have in the server? 128GB
how much ram did you...
Also noticed after recent updates that some containers max out ram usage.
Mar 23 08:02:47 proxmox10 kernel: [47306.772442] Call Trace:
Mar 23 08:02:47 proxmox10 kernel: [47306.772623] [<ffffffff810c5b9c>] ? __rwsem_do_wake+0x10c/0x140
Mar 23 08:02:47 proxmox10 kernel: [47306.772726]...
Does this indicate RAM issues? I noticed the below in the syslog
{code}
Mar 23 07:36:06 proxmox10 kernel: [45705.955755] dmidecode.sudo invoked oom-killer: gfp_mask=0xd0, order=0, oom_score_adj=0
Mar 23 07:36:06 proxmox10 kernel: [45705.955766] dmidecode.sudo cpuset=110 mems_allowed=0
Mar 23...
Any ideas anyone please? I also did notice that sometimes, the lxc containers freeze and I cannot SSH into them and even when attempting to force stop the container, it won't respond for a while. Could this be due to bad RAM, hard drive corruption or some kernel bug?
Good Afternoon wonderful folks. This is my first post on the forums. Could use any assistance available. I have a single node Proxmox instance hosting Windows server KVMs, Linux KVMs and some linux containers. I started noticing an issue over the weekend. I would look at the proxmox web...
I also ran into the same issue on a DELL R900 server. After dist-upgrade and a reboot, I didnt have network access to be able to complete the upgrade. Thank goodness for backups. I just reinstalled 3.4 and restored.
Why not have disk on LVM thru ISCSI or NFS so all Nodes have access to the storage? And by enabling HA, if disk is on a SAN and host node fails, do other nodes automatically take over or do you have to boot node up?
I ran into a problem. Upgrade went successful except after I ran upgrade-from-grub-legacy and I guess I selected the wrong device and now Server wont boot up. I booted from a live Ubuntu USB disk in order to correct GRUB but when going into /boot/grub i'm not seeing menu.lst. Please help
I used to have similar issue with one windows guest randomly crashing. I ended up fixing it by making a snapshot backup and then deleting the guest and restoring from backup. Havent had that happen in a few months now.
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.