Thanks guys. Since I'm not getting any "me too" responses, I'm guessing this is not a Proxmox issue, but something being configured wrong on our network. I'm moving this discussion to ServerFault: http://serverfault.com/questions/777103/linux-adding-a-bad-route-on-startup
And since posting...
I've got several Proxmox 3.1 machines that have a peculiar problem. After booting them up, an extra route is created that has to be deleted before the machine can talk to others on the same subnet. After booting, the routes look like this (public IPs masked):
# route
Kernel IP routing table...
I got distracted for a bit and am finally getting back to this.
Are you saying that lxc.mount.auto can be used in an openvz configuration file? In my googling, I'm only seeing examples of it being used in lxc container configs.
There must be some way to mount /proc in a secure way in...
Recently I discovered that the default /proc mount inside CentOS containers running under Proxmox 3.1 were mounted insecurely:
# mount|grep " /proc"
proc on /proc type proc (rw,relatime)
Under older versions of Proxmox, I was able to correct this by running:
mount -o remount,nosuid,noexec...
A quick follow up here... I was successfully able to get this done, but there was definitely more to it than what I originally thought. Here's roughly what I did:
Run from Node:
Remove the master's access from the node:
mv /root/.ssh/authorized_keys /root/.ssh/authorized_keys_proxmox
# save...
Yes, our main switch is managed, and for some reason it doesn't happen there. But we have sub-groups of servers that we wanted on private switches and we didn't need managed switches for them. If we had time to replace the switches with managed switches, we'd do that, but we don't have time for...
Thanks for the tip on how to upgrade the kernel. Since I'm not finding any evidence that another kernel upgrade will help, my first priority is to get these machines out of the cluster so it doesn't happen again. I've started a separate thread for that...
Due instability issues we're having with a Proxmox 3.1 cluster, we need to remove several servers from the cluster without migrating the OpenVZ containers off of them. This thread is not about the instability of the cluster, if you want to read about that, go here...
So, if I understand the link you provided, if we want to upgrade our kernel, we have to use the non-production "test" repo or their subscription based one. We don't have time for testing. We need a stable platform and I'm not interested in purchasing a subscription until after we've determined...
I thought Proxmox 3.1 would have included the most recent stable kernel. Where do I get the latest kernel? Also, is there a kernel change log somewhere?
Thanks again for your response.
It's an unmanaged switch, and so there are no settings to view. The feature list does not mention igmp snooping, however:
http://www.cisco.com/en/US/prod/collateral/switches/ps5718/ps10863/datasheet_C78-582017.html
Thanks for your thoughts. I'm also interested to know if anyone has any ideas...
Thanks for your response.
The NICs are onboard... Supermicro X9SCL-F. The motherboard specs say they are Intel 82579LM and 82574L (2x Gigabit Ethernet LAN ports).
The switch that the affected servers are connected to is a Cisco SG102-24. However, we are on our third switch... this same...
We have a couple dozen Proxmox servers, and about once a month, one of them will have a kernel panic and lock up. The worst part about these lock ups is that when it's a node that is on a separate switch, all other Proxmox servers on that switch will stop responding until we can find the server...
I wish I knew how to test it since I don't know how to reproduce it. I have a couple dozen servers running Proxmox... on average, the crashes happen on only about one server per month... no one server has crashed more than once and most have never crashed. I haven't heard of anyone...
We're running version 1.9. At the moment, I'm considering seeing what it would take to back port the patch myself since I've heard no indication that Proxmox 2.x addresses this bug. I thought perhaps there was something unique about my hardware, but Supermicro motherboards seems pretty...
Another Proxmox server crashed last night, same divide by zero issue. Does anyone know if the kernel fix for this has been back ported into the kernel for Proxmox 2.x? See:
http://serverfault.com/questions/495502/debian-server-locked-find-busiest-group
Thanks,
Curtis
Since Proxmox 2.x is based is based on pre-3.x kernels, I'm assuming that it's subject to the same divide by zero crashes as Proxmox 1.x... is that right? See:
http://forum.proxmox.com/threads/13392-Divide-by-zero-kernel-crashes
Curtis
Are you saying that the kernel fix has been back ported into the kernel used in Proxmox 2.x? My understanding is that the fix is only in the 3.x kernels, so if the fix hasn't been back ported, switching to Proxmox 2.x isn't going to help.
Curtis
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.