In terms of memory allocation on this node, the total memory size allotted to the active VMs is greater than the available RAM on the node. However, the host memory usage is still below 50%, so it doesn't seem like reducing VM memory allocation would really have an impact here
I think you're probably right that what I'm seeing here is a slow-creep of file-based as opposed to anonymous page swapping. If the Proxmox kernel does indeed use the Control Group swappiness value rather than the global swappiness value, this would seem to make sense:
root@node:~# cat...
I suppose if I were to boil this down to a concrete problem it would be this:
The constant utilization of ~100% of any server resource indicates to me a need to either upgrade or reconfigure the server. In this case, my concern is that keeping the swap usage pegged will lead to perpetual IO...
I have a Proxmox node currently configured with 126 GB RAM, actively running 10 Linux and Windows VMs. Our users were reporting issues with some of these VMs which seemed to coincide with high SWAP usage on the node, even though total RAM usage was below 50%. To ease the SWAP load, I made two...
I was noticing somewhat better outcomes on this node when I was running VMs and installer ISOs from local storage vs over a bonded 10GbE NFS connection (i.e no sudden node reboots, but numerous hangs and segfaults when installing new Linux VMs). In an attempt to remove as many variables as...
That is not something that had ever occurred to me to check. Thank you for shedding light on this aspect of Proxmox clustering and configuration. However, our setup is not clustered and the problematic node is not communicating with any other nodes on the network.
(We are planning on getting...
I've been diagnosing a Proxmox node on our network that began acting up about a year ago with the primary symptom being spontaneous and frequent node reboots while under light load. I've offloaded every VM and container from the node and reinstalled the latest version of PVE 6 to create a clean...
I think I've narrowed this issue down. The reboots started after I migrated a VM from a different node and stopped after I removed the VM. The two nodes have different hardware configurations. This Supermicro has an AMD EPYC CPU and the other node has an Intel Xeon CPU. It seemed like a...
So...this was definitely a case of user error.
The reason for the "NO-CARRIER" interface status was, embarrassingly, that I had configured the bond to use NIC interfaces from an older 10GbE card in the same server and then I plugged the Ethernet cables into the interfaces on my new card...
I ran Supermicro offline diagnostics on this node and everything passed, including memory and PSUs. It looks like the hardware angle may have been a dead end.
After running diagnostics, I rebooted the node back into Proxmox and powered down all VMs to see if the problem resurfaced. As of today...
The server in question is a Supermicro 819UAS-10 and it does not appear to have any configured power management devices from what I can tell remotely. Supermicro does provide an offline diagnostics tool that includes memory and PSU testing that runs on a USB drive. It looks like I'm straying way...
For additional info, here's what I currently have for the relevant interfaces under /etc/network/interfaces. All of these settings were configured in the Proxmox GUI:
...
...
I've replaced a 10GbE adapter in one of our Proxmox nodes that was running a bonded pair of Cat6 cables to an aggregate switch. The old NIC and its replacement are both Intel Ethernet Controllers (10-Gigabit X540-AT2) so I assumed (probably incorrectly) that all I'd need to do was plug it in...
Not sure if the backup is the culprit, looking at the correlation between the backups scheduled and the times of the reboots, sometimes less than 5 minutes apart. I manually ran the backup scheduled at 04:00 and got a task failure notification in the task log, which makes sense as I've migrated...
That did the trick. I removed the pre-up line from the /etc/network/interfaces and set the MTU of bond0 to 9000. After applying the configuration, the slave interfaces successfully inherited the bond's MTU value. Thanks for the tip!
I'm rebuilding network connections on a Proxmox node set up by a previous sysadmin after I replaced a failing NIC. While configuring the new NIC's interfaces in the web GUI, I noticed that there were some customizations that appeared to be deeper than I could touch through the GUI. Specifically...
Here's an excerpt from the syslog for a reboot that began at 04:13 today. I notice a string of ascii NULL characters in the log at the time of reboot which I interpret as an error of the log file getting interrupted as it was writing new entries from memory. The most recent prior entry was about...
I recently migrated a number of VMs to another existing node on my network in order to perform hardware maintenance (See thread: Manually move VM to a different Node). The migrated VMs are running just fine, however, I've discovered that the new node is now automatically rebooting at decreasing...
Thank you, kindly! I was able to transfer the config files to the new node and was impressed to see the VMs populate almost immediately. Taking your tip to remove the config files from the former node to prevent any collisions if both instances were to boot at the same time, here are the steps I...
All of our nodes have access to the same pool of NFS drives under Datacenter/Storage in the GUI so nodes X and Y can both see the storage location for these VMs. How does one go about moving the config files?
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.