Just want to chime in on the system not booting after updates, boot issue could possibly be related to an issue beeing discussed in this topic.
I've noticed OP also has AMD hardware and a good couple of people (myself included) seem to experience boot issues with the latest production kernel...
Just adding info, i've been able to extract the following from kern.log on my box:
Dec 6 20:24:45 arcturus kernel: [ 20.359574] amdgpu 0000:00:01.0: amdgpu: amdgpu_device_ip_init failed
Dec 6 20:24:45 arcturus kernel: [ 20.359580] amdgpu 0000:00:01.0: amdgpu: Fatal error during GPU init...
My tip is entirely seperate from the path's outlined earlier in this topic.
If you run into this issue and you have not done anything yet, only my suggested GRUB config change is enough to have a bootable scenario again.
Deleting the offending kernel and/or excluding it from updates is not...
If you've already removed the problematic kernel and you auto-boot correctly now there's not real value in doing this i think.
But if this is the latest kernel you have installed then 1>0 would be correct, assuming 1 (beeing the second menu option in the first menu) actually opens the advanced...
Running into this issue as well. Specifically with the -2 kernel. -1 Boots just fine.
Also put a Gen 10 Hp Proliant Microserver with AMD Opteron Dual Core on the stack of affected machines in the same veign.
This is my home box with community repo. This also a thing on enterprise?
I'm looking at working on my cluster security somewhat and for that end want to utilize the pve firewall.
Looking through the instructions here i read that if I want to administer it remotely I need to add exceptions for it in order not to lose access as it claims only 22 and 8006 from it's...
Dove into SMART codes a little more.
https://en.wikipedia.org/wiki/S.M.A.R.T. where 0x04 referes Start/Stop Count.
Using smartctl -a -d megaraid,<disk#> /dev/sda I can get some individual disk data.
On both servers with disk 0 they log:
(other drives report Health Status OK)
We have a 9 node cluster.
Now I have two nodes that start sending me SMART mails.
One of which started doing this this morning, after I installed the latest updates and rebooted it yesterday. The other one started logging this about a month ago.
Both of these nodes are PowerEdge R620 machines...
We've chosen to re-install this node as i'm running out of time and patience to further troubleshoot this.
Gave it a new name and new IP.
Removed old node from cluster.
Added this new-new node, also fully updated, to the same existing "not quite yet updated" node in our cluster, however now...
I've found out latency may cause a problem. Even though I haven't seperated corosync traffic from the rest, cluster has it's own switch for interconnectivity and pings between them are reliably below 0.260 ms.
Furthermore, pveversion -v output of working node vs new node:
I tried adding a node to our (existing) cluster of currently four machines.
We tried to do this through GUI.
GUI on new node stopped responding after it was restarting pve-cluster...something. (didn't grab a screenshot)
GUI didn't come back. Server is still reachable over SSH.
Node has been...
Commenting to point out that fstab indeed contains a devpts rule if you migrate a centos 6 openvz container to proxmox 5.3. Having this rule was no problem untill recently but commenting this line out works seems to aid in a solution.
Sometimes you gotta say stuff alound to think of other things.
Tried directly searching this forum.
Found this in another recent topic here:
Commenting out this devpts rule in...
Some recent-ish Proxmox 5 updates seems to have caused an issue on my end with Centos 6 containers.
I have a couple of Centos 6 containers still that have been migrated from Proxmox 3.5 to 5. (so openvz to lxc)
These used to work fine, but in a recent update round SSH "became broken", getting an...
Little point to note up front: I come from Proxmox 3 and I'm used to all VM files just sitting some vz directory somewhere in var. These directories on PX5 are empty. (this is a default install)
So now i'm trying to move/migrate a VMWare machine to Proxmox, the part that involves converting the...
Also ran into this issue at about the fifteenth container I was trying to deploy/restore. (during migration from PX3) which is not a lot in our use case. (fairly simple single purpose webserver contianers)
I would definitely say this limit needs increasing OOTB...