Boot failed after upgrading to 7.0-11

Aug 20, 2021
Wuppertal, Germany
I upgraded my system today from 6 to 7.
I do the instructions from and all looking fine.
Furthermore, I have run the script pve6to7 with --full and without --full and no Errors or something else was reported.
So run the rest and upgraded the system from buster to bullseye. Everything looks good, no errors was reported. My VMs and Containers was stopped. So I reboot the system.
After a while, the system doesn't come up.
By default, the new Kernel is 5.11.22-3-pve, and this Kernel doesn't run. I switched manually in to the old Kernel 5.4.128-1-pve and the system boot perfectly and my VMs and Containers come normally up.
Where are the difference between 5.11.22-3 and 5.4.128-1? Can someone help me?
Can you be more specific about how "it doesn't come up"? For example, at which line does the boot sequence get stuck, or what do you observe in a faulty startup?
At grub screen, you can press e to edit grub configuration before booting. At that point, you can remove quiet from the kernel command line and then boot. There should be more output available, usually enough for you to locate the issue, like missing files or initramfs modules.

Alternatively, you can reboot into a running kernel (as described yourself, 5.4), edit /etc/default/grub and remove quiet from GRUB_CMDLINE_LINUX, then run update-grub and reboot. This achieves the same result (make Linux generate more output at startup) and is a persistent change.
Last edited:


The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!