For those with dkms troubles in this kernel: a newer 5.11 build is available, at least on the pve-no-subscription repository, its pve-headers-5.11.17-1-pve package should improve the modules.lds linker script situation required for external module builds.
For those with dkms troubles in this kernel: a newer 5.11 build is available, at least on the pve-no-subscription repository, its pve-headers-5.11.17-1-pve package should improve the modules.lds linker script situation required for external module builds.
For those with dkms troubles in this kernel: a newer 5.11 build is available, at least on the pve-no-subscription repository, its pve-headers-5.11.17-1-pve package should improve the modules.lds linker script situation required for external module builds.
Ubuntu sunset their natively integrated Aufs support for their 21.04 Release, which the 5.11 Kernel comes from, only enable it for backports. So, enabling it for the 5.11 backport to the Proxmox VE 6.4 release could be done, but would be a very temporary measurement. We do not currently plan to re-add support for aufs in future releases, so alternatives for you would be:
building aufs yourself
the easiest way using an aufs-dkms package seems sadly not useable any more as that Debian package has be abandoned
using another FS as base for the applications requiring overlays, either through a loop device or by using (pasing through) a zvol into the CT and format it with ext4, xfs, ...
using a VM to manage those applications
Ideally, the kernel built-in overlayfs and ZFS would support each other avoiding more external modules completely, but that does not seems to be very close to happen.
Last Edit ... nothing to do with kernel 5.11?
Problem was there before (SATA SSD connected as UDMA/33 (slow speed)) but kernel 5.4 never told me anything.
Kernel 5.11 was giving me errors that gave me enough info to find an alternative BIOS (modded) for my NL40 to accept my SSD at correct speed.
To keep this thread clean i'll "spoiler" all the rest of my text because it's not relevant here.
Error was :
Code:
May 18 10:04:53 pve-backup kernel: ata1.01: ATA-9: SanDisk Ultra II 240GB, X41100RL, max UDMA/133
May 18 10:04:53 pve-backup kernel: ata1.01: 468862128 sectors, multi 1: LBA48 NCQ (depth 0/32)
May 18 10:04:53 pve-backup kernel: ata1.01: limited to UDMA/33 due to 40-wire cable
May 18 10:19:44 pve-backup kernel: ata1: lost interrupt (Status 0x58)
May 18 10:19:44 pve-backup kernel: ata1.01: exception Emask 0x0 SAct 0x0 SErr 0x0 action 0x6 frozen
May 18 10:19:44 pve-backup kernel: ata1.01: failed command: READ DMA
May 18 10:19:44 pve-backup kernel: ata1.01: cmd c8/00:01:7f:c0:55/00:00:00:00:00/f8 tag 0 dma 512 in
res 40/00:00:01:4f:c2/00:00:00:00:00/10 Emask 0x4 (timeout)
May 18 10:19:44 pve-backup kernel: ata1.01: status: { DRDY }
May 18 10:19:44 pve-backup kernel: ata1: soft resetting link
May 18 10:19:44 pve-backup kernel: ata1.01: configured for UDMA/33
May 18 10:19:44 pve-backup kernel: ata1: EH complete
Just updated one of my proxmox (on an old hardware) to the 5.11 kernel and I have a strange messages in logs regarding apparently my disks.
My server spec :
HP Proliant Microserver NL40
CPU(s) AMD Turion(tm) II Neo (1 Core/2 Threads)
8Go RAM
1* SSD SATA
4* HDD SATA
Update went well. The server is booting without any errors.
On this server I run 2 VM. One pfSense VM and ont PBS VM.
If i keep the server with only pfSense VM up, no error messages, everything seems fine.
The errors comes when I start the PBS VM. After like 1 or 2 minutes I have strange log messages.
So first, here is my PBS VM config :
While pasedthrough the VM, those 4 disks are excluded from pvestatd (like that I can spin them down).
When i run the VM the disks are spinned up and running.
And now the errors messages.
I have 2 kind of error messages.
Those errors messages may appear 1 to 4 times after the boot of the VM.
Some are directly on the screen of the pve :
Some are in the logs that I can see in pve WebGUI :
May 18 10:04:53 pve-backup kernel: ata1.01: ATA-9: SanDisk Ultra II 240GB, X41100RL, max UDMA/133
May 18 10:04:53 pve-backup kernel: ata1.01: 468862128 sectors, multi 1: LBA48 NCQ (depth 0/32)
May 18 10:04:53 pve-backup kernel: ata1.01: limited to UDMA/33 due to 40-wire cable
May 18 10:19:44 pve-backup kernel: ata1: lost interrupt (Status 0x58)
May 18 10:19:44 pve-backup kernel: ata1.01: exception Emask 0x0 SAct 0x0 SErr 0x0 action 0x6 frozen
May 18 10:19:44 pve-backup kernel: ata1.01: failed command: READ DMA
May 18 10:19:44 pve-backup kernel: ata1.01: cmd c8/00:01:7f:c0:55/00:00:00:00:00/f8 tag 0 dma 512 in
res 40/00:00:01:4f:c2/00:00:00:00:00/10 Emask 0x4 (timeout)
May 18 10:19:44 pve-backup kernel: ata1.01: status: { DRDY }
May 18 10:19:44 pve-backup kernel: ata1: soft resetting link
May 18 10:19:44 pve-backup kernel: ata1.01: configured for UDMA/33
May 18 10:19:44 pve-backup kernel: ata1: EH complete
What intrigates me is this limited to UDMA/33 due to 40-wire cable and the fact that the error seems to be on my SSD and not on the HDD that i passthrough on my PBS VM.
Things I tested :
- I booted back on the 5.4 kernel and everything went fine.
- I stopped the PBS VM to run automatically at boot and staarted it manually 20 minutes after the server boot (with 5.11 kernel). Same behavior at start.
- When the PBS VM is offline : no error messages. (even after like 1h).
- /etc/multipath.conf not present on the system (to answer a question).
Other info :
- I don't think it's hardware / cable related because those disks are used to backup my system ... and backups take like 3h to 4h to do so i think that if I had intermitent sata cable problem backups/restoration whould have showed problems.
- One really strange thing is ... when i start my PBS VM, graphs on the pve WebGUI are like freezed, I can not access the console of the PBS VM via WebGUI either
- Error messages are always on the ata1.01 wich is my SSD where proxmox is installed ...
@t.lamprecht to keep it "clean" in this thread I updated this post. If you prefer I can delete the post and create a new thread. Just tell me.
Edit :
So it seems that the problem is on my SSD and not my storage disks. Logs from pve WebGUI seems to show it.
That's really strange because ... well ... why does it only happen when i run my PBS VM ... the pfSense VM is olso on the SSD ... i'm lost.
Edit 2 :
Ok so apparently it's linked to my MB / Bios. In fact my SSD is seen as udma2 and not udma6 on kernel 5.4 and 5.11.
The only difference is that kernel 5.11 is showing log messages apparently.
I'll continue my search to confirm that.
In general new messages do not immediately have to mean that the new kernel is at fault, as the issue, if even one, could have been there forever and the newer kernel just exposes (logs) it...
For permanently switching back: Remove the installed pve-kernel-5.11-.... kernel package with apt remove PKG, that should remove the meta package too and thus only the in PVE 6.4 still default 5.4 kernel should be available, then reboot.
For a single boot to confirm if this is some effect of the new kernel: choose the old one on boot (e.g. "Advanced" menu in GRUB boot loader).
Is there iscsi/multipath setup on the system? If, then maybe the device paths changed due to the newer kernel, and the multipathd conf does not exclude those local devices anymore, so it tries to access it, which fails for local devices. In that case check if any device path changed and add the new one also to the blacklist in /etc/multipath.conf.
Another potential source of such link-resets could mean that there's a (slightly/semi-) broken SATA cable causing this, e.g., loose connection, would be a bit of a weird coincidence though.
If none of above seem to apply, and you still want to investigate this, I'd suggest opening a new thread with more info on your setup.
I think because proxmox is basically reusing the ubuntu 21.04 kernel which is 5.11 and not rolling this out completely on their own, but would be interested in their answer.
For my part, I would have preferred a 5.10 LTS kernel rather than the ubuntu one. Who precisely does everything against the grain they are never in LTS that's why I left ubuntu a long time ago.
One can port stable kernel patches to other kernel release tree, that's a need in general anyway, not everything can be immediately upstreamed and made available in a LTS.
FWIW, Proxmox VE based its Kernel on the Ubuntu Kernel since Proxmox VE 4.0, and that worked out well in general, at least I do not remember a time when a real issue popped up in our/ubuntu's kernel which would have been avoided in the respective LTS tree.
I just moved to 5.11 and noticed that the CPU counter for "guest" has stopped working. I have a telegraf + grafana setup which is where I first saw this but looking at /proc/stat shows all the values for guest usage at 0. Does anyone else see the same behavior?
For those with dkms troubles in this kernel: a newer 5.11 build is available, at least on the pve-no-subscription repository, its pve-headers-5.11.17-1-pve package should improve the modules.lds linker script situation required for external module builds.
i’ll wager the answer is in the post up above, linked below. They will use what Ubuntu puts out and switch to 5.12 when Ubuntu does and until then can include backports to the 5.11 kernel for any necessary 5.12 updates.
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.