Opt-in Linux 6.17 Kernel for Proxmox VE 9 available on test & no-subscription

RIP: 0010:megasas_build_and_issue_cmd_fusion+0xeaa/0x1870 [megaraid_sas][ 28.571290] Code: 20 48 89 d1 48 83 e1 fc 83 e2 01 48 0f 45 d9 4c 8b 73 10 44 8b 6b 18 4c 89 f9 4c 8d 79 08 45 85 fa 0f 84 fd 03 00 00 45 29 cc <4c>
similar trace as in https://forum.proxmox.com/threads/ceph-osd-crashes-with-kernel-6-17-2-1-pve-on-dell-system.176725/
(different system vendor, but also a newer machine)
-> I'll update the thread there as it seems a good fit (more targetet than the general thread for the kernel)
 
I can confirm this: BOSS-S1 controllers with Intel SSDs do not work with the 6.17 kernel. In contrast, BOSS-S1/2 controllers with Micron SSDs work without any issues. Same behavior as described.
Is there any idea what the problem could be? Is there a bug report or something similar about this? I haven’t found anything about it yet. We have entire clusters with this combination, so I hope the workaround won’t end up being to swap out the SSDs
 
Is there any idea what the problem could be? Is there a bug report or something similar about this? I haven’t found anything about it yet. We have entire clusters with this combination, so I hope the workaround won’t end up being to swap out the SSDs

I had to actually pin the 6.17.2-1 kernel on a R530 BOSS-S1 PBS instance. It locks up with the 6.17.2-2 kernel. Something obviously changed with 6.17.2-2.
 
I had to actually pin the 6.17.2-1 kernel on a R530 BOSS-S1 PBS instance. It locks up with the 6.17.2-2 kernel. Something obviously changed with 6.17.2-2.

6.17.4-1-pve is out now. It is based on the updated Ubuntu kernel of Ubuntu-6.17.0-9.9 kernel. So there should be new hardware updates.
 
I'm running 6.17.2-2-pve and I noticed it broke PBS, so I have rolled back to 6.14.11-4-pve as all backups were stalling. it appears the TCP receive window too small.

I'm still running 6.17.2-2-pve on one cluster of PVE9's and wondering if I need to pull the kernels back on them too, as I noticed live migration between hosts is slower than it used to be, again I suspect window size issues with the kernel?
 
I'm running 6.17.2-2-pve and I noticed it broke PBS, so I have rolled back to 6.14.11-4-pve as all backups were stalling. it appears the TCP receive window too small.

I'm still running 6.17.2-2-pve on one cluster of PVE9's and wondering if I need to pull the kernels back on them too, as I noticed live migration between hosts is slower than it used to be, again I suspect window size issues with the kernel?
Crosspost reply here. I also noticed problems with 6.17.2-2 which are not an issue on 6.17.2-1: https://forum.proxmox.com/threads/s...o-pve-9-1-1-and-pbs-4-0-20.176444/post-822997

On top of that, in that treat it does not look like 6.17.4-1 solves it completely.
 
Actually had to pin 6.14.11-4 kernel on PBS instances. 6.17.2-1 was giving intermittent issues on BOSS-S1.
 
  • Like
Reactions: carles89
Hi

Just a small comment.

After I updated the Proxmox kernel to 6.17.2-1 , my Plex container stopped working with hw transcoding . Intel_gpu_top show Plex transcoding with hw , but video enhance and render is “dead” . Seems that the server is doing software transcoding, instead of using the GPU.

Cpu is Intel Core Ultra 9 285T

Pinned kernel 6.14.11-4 everything works again before I updated.

Since im a total noob at this kernel thingy, I had to Google gemini my way back to the “old” kernel. Didnt know that I could do that, but I pinned and rebooted and its running perfectly :)

Dont know if its the right place to post , but I want to give it a go :)

Thank you

I tried to use kernel 6.17.2-2-pve on a lxc ubuntu 24.04.03 container. HW transcoding works again :)
 
  • Like
Reactions: SInisterPisces
We recently uploaded the 6.17 kernel to our repositories. The current default kernel for the Proxmox VE 9 series is still 6.14, but 6.17 is now an option.

We plan to use the 6.17 kernel as the new default for the Proxmox VE 9.1 release later in Q4.
This follows our tradition of upgrading the Proxmox VE kernel to match the current Ubuntu version until we reach an Ubuntu LTS release, at which point we will only provide newer kernels as an opt-in option. The 6.17 kernel is based on the Ubuntu 25.10 Questing release.

We have run this kernel on some of our test setups over the last few days without encountering any significant issues. However, for production setups, we strongly recommend either using the 6.14-based kernel or testing on similar hardware/setups before upgrading any production nodes to 6.17.


Guys there is a kernel memory leak in the in kernel CephFS driver for all kernels after 6.15. That means 6.17 WILL leak kernel memory over time if you use cephFS

https://tracker.ceph.com/issues/74156

It's a slow leak but you cannot get the memory to free once it has leaked. You probably want to look into that before moving to anything like production with this kernel.
 
Guys there is a kernel memory leak in the in kernel CephFS driver for all kernels after 6.15. That means 6.17 WILL leak kernel memory over time if you use cephFS

https://tracker.ceph.com/issues/74156

It's a slow leak but you cannot get the memory to free once it has leaked. You probably want to look into that before moving to anything like production with this kernel.
Developers don't always see posts like this on the forums. Submit this on https://bugzilla.proxmox.com/ for it to get the best attention.
 
There may be a bug with the cpu-scheduler:
  • 6.14.8-2-pve works
  • booting 6.17.1 or 6.17.2 only one cpu-core (0 out of 31) is used
  • CPU is AMD Opteron(tm) Processor 6262 HE
  • mpstat -P ALL 1
    shows that only CPU 0 is used, other CPUs sometimes shows 1% or to 2% usage, seems to be because of interrupts.