Proxmox 9.0.5 Ceph 19.2.3 Kubernetes ceph-csi-cephfs Hang

[SOLVED] Huh, I must have not rebooted after installing a LTS kernel. I can verify that 6.12.43 works just fine, while 6.16.3 (and at least a few releases before that) does not. That's what I get for living on the bleeding edge :( Nothing else is different but the kernel version, so there must be some issue in the ceph module in the kernel.
 
Damn, thanks for the follow-up! I'm glad it works now.

I'll see if I can maybe find something related to this upstream; maybe it's a known issue already. Otherwise I'll see if I can file a bug... somewhere.
 
Yeah, I couldn't find anything upstream at my searches. I don't think too many people are using Ceph clients with the bleeding edge kernel. Hopefully it can be found and squashed before the next LTS kernel. I'm not sure what exactly triggered it, I tried to reproduce it with fio with multiple mounts in different containers, and transferring between the two, but I could only trigger it with a real workload (only took a few minutes after that). Annoyingly there's no logs anywhere, the kernel just goes poof.

Thanks again!
 
  • Like
Reactions: Max Carrara