Hi Proxmox community!
I have a problem with my setup, which I know it is not ideal.
PVE0: SSD zfs pool with glusterfs
PVE1: SSD zfs pool with glusterfs, HDD zfs pool shared via NFS
PVE2: Arbiter glusterfs + "compute node"
PVE3: "compute node"
Since I updated every node from 6 to 7 I got random panics on PVE1.
Glusterfs is not connected anymore, but via console I can write and read from the local zfs mount (ssd and hdd pool).
Since then I tried to pin point what is going on but cant find it exactly.
On PVE1 I run a VM with Proxmox Backup Server with a disk on the NFS mount (PVE1) which always worked. Since the upgrade I can't run any backups.
Also when syncing up with the other glusterfs node it will eventually panic and crash the whole cluster. It also does something with the pvestatd. When restarting that progress the vm's will run fine then. I tried every zfs pool seperate to make sure it was not my HBA for HDD's nor the SSD pool. But on all the pools it gave me those panics.
Don't know which kernel it was, but the latest free kernel, not 5.15, gave me also some NFS kernel panics. But after upgrading to 5.15 kernel that panic went away. Also tried to reinstall the PVE1 node but that didn't gave me the solution.
I don't know what more information you need but I can give all details.
Here is the kernel panic from what I can get.
Sorry for my explanation.
I have a problem with my setup, which I know it is not ideal.
PVE0: SSD zfs pool with glusterfs
PVE1: SSD zfs pool with glusterfs, HDD zfs pool shared via NFS
PVE2: Arbiter glusterfs + "compute node"
PVE3: "compute node"
Since I updated every node from 6 to 7 I got random panics on PVE1.
Glusterfs is not connected anymore, but via console I can write and read from the local zfs mount (ssd and hdd pool).
Since then I tried to pin point what is going on but cant find it exactly.
On PVE1 I run a VM with Proxmox Backup Server with a disk on the NFS mount (PVE1) which always worked. Since the upgrade I can't run any backups.
Also when syncing up with the other glusterfs node it will eventually panic and crash the whole cluster. It also does something with the pvestatd. When restarting that progress the vm's will run fine then. I tried every zfs pool seperate to make sure it was not my HBA for HDD's nor the SSD pool. But on all the pools it gave me those panics.
Don't know which kernel it was, but the latest free kernel, not 5.15, gave me also some NFS kernel panics. But after upgrading to 5.15 kernel that panic went away. Also tried to reinstall the PVE1 node but that didn't gave me the solution.
I don't know what more information you need but I can give all details.
Here is the kernel panic from what I can get.
Sorry for my explanation.
Attachments
Last edited: